The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    ArrowNotImplementedError
Message:      Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1870, in _prepare_split_single
                  writer.write_table(table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 620, in write_table
                  self._build_writer(inferred_schema=pa_table.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1886, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 639, in finalize
                  self._build_writer(self.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1417, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1049, in convert_to_parquet
                  builder.download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare
                  self._download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1000, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1741, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1897, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

config
dict
report
dict
name
string
backend
dict
scenario
dict
launcher
dict
environment
dict
print_report
bool
log_report
bool
overall
dict
warmup
dict
train
dict
{ "name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6863499510000111, 0.5503289600000016, 0.5707251910000082, 0.5524302560000365, 0.5464166680000062 ], "count": 5, "total": 2.9062510260000636, "mean": 0.5812502052000127, "p50": 0.5524302560000365, "p90": 0.6401000470000099, "p95": 0.6632249990000104, "p99": 0.681724960600011, "stdev": 0.05321056414210851, "stdev_": 9.154502427022514 }, "throughput": { "unit": "samples/s", "value": 17.20429500159733 }, "energy": { "unit": "kWh", "cpu": 0.00012068394508333333, "ram": 0.000005047231533780827, "gpu": 0, "total": 0.00012573117661711416 }, "efficiency": { "unit": "samples/kWh", "value": 79534.76829738687 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6863499510000111, 0.5503289600000016 ], "count": 2, "total": 1.2366789110000127, "mean": 0.6183394555000064, "p50": 0.6183394555000064, "p90": 0.6727478519000101, "p95": 0.6795489014500106, "p99": 0.684989741090011, "stdev": 0.06801049550000471, "stdev_": 10.99889306675563 }, "throughput": { "unit": "samples/s", "value": 6.468938645950532 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.5707251910000082, 0.5524302560000365, 0.5464166680000062 ], "count": 3, "total": 1.6695721150000509, "mean": 0.5565240383333503, "p50": 0.5524302560000365, "p90": 0.5670662040000138, "p95": 0.568895697500011, "p99": 0.5703592923000087, "stdev": 0.010337483992887902, "stdev_": 1.8575089808961478 }, "throughput": { "unit": "samples/s", "value": 10.781205458740818 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_fill-mask_google-bert/bert-base-uncased
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6863499510000111, 0.5503289600000016, 0.5707251910000082, 0.5524302560000365, 0.5464166680000062 ], "count": 5, "total": 2.9062510260000636, "mean": 0.5812502052000127, "p50": 0.5524302560000365, "p90": 0.6401000470000099, "p95": 0.6632249990000104, "p99": 0.681724960600011, "stdev": 0.05321056414210851, "stdev_": 9.154502427022514 }, "throughput": { "unit": "samples/s", "value": 17.20429500159733 }, "energy": { "unit": "kWh", "cpu": 0.00012068394508333333, "ram": 0.000005047231533780827, "gpu": 0, "total": 0.00012573117661711416 }, "efficiency": { "unit": "samples/kWh", "value": 79534.76829738687 } }
{ "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6863499510000111, 0.5503289600000016 ], "count": 2, "total": 1.2366789110000127, "mean": 0.6183394555000064, "p50": 0.6183394555000064, "p90": 0.6727478519000101, "p95": 0.6795489014500106, "p99": 0.684989741090011, "stdev": 0.06801049550000471, "stdev_": 10.99889306675563 }, "throughput": { "unit": "samples/s", "value": 6.468938645950532 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 2519.744512, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.5707251910000082, 0.5524302560000365, 0.5464166680000062 ], "count": 3, "total": 1.6695721150000509, "mean": 0.5565240383333503, "p50": 0.5524302560000365, "p90": 0.5670662040000138, "p95": 0.568895697500011, "p99": 0.5703592923000087, "stdev": 0.010337483992887902, "stdev_": 1.8575089808961478 }, "throughput": { "unit": "samples/s", "value": 10.781205458740818 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "model": "google-bert/bert-base-uncased", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2488.782848, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 2.738775005999969, "mean": 0.5477550011999938, "stdev": 0.03693447784258994, "p50": 0.5307143729999666, "p90": 0.5856752317999963, "p95": 0.6036043034000045, "p99": 0.6179475606800111, "values": [ 0.6215333750000127, 0.5307143729999666, 0.5270229210000252, 0.5276163199999928, 0.5318880169999716 ] }, "throughput": { "unit": "samples/s", "value": 18.25633719106628 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 2488.782848, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 1.1522477479999793, "mean": 0.5761238739999897, "stdev": 0.045409501000023056, "p50": 0.5761238739999897, "p90": 0.612451474800008, "p95": 0.6169924249000104, "p99": 0.6206251849800123, "values": [ 0.6215333750000127, 0.5307143729999666 ] }, "throughput": { "unit": "samples/s", "value": 6.942951300088064 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2488.782848, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 1.5865272579999896, "mean": 0.5288424193333299, "stdev": 0.0021671455040491463, "p50": 0.5276163199999928, "p90": 0.5310336775999758, "p95": 0.5314608472999737, "p99": 0.531802583059972, "values": [ 0.5270229210000252, 0.5276163199999928, 0.5318880169999716 ] }, "throughput": { "unit": "samples/s", "value": 11.345534663357242 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cpu_training_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "model": "hf-internal-testing/tiny-random-BertModel", "processor": "hf-internal-testing/tiny-random-BertModel", "task": "fill-mask", "library": "transformers", "model_type": "bert", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "7cec62e016d76fe612308e4c2c074fc7f09289fd", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.1", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.01592566299999021, 0.012209281999986388, 0.012050035000015669, 0.01193647200000214, 0.012116388000009692 ], "count": 5, "total": 0.0642378400000041, "mean": 0.012847568000000819, "p50": 0.012116388000009692, "p90": 0.01443911059998868, "p95": 0.015182386799989444, "p99": 0.015777007759990056, "stdev": 0.001541613235960306, "stdev_": 11.999261151684177 }, "throughput": { "unit": "samples/s", "value": 778.357429203672 }, "energy": { "unit": "kWh", "cpu": 0.000008562100477777949, "ram": 3.576500245353414e-7, "gpu": 0, "total": 0.00000891975050231329 }, "efficiency": { "unit": "samples/kWh", "value": 1121107.5912276418 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.01592566299999021, 0.012209281999986388 ], "count": 2, "total": 0.028134944999976597, "mean": 0.014067472499988298, "p50": 0.014067472499988298, "p90": 0.015554024899989827, "p95": 0.015739843949990017, "p99": 0.01588849918999017, "stdev": 0.0018581905000019106, "stdev_": 13.209128363346409 }, "throughput": { "unit": "samples/s", "value": 284.34390044148495 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.012050035000015669, 0.01193647200000214, 0.012116388000009692 ], "count": 3, "total": 0.0361028950000275, "mean": 0.0120342983333425, "p50": 0.012050035000015669, "p90": 0.012103117400010887, "p95": 0.01210975270001029, "p99": 0.012115060940009812, "stdev": 0.00007428850863655893, "stdev_": 0.6173065232289738 }, "throughput": { "unit": "samples/s", "value": 498.5749757737237 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "model": "hf-internal-testing/tiny-random-BertModel", "processor": "hf-internal-testing/tiny-random-BertModel", "task": "fill-mask", "library": "transformers", "model_type": "bert", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "7cec62e016d76fe612308e4c2c074fc7f09289fd", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.1", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.01592566299999021, 0.012209281999986388, 0.012050035000015669, 0.01193647200000214, 0.012116388000009692 ], "count": 5, "total": 0.0642378400000041, "mean": 0.012847568000000819, "p50": 0.012116388000009692, "p90": 0.01443911059998868, "p95": 0.015182386799989444, "p99": 0.015777007759990056, "stdev": 0.001541613235960306, "stdev_": 11.999261151684177 }, "throughput": { "unit": "samples/s", "value": 778.357429203672 }, "energy": { "unit": "kWh", "cpu": 0.000008562100477777949, "ram": 3.576500245353414e-7, "gpu": 0, "total": 0.00000891975050231329 }, "efficiency": { "unit": "samples/kWh", "value": 1121107.5912276418 } }
{ "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.01592566299999021, 0.012209281999986388 ], "count": 2, "total": 0.028134944999976597, "mean": 0.014067472499988298, "p50": 0.014067472499988298, "p90": 0.015554024899989827, "p95": 0.015739843949990017, "p99": 0.01588849918999017, "stdev": 0.0018581905000019106, "stdev_": 13.209128363346409 }, "throughput": { "unit": "samples/s", "value": 284.34390044148495 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 477.351936, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.012050035000015669, 0.01193647200000214, 0.012116388000009692 ], "count": 3, "total": 0.0361028950000275, "mean": 0.0120342983333425, "p50": 0.012050035000015669, "p90": 0.012103117400010887, "p95": 0.01210975270001029, "p99": 0.012115060940009812, "stdev": 0.00007428850863655893, "stdev_": 0.6173065232289738 }, "throughput": { "unit": "samples/s", "value": 498.5749757737237 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.6716150309999875, 1.5119543120000003, 1.5309882200000402, 1.5602752419999888, 1.5229964970000083 ], "count": 5, "total": 7.797829302000025, "mean": 1.559565860400005, "p50": 1.5309882200000402, "p90": 1.627079115399988, "p95": 1.6493470731999877, "p99": 1.6671614394399876, "stdev": 0.058268995642202545, "stdev_": 3.736231801538502 }, "throughput": { "unit": "samples/s", "value": 6.412040846697652 }, "energy": { "unit": "kWh", "cpu": 0.0003162784703944444, "ram": 0.000013228306043901064, "gpu": 0, "total": 0.00032950677643834545 }, "efficiency": { "unit": "samples/kWh", "value": 30348.38951747967 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.6716150309999875, 1.5119543120000003 ], "count": 2, "total": 3.183569342999988, "mean": 1.591784671499994, "p50": 1.591784671499994, "p90": 1.6556489590999888, "p95": 1.6636319950499883, "p99": 1.6700184238099878, "stdev": 0.07983035949999362, "stdev_": 5.015148149703358 }, "throughput": { "unit": "samples/s", "value": 2.5129027007344287 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.5309882200000402, 1.5602752419999888, 1.5229964970000083 ], "count": 3, "total": 4.614259959000037, "mean": 1.5380866530000123, "p50": 1.5309882200000402, "p90": 1.554417837599999, "p95": 1.557346539799994, "p99": 1.5596895015599899, "stdev": 0.016025334535971766, "stdev_": 1.0419006305473504 }, "throughput": { "unit": "samples/s", "value": 3.9009505662747284 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_image-classification_google/vit-base-patch16-224
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.6716150309999875, 1.5119543120000003, 1.5309882200000402, 1.5602752419999888, 1.5229964970000083 ], "count": 5, "total": 7.797829302000025, "mean": 1.559565860400005, "p50": 1.5309882200000402, "p90": 1.627079115399988, "p95": 1.6493470731999877, "p99": 1.6671614394399876, "stdev": 0.058268995642202545, "stdev_": 3.736231801538502 }, "throughput": { "unit": "samples/s", "value": 6.412040846697652 }, "energy": { "unit": "kWh", "cpu": 0.0003162784703944444, "ram": 0.000013228306043901064, "gpu": 0, "total": 0.00032950677643834545 }, "efficiency": { "unit": "samples/kWh", "value": 30348.38951747967 } }
{ "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.6716150309999875, 1.5119543120000003 ], "count": 2, "total": 3.183569342999988, "mean": 1.591784671499994, "p50": 1.591784671499994, "p90": 1.6556489590999888, "p95": 1.6636319950499883, "p99": 1.6700184238099878, "stdev": 0.07983035949999362, "stdev_": 5.015148149703358 }, "throughput": { "unit": "samples/s", "value": 2.5129027007344287 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 2503.512064, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.5309882200000402, 1.5602752419999888, 1.5229964970000083 ], "count": 3, "total": 4.614259959000037, "mean": 1.5380866530000123, "p50": 1.5309882200000402, "p90": 1.554417837599999, "p95": 1.557346539799994, "p99": 1.5596895015599899, "stdev": 0.016025334535971766, "stdev_": 1.0419006305473504 }, "throughput": { "unit": "samples/s", "value": 3.9009505662747284 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "model": "google/vit-base-patch16-224", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2442.985472, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 7.2970974209999895, "mean": 1.459419484199998, "stdev": 0.05210139006345095, "p50": 1.4401334369999859, "p90": 1.521250764199999, "p95": 1.5379663595999886, "p99": 1.5513388359199802, "values": [ 1.5546819549999782, 1.4241662819999874, 1.4401334369999859, 1.4711039780000306, 1.4070117690000075 ] }, "throughput": { "unit": "samples/s", "value": 6.852039532336137 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 2442.985472, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 2.9788482369999656, "mean": 1.4894241184999828, "stdev": 0.06525783649999539, "p50": 1.4894241184999828, "p90": 1.541630387699979, "p95": 1.5481561713499787, "p99": 1.5533767982699782, "values": [ 1.5546819549999782, 1.4241662819999874 ] }, "throughput": { "unit": "samples/s", "value": 2.6856017371522416 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2442.985472, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 4.318249184000024, "mean": 1.4394163946666747, "stdev": 0.02617044676610726, "p50": 1.4401334369999859, "p90": 1.4649098698000216, "p95": 1.4680069239000262, "p99": 1.4704845671800297, "values": [ 1.4401334369999859, 1.4711039780000306, 1.4070117690000075 ] }, "throughput": { "unit": "samples/s", "value": 4.1683560241714614 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.8422577439999941, 0.7105179889999818, 0.7316318399999773, 0.7052919429999633, 0.7076670829999898 ], "count": 5, "total": 3.6973665989999063, "mean": 0.7394733197999812, "p50": 0.7105179889999818, "p90": 0.7980073823999874, "p95": 0.8201325631999907, "p99": 0.8378327078399934, "stdev": 0.05223895302289545, "stdev_": 7.064345882962413 }, "throughput": { "unit": "samples/s", "value": 13.523138336762281 }, "energy": { "unit": "kWh", "cpu": 0.00015092970748333288, "ram": 0.000006312406346927238, "gpu": 0, "total": 0.00015724211383026011 }, "efficiency": { "unit": "samples/kWh", "value": 63596.19415187213 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.8422577439999941, 0.7105179889999818 ], "count": 2, "total": 1.5527757329999758, "mean": 0.7763878664999879, "p50": 0.7763878664999879, "p90": 0.8290837684999929, "p95": 0.8356707562499934, "p99": 0.8409403464499939, "stdev": 0.06586987750000617, "stdev_": 8.484145662522046 }, "throughput": { "unit": "samples/s", "value": 5.152064029583933 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7316318399999773, 0.7052919429999633, 0.7076670829999898 ], "count": 3, "total": 2.1445908659999304, "mean": 0.7148636219999768, "p50": 0.7076670829999898, "p90": 0.7268388885999798, "p95": 0.7292353642999785, "p99": 0.7311525448599775, "stdev": 0.011896502950491687, "stdev_": 1.6641639865809361 }, "throughput": { "unit": "samples/s", "value": 8.393209299437808 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_multiple-choice_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.8422577439999941, 0.7105179889999818, 0.7316318399999773, 0.7052919429999633, 0.7076670829999898 ], "count": 5, "total": 3.6973665989999063, "mean": 0.7394733197999812, "p50": 0.7105179889999818, "p90": 0.7980073823999874, "p95": 0.8201325631999907, "p99": 0.8378327078399934, "stdev": 0.05223895302289545, "stdev_": 7.064345882962413 }, "throughput": { "unit": "samples/s", "value": 13.523138336762281 }, "energy": { "unit": "kWh", "cpu": 0.00015092970748333288, "ram": 0.000006312406346927238, "gpu": 0, "total": 0.00015724211383026011 }, "efficiency": { "unit": "samples/kWh", "value": 63596.19415187213 } }
{ "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.8422577439999941, 0.7105179889999818 ], "count": 2, "total": 1.5527757329999758, "mean": 0.7763878664999879, "p50": 0.7763878664999879, "p90": 0.8290837684999929, "p95": 0.8356707562499934, "p99": 0.8409403464499939, "stdev": 0.06586987750000617, "stdev_": 8.484145662522046 }, "throughput": { "unit": "samples/s", "value": 5.152064029583933 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 2909.048832, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7316318399999773, 0.7052919429999633, 0.7076670829999898 ], "count": 3, "total": 2.1445908659999304, "mean": 0.7148636219999768, "p50": 0.7076670829999898, "p90": 0.7268388885999798, "p95": 0.7292353642999785, "p99": 0.7311525448599775, "stdev": 0.011896502950491687, "stdev_": 1.6641639865809361 }, "throughput": { "unit": "samples/s", "value": 8.393209299437808 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2845.749248, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 3.581090587999995, "mean": 0.716218117599999, "stdev": 0.043372798377969854, "p50": 0.697155070000008, "p90": 0.7645524997999928, "p95": 0.7826806403999967, "p99": 0.7971831528799999, "values": [ 0.8008087810000006, 0.710168077999981, 0.6928677590000234, 0.697155070000008, 0.680090899999982 ] }, "throughput": { "unit": "samples/s", "value": 13.962227084549829 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 2845.749248, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 1.5109768589999817, "mean": 0.7554884294999908, "stdev": 0.04532035150000979, "p50": 0.7554884294999908, "p90": 0.7917447106999986, "p95": 0.7962767458499996, "p99": 0.7999023739700004, "values": [ 0.8008087810000006, 0.710168077999981 ] }, "throughput": { "unit": "samples/s", "value": 5.294588035778844 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2845.749248, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 2.0701137290000133, "mean": 0.6900379096666711, "stdev": 0.007248103654729414, "p50": 0.6928677590000234, "p90": 0.696297607800011, "p95": 0.6967263389000096, "p99": 0.6970693237800083, "values": [ 0.6928677590000234, 0.697155070000008, 0.680090899999982 ] }, "throughput": { "unit": "samples/s", "value": 8.695174447587021 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7184868769999753, 0.5956158320000213, 0.5611470650000001, 0.6053455429999985, 0.5643365999999901 ], "count": 5, "total": 3.0449319169999853, "mean": 0.6089863833999971, "p50": 0.5956158320000213, "p90": 0.6732303433999846, "p95": 0.6958586101999799, "p99": 0.7139612236399762, "stdev": 0.057383950726384025, "stdev_": 9.422862692923767 }, "throughput": { "unit": "samples/s", "value": 16.42072839817792 }, "energy": { "unit": "kWh", "cpu": 0.00012582436187222188, "ram": 0.0000052623983356353215, "gpu": 0, "total": 0.0001310867602078572 }, "efficiency": { "unit": "samples/kWh", "value": 76285.35470816075 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7184868769999753, 0.5956158320000213 ], "count": 2, "total": 1.3141027089999966, "mean": 0.6570513544999983, "p50": 0.6570513544999983, "p90": 0.7061997724999799, "p95": 0.7123433247499775, "p99": 0.7172581665499758, "stdev": 0.06143552249997697, "stdev_": 9.350185808037494 }, "throughput": { "unit": "samples/s", "value": 6.087804206786717 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.5611470650000001, 0.6053455429999985, 0.5643365999999901 ], "count": 3, "total": 1.7308292079999887, "mean": 0.5769430693333296, "p50": 0.5643365999999901, "p90": 0.5971437543999969, "p95": 0.6012446486999977, "p99": 0.6045253641399984, "stdev": 0.020125749115141166, "stdev_": 3.4883422966492885 }, "throughput": { "unit": "samples/s", "value": 10.399639615973083 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_text-classification_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7184868769999753, 0.5956158320000213, 0.5611470650000001, 0.6053455429999985, 0.5643365999999901 ], "count": 5, "total": 3.0449319169999853, "mean": 0.6089863833999971, "p50": 0.5956158320000213, "p90": 0.6732303433999846, "p95": 0.6958586101999799, "p99": 0.7139612236399762, "stdev": 0.057383950726384025, "stdev_": 9.422862692923767 }, "throughput": { "unit": "samples/s", "value": 16.42072839817792 }, "energy": { "unit": "kWh", "cpu": 0.00012582436187222188, "ram": 0.0000052623983356353215, "gpu": 0, "total": 0.0001310867602078572 }, "efficiency": { "unit": "samples/kWh", "value": 76285.35470816075 } }
{ "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7184868769999753, 0.5956158320000213 ], "count": 2, "total": 1.3141027089999966, "mean": 0.6570513544999983, "p50": 0.6570513544999983, "p90": 0.7061997724999799, "p95": 0.7123433247499775, "p99": 0.7172581665499758, "stdev": 0.06143552249997697, "stdev_": 9.350185808037494 }, "throughput": { "unit": "samples/s", "value": 6.087804206786717 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 2841.86624, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.5611470650000001, 0.6053455429999985, 0.5643365999999901 ], "count": 3, "total": 1.7308292079999887, "mean": 0.5769430693333296, "p50": 0.5643365999999901, "p90": 0.5971437543999969, "p95": 0.6012446486999977, "p99": 0.6045253641399984, "stdev": 0.020125749115141166, "stdev_": 3.4883422966492885 }, "throughput": { "unit": "samples/s", "value": 10.399639615973083 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2826.752, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 2.882509665999976, "mean": 0.5765019331999952, "stdev": 0.04978939696949581, "p50": 0.5569985249999831, "p90": 0.6300386333999881, "p95": 0.6525100941999881, "p99": 0.670487262839988, "values": [ 0.674981554999988, 0.5569985249999831, 0.5424031590000027, 0.5626242509999884, 0.5455021760000136 ] }, "throughput": { "unit": "samples/s", "value": 17.34599560576128 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 2826.752, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 1.2319800799999712, "mean": 0.6159900399999856, "stdev": 0.05899151500000244, "p50": 0.6159900399999856, "p90": 0.6631832519999875, "p95": 0.6690824034999878, "p99": 0.673801724699988, "values": [ 0.674981554999988, 0.5569985249999831 ] }, "throughput": { "unit": "samples/s", "value": 6.493611487614465 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2826.752, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 1.6505295860000047, "mean": 0.5501765286666682, "stdev": 0.00889233078021606, "p50": 0.5455021760000136, "p90": 0.5591998359999935, "p95": 0.5609120434999909, "p99": 0.5622818094999888, "values": [ 0.5424031590000027, 0.5626242509999884, 0.5455021760000136 ] }, "throughput": { "unit": "samples/s", "value": 10.905590637500968 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "model": "hf-internal-testing/tiny-random-LlamaForCausalLM", "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM", "task": "text-generation", "library": "transformers", "model_type": "llama", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "7cec62e016d76fe612308e4c2c074fc7f09289fd", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.1", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.019453390999984776, 0.012075559000010117, 0.011965802999981179, 0.011934053999993921, 0.011987785000002305 ], "count": 5, "total": 0.0674165919999723, "mean": 0.013483318399994459, "p50": 0.011987785000002305, "p90": 0.016502258199994912, "p95": 0.017977824599989842, "p99": 0.01915827771998579, "stdev": 0.0029854060594256856, "stdev_": 22.141478609797662 }, "throughput": { "unit": "samples/s", "value": 741.6571872992415 }, "energy": { "unit": "kWh", "cpu": 0.000008703254777777798, "ram": 3.635053251718738e-7, "gpu": 0, "total": 0.000009066760102949672 }, "efficiency": { "unit": "samples/kWh", "value": 1102929.8102578802 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.019453390999984776, 0.012075559000010117 ], "count": 2, "total": 0.03152894999999489, "mean": 0.015764474999997447, "p50": 0.015764474999997447, "p90": 0.01871560779998731, "p95": 0.019084499399986042, "p99": 0.01937961267998503, "stdev": 0.00368891599998733, "stdev_": 23.400183006334988 }, "throughput": { "unit": "samples/s", "value": 253.73505936611576 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.011965802999981179, 0.011934053999993921, 0.011987785000002305 ], "count": 3, "total": 0.035887641999977404, "mean": 0.011962547333325801, "p50": 0.011965802999981179, "p90": 0.01198338859999808, "p95": 0.011985586800000192, "p99": 0.011987345360001882, "stdev": 0.000022056059099537384, "stdev_": 0.184375940048259 }, "throughput": { "unit": "samples/s", "value": 501.56541352065796 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "model": "hf-internal-testing/tiny-random-LlamaForCausalLM", "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM", "task": "text-generation", "library": "transformers", "model_type": "llama", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "7cec62e016d76fe612308e4c2c074fc7f09289fd", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.1", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.019453390999984776, 0.012075559000010117, 0.011965802999981179, 0.011934053999993921, 0.011987785000002305 ], "count": 5, "total": 0.0674165919999723, "mean": 0.013483318399994459, "p50": 0.011987785000002305, "p90": 0.016502258199994912, "p95": 0.017977824599989842, "p99": 0.01915827771998579, "stdev": 0.0029854060594256856, "stdev_": 22.141478609797662 }, "throughput": { "unit": "samples/s", "value": 741.6571872992415 }, "energy": { "unit": "kWh", "cpu": 0.000008703254777777798, "ram": 3.635053251718738e-7, "gpu": 0, "total": 0.000009066760102949672 }, "efficiency": { "unit": "samples/kWh", "value": 1102929.8102578802 } }
{ "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.019453390999984776, 0.012075559000010117 ], "count": 2, "total": 0.03152894999999489, "mean": 0.015764474999997447, "p50": 0.015764474999997447, "p90": 0.01871560779998731, "p95": 0.019084499399986042, "p99": 0.01937961267998503, "stdev": 0.00368891599998733, "stdev_": 23.400183006334988 }, "throughput": { "unit": "samples/s", "value": 253.73505936611576 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 522.866688, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.011965802999981179, 0.011934053999993921, 0.011987785000002305 ], "count": 3, "total": 0.035887641999977404, "mean": 0.011962547333325801, "p50": 0.011965802999981179, "p90": 0.01198338859999808, "p95": 0.011985586800000192, "p99": 0.011987345360001882, "stdev": 0.000022056059099537384, "stdev_": 0.184375940048259 }, "throughput": { "unit": "samples/s", "value": 501.56541352065796 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7413827390000165, 0.6272131300000012, 0.6310885370000108, 0.6182816500000001, 0.6157634749999943 ], "count": 5, "total": 3.233729531000023, "mean": 0.6467459062000046, "p50": 0.6272131300000012, "p90": 0.6972650582000142, "p95": 0.7193238986000153, "p99": 0.7369709709200163, "stdev": 0.0476506822676121, "stdev_": 7.367759395275746 }, "throughput": { "unit": "samples/s", "value": 15.462022881220255 }, "energy": { "unit": "kWh", "cpu": 0.0001328167573444445, "ram": 0.000005554786473157777, "gpu": 0, "total": 0.00013837154381760226 }, "efficiency": { "unit": "samples/kWh", "value": 72269.19440301784 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7413827390000165, 0.6272131300000012 ], "count": 2, "total": 1.3685958690000177, "mean": 0.6842979345000089, "p50": 0.6842979345000089, "p90": 0.729965778100015, "p95": 0.7356742585500158, "p99": 0.7402410429100164, "stdev": 0.057084804500007635, "stdev_": 8.342098028063958 }, "throughput": { "unit": "samples/s", "value": 5.845407092924592 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6310885370000108, 0.6182816500000001, 0.6157634749999943 ], "count": 3, "total": 1.8651336620000052, "mean": 0.6217112206666684, "p50": 0.6182816500000001, "p90": 0.6285271596000086, "p95": 0.6298078483000097, "p99": 0.6308323992600106, "stdev": 0.006709984974863501, "stdev_": 1.0792768011599185 }, "throughput": { "unit": "samples/s", "value": 9.650782872418047 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_text-generation_openai-community/gpt2
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7413827390000165, 0.6272131300000012, 0.6310885370000108, 0.6182816500000001, 0.6157634749999943 ], "count": 5, "total": 3.233729531000023, "mean": 0.6467459062000046, "p50": 0.6272131300000012, "p90": 0.6972650582000142, "p95": 0.7193238986000153, "p99": 0.7369709709200163, "stdev": 0.0476506822676121, "stdev_": 7.367759395275746 }, "throughput": { "unit": "samples/s", "value": 15.462022881220255 }, "energy": { "unit": "kWh", "cpu": 0.0001328167573444445, "ram": 0.000005554786473157777, "gpu": 0, "total": 0.00013837154381760226 }, "efficiency": { "unit": "samples/kWh", "value": 72269.19440301784 } }
{ "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.7413827390000165, 0.6272131300000012 ], "count": 2, "total": 1.3685958690000177, "mean": 0.6842979345000089, "p50": 0.6842979345000089, "p90": 0.729965778100015, "p95": 0.7356742585500158, "p99": 0.7402410429100164, "stdev": 0.057084804500007635, "stdev_": 8.342098028063958 }, "throughput": { "unit": "samples/s", "value": 5.845407092924592 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 2837.745664, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.6310885370000108, 0.6182816500000001, 0.6157634749999943 ], "count": 3, "total": 1.8651336620000052, "mean": 0.6217112206666684, "p50": 0.6182816500000001, "p90": 0.6285271596000086, "p95": 0.6298078483000097, "p99": 0.6308323992600106, "stdev": 0.006709984974863501, "stdev_": 1.0792768011599185 }, "throughput": { "unit": "samples/s", "value": 9.650782872418047 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "model": "openai-community/gpt2", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 2827.354112, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 3.1791685380000274, "mean": 0.6358337076000055, "stdev": 0.07846941233493662, "p50": 0.596941285000014, "p90": 0.7161873142000047, "p95": 0.7544328206000045, "p99": 0.7850292257200044, "values": [ 0.7926783270000044, 0.6014507950000052, 0.596941285000014, 0.593667699000008, 0.5944304319999958 ] }, "throughput": { "unit": "samples/s", "value": 15.727382616668173 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 2827.354112, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 1.3941291220000096, "mean": 0.6970645610000048, "stdev": 0.0956137659999996, "p50": 0.6970645610000048, "p90": 0.7735555738000045, "p95": 0.7831169504000044, "p99": 0.7907660516800044, "values": [ 0.7926783270000044, 0.6014507950000052 ] }, "throughput": { "unit": "samples/s", "value": 5.738349392288174 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 2827.354112, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 1.7850394160000178, "mean": 0.5950131386666726, "stdev": 0.0013985114990352936, "p50": 0.5944304319999958, "p90": 0.5964391144000103, "p95": 0.5966901997000121, "p99": 0.5968910679400136, "values": [ 0.596941285000014, 0.593667699000008, 0.5944304319999958 ] }, "throughput": { "unit": "samples/s", "value": 10.083810944822195 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.3893911569999773, 1.1545377269999904, 1.156441610999991, 1.1526474670000084, 1.1748442610000325 ], "count": 5, "total": 6.027862223, "mean": 1.2055724446, "p50": 1.156441610999991, "p90": 1.3035723985999994, "p95": 1.3464817777999882, "p99": 1.3808092811599795, "stdev": 0.09225288587403972, "stdev_": 7.652205911578259 }, "throughput": { "unit": "samples/s", "value": 8.294814670650444 }, "energy": { "unit": "kWh", "cpu": 0.00024157405844444467, "ram": 0.00001010374289347771, "gpu": 0, "total": 0.00025167780133792235 }, "efficiency": { "unit": "samples/kWh", "value": 39733.34138664544 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.3893911569999773, 1.1545377269999904 ], "count": 2, "total": 2.5439288839999676, "mean": 1.2719644419999838, "p50": 1.2719644419999838, "p90": 1.3659058139999787, "p95": 1.377648485499978, "p99": 1.3870426226999775, "stdev": 0.11742671499999346, "stdev_": 9.231918057029691 }, "throughput": { "unit": "samples/s", "value": 3.1447419974339588 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.156441610999991, 1.1526474670000084, 1.1748442610000325 ], "count": 3, "total": 3.483933339000032, "mean": 1.1613111130000107, "p50": 1.156441610999991, "p90": 1.1711637310000242, "p95": 1.1730039960000282, "p99": 1.1744762080000317, "stdev": 0.009693931203456517, "stdev_": 0.834740242725674 }, "throughput": { "unit": "samples/s", "value": 5.16657417020683 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cpu_training_transformers_token-classification_microsoft/deberta-v3-base
{ "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16766.779392, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.8.0-1017-azure-x86_64-with-glibc2.39", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "12b33198089c125596a4363f88200dc72e18e532", "transformers_version": "4.47.0", "transformers_commit": null, "accelerate_version": "1.2.0", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.12", "timm_commit": null, "peft_version": null, "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.3893911569999773, 1.1545377269999904, 1.156441610999991, 1.1526474670000084, 1.1748442610000325 ], "count": 5, "total": 6.027862223, "mean": 1.2055724446, "p50": 1.156441610999991, "p90": 1.3035723985999994, "p95": 1.3464817777999882, "p99": 1.3808092811599795, "stdev": 0.09225288587403972, "stdev_": 7.652205911578259 }, "throughput": { "unit": "samples/s", "value": 8.294814670650444 }, "energy": { "unit": "kWh", "cpu": 0.00024157405844444467, "ram": 0.00001010374289347771, "gpu": 0, "total": 0.00025167780133792235 }, "efficiency": { "unit": "samples/kWh", "value": 39733.34138664544 } }
{ "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.3893911569999773, 1.1545377269999904 ], "count": 2, "total": 2.5439288839999676, "mean": 1.2719644419999838, "p50": 1.2719644419999838, "p90": 1.3659058139999787, "p95": 1.377648485499978, "p99": 1.3870426226999775, "stdev": 0.11742671499999346, "stdev_": 9.231918057029691 }, "throughput": { "unit": "samples/s", "value": 3.1447419974339588 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 4362.715136, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.156441610999991, 1.1526474670000084, 1.1748442610000325 ], "count": 3, "total": 3.483933339000032, "mean": 1.1613111130000107, "p50": 1.156441610999991, "p90": 1.1711637310000242, "p95": 1.1730039960000282, "p99": 1.1744762080000317, "stdev": 0.009693931203456517, "stdev_": 0.834740242725674 }, "throughput": { "unit": "samples/s", "value": 5.16657417020683 }, "energy": null, "efficiency": null }
{ "name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.3.0+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "model": "microsoft/deberta-v3-base", "library": "transformers", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.346304, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8", "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 4374.970368, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 5, "total": 5.583659903000068, "mean": 1.1167319806000138, "stdev": 0.07853258776030796, "p50": 1.0731834320000075, "p90": 1.2025418994000006, "p95": 1.2374836681999908, "p99": 1.265437083239983, "values": [ 1.2724254369999812, 1.0731834320000075, 1.0977165930000297, 1.0711078369999996, 1.0692266040000504 ] }, "throughput": { "unit": "samples/s", "value": 8.954700119385008 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 4374.970368, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 2, "total": 2.3456088689999888, "mean": 1.1728044344999944, "stdev": 0.09962100249998684, "p50": 1.1728044344999944, "p90": 1.252501236499984, "p95": 1.2624633367499825, "p99": 1.2704330169499816, "values": [ 1.2724254369999812, 1.0731834320000075 ] }, "throughput": { "unit": "samples/s", "value": 3.4106283045436583 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 4374.970368, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 3, "total": 3.2380510340000797, "mean": 1.0793503446666932, "stdev": 0.01300958794585394, "p50": 1.0711078369999996, "p90": 1.0923948418000236, "p95": 1.0950557174000266, "p99": 1.097184417880029, "values": [ 1.0977165930000297, 1.0711078369999996, 1.0692266040000504 ] }, "throughput": { "unit": "samples/s", "value": 5.558899415419021 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null