hf-transformers-bot commited on
Commit
9d42ac7
1 Parent(s): 4941993

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/multirun.yaml +234 -0
  31. 2024-09-13/summaries.json +41 -0
  32. 2024-09-13/summary.json +40 -0
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: a05ce550bf1cda8ae3bffcc62607597770460a42
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42
88
+ - hydra.run.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1333.182464,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.3963779296875,
120
+ "mean": 11.3963779296875,
121
+ "stdev": 0.0,
122
+ "p50": 11.3963779296875,
123
+ "p90": 11.3963779296875,
124
+ "p95": 11.3963779296875,
125
+ "p99": 11.3963779296875,
126
+ "values": [
127
+ 11.3963779296875
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1800.712192,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.04140790367126465,
147
+ "mean": 0.020703951835632325,
148
+ "stdev": 5.799961090087978e-05,
149
+ "p50": 0.020703951835632325,
150
+ "p90": 0.020750351524353026,
151
+ "p95": 0.020756151485443117,
152
+ "p99": 0.020760791454315187,
153
+ "values": [
154
+ 0.020645952224731444,
155
+ 0.020761951446533203
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 338.09970461546
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1859.125248,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.053552978515625,
178
+ "mean": 2.5267764892578124,
179
+ "stdev": 0.012900024414062461,
180
+ "p50": 2.5267764892578124,
181
+ "p90": 2.5370965087890625,
182
+ "p95": 2.5383865112304687,
183
+ "p99": 2.539418513183594,
184
+ "values": [
185
+ 2.51387646484375,
186
+ 2.539676513671875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 50.26166759898244
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.0532259902954095,
202
+ "mean": 0.019894590512974056,
203
+ "stdev": 0.0003843963811459021,
204
+ "p50": 0.02006169605255127,
205
+ "p90": 0.02019676170349121,
206
+ "p95": 0.02027335681915283,
207
+ "p99": 0.020557331657409667,
208
+ "values": [
209
+ 0.019119104385375976,
210
+ 0.019130367279052735,
211
+ 0.019183616638183593,
212
+ 0.01921433639526367,
213
+ 0.019184640884399414,
214
+ 0.01919795227050781,
215
+ 0.019215360641479492,
216
+ 0.019146751403808594,
217
+ 0.01918156814575195,
218
+ 0.01921331214904785,
219
+ 0.019180543899536134,
220
+ 0.019158016204833983,
221
+ 0.019144704818725586,
222
+ 0.019190784454345702,
223
+ 0.019207168579101562,
224
+ 0.019188735961914064,
225
+ 0.019203071594238282,
226
+ 0.01921023941040039,
227
+ 0.01923072052001953,
228
+ 0.01923276710510254,
229
+ 0.01921843147277832,
230
+ 0.019269632339477538,
231
+ 0.01924198341369629,
232
+ 0.019186687469482423,
233
+ 0.01922559928894043,
234
+ 0.019166208267211913,
235
+ 0.019215360641479492,
236
+ 0.019183616638183593,
237
+ 0.019186687469482423,
238
+ 0.01922867202758789,
239
+ 0.019184640884399414,
240
+ 0.01924095916748047,
241
+ 0.01913555145263672,
242
+ 0.019239871978759766,
243
+ 0.01922559928894043,
244
+ 0.019180543899536134,
245
+ 0.019176448822021484,
246
+ 0.019167232513427734,
247
+ 0.019176448822021484,
248
+ 0.0192993278503418,
249
+ 0.01964339256286621,
250
+ 0.019487743377685548,
251
+ 0.019679231643676756,
252
+ 0.019330047607421876,
253
+ 0.019149824142456053,
254
+ 0.019152896881103516,
255
+ 0.019183616638183593,
256
+ 0.01975503921508789,
257
+ 0.019796960830688475,
258
+ 0.019758079528808595,
259
+ 0.019578880310058593,
260
+ 0.020195327758789062,
261
+ 0.020588544845581053,
262
+ 0.020090879440307616,
263
+ 0.019920896530151368,
264
+ 0.019997695922851562,
265
+ 0.02000588798522949,
266
+ 0.020110336303710938,
267
+ 0.02005504035949707,
268
+ 0.02021683120727539,
269
+ 0.020158464431762696,
270
+ 0.020090879440307616,
271
+ 0.020154367446899413,
272
+ 0.020121599197387697,
273
+ 0.02036534309387207,
274
+ 0.020828128814697266,
275
+ 0.020256767272949217,
276
+ 0.01987174415588379,
277
+ 0.019903488159179687,
278
+ 0.020076543807983398,
279
+ 0.020157440185546875,
280
+ 0.02012876892089844,
281
+ 0.0200949764251709,
282
+ 0.020184064865112306,
283
+ 0.020155391693115234,
284
+ 0.02017791938781738,
285
+ 0.02011238479614258,
286
+ 0.020134912490844727,
287
+ 0.020106239318847655,
288
+ 0.020161535263061522,
289
+ 0.02020249557495117,
290
+ 0.020549631118774413,
291
+ 0.020100095748901366,
292
+ 0.02004275131225586,
293
+ 0.02006937599182129,
294
+ 0.020198400497436524,
295
+ 0.020142080307006836,
296
+ 0.020161535263061522,
297
+ 0.020146175384521483,
298
+ 0.020124671936035156,
299
+ 0.020197376251220703,
300
+ 0.020133888244628906,
301
+ 0.02024550437927246,
302
+ 0.020152320861816408,
303
+ 0.020110336303710938,
304
+ 0.02003046417236328,
305
+ 0.020114431381225584,
306
+ 0.020090879440307616,
307
+ 0.02011238479614258,
308
+ 0.02026905632019043,
309
+ 0.020131839752197265,
310
+ 0.020107263565063475,
311
+ 0.020155391693115234,
312
+ 0.020102144241333008,
313
+ 0.020145151138305666,
314
+ 0.020189184188842774,
315
+ 0.020125696182250977,
316
+ 0.02008780860900879,
317
+ 0.020167680740356447,
318
+ 0.020281343460083007,
319
+ 0.020144128799438478,
320
+ 0.020164608001708984,
321
+ 0.020124671936035156,
322
+ 0.020082687377929686,
323
+ 0.020123647689819335,
324
+ 0.020175872802734376,
325
+ 0.020143104553222657,
326
+ 0.020173824310302735,
327
+ 0.02006630325317383,
328
+ 0.020101119995117187,
329
+ 0.020109312057495117,
330
+ 0.020141056060791016,
331
+ 0.020150272369384766,
332
+ 0.020134912490844727,
333
+ 0.020102144241333008,
334
+ 0.020129791259765627,
335
+ 0.020131839752197265,
336
+ 0.01924095916748047,
337
+ 0.019896320343017578,
338
+ 0.019543039321899415,
339
+ 0.01940787124633789,
340
+ 0.019959808349609375,
341
+ 0.01992192077636719,
342
+ 0.019975168228149414,
343
+ 0.02002943992614746,
344
+ 0.019928064346313477,
345
+ 0.019991552352905274,
346
+ 0.019952640533447266,
347
+ 0.01991372871398926,
348
+ 0.02002739143371582,
349
+ 0.01998847961425781,
350
+ 0.020195327758789062,
351
+ 0.020067327499389647,
352
+ 0.019944448471069336,
353
+ 0.0202608642578125,
354
+ 0.02005401611328125,
355
+ 0.019990528106689453,
356
+ 0.019985408782958985,
357
+ 0.01998028755187988,
358
+ 0.019915775299072267,
359
+ 0.019934207916259765,
360
+ 0.020008960723876954,
361
+ 0.02008576011657715,
362
+ 0.01961881637573242,
363
+ 0.019957759857177734,
364
+ 0.01990553665161133,
365
+ 0.020033536911010744,
366
+ 0.020117504119873047,
367
+ 0.019853311538696287,
368
+ 0.01968435287475586,
369
+ 0.01963212776184082,
370
+ 0.019679231643676756,
371
+ 0.020162559509277343,
372
+ 0.019949567794799804,
373
+ 0.019976192474365235,
374
+ 0.019942399978637695,
375
+ 0.01999564743041992,
376
+ 0.019955711364746095,
377
+ 0.019968000411987305,
378
+ 0.019955711364746095,
379
+ 0.019994623184204103,
380
+ 0.02004582405090332,
381
+ 0.01992192077636719,
382
+ 0.020037696838378905,
383
+ 0.020002752304077148,
384
+ 0.019893312454223634,
385
+ 0.020054975509643556,
386
+ 0.020015104293823242,
387
+ 0.020460544586181642,
388
+ 0.020347904205322266,
389
+ 0.02007244873046875,
390
+ 0.020039680480957032,
391
+ 0.019945472717285157,
392
+ 0.019977216720581056,
393
+ 0.02004275131225586,
394
+ 0.02007142448425293,
395
+ 0.020205568313598633,
396
+ 0.020108287811279296,
397
+ 0.020110336303710938,
398
+ 0.020107263565063475,
399
+ 0.020114431381225584,
400
+ 0.02046771240234375,
401
+ 0.020141056060791016,
402
+ 0.02017791938781738,
403
+ 0.020107263565063475,
404
+ 0.020130815505981444,
405
+ 0.020090879440307616,
406
+ 0.020135936737060548,
407
+ 0.019329023361206055,
408
+ 0.019272703170776367,
409
+ 0.019725311279296876,
410
+ 0.02006937599182129,
411
+ 0.020115455627441405,
412
+ 0.02020147132873535,
413
+ 0.02009702491760254,
414
+ 0.01940787124633789,
415
+ 0.019679231643676756,
416
+ 0.02008576011657715,
417
+ 0.020313087463378905,
418
+ 0.020264959335327147,
419
+ 0.020155391693115234,
420
+ 0.02020147132873535,
421
+ 0.020146175384521483,
422
+ 0.019312639236450196,
423
+ 0.01927884864807129,
424
+ 0.020031488418579102,
425
+ 0.020159488677978517,
426
+ 0.02005606460571289,
427
+ 0.020150272369384766,
428
+ 0.02008883285522461,
429
+ 0.02006118392944336,
430
+ 0.0200949764251709,
431
+ 0.020023296356201172,
432
+ 0.02006220817565918,
433
+ 0.020121599197387697,
434
+ 0.02006220817565918,
435
+ 0.02024345588684082,
436
+ 0.020109312057495117,
437
+ 0.02043391990661621,
438
+ 0.020090879440307616,
439
+ 0.019363840103149413,
440
+ 0.019595264434814453,
441
+ 0.020121599197387697,
442
+ 0.020125696182250977,
443
+ 0.02011955261230469,
444
+ 0.020150272369384766,
445
+ 0.020566015243530272,
446
+ 0.020142080307006836,
447
+ 0.02011955261230469,
448
+ 0.019311616897583008,
449
+ 0.019339263916015623,
450
+ 0.020011007308959963,
451
+ 0.020405248641967775,
452
+ 0.020107263565063475,
453
+ 0.020379648208618165,
454
+ 0.020174848556518556,
455
+ 0.020025344848632814,
456
+ 0.020106239318847655,
457
+ 0.02008166313171387,
458
+ 0.020108287811279296,
459
+ 0.020142080307006836,
460
+ 0.020057088851928712,
461
+ 0.02001408004760742,
462
+ 0.02009702491760254
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 50.26491997147969
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1333.182464,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.3963779296875,
15
+ "mean": 11.3963779296875,
16
+ "stdev": 0.0,
17
+ "p50": 11.3963779296875,
18
+ "p90": 11.3963779296875,
19
+ "p95": 11.3963779296875,
20
+ "p99": 11.3963779296875,
21
+ "values": [
22
+ 11.3963779296875
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1800.712192,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.04140790367126465,
42
+ "mean": 0.020703951835632325,
43
+ "stdev": 5.799961090087978e-05,
44
+ "p50": 0.020703951835632325,
45
+ "p90": 0.020750351524353026,
46
+ "p95": 0.020756151485443117,
47
+ "p99": 0.020760791454315187,
48
+ "values": [
49
+ 0.020645952224731444,
50
+ 0.020761951446533203
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 338.09970461546
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1859.125248,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.053552978515625,
73
+ "mean": 2.5267764892578124,
74
+ "stdev": 0.012900024414062461,
75
+ "p50": 2.5267764892578124,
76
+ "p90": 2.5370965087890625,
77
+ "p95": 2.5383865112304687,
78
+ "p99": 2.539418513183594,
79
+ "values": [
80
+ 2.51387646484375,
81
+ 2.539676513671875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 50.26166759898244
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.0532259902954095,
97
+ "mean": 0.019894590512974056,
98
+ "stdev": 0.0003843963811459021,
99
+ "p50": 0.02006169605255127,
100
+ "p90": 0.02019676170349121,
101
+ "p95": 0.02027335681915283,
102
+ "p99": 0.020557331657409667,
103
+ "values": [
104
+ 0.019119104385375976,
105
+ 0.019130367279052735,
106
+ 0.019183616638183593,
107
+ 0.01921433639526367,
108
+ 0.019184640884399414,
109
+ 0.01919795227050781,
110
+ 0.019215360641479492,
111
+ 0.019146751403808594,
112
+ 0.01918156814575195,
113
+ 0.01921331214904785,
114
+ 0.019180543899536134,
115
+ 0.019158016204833983,
116
+ 0.019144704818725586,
117
+ 0.019190784454345702,
118
+ 0.019207168579101562,
119
+ 0.019188735961914064,
120
+ 0.019203071594238282,
121
+ 0.01921023941040039,
122
+ 0.01923072052001953,
123
+ 0.01923276710510254,
124
+ 0.01921843147277832,
125
+ 0.019269632339477538,
126
+ 0.01924198341369629,
127
+ 0.019186687469482423,
128
+ 0.01922559928894043,
129
+ 0.019166208267211913,
130
+ 0.019215360641479492,
131
+ 0.019183616638183593,
132
+ 0.019186687469482423,
133
+ 0.01922867202758789,
134
+ 0.019184640884399414,
135
+ 0.01924095916748047,
136
+ 0.01913555145263672,
137
+ 0.019239871978759766,
138
+ 0.01922559928894043,
139
+ 0.019180543899536134,
140
+ 0.019176448822021484,
141
+ 0.019167232513427734,
142
+ 0.019176448822021484,
143
+ 0.0192993278503418,
144
+ 0.01964339256286621,
145
+ 0.019487743377685548,
146
+ 0.019679231643676756,
147
+ 0.019330047607421876,
148
+ 0.019149824142456053,
149
+ 0.019152896881103516,
150
+ 0.019183616638183593,
151
+ 0.01975503921508789,
152
+ 0.019796960830688475,
153
+ 0.019758079528808595,
154
+ 0.019578880310058593,
155
+ 0.020195327758789062,
156
+ 0.020588544845581053,
157
+ 0.020090879440307616,
158
+ 0.019920896530151368,
159
+ 0.019997695922851562,
160
+ 0.02000588798522949,
161
+ 0.020110336303710938,
162
+ 0.02005504035949707,
163
+ 0.02021683120727539,
164
+ 0.020158464431762696,
165
+ 0.020090879440307616,
166
+ 0.020154367446899413,
167
+ 0.020121599197387697,
168
+ 0.02036534309387207,
169
+ 0.020828128814697266,
170
+ 0.020256767272949217,
171
+ 0.01987174415588379,
172
+ 0.019903488159179687,
173
+ 0.020076543807983398,
174
+ 0.020157440185546875,
175
+ 0.02012876892089844,
176
+ 0.0200949764251709,
177
+ 0.020184064865112306,
178
+ 0.020155391693115234,
179
+ 0.02017791938781738,
180
+ 0.02011238479614258,
181
+ 0.020134912490844727,
182
+ 0.020106239318847655,
183
+ 0.020161535263061522,
184
+ 0.02020249557495117,
185
+ 0.020549631118774413,
186
+ 0.020100095748901366,
187
+ 0.02004275131225586,
188
+ 0.02006937599182129,
189
+ 0.020198400497436524,
190
+ 0.020142080307006836,
191
+ 0.020161535263061522,
192
+ 0.020146175384521483,
193
+ 0.020124671936035156,
194
+ 0.020197376251220703,
195
+ 0.020133888244628906,
196
+ 0.02024550437927246,
197
+ 0.020152320861816408,
198
+ 0.020110336303710938,
199
+ 0.02003046417236328,
200
+ 0.020114431381225584,
201
+ 0.020090879440307616,
202
+ 0.02011238479614258,
203
+ 0.02026905632019043,
204
+ 0.020131839752197265,
205
+ 0.020107263565063475,
206
+ 0.020155391693115234,
207
+ 0.020102144241333008,
208
+ 0.020145151138305666,
209
+ 0.020189184188842774,
210
+ 0.020125696182250977,
211
+ 0.02008780860900879,
212
+ 0.020167680740356447,
213
+ 0.020281343460083007,
214
+ 0.020144128799438478,
215
+ 0.020164608001708984,
216
+ 0.020124671936035156,
217
+ 0.020082687377929686,
218
+ 0.020123647689819335,
219
+ 0.020175872802734376,
220
+ 0.020143104553222657,
221
+ 0.020173824310302735,
222
+ 0.02006630325317383,
223
+ 0.020101119995117187,
224
+ 0.020109312057495117,
225
+ 0.020141056060791016,
226
+ 0.020150272369384766,
227
+ 0.020134912490844727,
228
+ 0.020102144241333008,
229
+ 0.020129791259765627,
230
+ 0.020131839752197265,
231
+ 0.01924095916748047,
232
+ 0.019896320343017578,
233
+ 0.019543039321899415,
234
+ 0.01940787124633789,
235
+ 0.019959808349609375,
236
+ 0.01992192077636719,
237
+ 0.019975168228149414,
238
+ 0.02002943992614746,
239
+ 0.019928064346313477,
240
+ 0.019991552352905274,
241
+ 0.019952640533447266,
242
+ 0.01991372871398926,
243
+ 0.02002739143371582,
244
+ 0.01998847961425781,
245
+ 0.020195327758789062,
246
+ 0.020067327499389647,
247
+ 0.019944448471069336,
248
+ 0.0202608642578125,
249
+ 0.02005401611328125,
250
+ 0.019990528106689453,
251
+ 0.019985408782958985,
252
+ 0.01998028755187988,
253
+ 0.019915775299072267,
254
+ 0.019934207916259765,
255
+ 0.020008960723876954,
256
+ 0.02008576011657715,
257
+ 0.01961881637573242,
258
+ 0.019957759857177734,
259
+ 0.01990553665161133,
260
+ 0.020033536911010744,
261
+ 0.020117504119873047,
262
+ 0.019853311538696287,
263
+ 0.01968435287475586,
264
+ 0.01963212776184082,
265
+ 0.019679231643676756,
266
+ 0.020162559509277343,
267
+ 0.019949567794799804,
268
+ 0.019976192474365235,
269
+ 0.019942399978637695,
270
+ 0.01999564743041992,
271
+ 0.019955711364746095,
272
+ 0.019968000411987305,
273
+ 0.019955711364746095,
274
+ 0.019994623184204103,
275
+ 0.02004582405090332,
276
+ 0.01992192077636719,
277
+ 0.020037696838378905,
278
+ 0.020002752304077148,
279
+ 0.019893312454223634,
280
+ 0.020054975509643556,
281
+ 0.020015104293823242,
282
+ 0.020460544586181642,
283
+ 0.020347904205322266,
284
+ 0.02007244873046875,
285
+ 0.020039680480957032,
286
+ 0.019945472717285157,
287
+ 0.019977216720581056,
288
+ 0.02004275131225586,
289
+ 0.02007142448425293,
290
+ 0.020205568313598633,
291
+ 0.020108287811279296,
292
+ 0.020110336303710938,
293
+ 0.020107263565063475,
294
+ 0.020114431381225584,
295
+ 0.02046771240234375,
296
+ 0.020141056060791016,
297
+ 0.02017791938781738,
298
+ 0.020107263565063475,
299
+ 0.020130815505981444,
300
+ 0.020090879440307616,
301
+ 0.020135936737060548,
302
+ 0.019329023361206055,
303
+ 0.019272703170776367,
304
+ 0.019725311279296876,
305
+ 0.02006937599182129,
306
+ 0.020115455627441405,
307
+ 0.02020147132873535,
308
+ 0.02009702491760254,
309
+ 0.01940787124633789,
310
+ 0.019679231643676756,
311
+ 0.02008576011657715,
312
+ 0.020313087463378905,
313
+ 0.020264959335327147,
314
+ 0.020155391693115234,
315
+ 0.02020147132873535,
316
+ 0.020146175384521483,
317
+ 0.019312639236450196,
318
+ 0.01927884864807129,
319
+ 0.020031488418579102,
320
+ 0.020159488677978517,
321
+ 0.02005606460571289,
322
+ 0.020150272369384766,
323
+ 0.02008883285522461,
324
+ 0.02006118392944336,
325
+ 0.0200949764251709,
326
+ 0.020023296356201172,
327
+ 0.02006220817565918,
328
+ 0.020121599197387697,
329
+ 0.02006220817565918,
330
+ 0.02024345588684082,
331
+ 0.020109312057495117,
332
+ 0.02043391990661621,
333
+ 0.020090879440307616,
334
+ 0.019363840103149413,
335
+ 0.019595264434814453,
336
+ 0.020121599197387697,
337
+ 0.020125696182250977,
338
+ 0.02011955261230469,
339
+ 0.020150272369384766,
340
+ 0.020566015243530272,
341
+ 0.020142080307006836,
342
+ 0.02011955261230469,
343
+ 0.019311616897583008,
344
+ 0.019339263916015623,
345
+ 0.020011007308959963,
346
+ 0.020405248641967775,
347
+ 0.020107263565063475,
348
+ 0.020379648208618165,
349
+ 0.020174848556518556,
350
+ 0.020025344848632814,
351
+ 0.020106239318847655,
352
+ 0.02008166313171387,
353
+ 0.020108287811279296,
354
+ 0.020142080307006836,
355
+ 0.020057088851928712,
356
+ 0.02001408004760742,
357
+ 0.02009702491760254
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 50.26491997147969
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.020703951835632325,
7
+ "prefill.throughput.value": 338.09970461546,
8
+ "decode.latency.mean": 2.5267764892578124,
9
+ "decode.throughput.value": 50.26166759898244,
10
+ "per_token.latency.mean": 0.019894590512974056,
11
+ "per_token.throughput.value": 50.26491997147969
12
+ }
13
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: a05ce550bf1cda8ae3bffcc62607597770460a42
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42
88
+ - hydra.run.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-09-13 02:44:05,428][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-09-13 02:44:07,802][process][ERROR] - + Received traceback from isolated process
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: a05ce550bf1cda8ae3bffcc62607597770460a42
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42
88
+ - hydra.run.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1333.198848,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.347775390625,
120
+ "mean": 11.347775390625,
121
+ "stdev": 0.0,
122
+ "p50": 11.347775390625,
123
+ "p90": 11.347775390625,
124
+ "p95": 11.347775390625,
125
+ "p99": 11.347775390625,
126
+ "values": [
127
+ 11.347775390625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1817.04704,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.0448658561706543,
147
+ "mean": 0.02243292808532715,
148
+ "stdev": 5.2160263061522966e-05,
149
+ "p50": 0.02243292808532715,
150
+ "p90": 0.022474656295776368,
151
+ "p95": 0.02247987232208252,
152
+ "p99": 0.02248404514312744,
153
+ "values": [
154
+ 0.022380767822265626,
155
+ 0.022485088348388672
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 312.0412981031458
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1843.625984,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.290689208984375,
178
+ "mean": 2.6453446044921876,
179
+ "stdev": 0.01986877441406243,
180
+ "p50": 2.6453446044921876,
181
+ "p90": 2.6612396240234375,
182
+ "p95": 2.6632265014648437,
183
+ "p99": 2.6648160034179686,
184
+ "values": [
185
+ 2.625475830078125,
186
+ 2.66521337890625
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 48.0088680258652
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.2904243144989005,
202
+ "mean": 0.02082844218306654,
203
+ "stdev": 0.00032485950454247533,
204
+ "p50": 0.020652031898498534,
205
+ "p90": 0.021287628555297852,
206
+ "p95": 0.021401958274841308,
207
+ "p99": 0.02179714044570923,
208
+ "values": [
209
+ 0.02065100860595703,
210
+ 0.020617216110229493,
211
+ 0.02062233543395996,
212
+ 0.020591615676879883,
213
+ 0.02070425605773926,
214
+ 0.020956159591674805,
215
+ 0.021275648117065428,
216
+ 0.02083839988708496,
217
+ 0.02066329574584961,
218
+ 0.020594688415527345,
219
+ 0.02067148780822754,
220
+ 0.02100223922729492,
221
+ 0.02081996726989746,
222
+ 0.020578304290771485,
223
+ 0.020612096786499022,
224
+ 0.020584447860717774,
225
+ 0.020590591430664062,
226
+ 0.020579328536987306,
227
+ 0.020588544845581053,
228
+ 0.020585472106933594,
229
+ 0.020585472106933594,
230
+ 0.020576255798339844,
231
+ 0.020594688415527345,
232
+ 0.020603904724121092,
233
+ 0.020595712661743162,
234
+ 0.020603904724121092,
235
+ 0.020599807739257812,
236
+ 0.0206376953125,
237
+ 0.020634624481201173,
238
+ 0.02117433547973633,
239
+ 0.020584384918212892,
240
+ 0.0206059513092041,
241
+ 0.021173248291015623,
242
+ 0.020595712661743162,
243
+ 0.020579328536987306,
244
+ 0.02058956718444824,
245
+ 0.020553728103637696,
246
+ 0.020546592712402344,
247
+ 0.020596704483032226,
248
+ 0.0206059513092041,
249
+ 0.020583423614501953,
250
+ 0.02064588737487793,
251
+ 0.020582399368286132,
252
+ 0.020403200149536133,
253
+ 0.020398080825805662,
254
+ 0.0204769287109375,
255
+ 0.020585472106933594,
256
+ 0.020560895919799805,
257
+ 0.020600831985473633,
258
+ 0.020555776596069338,
259
+ 0.02065715217590332,
260
+ 0.02053222465515137,
261
+ 0.02061414337158203,
262
+ 0.020585472106933594,
263
+ 0.020632575988769532,
264
+ 0.020653055191040038,
265
+ 0.021574655532836915,
266
+ 0.020926464080810548,
267
+ 0.02085171127319336,
268
+ 0.02119987106323242,
269
+ 0.02060697555541992,
270
+ 0.02061414337158203,
271
+ 0.020593664169311524,
272
+ 0.020545536041259766,
273
+ 0.020632575988769532,
274
+ 0.02064691162109375,
275
+ 0.020563968658447264,
276
+ 0.020592639923095703,
277
+ 0.020594688415527345,
278
+ 0.02061414337158203,
279
+ 0.020585472106933594,
280
+ 0.020609024047851563,
281
+ 0.020587520599365236,
282
+ 0.020609024047851563,
283
+ 0.02065100860595703,
284
+ 0.0206059513092041,
285
+ 0.020599807739257812,
286
+ 0.020583423614501953,
287
+ 0.020599807739257812,
288
+ 0.020618240356445314,
289
+ 0.02065407943725586,
290
+ 0.02059775924682617,
291
+ 0.02061414337158203,
292
+ 0.02068079948425293,
293
+ 0.020649887084960936,
294
+ 0.020610048294067384,
295
+ 0.02063052749633789,
296
+ 0.020593664169311524,
297
+ 0.020562944412231447,
298
+ 0.02062745666503906,
299
+ 0.02064076805114746,
300
+ 0.020915199279785156,
301
+ 0.02124799919128418,
302
+ 0.02124185562133789,
303
+ 0.021165056228637694,
304
+ 0.020731903076171874,
305
+ 0.020632575988769532,
306
+ 0.02062950325012207,
307
+ 0.02061516761779785,
308
+ 0.020620288848876952,
309
+ 0.02065100860595703,
310
+ 0.020641792297363282,
311
+ 0.020692991256713866,
312
+ 0.020619264602661135,
313
+ 0.02063667106628418,
314
+ 0.02062950325012207,
315
+ 0.02124185562133789,
316
+ 0.021009408950805664,
317
+ 0.020644863128662108,
318
+ 0.020632575988769532,
319
+ 0.020599807739257812,
320
+ 0.020624383926391602,
321
+ 0.02064384078979492,
322
+ 0.020600831985473633,
323
+ 0.02061414337158203,
324
+ 0.020583423614501953,
325
+ 0.02061311912536621,
326
+ 0.020591615676879883,
327
+ 0.02063155174255371,
328
+ 0.02065715217590332,
329
+ 0.020617216110229493,
330
+ 0.02062950325012207,
331
+ 0.02065920066833496,
332
+ 0.020577280044555665,
333
+ 0.02062131118774414,
334
+ 0.020603904724121092,
335
+ 0.020576255798339844,
336
+ 0.020447231292724608,
337
+ 0.020446207046508787,
338
+ 0.02045952033996582,
339
+ 0.02063974380493164,
340
+ 0.02059878349304199,
341
+ 0.02061516761779785,
342
+ 0.020601856231689454,
343
+ 0.020616191864013672,
344
+ 0.020590591430664062,
345
+ 0.020591615676879883,
346
+ 0.020580352783203124,
347
+ 0.020569087982177735,
348
+ 0.020551679611206054,
349
+ 0.0206561279296875,
350
+ 0.020503551483154296,
351
+ 0.020410367965698242,
352
+ 0.020702207565307617,
353
+ 0.02205388832092285,
354
+ 0.020847616195678712,
355
+ 0.021394432067871092,
356
+ 0.020968448638916014,
357
+ 0.021308416366577147,
358
+ 0.021179391860961915,
359
+ 0.02122956848144531,
360
+ 0.02127462387084961,
361
+ 0.021415935516357423,
362
+ 0.020926464080810548,
363
+ 0.020684799194335936,
364
+ 0.021353471755981446,
365
+ 0.020954111099243163,
366
+ 0.020634624481201173,
367
+ 0.020785152435302736,
368
+ 0.021356544494628905,
369
+ 0.021259263992309572,
370
+ 0.021109760284423826,
371
+ 0.02062950325012207,
372
+ 0.020578304290771485,
373
+ 0.020619264602661135,
374
+ 0.0206059513092041,
375
+ 0.021145599365234375,
376
+ 0.021210111618041993,
377
+ 0.021226495742797852,
378
+ 0.02064384078979492,
379
+ 0.020603904724121092,
380
+ 0.02067353630065918,
381
+ 0.0206059513092041,
382
+ 0.021115903854370118,
383
+ 0.021203968048095705,
384
+ 0.021226495742797852,
385
+ 0.02066227149963379,
386
+ 0.021189632415771483,
387
+ 0.02143948745727539,
388
+ 0.021209087371826172,
389
+ 0.021234687805175782,
390
+ 0.02122444725036621,
391
+ 0.021621759414672852,
392
+ 0.021206016540527343,
393
+ 0.020666368484497072,
394
+ 0.02066022491455078,
395
+ 0.020974592208862306,
396
+ 0.02129408073425293,
397
+ 0.021149696350097655,
398
+ 0.02065510368347168,
399
+ 0.02064076805114746,
400
+ 0.020677631378173827,
401
+ 0.02063155174255371,
402
+ 0.020715520858764647,
403
+ 0.02085683250427246,
404
+ 0.021334016799926758,
405
+ 0.02145280075073242,
406
+ 0.02128486442565918,
407
+ 0.021819391250610352,
408
+ 0.02150809669494629,
409
+ 0.020967424392700194,
410
+ 0.020945920944213867,
411
+ 0.02089472007751465,
412
+ 0.020755456924438476,
413
+ 0.021032960891723632,
414
+ 0.02168217658996582,
415
+ 0.021373952865600586,
416
+ 0.020736000061035157,
417
+ 0.02066329574584961,
418
+ 0.020649984359741212,
419
+ 0.02104217529296875,
420
+ 0.021005311965942384,
421
+ 0.020706304550170897,
422
+ 0.021313535690307618,
423
+ 0.02123776054382324,
424
+ 0.021279743194580078,
425
+ 0.021287935256958008,
426
+ 0.021200895309448242,
427
+ 0.0212541446685791,
428
+ 0.020676607131958007,
429
+ 0.020787200927734374,
430
+ 0.02125209617614746,
431
+ 0.02066534423828125,
432
+ 0.02068172836303711,
433
+ 0.02090598487854004,
434
+ 0.02122854423522949,
435
+ 0.021337120056152344,
436
+ 0.02067555236816406,
437
+ 0.02070528030395508,
438
+ 0.02088755226135254,
439
+ 0.02128691291809082,
440
+ 0.020720640182495118,
441
+ 0.020699136734008788,
442
+ 0.020661247253417968,
443
+ 0.020978687286376953,
444
+ 0.02122547149658203,
445
+ 0.02062131118774414,
446
+ 0.02064588737487793,
447
+ 0.02106675148010254,
448
+ 0.02111692810058594,
449
+ 0.02083430480957031,
450
+ 0.02082918357849121,
451
+ 0.021217279434204102,
452
+ 0.021385215759277345,
453
+ 0.021271551132202148,
454
+ 0.02085273551940918,
455
+ 0.02066022491455078,
456
+ 0.021340160369873046,
457
+ 0.02203647994995117,
458
+ 0.021612543106079102,
459
+ 0.021777408599853516,
460
+ 0.021381120681762695,
461
+ 0.02144767951965332,
462
+ 0.02106265640258789
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 48.01127185656721
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1333.198848,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.347775390625,
15
+ "mean": 11.347775390625,
16
+ "stdev": 0.0,
17
+ "p50": 11.347775390625,
18
+ "p90": 11.347775390625,
19
+ "p95": 11.347775390625,
20
+ "p99": 11.347775390625,
21
+ "values": [
22
+ 11.347775390625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1817.04704,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.0448658561706543,
42
+ "mean": 0.02243292808532715,
43
+ "stdev": 5.2160263061522966e-05,
44
+ "p50": 0.02243292808532715,
45
+ "p90": 0.022474656295776368,
46
+ "p95": 0.02247987232208252,
47
+ "p99": 0.02248404514312744,
48
+ "values": [
49
+ 0.022380767822265626,
50
+ 0.022485088348388672
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 312.0412981031458
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1843.625984,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.290689208984375,
73
+ "mean": 2.6453446044921876,
74
+ "stdev": 0.01986877441406243,
75
+ "p50": 2.6453446044921876,
76
+ "p90": 2.6612396240234375,
77
+ "p95": 2.6632265014648437,
78
+ "p99": 2.6648160034179686,
79
+ "values": [
80
+ 2.625475830078125,
81
+ 2.66521337890625
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 48.0088680258652
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.2904243144989005,
97
+ "mean": 0.02082844218306654,
98
+ "stdev": 0.00032485950454247533,
99
+ "p50": 0.020652031898498534,
100
+ "p90": 0.021287628555297852,
101
+ "p95": 0.021401958274841308,
102
+ "p99": 0.02179714044570923,
103
+ "values": [
104
+ 0.02065100860595703,
105
+ 0.020617216110229493,
106
+ 0.02062233543395996,
107
+ 0.020591615676879883,
108
+ 0.02070425605773926,
109
+ 0.020956159591674805,
110
+ 0.021275648117065428,
111
+ 0.02083839988708496,
112
+ 0.02066329574584961,
113
+ 0.020594688415527345,
114
+ 0.02067148780822754,
115
+ 0.02100223922729492,
116
+ 0.02081996726989746,
117
+ 0.020578304290771485,
118
+ 0.020612096786499022,
119
+ 0.020584447860717774,
120
+ 0.020590591430664062,
121
+ 0.020579328536987306,
122
+ 0.020588544845581053,
123
+ 0.020585472106933594,
124
+ 0.020585472106933594,
125
+ 0.020576255798339844,
126
+ 0.020594688415527345,
127
+ 0.020603904724121092,
128
+ 0.020595712661743162,
129
+ 0.020603904724121092,
130
+ 0.020599807739257812,
131
+ 0.0206376953125,
132
+ 0.020634624481201173,
133
+ 0.02117433547973633,
134
+ 0.020584384918212892,
135
+ 0.0206059513092041,
136
+ 0.021173248291015623,
137
+ 0.020595712661743162,
138
+ 0.020579328536987306,
139
+ 0.02058956718444824,
140
+ 0.020553728103637696,
141
+ 0.020546592712402344,
142
+ 0.020596704483032226,
143
+ 0.0206059513092041,
144
+ 0.020583423614501953,
145
+ 0.02064588737487793,
146
+ 0.020582399368286132,
147
+ 0.020403200149536133,
148
+ 0.020398080825805662,
149
+ 0.0204769287109375,
150
+ 0.020585472106933594,
151
+ 0.020560895919799805,
152
+ 0.020600831985473633,
153
+ 0.020555776596069338,
154
+ 0.02065715217590332,
155
+ 0.02053222465515137,
156
+ 0.02061414337158203,
157
+ 0.020585472106933594,
158
+ 0.020632575988769532,
159
+ 0.020653055191040038,
160
+ 0.021574655532836915,
161
+ 0.020926464080810548,
162
+ 0.02085171127319336,
163
+ 0.02119987106323242,
164
+ 0.02060697555541992,
165
+ 0.02061414337158203,
166
+ 0.020593664169311524,
167
+ 0.020545536041259766,
168
+ 0.020632575988769532,
169
+ 0.02064691162109375,
170
+ 0.020563968658447264,
171
+ 0.020592639923095703,
172
+ 0.020594688415527345,
173
+ 0.02061414337158203,
174
+ 0.020585472106933594,
175
+ 0.020609024047851563,
176
+ 0.020587520599365236,
177
+ 0.020609024047851563,
178
+ 0.02065100860595703,
179
+ 0.0206059513092041,
180
+ 0.020599807739257812,
181
+ 0.020583423614501953,
182
+ 0.020599807739257812,
183
+ 0.020618240356445314,
184
+ 0.02065407943725586,
185
+ 0.02059775924682617,
186
+ 0.02061414337158203,
187
+ 0.02068079948425293,
188
+ 0.020649887084960936,
189
+ 0.020610048294067384,
190
+ 0.02063052749633789,
191
+ 0.020593664169311524,
192
+ 0.020562944412231447,
193
+ 0.02062745666503906,
194
+ 0.02064076805114746,
195
+ 0.020915199279785156,
196
+ 0.02124799919128418,
197
+ 0.02124185562133789,
198
+ 0.021165056228637694,
199
+ 0.020731903076171874,
200
+ 0.020632575988769532,
201
+ 0.02062950325012207,
202
+ 0.02061516761779785,
203
+ 0.020620288848876952,
204
+ 0.02065100860595703,
205
+ 0.020641792297363282,
206
+ 0.020692991256713866,
207
+ 0.020619264602661135,
208
+ 0.02063667106628418,
209
+ 0.02062950325012207,
210
+ 0.02124185562133789,
211
+ 0.021009408950805664,
212
+ 0.020644863128662108,
213
+ 0.020632575988769532,
214
+ 0.020599807739257812,
215
+ 0.020624383926391602,
216
+ 0.02064384078979492,
217
+ 0.020600831985473633,
218
+ 0.02061414337158203,
219
+ 0.020583423614501953,
220
+ 0.02061311912536621,
221
+ 0.020591615676879883,
222
+ 0.02063155174255371,
223
+ 0.02065715217590332,
224
+ 0.020617216110229493,
225
+ 0.02062950325012207,
226
+ 0.02065920066833496,
227
+ 0.020577280044555665,
228
+ 0.02062131118774414,
229
+ 0.020603904724121092,
230
+ 0.020576255798339844,
231
+ 0.020447231292724608,
232
+ 0.020446207046508787,
233
+ 0.02045952033996582,
234
+ 0.02063974380493164,
235
+ 0.02059878349304199,
236
+ 0.02061516761779785,
237
+ 0.020601856231689454,
238
+ 0.020616191864013672,
239
+ 0.020590591430664062,
240
+ 0.020591615676879883,
241
+ 0.020580352783203124,
242
+ 0.020569087982177735,
243
+ 0.020551679611206054,
244
+ 0.0206561279296875,
245
+ 0.020503551483154296,
246
+ 0.020410367965698242,
247
+ 0.020702207565307617,
248
+ 0.02205388832092285,
249
+ 0.020847616195678712,
250
+ 0.021394432067871092,
251
+ 0.020968448638916014,
252
+ 0.021308416366577147,
253
+ 0.021179391860961915,
254
+ 0.02122956848144531,
255
+ 0.02127462387084961,
256
+ 0.021415935516357423,
257
+ 0.020926464080810548,
258
+ 0.020684799194335936,
259
+ 0.021353471755981446,
260
+ 0.020954111099243163,
261
+ 0.020634624481201173,
262
+ 0.020785152435302736,
263
+ 0.021356544494628905,
264
+ 0.021259263992309572,
265
+ 0.021109760284423826,
266
+ 0.02062950325012207,
267
+ 0.020578304290771485,
268
+ 0.020619264602661135,
269
+ 0.0206059513092041,
270
+ 0.021145599365234375,
271
+ 0.021210111618041993,
272
+ 0.021226495742797852,
273
+ 0.02064384078979492,
274
+ 0.020603904724121092,
275
+ 0.02067353630065918,
276
+ 0.0206059513092041,
277
+ 0.021115903854370118,
278
+ 0.021203968048095705,
279
+ 0.021226495742797852,
280
+ 0.02066227149963379,
281
+ 0.021189632415771483,
282
+ 0.02143948745727539,
283
+ 0.021209087371826172,
284
+ 0.021234687805175782,
285
+ 0.02122444725036621,
286
+ 0.021621759414672852,
287
+ 0.021206016540527343,
288
+ 0.020666368484497072,
289
+ 0.02066022491455078,
290
+ 0.020974592208862306,
291
+ 0.02129408073425293,
292
+ 0.021149696350097655,
293
+ 0.02065510368347168,
294
+ 0.02064076805114746,
295
+ 0.020677631378173827,
296
+ 0.02063155174255371,
297
+ 0.020715520858764647,
298
+ 0.02085683250427246,
299
+ 0.021334016799926758,
300
+ 0.02145280075073242,
301
+ 0.02128486442565918,
302
+ 0.021819391250610352,
303
+ 0.02150809669494629,
304
+ 0.020967424392700194,
305
+ 0.020945920944213867,
306
+ 0.02089472007751465,
307
+ 0.020755456924438476,
308
+ 0.021032960891723632,
309
+ 0.02168217658996582,
310
+ 0.021373952865600586,
311
+ 0.020736000061035157,
312
+ 0.02066329574584961,
313
+ 0.020649984359741212,
314
+ 0.02104217529296875,
315
+ 0.021005311965942384,
316
+ 0.020706304550170897,
317
+ 0.021313535690307618,
318
+ 0.02123776054382324,
319
+ 0.021279743194580078,
320
+ 0.021287935256958008,
321
+ 0.021200895309448242,
322
+ 0.0212541446685791,
323
+ 0.020676607131958007,
324
+ 0.020787200927734374,
325
+ 0.02125209617614746,
326
+ 0.02066534423828125,
327
+ 0.02068172836303711,
328
+ 0.02090598487854004,
329
+ 0.02122854423522949,
330
+ 0.021337120056152344,
331
+ 0.02067555236816406,
332
+ 0.02070528030395508,
333
+ 0.02088755226135254,
334
+ 0.02128691291809082,
335
+ 0.020720640182495118,
336
+ 0.020699136734008788,
337
+ 0.020661247253417968,
338
+ 0.020978687286376953,
339
+ 0.02122547149658203,
340
+ 0.02062131118774414,
341
+ 0.02064588737487793,
342
+ 0.02106675148010254,
343
+ 0.02111692810058594,
344
+ 0.02083430480957031,
345
+ 0.02082918357849121,
346
+ 0.021217279434204102,
347
+ 0.021385215759277345,
348
+ 0.021271551132202148,
349
+ 0.02085273551940918,
350
+ 0.02066022491455078,
351
+ 0.021340160369873046,
352
+ 0.02203647994995117,
353
+ 0.021612543106079102,
354
+ 0.021777408599853516,
355
+ 0.021381120681762695,
356
+ 0.02144767951965332,
357
+ 0.02106265640258789
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 48.01127185656721
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02243292808532715,
7
+ "prefill.throughput.value": 312.0412981031458,
8
+ "decode.latency.mean": 2.6453446044921876,
9
+ "decode.throughput.value": 48.0088680258652,
10
+ "per_token.latency.mean": 0.02082844218306654,
11
+ "per_token.throughput.value": 48.01127185656721
12
+ }
13
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: a05ce550bf1cda8ae3bffcc62607597770460a42
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42
88
+ - hydra.run.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1353.216,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.391537109375,
120
+ "mean": 12.391537109375,
121
+ "stdev": 0.0,
122
+ "p50": 12.391537109375,
123
+ "p90": 12.391537109375,
124
+ "p95": 12.391537109375,
125
+ "p99": 12.391537109375,
126
+ "values": [
127
+ 12.391537109375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1924.329472,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.02835264015197754,
147
+ "mean": 0.01417632007598877,
148
+ "stdev": 0.00015244770050048834,
149
+ "p50": 0.01417632007598877,
150
+ "p90": 0.01429827823638916,
151
+ "p95": 0.014313523006439208,
152
+ "p99": 0.014325718822479247,
153
+ "values": [
154
+ 0.01402387237548828,
155
+ 0.014328767776489258
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 493.7811761076341
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2097.610752,
169
+ "max_global_vram": 6060.244992,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.1297249755859378,
178
+ "mean": 1.5648624877929689,
179
+ "stdev": 3.4484863281281086e-05,
180
+ "p50": 1.5648624877929689,
181
+ "p90": 1.5648900756835937,
182
+ "p95": 1.564893524169922,
183
+ "p99": 1.5648962829589845,
184
+ "values": [
185
+ 1.56489697265625,
186
+ 1.5648280029296875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.1572908103361
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.1294433279037492,
202
+ "mean": 0.012320643023243101,
203
+ "stdev": 6.516712266289387e-05,
204
+ "p50": 0.012308480262756348,
205
+ "p90": 0.012332032203674317,
206
+ "p95": 0.012361421060562134,
207
+ "p99": 0.012632555418014526,
208
+ "values": [
209
+ 0.01235865592956543,
210
+ 0.012311552047729492,
211
+ 0.012301312446594239,
212
+ 0.012321791648864745,
213
+ 0.012305407524108887,
214
+ 0.012299263954162597,
215
+ 0.012301312446594239,
216
+ 0.012308480262756348,
217
+ 0.012294143676757812,
218
+ 0.012307456016540527,
219
+ 0.012332032203674317,
220
+ 0.012307456016540527,
221
+ 0.012297216415405274,
222
+ 0.012297216415405274,
223
+ 0.012302335739135742,
224
+ 0.012306431770324706,
225
+ 0.012322815895080566,
226
+ 0.012308480262756348,
227
+ 0.012303359985351562,
228
+ 0.012303359985351562,
229
+ 0.012301312446594239,
230
+ 0.012312576293945313,
231
+ 0.012303359985351562,
232
+ 0.012319744110107422,
233
+ 0.012311552047729492,
234
+ 0.012307456016540527,
235
+ 0.012313599586486817,
236
+ 0.012303359985351562,
237
+ 0.012298239707946777,
238
+ 0.012312576293945313,
239
+ 0.012316672325134278,
240
+ 0.012309503555297852,
241
+ 0.012305407524108887,
242
+ 0.012294143676757812,
243
+ 0.012319744110107422,
244
+ 0.012311552047729492,
245
+ 0.012324864387512208,
246
+ 0.012465151786804199,
247
+ 0.01238425636291504,
248
+ 0.013000703811645508,
249
+ 0.012379136085510254,
250
+ 0.012313599586486817,
251
+ 0.012297216415405274,
252
+ 0.012316672325134278,
253
+ 0.01234329605102539,
254
+ 0.012379136085510254,
255
+ 0.012321791648864745,
256
+ 0.012318719863891601,
257
+ 0.012301312446594239,
258
+ 0.012308480262756348,
259
+ 0.012302335739135742,
260
+ 0.012305407524108887,
261
+ 0.012303359985351562,
262
+ 0.012310527801513671,
263
+ 0.012303359985351562,
264
+ 0.012304384231567383,
265
+ 0.012304384231567383,
266
+ 0.012307456016540527,
267
+ 0.012339200019836426,
268
+ 0.012317695617675782,
269
+ 0.012302335739135742,
270
+ 0.012304384231567383,
271
+ 0.012307456016540527,
272
+ 0.012305407524108887,
273
+ 0.012317695617675782,
274
+ 0.012317695617675782,
275
+ 0.012543999671936035,
276
+ 0.012340224266052247,
277
+ 0.012308480262756348,
278
+ 0.012304384231567383,
279
+ 0.012317695617675782,
280
+ 0.012323840141296387,
281
+ 0.012301312446594239,
282
+ 0.012308480262756348,
283
+ 0.012311552047729492,
284
+ 0.012304384231567383,
285
+ 0.012311552047729492,
286
+ 0.012314623832702636,
287
+ 0.01229312038421631,
288
+ 0.012295167922973632,
289
+ 0.012294143676757812,
290
+ 0.012321791648864745,
291
+ 0.012298239707946777,
292
+ 0.012319744110107422,
293
+ 0.012368895530700684,
294
+ 0.012311552047729492,
295
+ 0.012297216415405274,
296
+ 0.012301312446594239,
297
+ 0.012305407524108887,
298
+ 0.012311552047729492,
299
+ 0.012321791648864745,
300
+ 0.012312576293945313,
301
+ 0.012308480262756348,
302
+ 0.012308480262756348,
303
+ 0.012319744110107422,
304
+ 0.012297216415405274,
305
+ 0.012370944023132324,
306
+ 0.012326911926269531,
307
+ 0.012308480262756348,
308
+ 0.012303359985351562,
309
+ 0.012300288200378418,
310
+ 0.012299263954162597,
311
+ 0.012307456016540527,
312
+ 0.012301312446594239,
313
+ 0.012298239707946777,
314
+ 0.012304384231567383,
315
+ 0.012303359985351562,
316
+ 0.012298239707946777,
317
+ 0.012311552047729492,
318
+ 0.012320799827575684,
319
+ 0.012305376052856445,
320
+ 0.012304384231567383,
321
+ 0.012301312446594239,
322
+ 0.012302335739135742,
323
+ 0.012308480262756348,
324
+ 0.012312576293945313,
325
+ 0.012309503555297852,
326
+ 0.012299263954162597,
327
+ 0.012309503555297852,
328
+ 0.012310527801513671,
329
+ 0.012296192169189453,
330
+ 0.012377087593078612,
331
+ 0.012331040382385253,
332
+ 0.01230844783782959,
333
+ 0.012300288200378418,
334
+ 0.012299263954162597,
335
+ 0.012313599586486817,
336
+ 0.01235865592956543,
337
+ 0.012309503555297852,
338
+ 0.012299263954162597,
339
+ 0.012307456016540527,
340
+ 0.012317695617675782,
341
+ 0.012311552047729492,
342
+ 0.012315648078918457,
343
+ 0.012299263954162597,
344
+ 0.012306431770324706,
345
+ 0.012305407524108887,
346
+ 0.012307456016540527,
347
+ 0.012299263954162597,
348
+ 0.01233510398864746,
349
+ 0.012329983711242675,
350
+ 0.012320768356323243,
351
+ 0.012296192169189453,
352
+ 0.012296192169189453,
353
+ 0.012306431770324706,
354
+ 0.012306464195251466,
355
+ 0.012311519622802735,
356
+ 0.012306431770324706,
357
+ 0.012305407524108887,
358
+ 0.012306431770324706,
359
+ 0.012294143676757812,
360
+ 0.012304384231567383,
361
+ 0.012328960418701173,
362
+ 0.012315648078918457,
363
+ 0.012292096138000488,
364
+ 0.012308480262756348,
365
+ 0.012303359985351562,
366
+ 0.012301312446594239,
367
+ 0.01235865592956543,
368
+ 0.012309503555297852,
369
+ 0.012303359985351562,
370
+ 0.01232588768005371,
371
+ 0.012299263954162597,
372
+ 0.012297216415405274,
373
+ 0.012291071891784668,
374
+ 0.012321791648864745,
375
+ 0.012314623832702636,
376
+ 0.012302335739135742,
377
+ 0.012304384231567383,
378
+ 0.012306431770324706,
379
+ 0.012295167922973632,
380
+ 0.012311552047729492,
381
+ 0.012312576293945313,
382
+ 0.012297216415405274,
383
+ 0.012299263954162597,
384
+ 0.012301312446594239,
385
+ 0.012314623832702636,
386
+ 0.012322815895080566,
387
+ 0.012327936172485352,
388
+ 0.012317695617675782,
389
+ 0.012362751960754394,
390
+ 0.012306431770324706,
391
+ 0.012300288200378418,
392
+ 0.012310527801513671,
393
+ 0.012302335739135742,
394
+ 0.012298239707946777,
395
+ 0.012305407524108887,
396
+ 0.012291071891784668,
397
+ 0.012304384231567383,
398
+ 0.012302335739135742,
399
+ 0.012309503555297852,
400
+ 0.012319744110107422,
401
+ 0.012318719863891601,
402
+ 0.012300288200378418,
403
+ 0.012306431770324706,
404
+ 0.01229312038421631,
405
+ 0.012311552047729492,
406
+ 0.012311552047729492,
407
+ 0.012306431770324706,
408
+ 0.012306431770324706,
409
+ 0.012297216415405274,
410
+ 0.012360704421997071,
411
+ 0.012299263954162597,
412
+ 0.012309503555297852,
413
+ 0.012332032203674317,
414
+ 0.012313599586486817,
415
+ 0.012827648162841796,
416
+ 0.012732416152954102,
417
+ 0.012507136344909669,
418
+ 0.012355584144592285,
419
+ 0.012318719863891601,
420
+ 0.012302335739135742,
421
+ 0.012323840141296387,
422
+ 0.012333056449890138,
423
+ 0.012308480262756348,
424
+ 0.012308480262756348,
425
+ 0.012297216415405274,
426
+ 0.012316672325134278,
427
+ 0.012327936172485352,
428
+ 0.012321791648864745,
429
+ 0.012303359985351562,
430
+ 0.012296192169189453,
431
+ 0.012299263954162597,
432
+ 0.012306431770324706,
433
+ 0.012301312446594239,
434
+ 0.012320768356323243,
435
+ 0.012317695617675782,
436
+ 0.012310527801513671,
437
+ 0.012291071891784668,
438
+ 0.012304384231567383,
439
+ 0.01233510398864746,
440
+ 0.012322815895080566,
441
+ 0.012315648078918457,
442
+ 0.012301312446594239,
443
+ 0.012301312446594239,
444
+ 0.012306431770324706,
445
+ 0.012315648078918457,
446
+ 0.012315648078918457,
447
+ 0.012324864387512208,
448
+ 0.012320768356323243,
449
+ 0.012307456016540527,
450
+ 0.012289024353027344,
451
+ 0.012322815895080566,
452
+ 0.012307456016540527,
453
+ 0.012322815895080566,
454
+ 0.012312576293945313,
455
+ 0.012309503555297852,
456
+ 0.012306431770324706,
457
+ 0.012300288200378418,
458
+ 0.012302335739135742,
459
+ 0.012317695617675782,
460
+ 0.012321791648864745,
461
+ 0.012340224266052247,
462
+ 0.012310527801513671
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.16459490900623
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1353.216,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.391537109375,
15
+ "mean": 12.391537109375,
16
+ "stdev": 0.0,
17
+ "p50": 12.391537109375,
18
+ "p90": 12.391537109375,
19
+ "p95": 12.391537109375,
20
+ "p99": 12.391537109375,
21
+ "values": [
22
+ 12.391537109375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1924.329472,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.02835264015197754,
42
+ "mean": 0.01417632007598877,
43
+ "stdev": 0.00015244770050048834,
44
+ "p50": 0.01417632007598877,
45
+ "p90": 0.01429827823638916,
46
+ "p95": 0.014313523006439208,
47
+ "p99": 0.014325718822479247,
48
+ "values": [
49
+ 0.01402387237548828,
50
+ 0.014328767776489258
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 493.7811761076341
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2097.610752,
64
+ "max_global_vram": 6060.244992,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.1297249755859378,
73
+ "mean": 1.5648624877929689,
74
+ "stdev": 3.4484863281281086e-05,
75
+ "p50": 1.5648624877929689,
76
+ "p90": 1.5648900756835937,
77
+ "p95": 1.564893524169922,
78
+ "p99": 1.5648962829589845,
79
+ "values": [
80
+ 1.56489697265625,
81
+ 1.5648280029296875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.1572908103361
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.1294433279037492,
97
+ "mean": 0.012320643023243101,
98
+ "stdev": 6.516712266289387e-05,
99
+ "p50": 0.012308480262756348,
100
+ "p90": 0.012332032203674317,
101
+ "p95": 0.012361421060562134,
102
+ "p99": 0.012632555418014526,
103
+ "values": [
104
+ 0.01235865592956543,
105
+ 0.012311552047729492,
106
+ 0.012301312446594239,
107
+ 0.012321791648864745,
108
+ 0.012305407524108887,
109
+ 0.012299263954162597,
110
+ 0.012301312446594239,
111
+ 0.012308480262756348,
112
+ 0.012294143676757812,
113
+ 0.012307456016540527,
114
+ 0.012332032203674317,
115
+ 0.012307456016540527,
116
+ 0.012297216415405274,
117
+ 0.012297216415405274,
118
+ 0.012302335739135742,
119
+ 0.012306431770324706,
120
+ 0.012322815895080566,
121
+ 0.012308480262756348,
122
+ 0.012303359985351562,
123
+ 0.012303359985351562,
124
+ 0.012301312446594239,
125
+ 0.012312576293945313,
126
+ 0.012303359985351562,
127
+ 0.012319744110107422,
128
+ 0.012311552047729492,
129
+ 0.012307456016540527,
130
+ 0.012313599586486817,
131
+ 0.012303359985351562,
132
+ 0.012298239707946777,
133
+ 0.012312576293945313,
134
+ 0.012316672325134278,
135
+ 0.012309503555297852,
136
+ 0.012305407524108887,
137
+ 0.012294143676757812,
138
+ 0.012319744110107422,
139
+ 0.012311552047729492,
140
+ 0.012324864387512208,
141
+ 0.012465151786804199,
142
+ 0.01238425636291504,
143
+ 0.013000703811645508,
144
+ 0.012379136085510254,
145
+ 0.012313599586486817,
146
+ 0.012297216415405274,
147
+ 0.012316672325134278,
148
+ 0.01234329605102539,
149
+ 0.012379136085510254,
150
+ 0.012321791648864745,
151
+ 0.012318719863891601,
152
+ 0.012301312446594239,
153
+ 0.012308480262756348,
154
+ 0.012302335739135742,
155
+ 0.012305407524108887,
156
+ 0.012303359985351562,
157
+ 0.012310527801513671,
158
+ 0.012303359985351562,
159
+ 0.012304384231567383,
160
+ 0.012304384231567383,
161
+ 0.012307456016540527,
162
+ 0.012339200019836426,
163
+ 0.012317695617675782,
164
+ 0.012302335739135742,
165
+ 0.012304384231567383,
166
+ 0.012307456016540527,
167
+ 0.012305407524108887,
168
+ 0.012317695617675782,
169
+ 0.012317695617675782,
170
+ 0.012543999671936035,
171
+ 0.012340224266052247,
172
+ 0.012308480262756348,
173
+ 0.012304384231567383,
174
+ 0.012317695617675782,
175
+ 0.012323840141296387,
176
+ 0.012301312446594239,
177
+ 0.012308480262756348,
178
+ 0.012311552047729492,
179
+ 0.012304384231567383,
180
+ 0.012311552047729492,
181
+ 0.012314623832702636,
182
+ 0.01229312038421631,
183
+ 0.012295167922973632,
184
+ 0.012294143676757812,
185
+ 0.012321791648864745,
186
+ 0.012298239707946777,
187
+ 0.012319744110107422,
188
+ 0.012368895530700684,
189
+ 0.012311552047729492,
190
+ 0.012297216415405274,
191
+ 0.012301312446594239,
192
+ 0.012305407524108887,
193
+ 0.012311552047729492,
194
+ 0.012321791648864745,
195
+ 0.012312576293945313,
196
+ 0.012308480262756348,
197
+ 0.012308480262756348,
198
+ 0.012319744110107422,
199
+ 0.012297216415405274,
200
+ 0.012370944023132324,
201
+ 0.012326911926269531,
202
+ 0.012308480262756348,
203
+ 0.012303359985351562,
204
+ 0.012300288200378418,
205
+ 0.012299263954162597,
206
+ 0.012307456016540527,
207
+ 0.012301312446594239,
208
+ 0.012298239707946777,
209
+ 0.012304384231567383,
210
+ 0.012303359985351562,
211
+ 0.012298239707946777,
212
+ 0.012311552047729492,
213
+ 0.012320799827575684,
214
+ 0.012305376052856445,
215
+ 0.012304384231567383,
216
+ 0.012301312446594239,
217
+ 0.012302335739135742,
218
+ 0.012308480262756348,
219
+ 0.012312576293945313,
220
+ 0.012309503555297852,
221
+ 0.012299263954162597,
222
+ 0.012309503555297852,
223
+ 0.012310527801513671,
224
+ 0.012296192169189453,
225
+ 0.012377087593078612,
226
+ 0.012331040382385253,
227
+ 0.01230844783782959,
228
+ 0.012300288200378418,
229
+ 0.012299263954162597,
230
+ 0.012313599586486817,
231
+ 0.01235865592956543,
232
+ 0.012309503555297852,
233
+ 0.012299263954162597,
234
+ 0.012307456016540527,
235
+ 0.012317695617675782,
236
+ 0.012311552047729492,
237
+ 0.012315648078918457,
238
+ 0.012299263954162597,
239
+ 0.012306431770324706,
240
+ 0.012305407524108887,
241
+ 0.012307456016540527,
242
+ 0.012299263954162597,
243
+ 0.01233510398864746,
244
+ 0.012329983711242675,
245
+ 0.012320768356323243,
246
+ 0.012296192169189453,
247
+ 0.012296192169189453,
248
+ 0.012306431770324706,
249
+ 0.012306464195251466,
250
+ 0.012311519622802735,
251
+ 0.012306431770324706,
252
+ 0.012305407524108887,
253
+ 0.012306431770324706,
254
+ 0.012294143676757812,
255
+ 0.012304384231567383,
256
+ 0.012328960418701173,
257
+ 0.012315648078918457,
258
+ 0.012292096138000488,
259
+ 0.012308480262756348,
260
+ 0.012303359985351562,
261
+ 0.012301312446594239,
262
+ 0.01235865592956543,
263
+ 0.012309503555297852,
264
+ 0.012303359985351562,
265
+ 0.01232588768005371,
266
+ 0.012299263954162597,
267
+ 0.012297216415405274,
268
+ 0.012291071891784668,
269
+ 0.012321791648864745,
270
+ 0.012314623832702636,
271
+ 0.012302335739135742,
272
+ 0.012304384231567383,
273
+ 0.012306431770324706,
274
+ 0.012295167922973632,
275
+ 0.012311552047729492,
276
+ 0.012312576293945313,
277
+ 0.012297216415405274,
278
+ 0.012299263954162597,
279
+ 0.012301312446594239,
280
+ 0.012314623832702636,
281
+ 0.012322815895080566,
282
+ 0.012327936172485352,
283
+ 0.012317695617675782,
284
+ 0.012362751960754394,
285
+ 0.012306431770324706,
286
+ 0.012300288200378418,
287
+ 0.012310527801513671,
288
+ 0.012302335739135742,
289
+ 0.012298239707946777,
290
+ 0.012305407524108887,
291
+ 0.012291071891784668,
292
+ 0.012304384231567383,
293
+ 0.012302335739135742,
294
+ 0.012309503555297852,
295
+ 0.012319744110107422,
296
+ 0.012318719863891601,
297
+ 0.012300288200378418,
298
+ 0.012306431770324706,
299
+ 0.01229312038421631,
300
+ 0.012311552047729492,
301
+ 0.012311552047729492,
302
+ 0.012306431770324706,
303
+ 0.012306431770324706,
304
+ 0.012297216415405274,
305
+ 0.012360704421997071,
306
+ 0.012299263954162597,
307
+ 0.012309503555297852,
308
+ 0.012332032203674317,
309
+ 0.012313599586486817,
310
+ 0.012827648162841796,
311
+ 0.012732416152954102,
312
+ 0.012507136344909669,
313
+ 0.012355584144592285,
314
+ 0.012318719863891601,
315
+ 0.012302335739135742,
316
+ 0.012323840141296387,
317
+ 0.012333056449890138,
318
+ 0.012308480262756348,
319
+ 0.012308480262756348,
320
+ 0.012297216415405274,
321
+ 0.012316672325134278,
322
+ 0.012327936172485352,
323
+ 0.012321791648864745,
324
+ 0.012303359985351562,
325
+ 0.012296192169189453,
326
+ 0.012299263954162597,
327
+ 0.012306431770324706,
328
+ 0.012301312446594239,
329
+ 0.012320768356323243,
330
+ 0.012317695617675782,
331
+ 0.012310527801513671,
332
+ 0.012291071891784668,
333
+ 0.012304384231567383,
334
+ 0.01233510398864746,
335
+ 0.012322815895080566,
336
+ 0.012315648078918457,
337
+ 0.012301312446594239,
338
+ 0.012301312446594239,
339
+ 0.012306431770324706,
340
+ 0.012315648078918457,
341
+ 0.012315648078918457,
342
+ 0.012324864387512208,
343
+ 0.012320768356323243,
344
+ 0.012307456016540527,
345
+ 0.012289024353027344,
346
+ 0.012322815895080566,
347
+ 0.012307456016540527,
348
+ 0.012322815895080566,
349
+ 0.012312576293945313,
350
+ 0.012309503555297852,
351
+ 0.012306431770324706,
352
+ 0.012300288200378418,
353
+ 0.012302335739135742,
354
+ 0.012317695617675782,
355
+ 0.012321791648864745,
356
+ 0.012340224266052247,
357
+ 0.012310527801513671
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.16459490900623
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.01417632007598877,
7
+ "prefill.throughput.value": 493.7811761076341,
8
+ "decode.latency.mean": 1.5648624877929689,
9
+ "decode.throughput.value": 81.1572908103361,
10
+ "per_token.latency.mean": 0.012320643023243101,
11
+ "per_token.throughput.value": 81.16459490900623
12
+ }
13
+ }
2024-09-13/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpmmt3h_ky/commit=a05ce550bf1cda8ae3bffcc62607597770460a42
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42
88
+ - hydra.run.dir=_benchmark/tmpmmt3h_ky/commit\=a05ce550bf1cda8ae3bffcc62607597770460a42/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: a05ce550bf1cda8ae3bffcc62607597770460a42
225
+ accelerate_version: 0.35.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.23.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-09-13/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.020703951835632325,
8
+ "prefill.throughput.value": 338.09970461546,
9
+ "decode.latency.mean": 2.5267764892578124,
10
+ "decode.throughput.value": 50.26166759898244,
11
+ "per_token.latency.mean": 0.019894590512974056,
12
+ "per_token.throughput.value": 50.26491997147969
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.02243292808532715,
21
+ "prefill.throughput.value": 312.0412981031458,
22
+ "decode.latency.mean": 2.6453446044921876,
23
+ "decode.throughput.value": 48.0088680258652,
24
+ "per_token.latency.mean": 0.02082844218306654,
25
+ "per_token.throughput.value": 48.01127185656721
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "a05ce550bf1cda8ae3bffcc62607597770460a42",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.01417632007598877,
34
+ "prefill.throughput.value": 493.7811761076341,
35
+ "decode.latency.mean": 1.5648624877929689,
36
+ "decode.throughput.value": 81.1572908103361,
37
+ "per_token.latency.mean": 0.012320643023243101,
38
+ "per_token.throughput.value": 81.16459490900623
39
+ }
40
+ }
41
+ ]
2024-09-13/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "a05ce550bf1cda8ae3bffcc62607597770460a42": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.020703951835632325,
7
+ "prefill.throughput.value": 338.09970461546,
8
+ "decode.latency.mean": 2.5267764892578124,
9
+ "decode.throughput.value": 50.26166759898244,
10
+ "per_token.latency.mean": 0.019894590512974056,
11
+ "per_token.throughput.value": 50.26491997147969
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "a05ce550bf1cda8ae3bffcc62607597770460a42": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.02243292808532715,
19
+ "prefill.throughput.value": 312.0412981031458,
20
+ "decode.latency.mean": 2.6453446044921876,
21
+ "decode.throughput.value": 48.0088680258652,
22
+ "per_token.latency.mean": 0.02082844218306654,
23
+ "per_token.throughput.value": 48.01127185656721
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "a05ce550bf1cda8ae3bffcc62607597770460a42": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.01417632007598877,
31
+ "prefill.throughput.value": 493.7811761076341,
32
+ "decode.latency.mean": 1.5648624877929689,
33
+ "decode.throughput.value": 81.1572908103361,
34
+ "per_token.latency.mean": 0.012320643023243101,
35
+ "per_token.throughput.value": 81.16459490900623
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }