hf-transformers-bot commited on
Commit
b07f607
1 Parent(s): 68163e0

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +91 -0
  2. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +443 -0
  5. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +102 -0
  7. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +339 -0
  8. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +91 -0
  10. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +1 -0
  13. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +102 -0
  14. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +91 -0
  15. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +443 -0
  18. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +102 -0
  20. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +339 -0
  21. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +91 -0
  23. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +443 -0
  26. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +102 -0
  28. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +339 -0
  29. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/multirun.yaml +232 -0
  31. 2024-06-09/summaries.json +41 -0
  32. 2024-06-09/summary.json +40 -0
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: 25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
88
+ - hydra.run.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1728.65536,
109
+ "max_global_vram": 6793.199616,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 6146.752512,
112
+ "max_allocated": 5034.576896
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.04538415908813477,
118
+ "mean": 0.022692079544067385,
119
+ "stdev": 0.003335311889648437,
120
+ "p50": 0.022692079544067385,
121
+ "p90": 0.025360329055786134,
122
+ "p95": 0.025693860244750975,
123
+ "p99": 0.02596068519592285,
124
+ "values": [
125
+ 0.02602739143371582,
126
+ 0.019356767654418947
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 308.47767770275067
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 1728.745472,
140
+ "max_global_vram": 6795.296768,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 6148.849664,
143
+ "max_allocated": 5034.577408
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 4.77087451171875,
149
+ "mean": 2.385437255859375,
150
+ "stdev": 0.001996337890624922,
151
+ "p50": 2.385437255859375,
152
+ "p90": 2.387034326171875,
153
+ "p95": 2.3872339599609376,
154
+ "p99": 2.3873936669921876,
155
+ "values": [
156
+ 2.38743359375,
157
+ 2.38344091796875
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 53.23971514574468
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 4.770945024490353,
173
+ "mean": 0.018857490215376905,
174
+ "stdev": 0.0012656888721164668,
175
+ "p50": 0.018661376953125,
176
+ "p90": 0.01934172248840332,
177
+ "p95": 0.019502080154418944,
178
+ "p99": 0.020219125366210934,
179
+ "values": [
180
+ 0.019984384536743165,
181
+ 0.020305919647216796,
182
+ 0.018746368408203123,
183
+ 0.018490367889404297,
184
+ 0.018481151580810547,
185
+ 0.018464767456054687,
186
+ 0.018486272811889647,
187
+ 0.018510847091674804,
188
+ 0.01845964813232422,
189
+ 0.018529279708862305,
190
+ 0.018509824752807616,
191
+ 0.01841459274291992,
192
+ 0.018298879623413086,
193
+ 0.018324480056762696,
194
+ 0.01846886444091797,
195
+ 0.018748416900634765,
196
+ 0.019137535095214844,
197
+ 0.019017728805541992,
198
+ 0.019188735961914064,
199
+ 0.01929523277282715,
200
+ 0.018540544509887694,
201
+ 0.018530303955078126,
202
+ 0.018387968063354493,
203
+ 0.018487295150756835,
204
+ 0.018587648391723634,
205
+ 0.01848531150817871,
206
+ 0.01852409553527832,
207
+ 0.018553855895996094,
208
+ 0.018515968322753908,
209
+ 0.018520063400268554,
210
+ 0.018471935272216796,
211
+ 0.018737152099609376,
212
+ 0.018405376434326173,
213
+ 0.018319360733032225,
214
+ 0.018663423538208008,
215
+ 0.01857535934448242,
216
+ 0.018498559951782227,
217
+ 0.018535423278808593,
218
+ 0.018510847091674804,
219
+ 0.018491392135620118,
220
+ 0.018587648391723634,
221
+ 0.018686975479125977,
222
+ 0.018532352447509767,
223
+ 0.019281919479370118,
224
+ 0.018699264526367186,
225
+ 0.01859686470031738,
226
+ 0.018505727767944336,
227
+ 0.01843097686767578,
228
+ 0.018566144943237304,
229
+ 0.018495487213134765,
230
+ 0.018502656936645507,
231
+ 0.018966527938842775,
232
+ 0.01948569679260254,
233
+ 0.019645471572875977,
234
+ 0.01967407989501953,
235
+ 0.019325952529907226,
236
+ 0.01940275192260742,
237
+ 0.0187064323425293,
238
+ 0.01866035270690918,
239
+ 0.018929664611816405,
240
+ 0.019466239929199217,
241
+ 0.01857535934448242,
242
+ 0.01845964813232422,
243
+ 0.01843916893005371,
244
+ 0.01845043182373047,
245
+ 0.018510847091674804,
246
+ 0.018618368148803712,
247
+ 0.01865011215209961,
248
+ 0.01863372802734375,
249
+ 0.018663423538208008,
250
+ 0.018720767974853517,
251
+ 0.01859686470031738,
252
+ 0.01878118324279785,
253
+ 0.01863577651977539,
254
+ 0.018691072463989256,
255
+ 0.01866547203063965,
256
+ 0.019335168838500977,
257
+ 0.019352575302124024,
258
+ 0.019316736221313476,
259
+ 0.019578880310058593,
260
+ 0.019360767364501954,
261
+ 0.019361791610717775,
262
+ 0.019343360900878907,
263
+ 0.018675712585449217,
264
+ 0.01844633674621582,
265
+ 0.01844428825378418,
266
+ 0.01883033561706543,
267
+ 0.018907136917114258,
268
+ 0.018997247695922852,
269
+ 0.018457599639892578,
270
+ 0.018456575393676757,
271
+ 0.018488319396972656,
272
+ 0.018569215774536133,
273
+ 0.01864192008972168,
274
+ 0.018651136398315428,
275
+ 0.018618368148803712,
276
+ 0.019882080078125,
277
+ 0.019469215393066407,
278
+ 0.018634752273559572,
279
+ 0.01865216064453125,
280
+ 0.018662399291992187,
281
+ 0.018678783416748047,
282
+ 0.018745344161987306,
283
+ 0.018611200332641603,
284
+ 0.018716672897338867,
285
+ 0.01862553596496582,
286
+ 0.018732032775878905,
287
+ 0.01945292854309082,
288
+ 0.018685951232910156,
289
+ 0.01863577651977539,
290
+ 0.01864806365966797,
291
+ 0.01923072052001953,
292
+ 0.019526655197143555,
293
+ 0.020535295486450195,
294
+ 0.019017728805541992,
295
+ 0.018740224838256835,
296
+ 0.01877299118041992,
297
+ 0.018661376953125,
298
+ 0.019096576690673828,
299
+ 0.019260416030883788,
300
+ 0.018651136398315428,
301
+ 0.018742271423339844,
302
+ 0.018725887298583984,
303
+ 0.01869824028015137,
304
+ 0.018655231475830078,
305
+ 0.018689023971557618,
306
+ 0.03800985717773438,
307
+ 0.018487295150756835,
308
+ 0.018457599639892578,
309
+ 0.018562047958374024,
310
+ 0.018456575393676757,
311
+ 0.018499584197998048,
312
+ 0.018694143295288086,
313
+ 0.018505727767944336,
314
+ 0.01849555206298828,
315
+ 0.01854867172241211,
316
+ 0.018491455078125,
317
+ 0.01854867172241211,
318
+ 0.018719743728637696,
319
+ 0.01849241638183594,
320
+ 0.018353151321411132,
321
+ 0.01836953544616699,
322
+ 0.018297855377197265,
323
+ 0.01836031913757324,
324
+ 0.018339839935302735,
325
+ 0.018363391876220703,
326
+ 0.018329599380493163,
327
+ 0.018342912673950194,
328
+ 0.01862041664123535,
329
+ 0.019191808700561523,
330
+ 0.01903104019165039,
331
+ 0.019002368927001953,
332
+ 0.01908531188964844,
333
+ 0.01904128074645996,
334
+ 0.019162111282348633,
335
+ 0.018985984802246093,
336
+ 0.018965503692626954,
337
+ 0.01901158332824707,
338
+ 0.018751487731933594,
339
+ 0.01845964813232422,
340
+ 0.018540544509887694,
341
+ 0.01843507194519043,
342
+ 0.018453504562377928,
343
+ 0.01846784019470215,
344
+ 0.018319360733032225,
345
+ 0.01864089584350586,
346
+ 0.019142656326293944,
347
+ 0.018729984283447267,
348
+ 0.01905971145629883,
349
+ 0.01906790351867676,
350
+ 0.018696191787719727,
351
+ 0.01875660705566406,
352
+ 0.019056640625,
353
+ 0.019099647521972657,
354
+ 0.018824192047119142,
355
+ 0.018364416122436524,
356
+ 0.018318336486816408,
357
+ 0.01835212707519531,
358
+ 0.018510847091674804,
359
+ 0.01847603225708008,
360
+ 0.01847500801086426,
361
+ 0.018472959518432617,
362
+ 0.01846784019470215,
363
+ 0.01867263984680176,
364
+ 0.01865318489074707,
365
+ 0.019129344940185547,
366
+ 0.018734079360961914,
367
+ 0.01862348747253418,
368
+ 0.018684928894042968,
369
+ 0.018671615600585938,
370
+ 0.018743295669555664,
371
+ 0.020139007568359374,
372
+ 0.01940275192260742,
373
+ 0.018686975479125977,
374
+ 0.01862860870361328,
375
+ 0.01863065528869629,
376
+ 0.018661376953125,
377
+ 0.01869004821777344,
378
+ 0.018654207229614257,
379
+ 0.018694143295288086,
380
+ 0.018491392135620118,
381
+ 0.018715648651123046,
382
+ 0.01984102439880371,
383
+ 0.0194334716796875,
384
+ 0.019323904037475585,
385
+ 0.01942630386352539,
386
+ 0.019312639236450196,
387
+ 0.019174400329589843,
388
+ 0.01923686408996582,
389
+ 0.019581951141357423,
390
+ 0.01979084777832031,
391
+ 0.018968576431274413,
392
+ 0.019305471420288087,
393
+ 0.019354623794555666,
394
+ 0.01922867202758789,
395
+ 0.018917375564575196,
396
+ 0.01864806365966797,
397
+ 0.01861222457885742,
398
+ 0.018755584716796874,
399
+ 0.01869004821777344,
400
+ 0.018646015167236327,
401
+ 0.018761728286743166,
402
+ 0.018711551666259766,
403
+ 0.018651136398315428,
404
+ 0.018712575912475587,
405
+ 0.01862348747253418,
406
+ 0.018704383850097657,
407
+ 0.018597888946533202,
408
+ 0.018594816207885743,
409
+ 0.01864908790588379,
410
+ 0.018659328460693358,
411
+ 0.01865830421447754,
412
+ 0.018695167541503906,
413
+ 0.018758655548095703,
414
+ 0.018973695755004884,
415
+ 0.019306495666503908,
416
+ 0.01923481559753418,
417
+ 0.019203071594238282,
418
+ 0.01863680076599121,
419
+ 0.01862144088745117,
420
+ 0.01859174346923828,
421
+ 0.018480127334594726,
422
+ 0.018493440628051756,
423
+ 0.018740224838256835,
424
+ 0.01920102310180664,
425
+ 0.0191016960144043,
426
+ 0.018726911544799805,
427
+ 0.018461696624755858,
428
+ 0.018529279708862305,
429
+ 0.01844121551513672,
430
+ 0.018662399291992187,
431
+ 0.01868083190917969,
432
+ 0.01864499282836914
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 53.029326202941526
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1728.65536,
6
+ "max_global_vram": 6793.199616,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6146.752512,
9
+ "max_allocated": 5034.576896
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.04538415908813477,
15
+ "mean": 0.022692079544067385,
16
+ "stdev": 0.003335311889648437,
17
+ "p50": 0.022692079544067385,
18
+ "p90": 0.025360329055786134,
19
+ "p95": 0.025693860244750975,
20
+ "p99": 0.02596068519592285,
21
+ "values": [
22
+ 0.02602739143371582,
23
+ 0.019356767654418947
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 308.47767770275067
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 1728.745472,
37
+ "max_global_vram": 6795.296768,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 6148.849664,
40
+ "max_allocated": 5034.577408
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 4.77087451171875,
46
+ "mean": 2.385437255859375,
47
+ "stdev": 0.001996337890624922,
48
+ "p50": 2.385437255859375,
49
+ "p90": 2.387034326171875,
50
+ "p95": 2.3872339599609376,
51
+ "p99": 2.3873936669921876,
52
+ "values": [
53
+ 2.38743359375,
54
+ 2.38344091796875
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 53.23971514574468
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 4.770945024490353,
70
+ "mean": 0.018857490215376905,
71
+ "stdev": 0.0012656888721164668,
72
+ "p50": 0.018661376953125,
73
+ "p90": 0.01934172248840332,
74
+ "p95": 0.019502080154418944,
75
+ "p99": 0.020219125366210934,
76
+ "values": [
77
+ 0.019984384536743165,
78
+ 0.020305919647216796,
79
+ 0.018746368408203123,
80
+ 0.018490367889404297,
81
+ 0.018481151580810547,
82
+ 0.018464767456054687,
83
+ 0.018486272811889647,
84
+ 0.018510847091674804,
85
+ 0.01845964813232422,
86
+ 0.018529279708862305,
87
+ 0.018509824752807616,
88
+ 0.01841459274291992,
89
+ 0.018298879623413086,
90
+ 0.018324480056762696,
91
+ 0.01846886444091797,
92
+ 0.018748416900634765,
93
+ 0.019137535095214844,
94
+ 0.019017728805541992,
95
+ 0.019188735961914064,
96
+ 0.01929523277282715,
97
+ 0.018540544509887694,
98
+ 0.018530303955078126,
99
+ 0.018387968063354493,
100
+ 0.018487295150756835,
101
+ 0.018587648391723634,
102
+ 0.01848531150817871,
103
+ 0.01852409553527832,
104
+ 0.018553855895996094,
105
+ 0.018515968322753908,
106
+ 0.018520063400268554,
107
+ 0.018471935272216796,
108
+ 0.018737152099609376,
109
+ 0.018405376434326173,
110
+ 0.018319360733032225,
111
+ 0.018663423538208008,
112
+ 0.01857535934448242,
113
+ 0.018498559951782227,
114
+ 0.018535423278808593,
115
+ 0.018510847091674804,
116
+ 0.018491392135620118,
117
+ 0.018587648391723634,
118
+ 0.018686975479125977,
119
+ 0.018532352447509767,
120
+ 0.019281919479370118,
121
+ 0.018699264526367186,
122
+ 0.01859686470031738,
123
+ 0.018505727767944336,
124
+ 0.01843097686767578,
125
+ 0.018566144943237304,
126
+ 0.018495487213134765,
127
+ 0.018502656936645507,
128
+ 0.018966527938842775,
129
+ 0.01948569679260254,
130
+ 0.019645471572875977,
131
+ 0.01967407989501953,
132
+ 0.019325952529907226,
133
+ 0.01940275192260742,
134
+ 0.0187064323425293,
135
+ 0.01866035270690918,
136
+ 0.018929664611816405,
137
+ 0.019466239929199217,
138
+ 0.01857535934448242,
139
+ 0.01845964813232422,
140
+ 0.01843916893005371,
141
+ 0.01845043182373047,
142
+ 0.018510847091674804,
143
+ 0.018618368148803712,
144
+ 0.01865011215209961,
145
+ 0.01863372802734375,
146
+ 0.018663423538208008,
147
+ 0.018720767974853517,
148
+ 0.01859686470031738,
149
+ 0.01878118324279785,
150
+ 0.01863577651977539,
151
+ 0.018691072463989256,
152
+ 0.01866547203063965,
153
+ 0.019335168838500977,
154
+ 0.019352575302124024,
155
+ 0.019316736221313476,
156
+ 0.019578880310058593,
157
+ 0.019360767364501954,
158
+ 0.019361791610717775,
159
+ 0.019343360900878907,
160
+ 0.018675712585449217,
161
+ 0.01844633674621582,
162
+ 0.01844428825378418,
163
+ 0.01883033561706543,
164
+ 0.018907136917114258,
165
+ 0.018997247695922852,
166
+ 0.018457599639892578,
167
+ 0.018456575393676757,
168
+ 0.018488319396972656,
169
+ 0.018569215774536133,
170
+ 0.01864192008972168,
171
+ 0.018651136398315428,
172
+ 0.018618368148803712,
173
+ 0.019882080078125,
174
+ 0.019469215393066407,
175
+ 0.018634752273559572,
176
+ 0.01865216064453125,
177
+ 0.018662399291992187,
178
+ 0.018678783416748047,
179
+ 0.018745344161987306,
180
+ 0.018611200332641603,
181
+ 0.018716672897338867,
182
+ 0.01862553596496582,
183
+ 0.018732032775878905,
184
+ 0.01945292854309082,
185
+ 0.018685951232910156,
186
+ 0.01863577651977539,
187
+ 0.01864806365966797,
188
+ 0.01923072052001953,
189
+ 0.019526655197143555,
190
+ 0.020535295486450195,
191
+ 0.019017728805541992,
192
+ 0.018740224838256835,
193
+ 0.01877299118041992,
194
+ 0.018661376953125,
195
+ 0.019096576690673828,
196
+ 0.019260416030883788,
197
+ 0.018651136398315428,
198
+ 0.018742271423339844,
199
+ 0.018725887298583984,
200
+ 0.01869824028015137,
201
+ 0.018655231475830078,
202
+ 0.018689023971557618,
203
+ 0.03800985717773438,
204
+ 0.018487295150756835,
205
+ 0.018457599639892578,
206
+ 0.018562047958374024,
207
+ 0.018456575393676757,
208
+ 0.018499584197998048,
209
+ 0.018694143295288086,
210
+ 0.018505727767944336,
211
+ 0.01849555206298828,
212
+ 0.01854867172241211,
213
+ 0.018491455078125,
214
+ 0.01854867172241211,
215
+ 0.018719743728637696,
216
+ 0.01849241638183594,
217
+ 0.018353151321411132,
218
+ 0.01836953544616699,
219
+ 0.018297855377197265,
220
+ 0.01836031913757324,
221
+ 0.018339839935302735,
222
+ 0.018363391876220703,
223
+ 0.018329599380493163,
224
+ 0.018342912673950194,
225
+ 0.01862041664123535,
226
+ 0.019191808700561523,
227
+ 0.01903104019165039,
228
+ 0.019002368927001953,
229
+ 0.01908531188964844,
230
+ 0.01904128074645996,
231
+ 0.019162111282348633,
232
+ 0.018985984802246093,
233
+ 0.018965503692626954,
234
+ 0.01901158332824707,
235
+ 0.018751487731933594,
236
+ 0.01845964813232422,
237
+ 0.018540544509887694,
238
+ 0.01843507194519043,
239
+ 0.018453504562377928,
240
+ 0.01846784019470215,
241
+ 0.018319360733032225,
242
+ 0.01864089584350586,
243
+ 0.019142656326293944,
244
+ 0.018729984283447267,
245
+ 0.01905971145629883,
246
+ 0.01906790351867676,
247
+ 0.018696191787719727,
248
+ 0.01875660705566406,
249
+ 0.019056640625,
250
+ 0.019099647521972657,
251
+ 0.018824192047119142,
252
+ 0.018364416122436524,
253
+ 0.018318336486816408,
254
+ 0.01835212707519531,
255
+ 0.018510847091674804,
256
+ 0.01847603225708008,
257
+ 0.01847500801086426,
258
+ 0.018472959518432617,
259
+ 0.01846784019470215,
260
+ 0.01867263984680176,
261
+ 0.01865318489074707,
262
+ 0.019129344940185547,
263
+ 0.018734079360961914,
264
+ 0.01862348747253418,
265
+ 0.018684928894042968,
266
+ 0.018671615600585938,
267
+ 0.018743295669555664,
268
+ 0.020139007568359374,
269
+ 0.01940275192260742,
270
+ 0.018686975479125977,
271
+ 0.01862860870361328,
272
+ 0.01863065528869629,
273
+ 0.018661376953125,
274
+ 0.01869004821777344,
275
+ 0.018654207229614257,
276
+ 0.018694143295288086,
277
+ 0.018491392135620118,
278
+ 0.018715648651123046,
279
+ 0.01984102439880371,
280
+ 0.0194334716796875,
281
+ 0.019323904037475585,
282
+ 0.01942630386352539,
283
+ 0.019312639236450196,
284
+ 0.019174400329589843,
285
+ 0.01923686408996582,
286
+ 0.019581951141357423,
287
+ 0.01979084777832031,
288
+ 0.018968576431274413,
289
+ 0.019305471420288087,
290
+ 0.019354623794555666,
291
+ 0.01922867202758789,
292
+ 0.018917375564575196,
293
+ 0.01864806365966797,
294
+ 0.01861222457885742,
295
+ 0.018755584716796874,
296
+ 0.01869004821777344,
297
+ 0.018646015167236327,
298
+ 0.018761728286743166,
299
+ 0.018711551666259766,
300
+ 0.018651136398315428,
301
+ 0.018712575912475587,
302
+ 0.01862348747253418,
303
+ 0.018704383850097657,
304
+ 0.018597888946533202,
305
+ 0.018594816207885743,
306
+ 0.01864908790588379,
307
+ 0.018659328460693358,
308
+ 0.01865830421447754,
309
+ 0.018695167541503906,
310
+ 0.018758655548095703,
311
+ 0.018973695755004884,
312
+ 0.019306495666503908,
313
+ 0.01923481559753418,
314
+ 0.019203071594238282,
315
+ 0.01863680076599121,
316
+ 0.01862144088745117,
317
+ 0.01859174346923828,
318
+ 0.018480127334594726,
319
+ 0.018493440628051756,
320
+ 0.018740224838256835,
321
+ 0.01920102310180664,
322
+ 0.0191016960144043,
323
+ 0.018726911544799805,
324
+ 0.018461696624755858,
325
+ 0.018529279708862305,
326
+ 0.01844121551513672,
327
+ 0.018662399291992187,
328
+ 0.01868083190917969,
329
+ 0.01864499282836914
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 53.029326202941526
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022692079544067385,
7
+ "prefill.throughput.value": 308.47767770275067,
8
+ "decode.latency.mean": 2.385437255859375,
9
+ "decode.throughput.value": 53.23971514574468,
10
+ "per_token.latency.mean": 0.018857490215376905,
11
+ "per_token.throughput.value": 53.029326202941526
12
+ }
13
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: 25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
88
+ - hydra.run.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [ISOLATED-PROCESS][2024-06-09 02:37:24,141][process][ERROR] - + Sending traceback to main process
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: 25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
88
+ - hydra.run.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1763.38944,
109
+ "max_global_vram": 6793.199616,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 6146.752512,
112
+ "max_allocated": 5036.936192
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.043408096313476564,
118
+ "mean": 0.021704048156738282,
119
+ "stdev": 0.00020817565917968794,
120
+ "p50": 0.021704048156738282,
121
+ "p90": 0.021870588684082034,
122
+ "p95": 0.021891406250000002,
123
+ "p99": 0.021908060302734376,
124
+ "values": [
125
+ 0.02191222381591797,
126
+ 0.021495872497558594
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 322.52047864291
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 1763.401728,
140
+ "max_global_vram": 6793.199616,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 6146.752512,
143
+ "max_allocated": 5036.936704
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 5.184416259765625,
149
+ "mean": 2.5922081298828124,
150
+ "stdev": 0.002266479492187612,
151
+ "p50": 2.5922081298828124,
152
+ "p90": 2.5940213134765626,
153
+ "p95": 2.5942479614257814,
154
+ "p99": 2.5944292797851563,
155
+ "values": [
156
+ 2.589941650390625,
157
+ 2.594474609375
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 48.99297958985314
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 5.185601533889769,
173
+ "mean": 0.020496448750552454,
174
+ "stdev": 0.001512014375107108,
175
+ "p50": 0.020572160720825194,
176
+ "p90": 0.020751974868774416,
177
+ "p95": 0.020997119903564454,
178
+ "p99": 0.022830817642211898,
179
+ "values": [
180
+ 0.02065407943725586,
181
+ 0.021000192642211913,
182
+ 0.020497407913208008,
183
+ 0.020410367965698242,
184
+ 0.02046463966369629,
185
+ 0.020350976943969725,
186
+ 0.02105548858642578,
187
+ 0.02065203285217285,
188
+ 0.020591615676879883,
189
+ 0.02031001663208008,
190
+ 0.020496383666992187,
191
+ 0.02048102378845215,
192
+ 0.02021683120727539,
193
+ 0.020540416717529295,
194
+ 0.020544511795043945,
195
+ 0.020539392471313478,
196
+ 0.02027212715148926,
197
+ 0.020220928192138672,
198
+ 0.020331520080566406,
199
+ 0.020358144760131838,
200
+ 0.02010419273376465,
201
+ 0.020139007568359374,
202
+ 0.02022707176208496,
203
+ 0.020168703079223634,
204
+ 0.020295679092407228,
205
+ 0.02023423957824707,
206
+ 0.02045644760131836,
207
+ 0.020641792297363282,
208
+ 0.02058137512207031,
209
+ 0.020248575210571287,
210
+ 0.02059775924682617,
211
+ 0.020368383407592772,
212
+ 0.020340736389160157,
213
+ 0.020207616806030275,
214
+ 0.02069811248779297,
215
+ 0.02086809539794922,
216
+ 0.02655232048034668,
217
+ 0.021336063385009766,
218
+ 0.020794368743896483,
219
+ 0.02109644889831543,
220
+ 0.021176319122314453,
221
+ 0.02062745666503906,
222
+ 0.020692991256713866,
223
+ 0.022148096084594726,
224
+ 0.023570432662963867,
225
+ 0.02123776054382324,
226
+ 0.020995071411132812,
227
+ 0.0208353271484375,
228
+ 0.020720640182495118,
229
+ 0.02062131118774414,
230
+ 0.020746240615844725,
231
+ 0.02062131118774414,
232
+ 0.020753408432006838,
233
+ 0.020619264602661135,
234
+ 0.02064588737487793,
235
+ 0.02068992042541504,
236
+ 0.02069196891784668,
237
+ 0.02066739273071289,
238
+ 0.0206878719329834,
239
+ 0.020816896438598635,
240
+ 0.02017791938781738,
241
+ 0.02004787254333496,
242
+ 0.019791872024536132,
243
+ 0.019763200759887696,
244
+ 0.019927040100097656,
245
+ 0.019727359771728514,
246
+ 0.019764223098754884,
247
+ 0.019785728454589844,
248
+ 0.019770368576049805,
249
+ 0.019729408264160156,
250
+ 0.019794944763183595,
251
+ 0.019778560638427735,
252
+ 0.01982771110534668,
253
+ 0.019710975646972655,
254
+ 0.01979084777832031,
255
+ 0.019720191955566405,
256
+ 0.01983180809020996,
257
+ 0.02089574432373047,
258
+ 0.020620288848876952,
259
+ 0.020602880477905275,
260
+ 0.02060697555541992,
261
+ 0.02041753578186035,
262
+ 0.020767744064331056,
263
+ 0.019778560638427735,
264
+ 0.020602880477905275,
265
+ 0.020145151138305666,
266
+ 0.020538368225097657,
267
+ 0.020548608779907225,
268
+ 0.020624383926391602,
269
+ 0.020184064865112306,
270
+ 0.019750911712646483,
271
+ 0.019711999893188475,
272
+ 0.019770368576049805,
273
+ 0.019719167709350584,
274
+ 0.019736576080322265,
275
+ 0.019741695404052736,
276
+ 0.019719167709350584,
277
+ 0.01986867141723633,
278
+ 0.019791872024536132,
279
+ 0.01971609687805176,
280
+ 0.019745792388916016,
281
+ 0.019743743896484374,
282
+ 0.019771392822265626,
283
+ 0.02011238479614258,
284
+ 0.020701183319091796,
285
+ 0.02064076805114746,
286
+ 0.020607999801635742,
287
+ 0.020542463302612304,
288
+ 0.020987903594970703,
289
+ 0.02168115234375,
290
+ 0.02080460739135742,
291
+ 0.020557823181152343,
292
+ 0.020601856231689454,
293
+ 0.019742719650268553,
294
+ 0.019770368576049805,
295
+ 0.019730432510375977,
296
+ 0.019747840881347657,
297
+ 0.019781631469726564,
298
+ 0.019772415161132813,
299
+ 0.019744768142700195,
300
+ 0.01990656089782715,
301
+ 0.019734527587890623,
302
+ 0.019732479095458985,
303
+ 0.019708927154541016,
304
+ 0.019717119216918946,
305
+ 0.019698688507080078,
306
+ 0.042347518920898435,
307
+ 0.01982054328918457,
308
+ 0.019703807830810546,
309
+ 0.020389888763427736,
310
+ 0.020534271240234374,
311
+ 0.020634624481201173,
312
+ 0.019793920516967774,
313
+ 0.020489215850830078,
314
+ 0.019524608612060547,
315
+ 0.019561471939086913,
316
+ 0.019937280654907227,
317
+ 0.019809280395507813,
318
+ 0.019711999893188475,
319
+ 0.02021683120727539,
320
+ 0.020575231552124023,
321
+ 0.02062131118774414,
322
+ 0.020544511795043945,
323
+ 0.020641792297363282,
324
+ 0.02062950325012207,
325
+ 0.020603904724121092,
326
+ 0.02062131118774414,
327
+ 0.020609024047851563,
328
+ 0.020587520599365236,
329
+ 0.020668415069580077,
330
+ 0.020649984359741212,
331
+ 0.020612096786499022,
332
+ 0.020620288848876952,
333
+ 0.020599807739257812,
334
+ 0.02065203285217285,
335
+ 0.02059775924682617,
336
+ 0.020584447860717774,
337
+ 0.02061414337158203,
338
+ 0.02065510368347168,
339
+ 0.02066227149963379,
340
+ 0.020536319732666015,
341
+ 0.020619264602661135,
342
+ 0.0206059513092041,
343
+ 0.020533248901367186,
344
+ 0.02061516761779785,
345
+ 0.020544511795043945,
346
+ 0.020555776596069338,
347
+ 0.02062335968017578,
348
+ 0.020420608520507814,
349
+ 0.02062950325012207,
350
+ 0.020602880477905275,
351
+ 0.020668415069580077,
352
+ 0.020582399368286132,
353
+ 0.02065920066833496,
354
+ 0.020661247253417968,
355
+ 0.020732927322387695,
356
+ 0.020593664169311524,
357
+ 0.020572160720825194,
358
+ 0.020714496612548827,
359
+ 0.020866111755371095,
360
+ 0.020652992248535156,
361
+ 0.02067046356201172,
362
+ 0.020633600234985353,
363
+ 0.02067148780822754,
364
+ 0.020736000061035157,
365
+ 0.02085171127319336,
366
+ 0.020501504898071288,
367
+ 0.02063052749633789,
368
+ 0.02066534423828125,
369
+ 0.020612096786499022,
370
+ 0.020591615676879883,
371
+ 0.020624383926391602,
372
+ 0.02044108772277832,
373
+ 0.020599807739257812,
374
+ 0.02064384078979492,
375
+ 0.02062131118774414,
376
+ 0.020685823440551757,
377
+ 0.02064691162109375,
378
+ 0.020588544845581053,
379
+ 0.020686847686767578,
380
+ 0.020719615936279297,
381
+ 0.020710399627685547,
382
+ 0.020690944671630858,
383
+ 0.020583423614501953,
384
+ 0.020588544845581053,
385
+ 0.01987276840209961,
386
+ 0.019733503341674806,
387
+ 0.019776512145996093,
388
+ 0.019786752700805665,
389
+ 0.0198154239654541,
390
+ 0.019741695404052736,
391
+ 0.019802112579345704,
392
+ 0.021840896606445313,
393
+ 0.01988812828063965,
394
+ 0.019746816635131836,
395
+ 0.019792896270751953,
396
+ 0.019734527587890623,
397
+ 0.01982156753540039,
398
+ 0.019792896270751953,
399
+ 0.019761152267456054,
400
+ 0.019734527587890623,
401
+ 0.019735551834106444,
402
+ 0.019701759338378907,
403
+ 0.019772415161132813,
404
+ 0.01982771110534668,
405
+ 0.019753984451293945,
406
+ 0.019757055282592775,
407
+ 0.019803136825561524,
408
+ 0.019730432510375977,
409
+ 0.020649984359741212,
410
+ 0.021424127578735352,
411
+ 0.020900863647460938,
412
+ 0.020674560546875,
413
+ 0.020706304550170897,
414
+ 0.020549631118774413,
415
+ 0.0206878719329834,
416
+ 0.020588544845581053,
417
+ 0.02069606399536133,
418
+ 0.020596736907958983,
419
+ 0.020742143630981445,
420
+ 0.02063564872741699,
421
+ 0.02037555122375488,
422
+ 0.020618240356445314,
423
+ 0.020684799194335936,
424
+ 0.020742143630981445,
425
+ 0.02066534423828125,
426
+ 0.020602880477905275,
427
+ 0.020529151916503906,
428
+ 0.02062335968017578,
429
+ 0.020566015243530272,
430
+ 0.0206059513092041,
431
+ 0.02063564872741699,
432
+ 0.020566015243530272
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 48.78893959486744
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1763.38944,
6
+ "max_global_vram": 6793.199616,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6146.752512,
9
+ "max_allocated": 5036.936192
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.043408096313476564,
15
+ "mean": 0.021704048156738282,
16
+ "stdev": 0.00020817565917968794,
17
+ "p50": 0.021704048156738282,
18
+ "p90": 0.021870588684082034,
19
+ "p95": 0.021891406250000002,
20
+ "p99": 0.021908060302734376,
21
+ "values": [
22
+ 0.02191222381591797,
23
+ 0.021495872497558594
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 322.52047864291
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 1763.401728,
37
+ "max_global_vram": 6793.199616,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 6146.752512,
40
+ "max_allocated": 5036.936704
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 5.184416259765625,
46
+ "mean": 2.5922081298828124,
47
+ "stdev": 0.002266479492187612,
48
+ "p50": 2.5922081298828124,
49
+ "p90": 2.5940213134765626,
50
+ "p95": 2.5942479614257814,
51
+ "p99": 2.5944292797851563,
52
+ "values": [
53
+ 2.589941650390625,
54
+ 2.594474609375
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 48.99297958985314
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 5.185601533889769,
70
+ "mean": 0.020496448750552454,
71
+ "stdev": 0.001512014375107108,
72
+ "p50": 0.020572160720825194,
73
+ "p90": 0.020751974868774416,
74
+ "p95": 0.020997119903564454,
75
+ "p99": 0.022830817642211898,
76
+ "values": [
77
+ 0.02065407943725586,
78
+ 0.021000192642211913,
79
+ 0.020497407913208008,
80
+ 0.020410367965698242,
81
+ 0.02046463966369629,
82
+ 0.020350976943969725,
83
+ 0.02105548858642578,
84
+ 0.02065203285217285,
85
+ 0.020591615676879883,
86
+ 0.02031001663208008,
87
+ 0.020496383666992187,
88
+ 0.02048102378845215,
89
+ 0.02021683120727539,
90
+ 0.020540416717529295,
91
+ 0.020544511795043945,
92
+ 0.020539392471313478,
93
+ 0.02027212715148926,
94
+ 0.020220928192138672,
95
+ 0.020331520080566406,
96
+ 0.020358144760131838,
97
+ 0.02010419273376465,
98
+ 0.020139007568359374,
99
+ 0.02022707176208496,
100
+ 0.020168703079223634,
101
+ 0.020295679092407228,
102
+ 0.02023423957824707,
103
+ 0.02045644760131836,
104
+ 0.020641792297363282,
105
+ 0.02058137512207031,
106
+ 0.020248575210571287,
107
+ 0.02059775924682617,
108
+ 0.020368383407592772,
109
+ 0.020340736389160157,
110
+ 0.020207616806030275,
111
+ 0.02069811248779297,
112
+ 0.02086809539794922,
113
+ 0.02655232048034668,
114
+ 0.021336063385009766,
115
+ 0.020794368743896483,
116
+ 0.02109644889831543,
117
+ 0.021176319122314453,
118
+ 0.02062745666503906,
119
+ 0.020692991256713866,
120
+ 0.022148096084594726,
121
+ 0.023570432662963867,
122
+ 0.02123776054382324,
123
+ 0.020995071411132812,
124
+ 0.0208353271484375,
125
+ 0.020720640182495118,
126
+ 0.02062131118774414,
127
+ 0.020746240615844725,
128
+ 0.02062131118774414,
129
+ 0.020753408432006838,
130
+ 0.020619264602661135,
131
+ 0.02064588737487793,
132
+ 0.02068992042541504,
133
+ 0.02069196891784668,
134
+ 0.02066739273071289,
135
+ 0.0206878719329834,
136
+ 0.020816896438598635,
137
+ 0.02017791938781738,
138
+ 0.02004787254333496,
139
+ 0.019791872024536132,
140
+ 0.019763200759887696,
141
+ 0.019927040100097656,
142
+ 0.019727359771728514,
143
+ 0.019764223098754884,
144
+ 0.019785728454589844,
145
+ 0.019770368576049805,
146
+ 0.019729408264160156,
147
+ 0.019794944763183595,
148
+ 0.019778560638427735,
149
+ 0.01982771110534668,
150
+ 0.019710975646972655,
151
+ 0.01979084777832031,
152
+ 0.019720191955566405,
153
+ 0.01983180809020996,
154
+ 0.02089574432373047,
155
+ 0.020620288848876952,
156
+ 0.020602880477905275,
157
+ 0.02060697555541992,
158
+ 0.02041753578186035,
159
+ 0.020767744064331056,
160
+ 0.019778560638427735,
161
+ 0.020602880477905275,
162
+ 0.020145151138305666,
163
+ 0.020538368225097657,
164
+ 0.020548608779907225,
165
+ 0.020624383926391602,
166
+ 0.020184064865112306,
167
+ 0.019750911712646483,
168
+ 0.019711999893188475,
169
+ 0.019770368576049805,
170
+ 0.019719167709350584,
171
+ 0.019736576080322265,
172
+ 0.019741695404052736,
173
+ 0.019719167709350584,
174
+ 0.01986867141723633,
175
+ 0.019791872024536132,
176
+ 0.01971609687805176,
177
+ 0.019745792388916016,
178
+ 0.019743743896484374,
179
+ 0.019771392822265626,
180
+ 0.02011238479614258,
181
+ 0.020701183319091796,
182
+ 0.02064076805114746,
183
+ 0.020607999801635742,
184
+ 0.020542463302612304,
185
+ 0.020987903594970703,
186
+ 0.02168115234375,
187
+ 0.02080460739135742,
188
+ 0.020557823181152343,
189
+ 0.020601856231689454,
190
+ 0.019742719650268553,
191
+ 0.019770368576049805,
192
+ 0.019730432510375977,
193
+ 0.019747840881347657,
194
+ 0.019781631469726564,
195
+ 0.019772415161132813,
196
+ 0.019744768142700195,
197
+ 0.01990656089782715,
198
+ 0.019734527587890623,
199
+ 0.019732479095458985,
200
+ 0.019708927154541016,
201
+ 0.019717119216918946,
202
+ 0.019698688507080078,
203
+ 0.042347518920898435,
204
+ 0.01982054328918457,
205
+ 0.019703807830810546,
206
+ 0.020389888763427736,
207
+ 0.020534271240234374,
208
+ 0.020634624481201173,
209
+ 0.019793920516967774,
210
+ 0.020489215850830078,
211
+ 0.019524608612060547,
212
+ 0.019561471939086913,
213
+ 0.019937280654907227,
214
+ 0.019809280395507813,
215
+ 0.019711999893188475,
216
+ 0.02021683120727539,
217
+ 0.020575231552124023,
218
+ 0.02062131118774414,
219
+ 0.020544511795043945,
220
+ 0.020641792297363282,
221
+ 0.02062950325012207,
222
+ 0.020603904724121092,
223
+ 0.02062131118774414,
224
+ 0.020609024047851563,
225
+ 0.020587520599365236,
226
+ 0.020668415069580077,
227
+ 0.020649984359741212,
228
+ 0.020612096786499022,
229
+ 0.020620288848876952,
230
+ 0.020599807739257812,
231
+ 0.02065203285217285,
232
+ 0.02059775924682617,
233
+ 0.020584447860717774,
234
+ 0.02061414337158203,
235
+ 0.02065510368347168,
236
+ 0.02066227149963379,
237
+ 0.020536319732666015,
238
+ 0.020619264602661135,
239
+ 0.0206059513092041,
240
+ 0.020533248901367186,
241
+ 0.02061516761779785,
242
+ 0.020544511795043945,
243
+ 0.020555776596069338,
244
+ 0.02062335968017578,
245
+ 0.020420608520507814,
246
+ 0.02062950325012207,
247
+ 0.020602880477905275,
248
+ 0.020668415069580077,
249
+ 0.020582399368286132,
250
+ 0.02065920066833496,
251
+ 0.020661247253417968,
252
+ 0.020732927322387695,
253
+ 0.020593664169311524,
254
+ 0.020572160720825194,
255
+ 0.020714496612548827,
256
+ 0.020866111755371095,
257
+ 0.020652992248535156,
258
+ 0.02067046356201172,
259
+ 0.020633600234985353,
260
+ 0.02067148780822754,
261
+ 0.020736000061035157,
262
+ 0.02085171127319336,
263
+ 0.020501504898071288,
264
+ 0.02063052749633789,
265
+ 0.02066534423828125,
266
+ 0.020612096786499022,
267
+ 0.020591615676879883,
268
+ 0.020624383926391602,
269
+ 0.02044108772277832,
270
+ 0.020599807739257812,
271
+ 0.02064384078979492,
272
+ 0.02062131118774414,
273
+ 0.020685823440551757,
274
+ 0.02064691162109375,
275
+ 0.020588544845581053,
276
+ 0.020686847686767578,
277
+ 0.020719615936279297,
278
+ 0.020710399627685547,
279
+ 0.020690944671630858,
280
+ 0.020583423614501953,
281
+ 0.020588544845581053,
282
+ 0.01987276840209961,
283
+ 0.019733503341674806,
284
+ 0.019776512145996093,
285
+ 0.019786752700805665,
286
+ 0.0198154239654541,
287
+ 0.019741695404052736,
288
+ 0.019802112579345704,
289
+ 0.021840896606445313,
290
+ 0.01988812828063965,
291
+ 0.019746816635131836,
292
+ 0.019792896270751953,
293
+ 0.019734527587890623,
294
+ 0.01982156753540039,
295
+ 0.019792896270751953,
296
+ 0.019761152267456054,
297
+ 0.019734527587890623,
298
+ 0.019735551834106444,
299
+ 0.019701759338378907,
300
+ 0.019772415161132813,
301
+ 0.01982771110534668,
302
+ 0.019753984451293945,
303
+ 0.019757055282592775,
304
+ 0.019803136825561524,
305
+ 0.019730432510375977,
306
+ 0.020649984359741212,
307
+ 0.021424127578735352,
308
+ 0.020900863647460938,
309
+ 0.020674560546875,
310
+ 0.020706304550170897,
311
+ 0.020549631118774413,
312
+ 0.0206878719329834,
313
+ 0.020588544845581053,
314
+ 0.02069606399536133,
315
+ 0.020596736907958983,
316
+ 0.020742143630981445,
317
+ 0.02063564872741699,
318
+ 0.02037555122375488,
319
+ 0.020618240356445314,
320
+ 0.020684799194335936,
321
+ 0.020742143630981445,
322
+ 0.02066534423828125,
323
+ 0.020602880477905275,
324
+ 0.020529151916503906,
325
+ 0.02062335968017578,
326
+ 0.020566015243530272,
327
+ 0.0206059513092041,
328
+ 0.02063564872741699,
329
+ 0.020566015243530272
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 48.78893959486744
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.021704048156738282,
7
+ "prefill.throughput.value": 322.52047864291,
8
+ "decode.latency.mean": 2.5922081298828124,
9
+ "decode.throughput.value": 48.99297958985314,
10
+ "per_token.latency.mean": 0.020496448750552454,
11
+ "per_token.throughput.value": 48.78893959486744
12
+ }
13
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: 25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
88
+ - hydra.run.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1926.950912,
109
+ "max_global_vram": 6094.848,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 5379.19488,
112
+ "max_allocated": 5021.251584
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.02810588836669922,
118
+ "mean": 0.01405294418334961,
119
+ "stdev": 0.00012724781036376961,
120
+ "p50": 0.01405294418334961,
121
+ "p90": 0.014154742431640626,
122
+ "p95": 0.014167467212677002,
123
+ "p99": 0.014177647037506104,
124
+ "values": [
125
+ 0.014180191993713379,
126
+ 0.01392569637298584
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 498.1162600997043
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 2270.900224,
140
+ "max_global_vram": 8907.128832,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 5379.19488,
143
+ "max_allocated": 5027.089408
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 3.1168135986328127,
149
+ "mean": 1.5584067993164064,
150
+ "stdev": 0.0010114135742187447,
151
+ "p50": 1.5584067993164064,
152
+ "p90": 1.5592159301757813,
153
+ "p95": 1.5593170715332032,
154
+ "p99": 1.5593979846191408,
155
+ "values": [
156
+ 1.559418212890625,
157
+ 1.5573953857421876
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 81.49348427875726
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 3.1178332242965703,
173
+ "mean": 0.012323451479433083,
174
+ "stdev": 0.0008897335078486659,
175
+ "p50": 0.01226137638092041,
176
+ "p90": 0.012298854446411133,
177
+ "p95": 0.012316467094421387,
178
+ "p99": 0.012717056045532226,
179
+ "values": [
180
+ 0.012692480087280274,
181
+ 0.01275391960144043,
182
+ 0.012743680000305176,
183
+ 0.012317695617675782,
184
+ 0.012254207611083985,
185
+ 0.012332032203674317,
186
+ 0.012283904075622559,
187
+ 0.012326975822448731,
188
+ 0.012251071929931641,
189
+ 0.012297216415405274,
190
+ 0.012315648078918457,
191
+ 0.012327936172485352,
192
+ 0.01225011157989502,
193
+ 0.012309503555297852,
194
+ 0.012266495704650878,
195
+ 0.012314623832702636,
196
+ 0.012246015548706055,
197
+ 0.01226956844329834,
198
+ 0.012254207611083985,
199
+ 0.012281855583190919,
200
+ 0.01233510398864746,
201
+ 0.012299263954162597,
202
+ 0.012278783798217773,
203
+ 0.012301312446594239,
204
+ 0.01226035213470459,
205
+ 0.01229312038421631,
206
+ 0.01224396800994873,
207
+ 0.012283904075622559,
208
+ 0.012233728408813477,
209
+ 0.012303359985351562,
210
+ 0.01225216007232666,
211
+ 0.012263423919677734,
212
+ 0.01222867202758789,
213
+ 0.012272576332092285,
214
+ 0.012221440315246582,
215
+ 0.012274687767028808,
216
+ 0.012231679916381836,
217
+ 0.012285951614379884,
218
+ 0.012227616310119628,
219
+ 0.012268511772155762,
220
+ 0.012216320037841797,
221
+ 0.012259327888488769,
222
+ 0.012225536346435547,
223
+ 0.012257280349731446,
224
+ 0.012223487854003906,
225
+ 0.012289024353027344,
226
+ 0.01225830364227295,
227
+ 0.012283904075622559,
228
+ 0.012233728408813477,
229
+ 0.012264448165893555,
230
+ 0.012232704162597656,
231
+ 0.012296192169189453,
232
+ 0.012229632377624512,
233
+ 0.012277759552001954,
234
+ 0.012231679916381836,
235
+ 0.012264448165893555,
236
+ 0.012226559638977052,
237
+ 0.012271615982055664,
238
+ 0.01223475170135498,
239
+ 0.012263423919677734,
240
+ 0.012233728408813477,
241
+ 0.012291071891784668,
242
+ 0.012223487854003906,
243
+ 0.01225830364227295,
244
+ 0.012215295791625976,
245
+ 0.012277759552001954,
246
+ 0.012223487854003906,
247
+ 0.012275712013244629,
248
+ 0.012228608131408691,
249
+ 0.012280832290649414,
250
+ 0.012230655670166016,
251
+ 0.012273664474487305,
252
+ 0.012232704162597656,
253
+ 0.012289024353027344,
254
+ 0.01223475170135498,
255
+ 0.012266495704650878,
256
+ 0.012224512100219726,
257
+ 0.012283904075622559,
258
+ 0.0122357759475708,
259
+ 0.012278783798217773,
260
+ 0.012214271545410157,
261
+ 0.012285951614379884,
262
+ 0.012239871978759765,
263
+ 0.012281855583190919,
264
+ 0.012248064041137695,
265
+ 0.012286975860595703,
266
+ 0.012244992256164551,
267
+ 0.012263423919677734,
268
+ 0.012218367576599122,
269
+ 0.012265472412109376,
270
+ 0.012216320037841797,
271
+ 0.01225830364227295,
272
+ 0.012239871978759765,
273
+ 0.012283904075622559,
274
+ 0.012233728408813477,
275
+ 0.012315648078918457,
276
+ 0.0122357759475708,
277
+ 0.012271615982055664,
278
+ 0.01224294376373291,
279
+ 0.01229312038421631,
280
+ 0.012236800193786621,
281
+ 0.012257280349731446,
282
+ 0.012254207611083985,
283
+ 0.012286975860595703,
284
+ 0.012270591735839843,
285
+ 0.012289024353027344,
286
+ 0.012224512100219726,
287
+ 0.012301312446594239,
288
+ 0.012253184318542481,
289
+ 0.012255231857299804,
290
+ 0.012236800193786621,
291
+ 0.01229312038421631,
292
+ 0.012273664474487305,
293
+ 0.01227673625946045,
294
+ 0.012230655670166016,
295
+ 0.01226854419708252,
296
+ 0.012226559638977052,
297
+ 0.01226035213470459,
298
+ 0.012222463607788087,
299
+ 0.012282879829406738,
300
+ 0.012241920471191407,
301
+ 0.012272640228271485,
302
+ 0.012238847732543945,
303
+ 0.012297216415405274,
304
+ 0.012254240036010743,
305
+ 0.012283871650695801,
306
+ 0.026412031173706055,
307
+ 0.01222758388519287,
308
+ 0.012278783798217773,
309
+ 0.01222758388519287,
310
+ 0.012304384231567383,
311
+ 0.012241920471191407,
312
+ 0.012279808044433594,
313
+ 0.012224512100219726,
314
+ 0.012278783798217773,
315
+ 0.012216320037841797,
316
+ 0.01227673625946045,
317
+ 0.012236800193786621,
318
+ 0.012559359550476074,
319
+ 0.012238847732543945,
320
+ 0.012264448165893555,
321
+ 0.012233728408813477,
322
+ 0.012288000106811523,
323
+ 0.012239871978759765,
324
+ 0.012274687767028808,
325
+ 0.012257280349731446,
326
+ 0.012278783798217773,
327
+ 0.01224396800994873,
328
+ 0.012273664474487305,
329
+ 0.01223475170135498,
330
+ 0.012295167922973632,
331
+ 0.012264448165893555,
332
+ 0.01226854419708252,
333
+ 0.012223487854003906,
334
+ 0.012277759552001954,
335
+ 0.012256256103515625,
336
+ 0.012262399673461915,
337
+ 0.012220447540283202,
338
+ 0.012261568069458008,
339
+ 0.012232512474060058,
340
+ 0.012286944389343262,
341
+ 0.01224396800994873,
342
+ 0.01225830364227295,
343
+ 0.01221939182281494,
344
+ 0.012277759552001954,
345
+ 0.012220416069030762,
346
+ 0.012271615982055664,
347
+ 0.01225216007232666,
348
+ 0.01229312038421631,
349
+ 0.012241920471191407,
350
+ 0.01226137638092041,
351
+ 0.012224512100219726,
352
+ 0.012289024353027344,
353
+ 0.01223475170135498,
354
+ 0.012271615982055664,
355
+ 0.012273664474487305,
356
+ 0.012294143676757812,
357
+ 0.012233728408813477,
358
+ 0.01227673625946045,
359
+ 0.012247039794921874,
360
+ 0.012305407524108887,
361
+ 0.012231807708740234,
362
+ 0.012311424255371094,
363
+ 0.012220416069030762,
364
+ 0.012393471717834472,
365
+ 0.01226035213470459,
366
+ 0.012280832290649414,
367
+ 0.012220416069030762,
368
+ 0.01226956844329834,
369
+ 0.012220416069030762,
370
+ 0.012265472412109376,
371
+ 0.012233728408813477,
372
+ 0.012272640228271485,
373
+ 0.012224512100219726,
374
+ 0.012271615982055664,
375
+ 0.01222758388519287,
376
+ 0.012277759552001954,
377
+ 0.012347392082214356,
378
+ 0.012498944282531739,
379
+ 0.01228492832183838,
380
+ 0.012302335739135742,
381
+ 0.012244992256164551,
382
+ 0.01228492832183838,
383
+ 0.012230655670166016,
384
+ 0.012272640228271485,
385
+ 0.012253184318542481,
386
+ 0.012274687767028808,
387
+ 0.012222623825073243,
388
+ 0.012265312194824218,
389
+ 0.012249088287353516,
390
+ 0.012297216415405274,
391
+ 0.01225216007232666,
392
+ 0.012265472412109376,
393
+ 0.012233728408813477,
394
+ 0.01226035213470459,
395
+ 0.012200960159301758,
396
+ 0.01226137638092041,
397
+ 0.012221440315246582,
398
+ 0.012273664474487305,
399
+ 0.01223475170135498,
400
+ 0.012288000106811523,
401
+ 0.012265472412109376,
402
+ 0.012262399673461915,
403
+ 0.012228608131408691,
404
+ 0.012273664474487305,
405
+ 0.012221440315246582,
406
+ 0.012273664474487305,
407
+ 0.012237824440002442,
408
+ 0.012253184318542481,
409
+ 0.012231679916381836,
410
+ 0.012263423919677734,
411
+ 0.01222758388519287,
412
+ 0.012259327888488769,
413
+ 0.012212224006652832,
414
+ 0.012264448165893555,
415
+ 0.012225536346435547,
416
+ 0.01226854419708252,
417
+ 0.012246015548706055,
418
+ 0.012259327888488769,
419
+ 0.012220416069030762,
420
+ 0.01225830364227295,
421
+ 0.012216320037841797,
422
+ 0.012262399673461915,
423
+ 0.012237824440002442,
424
+ 0.012280832290649414,
425
+ 0.012221440315246582,
426
+ 0.012270591735839843,
427
+ 0.012232704162597656,
428
+ 0.012271615982055664,
429
+ 0.012248064041137695,
430
+ 0.012279808044433594,
431
+ 0.012244992256164551,
432
+ 0.012300288200378418
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 81.14609788247432
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1926.950912,
6
+ "max_global_vram": 6094.848,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 5379.19488,
9
+ "max_allocated": 5021.251584
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.02810588836669922,
15
+ "mean": 0.01405294418334961,
16
+ "stdev": 0.00012724781036376961,
17
+ "p50": 0.01405294418334961,
18
+ "p90": 0.014154742431640626,
19
+ "p95": 0.014167467212677002,
20
+ "p99": 0.014177647037506104,
21
+ "values": [
22
+ 0.014180191993713379,
23
+ 0.01392569637298584
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 498.1162600997043
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 2270.900224,
37
+ "max_global_vram": 8907.128832,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 5379.19488,
40
+ "max_allocated": 5027.089408
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 3.1168135986328127,
46
+ "mean": 1.5584067993164064,
47
+ "stdev": 0.0010114135742187447,
48
+ "p50": 1.5584067993164064,
49
+ "p90": 1.5592159301757813,
50
+ "p95": 1.5593170715332032,
51
+ "p99": 1.5593979846191408,
52
+ "values": [
53
+ 1.559418212890625,
54
+ 1.5573953857421876
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 81.49348427875726
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 3.1178332242965703,
70
+ "mean": 0.012323451479433083,
71
+ "stdev": 0.0008897335078486659,
72
+ "p50": 0.01226137638092041,
73
+ "p90": 0.012298854446411133,
74
+ "p95": 0.012316467094421387,
75
+ "p99": 0.012717056045532226,
76
+ "values": [
77
+ 0.012692480087280274,
78
+ 0.01275391960144043,
79
+ 0.012743680000305176,
80
+ 0.012317695617675782,
81
+ 0.012254207611083985,
82
+ 0.012332032203674317,
83
+ 0.012283904075622559,
84
+ 0.012326975822448731,
85
+ 0.012251071929931641,
86
+ 0.012297216415405274,
87
+ 0.012315648078918457,
88
+ 0.012327936172485352,
89
+ 0.01225011157989502,
90
+ 0.012309503555297852,
91
+ 0.012266495704650878,
92
+ 0.012314623832702636,
93
+ 0.012246015548706055,
94
+ 0.01226956844329834,
95
+ 0.012254207611083985,
96
+ 0.012281855583190919,
97
+ 0.01233510398864746,
98
+ 0.012299263954162597,
99
+ 0.012278783798217773,
100
+ 0.012301312446594239,
101
+ 0.01226035213470459,
102
+ 0.01229312038421631,
103
+ 0.01224396800994873,
104
+ 0.012283904075622559,
105
+ 0.012233728408813477,
106
+ 0.012303359985351562,
107
+ 0.01225216007232666,
108
+ 0.012263423919677734,
109
+ 0.01222867202758789,
110
+ 0.012272576332092285,
111
+ 0.012221440315246582,
112
+ 0.012274687767028808,
113
+ 0.012231679916381836,
114
+ 0.012285951614379884,
115
+ 0.012227616310119628,
116
+ 0.012268511772155762,
117
+ 0.012216320037841797,
118
+ 0.012259327888488769,
119
+ 0.012225536346435547,
120
+ 0.012257280349731446,
121
+ 0.012223487854003906,
122
+ 0.012289024353027344,
123
+ 0.01225830364227295,
124
+ 0.012283904075622559,
125
+ 0.012233728408813477,
126
+ 0.012264448165893555,
127
+ 0.012232704162597656,
128
+ 0.012296192169189453,
129
+ 0.012229632377624512,
130
+ 0.012277759552001954,
131
+ 0.012231679916381836,
132
+ 0.012264448165893555,
133
+ 0.012226559638977052,
134
+ 0.012271615982055664,
135
+ 0.01223475170135498,
136
+ 0.012263423919677734,
137
+ 0.012233728408813477,
138
+ 0.012291071891784668,
139
+ 0.012223487854003906,
140
+ 0.01225830364227295,
141
+ 0.012215295791625976,
142
+ 0.012277759552001954,
143
+ 0.012223487854003906,
144
+ 0.012275712013244629,
145
+ 0.012228608131408691,
146
+ 0.012280832290649414,
147
+ 0.012230655670166016,
148
+ 0.012273664474487305,
149
+ 0.012232704162597656,
150
+ 0.012289024353027344,
151
+ 0.01223475170135498,
152
+ 0.012266495704650878,
153
+ 0.012224512100219726,
154
+ 0.012283904075622559,
155
+ 0.0122357759475708,
156
+ 0.012278783798217773,
157
+ 0.012214271545410157,
158
+ 0.012285951614379884,
159
+ 0.012239871978759765,
160
+ 0.012281855583190919,
161
+ 0.012248064041137695,
162
+ 0.012286975860595703,
163
+ 0.012244992256164551,
164
+ 0.012263423919677734,
165
+ 0.012218367576599122,
166
+ 0.012265472412109376,
167
+ 0.012216320037841797,
168
+ 0.01225830364227295,
169
+ 0.012239871978759765,
170
+ 0.012283904075622559,
171
+ 0.012233728408813477,
172
+ 0.012315648078918457,
173
+ 0.0122357759475708,
174
+ 0.012271615982055664,
175
+ 0.01224294376373291,
176
+ 0.01229312038421631,
177
+ 0.012236800193786621,
178
+ 0.012257280349731446,
179
+ 0.012254207611083985,
180
+ 0.012286975860595703,
181
+ 0.012270591735839843,
182
+ 0.012289024353027344,
183
+ 0.012224512100219726,
184
+ 0.012301312446594239,
185
+ 0.012253184318542481,
186
+ 0.012255231857299804,
187
+ 0.012236800193786621,
188
+ 0.01229312038421631,
189
+ 0.012273664474487305,
190
+ 0.01227673625946045,
191
+ 0.012230655670166016,
192
+ 0.01226854419708252,
193
+ 0.012226559638977052,
194
+ 0.01226035213470459,
195
+ 0.012222463607788087,
196
+ 0.012282879829406738,
197
+ 0.012241920471191407,
198
+ 0.012272640228271485,
199
+ 0.012238847732543945,
200
+ 0.012297216415405274,
201
+ 0.012254240036010743,
202
+ 0.012283871650695801,
203
+ 0.026412031173706055,
204
+ 0.01222758388519287,
205
+ 0.012278783798217773,
206
+ 0.01222758388519287,
207
+ 0.012304384231567383,
208
+ 0.012241920471191407,
209
+ 0.012279808044433594,
210
+ 0.012224512100219726,
211
+ 0.012278783798217773,
212
+ 0.012216320037841797,
213
+ 0.01227673625946045,
214
+ 0.012236800193786621,
215
+ 0.012559359550476074,
216
+ 0.012238847732543945,
217
+ 0.012264448165893555,
218
+ 0.012233728408813477,
219
+ 0.012288000106811523,
220
+ 0.012239871978759765,
221
+ 0.012274687767028808,
222
+ 0.012257280349731446,
223
+ 0.012278783798217773,
224
+ 0.01224396800994873,
225
+ 0.012273664474487305,
226
+ 0.01223475170135498,
227
+ 0.012295167922973632,
228
+ 0.012264448165893555,
229
+ 0.01226854419708252,
230
+ 0.012223487854003906,
231
+ 0.012277759552001954,
232
+ 0.012256256103515625,
233
+ 0.012262399673461915,
234
+ 0.012220447540283202,
235
+ 0.012261568069458008,
236
+ 0.012232512474060058,
237
+ 0.012286944389343262,
238
+ 0.01224396800994873,
239
+ 0.01225830364227295,
240
+ 0.01221939182281494,
241
+ 0.012277759552001954,
242
+ 0.012220416069030762,
243
+ 0.012271615982055664,
244
+ 0.01225216007232666,
245
+ 0.01229312038421631,
246
+ 0.012241920471191407,
247
+ 0.01226137638092041,
248
+ 0.012224512100219726,
249
+ 0.012289024353027344,
250
+ 0.01223475170135498,
251
+ 0.012271615982055664,
252
+ 0.012273664474487305,
253
+ 0.012294143676757812,
254
+ 0.012233728408813477,
255
+ 0.01227673625946045,
256
+ 0.012247039794921874,
257
+ 0.012305407524108887,
258
+ 0.012231807708740234,
259
+ 0.012311424255371094,
260
+ 0.012220416069030762,
261
+ 0.012393471717834472,
262
+ 0.01226035213470459,
263
+ 0.012280832290649414,
264
+ 0.012220416069030762,
265
+ 0.01226956844329834,
266
+ 0.012220416069030762,
267
+ 0.012265472412109376,
268
+ 0.012233728408813477,
269
+ 0.012272640228271485,
270
+ 0.012224512100219726,
271
+ 0.012271615982055664,
272
+ 0.01222758388519287,
273
+ 0.012277759552001954,
274
+ 0.012347392082214356,
275
+ 0.012498944282531739,
276
+ 0.01228492832183838,
277
+ 0.012302335739135742,
278
+ 0.012244992256164551,
279
+ 0.01228492832183838,
280
+ 0.012230655670166016,
281
+ 0.012272640228271485,
282
+ 0.012253184318542481,
283
+ 0.012274687767028808,
284
+ 0.012222623825073243,
285
+ 0.012265312194824218,
286
+ 0.012249088287353516,
287
+ 0.012297216415405274,
288
+ 0.01225216007232666,
289
+ 0.012265472412109376,
290
+ 0.012233728408813477,
291
+ 0.01226035213470459,
292
+ 0.012200960159301758,
293
+ 0.01226137638092041,
294
+ 0.012221440315246582,
295
+ 0.012273664474487305,
296
+ 0.01223475170135498,
297
+ 0.012288000106811523,
298
+ 0.012265472412109376,
299
+ 0.012262399673461915,
300
+ 0.012228608131408691,
301
+ 0.012273664474487305,
302
+ 0.012221440315246582,
303
+ 0.012273664474487305,
304
+ 0.012237824440002442,
305
+ 0.012253184318542481,
306
+ 0.012231679916381836,
307
+ 0.012263423919677734,
308
+ 0.01222758388519287,
309
+ 0.012259327888488769,
310
+ 0.012212224006652832,
311
+ 0.012264448165893555,
312
+ 0.012225536346435547,
313
+ 0.01226854419708252,
314
+ 0.012246015548706055,
315
+ 0.012259327888488769,
316
+ 0.012220416069030762,
317
+ 0.01225830364227295,
318
+ 0.012216320037841797,
319
+ 0.012262399673461915,
320
+ 0.012237824440002442,
321
+ 0.012280832290649414,
322
+ 0.012221440315246582,
323
+ 0.012270591735839843,
324
+ 0.012232704162597656,
325
+ 0.012271615982055664,
326
+ 0.012248064041137695,
327
+ 0.012279808044433594,
328
+ 0.012244992256164551,
329
+ 0.012300288200378418
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 81.14609788247432
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.01405294418334961,
7
+ "prefill.throughput.value": 498.1162600997043,
8
+ "decode.latency.mean": 1.5584067993164064,
9
+ "decode.throughput.value": 81.49348427875726,
10
+ "per_token.latency.mean": 0.012323451479433083,
11
+ "per_token.throughput.value": 81.14609788247432
12
+ }
13
+ }
2024-06-09/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/multirun.yaml ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpksh54epa/commit=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
88
+ - hydra.run.dir=_benchmark/tmpksh54epa/commit\=25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.3.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model: google/gemma-2b
150
+ processor: null
151
+ device: cuda
152
+ device_ids: '0'
153
+ seed: 42
154
+ inter_op_num_threads: null
155
+ intra_op_num_threads: null
156
+ model_kwargs: {}
157
+ processor_kwargs: {}
158
+ hub_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ latency: true
192
+ memory: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ start_method: spawn
206
+ environment:
207
+ cpu: ' AMD EPYC 7R32'
208
+ cpu_count: 16
209
+ cpu_ram_mb: 66697.29792
210
+ system: Linux
211
+ machine: x86_64
212
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
213
+ processor: x86_64
214
+ python_version: 3.8.10
215
+ gpu:
216
+ - NVIDIA A10G
217
+ gpu_count: 1
218
+ gpu_vram_mb: 24146608128
219
+ optimum_benchmark_version: 0.2.1
220
+ optimum_benchmark_commit: null
221
+ transformers_version: 4.42.0.dev0
222
+ transformers_commit: 25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5
223
+ accelerate_version: 0.32.0.dev0
224
+ accelerate_commit: null
225
+ diffusers_version: null
226
+ diffusers_commit: null
227
+ optimum_version: 1.21.0.dev0
228
+ optimum_commit: null
229
+ timm_version: 0.9.16
230
+ timm_commit: null
231
+ peft_version: 0.11.2.dev0
232
+ peft_commit: null
2024-06-09/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.022692079544067385,
8
+ "prefill.throughput.value": 308.47767770275067,
9
+ "decode.latency.mean": 2.385437255859375,
10
+ "decode.throughput.value": 53.23971514574468,
11
+ "per_token.latency.mean": 0.018857490215376905,
12
+ "per_token.throughput.value": 53.029326202941526
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.021704048156738282,
21
+ "prefill.throughput.value": 322.52047864291,
22
+ "decode.latency.mean": 2.5922081298828124,
23
+ "decode.throughput.value": 48.99297958985314,
24
+ "per_token.latency.mean": 0.020496448750552454,
25
+ "per_token.throughput.value": 48.78893959486744
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.01405294418334961,
34
+ "prefill.throughput.value": 498.1162600997043,
35
+ "decode.latency.mean": 1.5584067993164064,
36
+ "decode.throughput.value": 81.49348427875726,
37
+ "per_token.latency.mean": 0.012323451479433083,
38
+ "per_token.throughput.value": 81.14609788247432
39
+ }
40
+ }
41
+ ]
2024-06-09/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022692079544067385,
7
+ "prefill.throughput.value": 308.47767770275067,
8
+ "decode.latency.mean": 2.385437255859375,
9
+ "decode.throughput.value": 53.23971514574468,
10
+ "per_token.latency.mean": 0.018857490215376905,
11
+ "per_token.throughput.value": 53.029326202941526
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.021704048156738282,
19
+ "prefill.throughput.value": 322.52047864291,
20
+ "decode.latency.mean": 2.5922081298828124,
21
+ "decode.throughput.value": 48.99297958985314,
22
+ "per_token.latency.mean": 0.020496448750552454,
23
+ "per_token.throughput.value": 48.78893959486744
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.01405294418334961,
31
+ "prefill.throughput.value": 498.1162600997043,
32
+ "decode.latency.mean": 1.5584067993164064,
33
+ "decode.throughput.value": 81.49348427875726,
34
+ "per_token.latency.mean": 0.012323451479433083,
35
+ "per_token.throughput.value": 81.14609788247432
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }