hf-transformers-bot commited on
Commit
50b32bb
1 Parent(s): cca31db

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/multirun.yaml +234 -0
  31. 2024-09-16/summaries.json +41 -0
  32. 2024-09-16/summary.json +40 -0
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 95e816f2bca48de32167ce6243e6770dee23923d
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d
88
+ - hydra.run.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1333.161984,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.9035693359375,
120
+ "mean": 11.9035693359375,
121
+ "stdev": 0.0,
122
+ "p50": 11.9035693359375,
123
+ "p90": 11.9035693359375,
124
+ "p95": 11.9035693359375,
125
+ "p99": 11.9035693359375,
126
+ "values": [
127
+ 11.9035693359375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1782.960128,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.042188287734985355,
147
+ "mean": 0.021094143867492678,
148
+ "stdev": 0.00021910381317138787,
149
+ "p50": 0.021094143867492678,
150
+ "p90": 0.021269426918029785,
151
+ "p95": 0.021291337299346926,
152
+ "p99": 0.021308865604400637,
153
+ "values": [
154
+ 0.021313247680664064,
155
+ 0.020875040054321288
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 331.8456555512269
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1841.074176,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.093490966796875,
178
+ "mean": 2.5467454833984373,
179
+ "stdev": 0.0005391845703124343,
180
+ "p50": 2.5467454833984373,
181
+ "p90": 2.5471768310546876,
182
+ "p95": 2.547230749511719,
183
+ "p99": 2.547273884277344,
184
+ "values": [
185
+ 2.546206298828125,
186
+ 2.54728466796875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 49.867566597400305
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.093169151306151,
202
+ "mean": 0.020051847052386426,
203
+ "stdev": 0.00032461421230515656,
204
+ "p50": 0.020071375846862794,
205
+ "p90": 0.020297682762145997,
206
+ "p95": 0.020420659065246583,
207
+ "p99": 0.02087001085281372,
208
+ "values": [
209
+ 0.019268608093261717,
210
+ 0.01940377616882324,
211
+ 0.01984409523010254,
212
+ 0.019934207916259765,
213
+ 0.02005401611328125,
214
+ 0.019714048385620117,
215
+ 0.019959808349609375,
216
+ 0.019992576599121094,
217
+ 0.020241504669189454,
218
+ 0.020071327209472658,
219
+ 0.02030284881591797,
220
+ 0.020011007308959963,
221
+ 0.02007859230041504,
222
+ 0.019927167892456056,
223
+ 0.019912576675415038,
224
+ 0.02007040023803711,
225
+ 0.02027724838256836,
226
+ 0.020159488677978517,
227
+ 0.01927475166320801,
228
+ 0.019944448471069336,
229
+ 0.020113407135009767,
230
+ 0.02025267219543457,
231
+ 0.019927040100097656,
232
+ 0.02001408004760742,
233
+ 0.020158464431762696,
234
+ 0.019351551055908203,
235
+ 0.0200949764251709,
236
+ 0.020150272369384766,
237
+ 0.0198922233581543,
238
+ 0.020007936477661133,
239
+ 0.020378623962402344,
240
+ 0.0202106876373291,
241
+ 0.02042265510559082,
242
+ 0.020364288330078126,
243
+ 0.02085990333557129,
244
+ 0.020109376907348632,
245
+ 0.020300735473632814,
246
+ 0.02019327926635742,
247
+ 0.020023296356201172,
248
+ 0.02007859230041504,
249
+ 0.020015104293823242,
250
+ 0.020007936477661133,
251
+ 0.020157440185546875,
252
+ 0.01987174415588379,
253
+ 0.019993600845336915,
254
+ 0.019974143981933593,
255
+ 0.020031488418579102,
256
+ 0.020002815246582033,
257
+ 0.020183040618896485,
258
+ 0.020067327499389647,
259
+ 0.019978239059448243,
260
+ 0.020139007568359374,
261
+ 0.020000768661499024,
262
+ 0.020173824310302735,
263
+ 0.020048896789550782,
264
+ 0.019703807830810546,
265
+ 0.019598335266113282,
266
+ 0.019941375732421874,
267
+ 0.0200263671875,
268
+ 0.019931135177612306,
269
+ 0.0200130558013916,
270
+ 0.019940351486206053,
271
+ 0.020108287811279296,
272
+ 0.02040729522705078,
273
+ 0.02088140869140625,
274
+ 0.020304895401000975,
275
+ 0.02022707176208496,
276
+ 0.020340736389160157,
277
+ 0.020141056060791016,
278
+ 0.020279296875,
279
+ 0.01992192077636719,
280
+ 0.01939455986022949,
281
+ 0.019765247344970704,
282
+ 0.02023321533203125,
283
+ 0.02027212715148926,
284
+ 0.01945599937438965,
285
+ 0.01992192077636719,
286
+ 0.020198400497436524,
287
+ 0.02012057685852051,
288
+ 0.020666368484497072,
289
+ 0.02007244873046875,
290
+ 0.01930342483520508,
291
+ 0.020171775817871093,
292
+ 0.020494335174560546,
293
+ 0.020452352523803712,
294
+ 0.01967103958129883,
295
+ 0.020427776336669923,
296
+ 0.020180992126464844,
297
+ 0.0200263671875,
298
+ 0.020057088851928712,
299
+ 0.019941375732421874,
300
+ 0.02011238479614258,
301
+ 0.020006912231445313,
302
+ 0.020063232421875,
303
+ 0.02018611145019531,
304
+ 0.02022604751586914,
305
+ 0.020404224395751954,
306
+ 0.02011136054992676,
307
+ 0.020180992126464844,
308
+ 0.02021683120727539,
309
+ 0.02027519989013672,
310
+ 0.020179967880249023,
311
+ 0.02023129653930664,
312
+ 0.020210559844970704,
313
+ 0.019907583236694337,
314
+ 0.019922943115234376,
315
+ 0.019968000411987305,
316
+ 0.020388864517211915,
317
+ 0.020170751571655272,
318
+ 0.02042470359802246,
319
+ 0.01944268798828125,
320
+ 0.01925734329223633,
321
+ 0.019521535873413084,
322
+ 0.020023296356201172,
323
+ 0.020206592559814454,
324
+ 0.019944448471069336,
325
+ 0.019740671157836915,
326
+ 0.020100095748901366,
327
+ 0.020163583755493163,
328
+ 0.020118528366088868,
329
+ 0.020419584274291993,
330
+ 0.02031001663208008,
331
+ 0.02001919937133789,
332
+ 0.020204544067382812,
333
+ 0.02024140739440918,
334
+ 0.01927475166320801,
335
+ 0.01928704071044922,
336
+ 0.019269695281982423,
337
+ 0.019199935913085938,
338
+ 0.020333568572998048,
339
+ 0.02002022361755371,
340
+ 0.020001792907714845,
341
+ 0.019976192474365235,
342
+ 0.020007936477661133,
343
+ 0.020935680389404295,
344
+ 0.020172800064086914,
345
+ 0.0200130558013916,
346
+ 0.02003660774230957,
347
+ 0.0198656005859375,
348
+ 0.0198656005859375,
349
+ 0.02006220817565918,
350
+ 0.020143104553222657,
351
+ 0.019935232162475586,
352
+ 0.019805280685424805,
353
+ 0.01997711944580078,
354
+ 0.019891199111938478,
355
+ 0.020079744338989257,
356
+ 0.01991974449157715,
357
+ 0.020115455627441405,
358
+ 0.020173824310302735,
359
+ 0.02026393508911133,
360
+ 0.01989836883544922,
361
+ 0.02008883285522461,
362
+ 0.02006425666809082,
363
+ 0.020048896789550782,
364
+ 0.01980620765686035,
365
+ 0.02009702491760254,
366
+ 0.02000588798522949,
367
+ 0.019773439407348634,
368
+ 0.019973119735717772,
369
+ 0.01983795166015625,
370
+ 0.02003865623474121,
371
+ 0.019938304901123048,
372
+ 0.020031488418579102,
373
+ 0.01999564743041992,
374
+ 0.01987379264831543,
375
+ 0.019974143981933593,
376
+ 0.019959871292114257,
377
+ 0.020030399322509766,
378
+ 0.02004582405090332,
379
+ 0.02011552047729492,
380
+ 0.019853248596191406,
381
+ 0.01984614372253418,
382
+ 0.01984614372253418,
383
+ 0.019986431121826173,
384
+ 0.02003455924987793,
385
+ 0.019986431121826173,
386
+ 0.019991552352905274,
387
+ 0.019915775299072267,
388
+ 0.01979801559448242,
389
+ 0.01920204734802246,
390
+ 0.0192675838470459,
391
+ 0.01960550308227539,
392
+ 0.01922047996520996,
393
+ 0.01987993621826172,
394
+ 0.020156415939331054,
395
+ 0.02064691162109375,
396
+ 0.02024140739440918,
397
+ 0.019574783325195313,
398
+ 0.020075519561767577,
399
+ 0.02002022361755371,
400
+ 0.02024140739440918,
401
+ 0.02025574493408203,
402
+ 0.0200898551940918,
403
+ 0.020182016372680665,
404
+ 0.019943424224853516,
405
+ 0.020131839752197265,
406
+ 0.020171775817871093,
407
+ 0.019973119735717772,
408
+ 0.019923967361450197,
409
+ 0.02021887969970703,
410
+ 0.02002739143371582,
411
+ 0.020131839752197265,
412
+ 0.02007244873046875,
413
+ 0.019949567794799804,
414
+ 0.01987379264831543,
415
+ 0.02024550437927246,
416
+ 0.019982336044311523,
417
+ 0.020265983581542968,
418
+ 0.020125696182250977,
419
+ 0.02007756805419922,
420
+ 0.02026803207397461,
421
+ 0.020288511276245116,
422
+ 0.020143104553222657,
423
+ 0.020145151138305666,
424
+ 0.02007142448425293,
425
+ 0.020205568313598633,
426
+ 0.020208639144897463,
427
+ 0.01982156753540039,
428
+ 0.020199424743652345,
429
+ 0.02025164794921875,
430
+ 0.020122623443603514,
431
+ 0.0200949764251709,
432
+ 0.020129791259765627,
433
+ 0.020290559768676757,
434
+ 0.020167680740356447,
435
+ 0.019962879180908204,
436
+ 0.020155391693115234,
437
+ 0.020113407135009767,
438
+ 0.020200447082519533,
439
+ 0.020189184188842774,
440
+ 0.02010419273376465,
441
+ 0.020006912231445313,
442
+ 0.02007244873046875,
443
+ 0.020083711624145507,
444
+ 0.020404224395751954,
445
+ 0.02006630325317383,
446
+ 0.0206376953125,
447
+ 0.020556800842285155,
448
+ 0.020174848556518556,
449
+ 0.020190208435058594,
450
+ 0.019619840621948242,
451
+ 0.019527679443359376,
452
+ 0.02011136054992676,
453
+ 0.020110336303710938,
454
+ 0.020135936737060548,
455
+ 0.020082687377929686,
456
+ 0.020170751571655272,
457
+ 0.02005606460571289,
458
+ 0.020151296615600587,
459
+ 0.022767616271972657,
460
+ 0.020207616806030275,
461
+ 0.020007936477661133,
462
+ 0.020147199630737304
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 49.87071751482301
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1333.161984,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.9035693359375,
15
+ "mean": 11.9035693359375,
16
+ "stdev": 0.0,
17
+ "p50": 11.9035693359375,
18
+ "p90": 11.9035693359375,
19
+ "p95": 11.9035693359375,
20
+ "p99": 11.9035693359375,
21
+ "values": [
22
+ 11.9035693359375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1782.960128,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.042188287734985355,
42
+ "mean": 0.021094143867492678,
43
+ "stdev": 0.00021910381317138787,
44
+ "p50": 0.021094143867492678,
45
+ "p90": 0.021269426918029785,
46
+ "p95": 0.021291337299346926,
47
+ "p99": 0.021308865604400637,
48
+ "values": [
49
+ 0.021313247680664064,
50
+ 0.020875040054321288
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 331.8456555512269
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1841.074176,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.093490966796875,
73
+ "mean": 2.5467454833984373,
74
+ "stdev": 0.0005391845703124343,
75
+ "p50": 2.5467454833984373,
76
+ "p90": 2.5471768310546876,
77
+ "p95": 2.547230749511719,
78
+ "p99": 2.547273884277344,
79
+ "values": [
80
+ 2.546206298828125,
81
+ 2.54728466796875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 49.867566597400305
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.093169151306151,
97
+ "mean": 0.020051847052386426,
98
+ "stdev": 0.00032461421230515656,
99
+ "p50": 0.020071375846862794,
100
+ "p90": 0.020297682762145997,
101
+ "p95": 0.020420659065246583,
102
+ "p99": 0.02087001085281372,
103
+ "values": [
104
+ 0.019268608093261717,
105
+ 0.01940377616882324,
106
+ 0.01984409523010254,
107
+ 0.019934207916259765,
108
+ 0.02005401611328125,
109
+ 0.019714048385620117,
110
+ 0.019959808349609375,
111
+ 0.019992576599121094,
112
+ 0.020241504669189454,
113
+ 0.020071327209472658,
114
+ 0.02030284881591797,
115
+ 0.020011007308959963,
116
+ 0.02007859230041504,
117
+ 0.019927167892456056,
118
+ 0.019912576675415038,
119
+ 0.02007040023803711,
120
+ 0.02027724838256836,
121
+ 0.020159488677978517,
122
+ 0.01927475166320801,
123
+ 0.019944448471069336,
124
+ 0.020113407135009767,
125
+ 0.02025267219543457,
126
+ 0.019927040100097656,
127
+ 0.02001408004760742,
128
+ 0.020158464431762696,
129
+ 0.019351551055908203,
130
+ 0.0200949764251709,
131
+ 0.020150272369384766,
132
+ 0.0198922233581543,
133
+ 0.020007936477661133,
134
+ 0.020378623962402344,
135
+ 0.0202106876373291,
136
+ 0.02042265510559082,
137
+ 0.020364288330078126,
138
+ 0.02085990333557129,
139
+ 0.020109376907348632,
140
+ 0.020300735473632814,
141
+ 0.02019327926635742,
142
+ 0.020023296356201172,
143
+ 0.02007859230041504,
144
+ 0.020015104293823242,
145
+ 0.020007936477661133,
146
+ 0.020157440185546875,
147
+ 0.01987174415588379,
148
+ 0.019993600845336915,
149
+ 0.019974143981933593,
150
+ 0.020031488418579102,
151
+ 0.020002815246582033,
152
+ 0.020183040618896485,
153
+ 0.020067327499389647,
154
+ 0.019978239059448243,
155
+ 0.020139007568359374,
156
+ 0.020000768661499024,
157
+ 0.020173824310302735,
158
+ 0.020048896789550782,
159
+ 0.019703807830810546,
160
+ 0.019598335266113282,
161
+ 0.019941375732421874,
162
+ 0.0200263671875,
163
+ 0.019931135177612306,
164
+ 0.0200130558013916,
165
+ 0.019940351486206053,
166
+ 0.020108287811279296,
167
+ 0.02040729522705078,
168
+ 0.02088140869140625,
169
+ 0.020304895401000975,
170
+ 0.02022707176208496,
171
+ 0.020340736389160157,
172
+ 0.020141056060791016,
173
+ 0.020279296875,
174
+ 0.01992192077636719,
175
+ 0.01939455986022949,
176
+ 0.019765247344970704,
177
+ 0.02023321533203125,
178
+ 0.02027212715148926,
179
+ 0.01945599937438965,
180
+ 0.01992192077636719,
181
+ 0.020198400497436524,
182
+ 0.02012057685852051,
183
+ 0.020666368484497072,
184
+ 0.02007244873046875,
185
+ 0.01930342483520508,
186
+ 0.020171775817871093,
187
+ 0.020494335174560546,
188
+ 0.020452352523803712,
189
+ 0.01967103958129883,
190
+ 0.020427776336669923,
191
+ 0.020180992126464844,
192
+ 0.0200263671875,
193
+ 0.020057088851928712,
194
+ 0.019941375732421874,
195
+ 0.02011238479614258,
196
+ 0.020006912231445313,
197
+ 0.020063232421875,
198
+ 0.02018611145019531,
199
+ 0.02022604751586914,
200
+ 0.020404224395751954,
201
+ 0.02011136054992676,
202
+ 0.020180992126464844,
203
+ 0.02021683120727539,
204
+ 0.02027519989013672,
205
+ 0.020179967880249023,
206
+ 0.02023129653930664,
207
+ 0.020210559844970704,
208
+ 0.019907583236694337,
209
+ 0.019922943115234376,
210
+ 0.019968000411987305,
211
+ 0.020388864517211915,
212
+ 0.020170751571655272,
213
+ 0.02042470359802246,
214
+ 0.01944268798828125,
215
+ 0.01925734329223633,
216
+ 0.019521535873413084,
217
+ 0.020023296356201172,
218
+ 0.020206592559814454,
219
+ 0.019944448471069336,
220
+ 0.019740671157836915,
221
+ 0.020100095748901366,
222
+ 0.020163583755493163,
223
+ 0.020118528366088868,
224
+ 0.020419584274291993,
225
+ 0.02031001663208008,
226
+ 0.02001919937133789,
227
+ 0.020204544067382812,
228
+ 0.02024140739440918,
229
+ 0.01927475166320801,
230
+ 0.01928704071044922,
231
+ 0.019269695281982423,
232
+ 0.019199935913085938,
233
+ 0.020333568572998048,
234
+ 0.02002022361755371,
235
+ 0.020001792907714845,
236
+ 0.019976192474365235,
237
+ 0.020007936477661133,
238
+ 0.020935680389404295,
239
+ 0.020172800064086914,
240
+ 0.0200130558013916,
241
+ 0.02003660774230957,
242
+ 0.0198656005859375,
243
+ 0.0198656005859375,
244
+ 0.02006220817565918,
245
+ 0.020143104553222657,
246
+ 0.019935232162475586,
247
+ 0.019805280685424805,
248
+ 0.01997711944580078,
249
+ 0.019891199111938478,
250
+ 0.020079744338989257,
251
+ 0.01991974449157715,
252
+ 0.020115455627441405,
253
+ 0.020173824310302735,
254
+ 0.02026393508911133,
255
+ 0.01989836883544922,
256
+ 0.02008883285522461,
257
+ 0.02006425666809082,
258
+ 0.020048896789550782,
259
+ 0.01980620765686035,
260
+ 0.02009702491760254,
261
+ 0.02000588798522949,
262
+ 0.019773439407348634,
263
+ 0.019973119735717772,
264
+ 0.01983795166015625,
265
+ 0.02003865623474121,
266
+ 0.019938304901123048,
267
+ 0.020031488418579102,
268
+ 0.01999564743041992,
269
+ 0.01987379264831543,
270
+ 0.019974143981933593,
271
+ 0.019959871292114257,
272
+ 0.020030399322509766,
273
+ 0.02004582405090332,
274
+ 0.02011552047729492,
275
+ 0.019853248596191406,
276
+ 0.01984614372253418,
277
+ 0.01984614372253418,
278
+ 0.019986431121826173,
279
+ 0.02003455924987793,
280
+ 0.019986431121826173,
281
+ 0.019991552352905274,
282
+ 0.019915775299072267,
283
+ 0.01979801559448242,
284
+ 0.01920204734802246,
285
+ 0.0192675838470459,
286
+ 0.01960550308227539,
287
+ 0.01922047996520996,
288
+ 0.01987993621826172,
289
+ 0.020156415939331054,
290
+ 0.02064691162109375,
291
+ 0.02024140739440918,
292
+ 0.019574783325195313,
293
+ 0.020075519561767577,
294
+ 0.02002022361755371,
295
+ 0.02024140739440918,
296
+ 0.02025574493408203,
297
+ 0.0200898551940918,
298
+ 0.020182016372680665,
299
+ 0.019943424224853516,
300
+ 0.020131839752197265,
301
+ 0.020171775817871093,
302
+ 0.019973119735717772,
303
+ 0.019923967361450197,
304
+ 0.02021887969970703,
305
+ 0.02002739143371582,
306
+ 0.020131839752197265,
307
+ 0.02007244873046875,
308
+ 0.019949567794799804,
309
+ 0.01987379264831543,
310
+ 0.02024550437927246,
311
+ 0.019982336044311523,
312
+ 0.020265983581542968,
313
+ 0.020125696182250977,
314
+ 0.02007756805419922,
315
+ 0.02026803207397461,
316
+ 0.020288511276245116,
317
+ 0.020143104553222657,
318
+ 0.020145151138305666,
319
+ 0.02007142448425293,
320
+ 0.020205568313598633,
321
+ 0.020208639144897463,
322
+ 0.01982156753540039,
323
+ 0.020199424743652345,
324
+ 0.02025164794921875,
325
+ 0.020122623443603514,
326
+ 0.0200949764251709,
327
+ 0.020129791259765627,
328
+ 0.020290559768676757,
329
+ 0.020167680740356447,
330
+ 0.019962879180908204,
331
+ 0.020155391693115234,
332
+ 0.020113407135009767,
333
+ 0.020200447082519533,
334
+ 0.020189184188842774,
335
+ 0.02010419273376465,
336
+ 0.020006912231445313,
337
+ 0.02007244873046875,
338
+ 0.020083711624145507,
339
+ 0.020404224395751954,
340
+ 0.02006630325317383,
341
+ 0.0206376953125,
342
+ 0.020556800842285155,
343
+ 0.020174848556518556,
344
+ 0.020190208435058594,
345
+ 0.019619840621948242,
346
+ 0.019527679443359376,
347
+ 0.02011136054992676,
348
+ 0.020110336303710938,
349
+ 0.020135936737060548,
350
+ 0.020082687377929686,
351
+ 0.020170751571655272,
352
+ 0.02005606460571289,
353
+ 0.020151296615600587,
354
+ 0.022767616271972657,
355
+ 0.020207616806030275,
356
+ 0.020007936477661133,
357
+ 0.020147199630737304
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 49.87071751482301
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.021094143867492678,
7
+ "prefill.throughput.value": 331.8456555512269,
8
+ "decode.latency.mean": 2.5467454833984373,
9
+ "decode.throughput.value": 49.867566597400305,
10
+ "per_token.latency.mean": 0.020051847052386426,
11
+ "per_token.throughput.value": 49.87071751482301
12
+ }
13
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 95e816f2bca48de32167ce6243e6770dee23923d
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d
88
+ - hydra.run.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-09-16 08:55:05,873][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-09-16 08:55:08,253][process][ERROR] - + Received traceback from isolated process
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 95e816f2bca48de32167ce6243e6770dee23923d
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d
88
+ - hydra.run.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1335.320576,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 13.2067099609375,
120
+ "mean": 13.2067099609375,
121
+ "stdev": 0.0,
122
+ "p50": 13.2067099609375,
123
+ "p90": 13.2067099609375,
124
+ "p95": 13.2067099609375,
125
+ "p99": 13.2067099609375,
126
+ "values": [
127
+ 13.2067099609375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1809.002496,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.043523487091064454,
147
+ "mean": 0.021761743545532227,
148
+ "stdev": 0.00020244789123535084,
149
+ "p50": 0.021761743545532227,
150
+ "p90": 0.02192370185852051,
151
+ "p95": 0.02194394664764404,
152
+ "p99": 0.02196014247894287,
153
+ "values": [
154
+ 0.021964191436767578,
155
+ 0.021559295654296876
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 321.6654026527726
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1835.769856,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.16451806640625,
178
+ "mean": 2.582259033203125,
179
+ "stdev": 0.04007031249999993,
180
+ "p50": 2.582259033203125,
181
+ "p90": 2.614315283203125,
182
+ "p95": 2.618322314453125,
183
+ "p99": 2.621527939453125,
184
+ "values": [
185
+ 2.622329345703125,
186
+ 2.542188720703125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 49.18174294949206
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.164267509460451,
202
+ "mean": 0.020331761848269486,
203
+ "stdev": 0.0004959884821549371,
204
+ "p50": 0.02039193630218506,
205
+ "p90": 0.02071357479095459,
206
+ "p95": 0.020821042728424072,
207
+ "p99": 0.02127322088241577,
208
+ "values": [
209
+ 0.020847616195678712,
210
+ 0.02056809616088867,
211
+ 0.020595680236816405,
212
+ 0.020420608520507814,
213
+ 0.019934207916259765,
214
+ 0.02053222465515137,
215
+ 0.020362239837646484,
216
+ 0.020617216110229493,
217
+ 0.0206059513092041,
218
+ 0.020610048294067384,
219
+ 0.020625408172607423,
220
+ 0.020149248123168945,
221
+ 0.020528127670288086,
222
+ 0.020554752349853517,
223
+ 0.020758527755737305,
224
+ 0.020724735260009765,
225
+ 0.020610048294067384,
226
+ 0.02067353630065918,
227
+ 0.020091903686523437,
228
+ 0.020626432418823244,
229
+ 0.02087116813659668,
230
+ 0.020616191864013672,
231
+ 0.020455423355102538,
232
+ 0.0206561279296875,
233
+ 0.0206376953125,
234
+ 0.020545536041259766,
235
+ 0.02058137512207031,
236
+ 0.020584447860717774,
237
+ 0.02065715217590332,
238
+ 0.0206376953125,
239
+ 0.020669439315795898,
240
+ 0.02042265510559082,
241
+ 0.02043289566040039,
242
+ 0.02045747184753418,
243
+ 0.02048204803466797,
244
+ 0.020588544845581053,
245
+ 0.020577280044555665,
246
+ 0.02063871955871582,
247
+ 0.020694015502929687,
248
+ 0.02064896011352539,
249
+ 0.02081996726989746,
250
+ 0.020815872192382814,
251
+ 0.02067251205444336,
252
+ 0.020463615417480468,
253
+ 0.020701183319091796,
254
+ 0.020823040008544923,
255
+ 0.020546560287475587,
256
+ 0.02066431999206543,
257
+ 0.020642816543579103,
258
+ 0.020642816543579103,
259
+ 0.020580352783203124,
260
+ 0.02049126434326172,
261
+ 0.020568063735961914,
262
+ 0.02004991912841797,
263
+ 0.02023219108581543,
264
+ 0.020685823440551757,
265
+ 0.020572160720825194,
266
+ 0.020503551483154296,
267
+ 0.02064384078979492,
268
+ 0.02035916709899902,
269
+ 0.02104217529296875,
270
+ 0.02106060791015625,
271
+ 0.020570112228393556,
272
+ 0.020722688674926756,
273
+ 0.020742143630981445,
274
+ 0.020722688674926756,
275
+ 0.02044211196899414,
276
+ 0.0206376953125,
277
+ 0.020714496612548827,
278
+ 0.02059775924682617,
279
+ 0.020976640701293944,
280
+ 0.020505599975585938,
281
+ 0.020586496353149415,
282
+ 0.020649984359741212,
283
+ 0.02049126434326172,
284
+ 0.020595712661743162,
285
+ 0.020521984100341797,
286
+ 0.020632575988769532,
287
+ 0.02064896011352539,
288
+ 0.020607999801635742,
289
+ 0.02060492706298828,
290
+ 0.020593664169311524,
291
+ 0.020699136734008788,
292
+ 0.02057318305969238,
293
+ 0.020800512313842775,
294
+ 0.020710399627685547,
295
+ 0.02048409652709961,
296
+ 0.020544511795043945,
297
+ 0.020708351135253905,
298
+ 0.020711423873901368,
299
+ 0.02063564872741699,
300
+ 0.020393983840942383,
301
+ 0.02064384078979492,
302
+ 0.020579328536987306,
303
+ 0.02063155174255371,
304
+ 0.02122035217285156,
305
+ 0.020751359939575196,
306
+ 0.02057318305969238,
307
+ 0.02042163276672363,
308
+ 0.02064691162109375,
309
+ 0.02048307228088379,
310
+ 0.020446207046508787,
311
+ 0.02064896011352539,
312
+ 0.02062745666503906,
313
+ 0.020599807739257812,
314
+ 0.020587520599365236,
315
+ 0.020447231292724608,
316
+ 0.020534271240234374,
317
+ 0.02067353630065918,
318
+ 0.020684799194335936,
319
+ 0.020074495315551756,
320
+ 0.020184064865112306,
321
+ 0.020700159072875975,
322
+ 0.02066739273071289,
323
+ 0.020618240356445314,
324
+ 0.020578304290771485,
325
+ 0.020813823699951172,
326
+ 0.02006220817565918,
327
+ 0.020981760025024415,
328
+ 0.020509695053100584,
329
+ 0.02062233543395996,
330
+ 0.02128895950317383,
331
+ 0.02553856086730957,
332
+ 0.0208035831451416,
333
+ 0.020617216110229493,
334
+ 0.021737472534179687,
335
+ 0.020504575729370117,
336
+ 0.02025062370300293,
337
+ 0.02063155174255371,
338
+ 0.020594688415527345,
339
+ 0.020059135437011717,
340
+ 0.019842048645019532,
341
+ 0.0202926082611084,
342
+ 0.019939327239990236,
343
+ 0.019920896530151368,
344
+ 0.019937280654907227,
345
+ 0.019985408782958985,
346
+ 0.0198922233581543,
347
+ 0.019949567794799804,
348
+ 0.019974143981933593,
349
+ 0.020011007308959963,
350
+ 0.02049228858947754,
351
+ 0.021259263992309572,
352
+ 0.020953088760375976,
353
+ 0.02069606399536133,
354
+ 0.020773887634277344,
355
+ 0.019920896530151368,
356
+ 0.019939327239990236,
357
+ 0.019887104034423828,
358
+ 0.019885055541992186,
359
+ 0.019954687118530275,
360
+ 0.020247552871704103,
361
+ 0.020617216110229493,
362
+ 0.02001408004760742,
363
+ 0.019915775299072267,
364
+ 0.02042572784423828,
365
+ 0.0198154239654541,
366
+ 0.019932159423828123,
367
+ 0.019932159423828123,
368
+ 0.019943424224853516,
369
+ 0.019961856842041017,
370
+ 0.019977216720581056,
371
+ 0.019969024658203126,
372
+ 0.01982771110534668,
373
+ 0.020082687377929686,
374
+ 0.02001817512512207,
375
+ 0.02028646469116211,
376
+ 0.019912704467773438,
377
+ 0.019966976165771484,
378
+ 0.01991372871398926,
379
+ 0.019842048645019532,
380
+ 0.020016128540039063,
381
+ 0.019946495056152345,
382
+ 0.019915775299072267,
383
+ 0.019877887725830077,
384
+ 0.020076543807983398,
385
+ 0.019974143981933593,
386
+ 0.019747840881347657,
387
+ 0.019739648818969727,
388
+ 0.020040704727172853,
389
+ 0.01999667167663574,
390
+ 0.02002124786376953,
391
+ 0.019900415420532228,
392
+ 0.01989017677307129,
393
+ 0.019942399978637695,
394
+ 0.019989503860473632,
395
+ 0.02007142448425293,
396
+ 0.019990528106689453,
397
+ 0.019938304901123048,
398
+ 0.0200581111907959,
399
+ 0.019982336044311523,
400
+ 0.020000768661499024,
401
+ 0.02006220817565918,
402
+ 0.020548608779907225,
403
+ 0.01960140800476074,
404
+ 0.019989503860473632,
405
+ 0.01987379264831543,
406
+ 0.01998028755187988,
407
+ 0.019965951919555663,
408
+ 0.020039680480957032,
409
+ 0.01990656089782715,
410
+ 0.019955711364746095,
411
+ 0.02007756805419922,
412
+ 0.019978239059448243,
413
+ 0.019997695922851562,
414
+ 0.019954687118530275,
415
+ 0.019979263305664064,
416
+ 0.019940351486206053,
417
+ 0.019935232162475586,
418
+ 0.01990553665161133,
419
+ 0.02008166313171387,
420
+ 0.019908607482910155,
421
+ 0.019909631729125975,
422
+ 0.02002943992614746,
423
+ 0.02004684829711914,
424
+ 0.020389888763427736,
425
+ 0.019963903427124022,
426
+ 0.020025344848632814,
427
+ 0.02000588798522949,
428
+ 0.02002022361755371,
429
+ 0.019869695663452147,
430
+ 0.02000383949279785,
431
+ 0.019974143981933593,
432
+ 0.019843072891235353,
433
+ 0.019914751052856446,
434
+ 0.019976192474365235,
435
+ 0.019946495056152345,
436
+ 0.019787776947021486,
437
+ 0.019939327239990236,
438
+ 0.01984409523010254,
439
+ 0.019895296096801757,
440
+ 0.019765247344970704,
441
+ 0.0198973445892334,
442
+ 0.01998028755187988,
443
+ 0.019901439666748046,
444
+ 0.019915775299072267,
445
+ 0.02006425666809082,
446
+ 0.01982771110534668,
447
+ 0.019915775299072267,
448
+ 0.019993600845336915,
449
+ 0.019918848037719726,
450
+ 0.019932159423828123,
451
+ 0.01984614372253418,
452
+ 0.020101119995117187,
453
+ 0.019971071243286134,
454
+ 0.01993011283874512,
455
+ 0.019940351486206053,
456
+ 0.01987583923339844,
457
+ 0.019939327239990236,
458
+ 0.019941375732421874,
459
+ 0.019968000411987305,
460
+ 0.019924991607666014,
461
+ 0.01985638427734375,
462
+ 0.019826688766479493
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 49.18412912086681
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1335.320576,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 13.2067099609375,
15
+ "mean": 13.2067099609375,
16
+ "stdev": 0.0,
17
+ "p50": 13.2067099609375,
18
+ "p90": 13.2067099609375,
19
+ "p95": 13.2067099609375,
20
+ "p99": 13.2067099609375,
21
+ "values": [
22
+ 13.2067099609375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1809.002496,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.043523487091064454,
42
+ "mean": 0.021761743545532227,
43
+ "stdev": 0.00020244789123535084,
44
+ "p50": 0.021761743545532227,
45
+ "p90": 0.02192370185852051,
46
+ "p95": 0.02194394664764404,
47
+ "p99": 0.02196014247894287,
48
+ "values": [
49
+ 0.021964191436767578,
50
+ 0.021559295654296876
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 321.6654026527726
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1835.769856,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.16451806640625,
73
+ "mean": 2.582259033203125,
74
+ "stdev": 0.04007031249999993,
75
+ "p50": 2.582259033203125,
76
+ "p90": 2.614315283203125,
77
+ "p95": 2.618322314453125,
78
+ "p99": 2.621527939453125,
79
+ "values": [
80
+ 2.622329345703125,
81
+ 2.542188720703125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 49.18174294949206
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.164267509460451,
97
+ "mean": 0.020331761848269486,
98
+ "stdev": 0.0004959884821549371,
99
+ "p50": 0.02039193630218506,
100
+ "p90": 0.02071357479095459,
101
+ "p95": 0.020821042728424072,
102
+ "p99": 0.02127322088241577,
103
+ "values": [
104
+ 0.020847616195678712,
105
+ 0.02056809616088867,
106
+ 0.020595680236816405,
107
+ 0.020420608520507814,
108
+ 0.019934207916259765,
109
+ 0.02053222465515137,
110
+ 0.020362239837646484,
111
+ 0.020617216110229493,
112
+ 0.0206059513092041,
113
+ 0.020610048294067384,
114
+ 0.020625408172607423,
115
+ 0.020149248123168945,
116
+ 0.020528127670288086,
117
+ 0.020554752349853517,
118
+ 0.020758527755737305,
119
+ 0.020724735260009765,
120
+ 0.020610048294067384,
121
+ 0.02067353630065918,
122
+ 0.020091903686523437,
123
+ 0.020626432418823244,
124
+ 0.02087116813659668,
125
+ 0.020616191864013672,
126
+ 0.020455423355102538,
127
+ 0.0206561279296875,
128
+ 0.0206376953125,
129
+ 0.020545536041259766,
130
+ 0.02058137512207031,
131
+ 0.020584447860717774,
132
+ 0.02065715217590332,
133
+ 0.0206376953125,
134
+ 0.020669439315795898,
135
+ 0.02042265510559082,
136
+ 0.02043289566040039,
137
+ 0.02045747184753418,
138
+ 0.02048204803466797,
139
+ 0.020588544845581053,
140
+ 0.020577280044555665,
141
+ 0.02063871955871582,
142
+ 0.020694015502929687,
143
+ 0.02064896011352539,
144
+ 0.02081996726989746,
145
+ 0.020815872192382814,
146
+ 0.02067251205444336,
147
+ 0.020463615417480468,
148
+ 0.020701183319091796,
149
+ 0.020823040008544923,
150
+ 0.020546560287475587,
151
+ 0.02066431999206543,
152
+ 0.020642816543579103,
153
+ 0.020642816543579103,
154
+ 0.020580352783203124,
155
+ 0.02049126434326172,
156
+ 0.020568063735961914,
157
+ 0.02004991912841797,
158
+ 0.02023219108581543,
159
+ 0.020685823440551757,
160
+ 0.020572160720825194,
161
+ 0.020503551483154296,
162
+ 0.02064384078979492,
163
+ 0.02035916709899902,
164
+ 0.02104217529296875,
165
+ 0.02106060791015625,
166
+ 0.020570112228393556,
167
+ 0.020722688674926756,
168
+ 0.020742143630981445,
169
+ 0.020722688674926756,
170
+ 0.02044211196899414,
171
+ 0.0206376953125,
172
+ 0.020714496612548827,
173
+ 0.02059775924682617,
174
+ 0.020976640701293944,
175
+ 0.020505599975585938,
176
+ 0.020586496353149415,
177
+ 0.020649984359741212,
178
+ 0.02049126434326172,
179
+ 0.020595712661743162,
180
+ 0.020521984100341797,
181
+ 0.020632575988769532,
182
+ 0.02064896011352539,
183
+ 0.020607999801635742,
184
+ 0.02060492706298828,
185
+ 0.020593664169311524,
186
+ 0.020699136734008788,
187
+ 0.02057318305969238,
188
+ 0.020800512313842775,
189
+ 0.020710399627685547,
190
+ 0.02048409652709961,
191
+ 0.020544511795043945,
192
+ 0.020708351135253905,
193
+ 0.020711423873901368,
194
+ 0.02063564872741699,
195
+ 0.020393983840942383,
196
+ 0.02064384078979492,
197
+ 0.020579328536987306,
198
+ 0.02063155174255371,
199
+ 0.02122035217285156,
200
+ 0.020751359939575196,
201
+ 0.02057318305969238,
202
+ 0.02042163276672363,
203
+ 0.02064691162109375,
204
+ 0.02048307228088379,
205
+ 0.020446207046508787,
206
+ 0.02064896011352539,
207
+ 0.02062745666503906,
208
+ 0.020599807739257812,
209
+ 0.020587520599365236,
210
+ 0.020447231292724608,
211
+ 0.020534271240234374,
212
+ 0.02067353630065918,
213
+ 0.020684799194335936,
214
+ 0.020074495315551756,
215
+ 0.020184064865112306,
216
+ 0.020700159072875975,
217
+ 0.02066739273071289,
218
+ 0.020618240356445314,
219
+ 0.020578304290771485,
220
+ 0.020813823699951172,
221
+ 0.02006220817565918,
222
+ 0.020981760025024415,
223
+ 0.020509695053100584,
224
+ 0.02062233543395996,
225
+ 0.02128895950317383,
226
+ 0.02553856086730957,
227
+ 0.0208035831451416,
228
+ 0.020617216110229493,
229
+ 0.021737472534179687,
230
+ 0.020504575729370117,
231
+ 0.02025062370300293,
232
+ 0.02063155174255371,
233
+ 0.020594688415527345,
234
+ 0.020059135437011717,
235
+ 0.019842048645019532,
236
+ 0.0202926082611084,
237
+ 0.019939327239990236,
238
+ 0.019920896530151368,
239
+ 0.019937280654907227,
240
+ 0.019985408782958985,
241
+ 0.0198922233581543,
242
+ 0.019949567794799804,
243
+ 0.019974143981933593,
244
+ 0.020011007308959963,
245
+ 0.02049228858947754,
246
+ 0.021259263992309572,
247
+ 0.020953088760375976,
248
+ 0.02069606399536133,
249
+ 0.020773887634277344,
250
+ 0.019920896530151368,
251
+ 0.019939327239990236,
252
+ 0.019887104034423828,
253
+ 0.019885055541992186,
254
+ 0.019954687118530275,
255
+ 0.020247552871704103,
256
+ 0.020617216110229493,
257
+ 0.02001408004760742,
258
+ 0.019915775299072267,
259
+ 0.02042572784423828,
260
+ 0.0198154239654541,
261
+ 0.019932159423828123,
262
+ 0.019932159423828123,
263
+ 0.019943424224853516,
264
+ 0.019961856842041017,
265
+ 0.019977216720581056,
266
+ 0.019969024658203126,
267
+ 0.01982771110534668,
268
+ 0.020082687377929686,
269
+ 0.02001817512512207,
270
+ 0.02028646469116211,
271
+ 0.019912704467773438,
272
+ 0.019966976165771484,
273
+ 0.01991372871398926,
274
+ 0.019842048645019532,
275
+ 0.020016128540039063,
276
+ 0.019946495056152345,
277
+ 0.019915775299072267,
278
+ 0.019877887725830077,
279
+ 0.020076543807983398,
280
+ 0.019974143981933593,
281
+ 0.019747840881347657,
282
+ 0.019739648818969727,
283
+ 0.020040704727172853,
284
+ 0.01999667167663574,
285
+ 0.02002124786376953,
286
+ 0.019900415420532228,
287
+ 0.01989017677307129,
288
+ 0.019942399978637695,
289
+ 0.019989503860473632,
290
+ 0.02007142448425293,
291
+ 0.019990528106689453,
292
+ 0.019938304901123048,
293
+ 0.0200581111907959,
294
+ 0.019982336044311523,
295
+ 0.020000768661499024,
296
+ 0.02006220817565918,
297
+ 0.020548608779907225,
298
+ 0.01960140800476074,
299
+ 0.019989503860473632,
300
+ 0.01987379264831543,
301
+ 0.01998028755187988,
302
+ 0.019965951919555663,
303
+ 0.020039680480957032,
304
+ 0.01990656089782715,
305
+ 0.019955711364746095,
306
+ 0.02007756805419922,
307
+ 0.019978239059448243,
308
+ 0.019997695922851562,
309
+ 0.019954687118530275,
310
+ 0.019979263305664064,
311
+ 0.019940351486206053,
312
+ 0.019935232162475586,
313
+ 0.01990553665161133,
314
+ 0.02008166313171387,
315
+ 0.019908607482910155,
316
+ 0.019909631729125975,
317
+ 0.02002943992614746,
318
+ 0.02004684829711914,
319
+ 0.020389888763427736,
320
+ 0.019963903427124022,
321
+ 0.020025344848632814,
322
+ 0.02000588798522949,
323
+ 0.02002022361755371,
324
+ 0.019869695663452147,
325
+ 0.02000383949279785,
326
+ 0.019974143981933593,
327
+ 0.019843072891235353,
328
+ 0.019914751052856446,
329
+ 0.019976192474365235,
330
+ 0.019946495056152345,
331
+ 0.019787776947021486,
332
+ 0.019939327239990236,
333
+ 0.01984409523010254,
334
+ 0.019895296096801757,
335
+ 0.019765247344970704,
336
+ 0.0198973445892334,
337
+ 0.01998028755187988,
338
+ 0.019901439666748046,
339
+ 0.019915775299072267,
340
+ 0.02006425666809082,
341
+ 0.01982771110534668,
342
+ 0.019915775299072267,
343
+ 0.019993600845336915,
344
+ 0.019918848037719726,
345
+ 0.019932159423828123,
346
+ 0.01984614372253418,
347
+ 0.020101119995117187,
348
+ 0.019971071243286134,
349
+ 0.01993011283874512,
350
+ 0.019940351486206053,
351
+ 0.01987583923339844,
352
+ 0.019939327239990236,
353
+ 0.019941375732421874,
354
+ 0.019968000411987305,
355
+ 0.019924991607666014,
356
+ 0.01985638427734375,
357
+ 0.019826688766479493
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 49.18412912086681
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.021761743545532227,
7
+ "prefill.throughput.value": 321.6654026527726,
8
+ "decode.latency.mean": 2.582259033203125,
9
+ "decode.throughput.value": 49.18174294949206,
10
+ "per_token.latency.mean": 0.020331761848269486,
11
+ "per_token.throughput.value": 49.18412912086681
12
+ }
13
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 95e816f2bca48de32167ce6243e6770dee23923d
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d
88
+ - hydra.run.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1352.548352,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 13.4631787109375,
120
+ "mean": 13.4631787109375,
121
+ "stdev": 0.0,
122
+ "p50": 13.4631787109375,
123
+ "p90": 13.4631787109375,
124
+ "p95": 13.4631787109375,
125
+ "p99": 13.4631787109375,
126
+ "values": [
127
+ 13.4631787109375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1937.1008,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.028469568252563478,
147
+ "mean": 0.014234784126281739,
148
+ "stdev": 7.692813873291012e-05,
149
+ "p50": 0.014234784126281739,
150
+ "p90": 0.014296326637268066,
151
+ "p95": 0.014304019451141356,
152
+ "p99": 0.014310173702239989,
153
+ "values": [
154
+ 0.014157855987548828,
155
+ 0.014311712265014648
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 491.7531546597796
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2110.103552,
169
+ "max_global_vram": 6060.244992,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.1267911376953124,
178
+ "mean": 1.5633955688476562,
179
+ "stdev": 0.0011890258789062358,
180
+ "p50": 1.5633955688476562,
181
+ "p90": 1.5643467895507812,
182
+ "p95": 1.5644656921386717,
183
+ "p99": 1.5645608142089842,
184
+ "values": [
185
+ 1.5645845947265624,
186
+ 1.56220654296875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.23343991156943
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.126509575843812,
202
+ "mean": 0.012309092818282722,
203
+ "stdev": 5.2855735257225656e-05,
204
+ "p50": 0.012297216415405274,
205
+ "p90": 0.012330393695831299,
206
+ "p95": 0.012354559898376465,
207
+ "p99": 0.012616918916702271,
208
+ "values": [
209
+ 0.012327936172485352,
210
+ 0.01229312038421631,
211
+ 0.01235148811340332,
212
+ 0.012311552047729492,
213
+ 0.012292096138000488,
214
+ 0.012296192169189453,
215
+ 0.012299263954162597,
216
+ 0.012305407524108887,
217
+ 0.012311552047729492,
218
+ 0.012310527801513671,
219
+ 0.01245798397064209,
220
+ 0.012731391906738282,
221
+ 0.012596223831176758,
222
+ 0.012339200019836426,
223
+ 0.012354559898376465,
224
+ 0.012318719863891601,
225
+ 0.012310527801513671,
226
+ 0.012289024353027344,
227
+ 0.012297216415405274,
228
+ 0.012311552047729492,
229
+ 0.012314623832702636,
230
+ 0.01228492832183838,
231
+ 0.012299263954162597,
232
+ 0.012297216415405274,
233
+ 0.01229312038421631,
234
+ 0.012295167922973632,
235
+ 0.012296192169189453,
236
+ 0.012275712013244629,
237
+ 0.012326911926269531,
238
+ 0.012344320297241211,
239
+ 0.012304384231567383,
240
+ 0.012302335739135742,
241
+ 0.012296192169189453,
242
+ 0.012315648078918457,
243
+ 0.012285951614379884,
244
+ 0.012291071891784668,
245
+ 0.012297216415405274,
246
+ 0.012297216415405274,
247
+ 0.012299263954162597,
248
+ 0.012297216415405274,
249
+ 0.012306431770324706,
250
+ 0.012280832290649414,
251
+ 0.012279808044433594,
252
+ 0.012285951614379884,
253
+ 0.012301312446594239,
254
+ 0.012309503555297852,
255
+ 0.012298239707946777,
256
+ 0.012308480262756348,
257
+ 0.012305407524108887,
258
+ 0.012288000106811523,
259
+ 0.012304384231567383,
260
+ 0.012294143676757812,
261
+ 0.012302335739135742,
262
+ 0.012299263954162597,
263
+ 0.012289024353027344,
264
+ 0.012381183624267577,
265
+ 0.012296192169189453,
266
+ 0.012285951614379884,
267
+ 0.012313599586486817,
268
+ 0.012640255928039551,
269
+ 0.012662783622741699,
270
+ 0.012588031768798828,
271
+ 0.012317695617675782,
272
+ 0.012314623832702636,
273
+ 0.012304384231567383,
274
+ 0.012310527801513671,
275
+ 0.012274687767028808,
276
+ 0.012277759552001954,
277
+ 0.012331007957458496,
278
+ 0.012365823745727538,
279
+ 0.012307456016540527,
280
+ 0.012312576293945313,
281
+ 0.012298239707946777,
282
+ 0.012324864387512208,
283
+ 0.012300288200378418,
284
+ 0.012322815895080566,
285
+ 0.012274687767028808,
286
+ 0.012307456016540527,
287
+ 0.012297216415405274,
288
+ 0.012289024353027344,
289
+ 0.012332032203674317,
290
+ 0.012337151527404786,
291
+ 0.012300288200378418,
292
+ 0.012326911926269531,
293
+ 0.012291071891784668,
294
+ 0.012301312446594239,
295
+ 0.012291071891784668,
296
+ 0.012299263954162597,
297
+ 0.012307456016540527,
298
+ 0.012290047645568849,
299
+ 0.012285951614379884,
300
+ 0.012278783798217773,
301
+ 0.012283904075622559,
302
+ 0.012303359985351562,
303
+ 0.012378111839294433,
304
+ 0.012346367835998535,
305
+ 0.012311552047729492,
306
+ 0.012278783798217773,
307
+ 0.012291071891784668,
308
+ 0.012283904075622559,
309
+ 0.012285951614379884,
310
+ 0.012339200019836426,
311
+ 0.012295167922973632,
312
+ 0.012299263954162597,
313
+ 0.01229312038421631,
314
+ 0.012278783798217773,
315
+ 0.012313599586486817,
316
+ 0.012315648078918457,
317
+ 0.01232588768005371,
318
+ 0.012297216415405274,
319
+ 0.01228492832183838,
320
+ 0.012292096138000488,
321
+ 0.012295167922973632,
322
+ 0.012279808044433594,
323
+ 0.012290047645568849,
324
+ 0.012295167922973632,
325
+ 0.012294143676757812,
326
+ 0.01228492832183838,
327
+ 0.012305407524108887,
328
+ 0.012292096138000488,
329
+ 0.012295167922973632,
330
+ 0.01233510398864746,
331
+ 0.012318719863891601,
332
+ 0.01227673625946045,
333
+ 0.012277759552001954,
334
+ 0.012291071891784668,
335
+ 0.012302335739135742,
336
+ 0.012380160331726075,
337
+ 0.01229312038421631,
338
+ 0.012297216415405274,
339
+ 0.012400639533996583,
340
+ 0.012305407524108887,
341
+ 0.012290047645568849,
342
+ 0.012300288200378418,
343
+ 0.01228492832183838,
344
+ 0.012354559898376465,
345
+ 0.012301312446594239,
346
+ 0.012283904075622559,
347
+ 0.012300288200378418,
348
+ 0.012303359985351562,
349
+ 0.012297216415405274,
350
+ 0.012295167922973632,
351
+ 0.012299263954162597,
352
+ 0.01228492832183838,
353
+ 0.01229312038421631,
354
+ 0.012285951614379884,
355
+ 0.012273664474487305,
356
+ 0.012290047645568849,
357
+ 0.012290047645568849,
358
+ 0.01229312038421631,
359
+ 0.012292096138000488,
360
+ 0.012314623832702636,
361
+ 0.012311552047729492,
362
+ 0.012318719863891601,
363
+ 0.012296192169189453,
364
+ 0.012289024353027344,
365
+ 0.01229312038421631,
366
+ 0.012297280311584472,
367
+ 0.012277695655822754,
368
+ 0.012290047645568849,
369
+ 0.012304384231567383,
370
+ 0.012283904075622559,
371
+ 0.012283904075622559,
372
+ 0.012316672325134278,
373
+ 0.012301312446594239,
374
+ 0.012318719863891601,
375
+ 0.012324864387512208,
376
+ 0.01229312038421631,
377
+ 0.012308480262756348,
378
+ 0.012306431770324706,
379
+ 0.012278783798217773,
380
+ 0.012289024353027344,
381
+ 0.012297216415405274,
382
+ 0.01228492832183838,
383
+ 0.012310527801513671,
384
+ 0.01228492832183838,
385
+ 0.012348416328430176,
386
+ 0.012300288200378418,
387
+ 0.012267519950866699,
388
+ 0.012295167922973632,
389
+ 0.012340288162231445,
390
+ 0.012286911964416504,
391
+ 0.012304384231567383,
392
+ 0.012282879829406738,
393
+ 0.012292096138000488,
394
+ 0.012282879829406738,
395
+ 0.012289024353027344,
396
+ 0.012280832290649414,
397
+ 0.012278783798217773,
398
+ 0.012290047645568849,
399
+ 0.012299263954162597,
400
+ 0.012304384231567383,
401
+ 0.012312576293945313,
402
+ 0.012327936172485352,
403
+ 0.01229312038421631,
404
+ 0.012289024353027344,
405
+ 0.012288000106811523,
406
+ 0.012290047645568849,
407
+ 0.012300288200378418,
408
+ 0.012289024353027344,
409
+ 0.012288000106811523,
410
+ 0.01233510398864746,
411
+ 0.01229312038421631,
412
+ 0.012297216415405274,
413
+ 0.012288000106811523,
414
+ 0.012288000106811523,
415
+ 0.012323840141296387,
416
+ 0.012296192169189453,
417
+ 0.012286975860595703,
418
+ 0.012296192169189453,
419
+ 0.012298239707946777,
420
+ 0.012300288200378418,
421
+ 0.012316672325134278,
422
+ 0.012295167922973632,
423
+ 0.012291071891784668,
424
+ 0.012292096138000488,
425
+ 0.012363776206970215,
426
+ 0.012328960418701173,
427
+ 0.012289024353027344,
428
+ 0.012314623832702636,
429
+ 0.012299263954162597,
430
+ 0.012294143676757812,
431
+ 0.012280832290649414,
432
+ 0.012296192169189453,
433
+ 0.01228492832183838,
434
+ 0.012297216415405274,
435
+ 0.012304384231567383,
436
+ 0.012292096138000488,
437
+ 0.012294143676757812,
438
+ 0.012290047645568849,
439
+ 0.012280832290649414,
440
+ 0.012288000106811523,
441
+ 0.012304415702819824,
442
+ 0.012319711685180664,
443
+ 0.012296192169189453,
444
+ 0.012283904075622559,
445
+ 0.012278783798217773,
446
+ 0.01228492832183838,
447
+ 0.012275712013244629,
448
+ 0.012306431770324706,
449
+ 0.012301312446594239,
450
+ 0.012316672325134278,
451
+ 0.012304384231567383,
452
+ 0.012280832290649414,
453
+ 0.012291071891784668,
454
+ 0.012294143676757812,
455
+ 0.012313599586486817,
456
+ 0.012319744110107422,
457
+ 0.012296192169189453,
458
+ 0.012280832290649414,
459
+ 0.012323840141296387,
460
+ 0.012290047645568849,
461
+ 0.012295167922973632,
462
+ 0.012323840141296387
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.24075549375156
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "95e816f2bca48de32167ce6243e6770dee23923d",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1352.548352,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 13.4631787109375,
15
+ "mean": 13.4631787109375,
16
+ "stdev": 0.0,
17
+ "p50": 13.4631787109375,
18
+ "p90": 13.4631787109375,
19
+ "p95": 13.4631787109375,
20
+ "p99": 13.4631787109375,
21
+ "values": [
22
+ 13.4631787109375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1937.1008,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.028469568252563478,
42
+ "mean": 0.014234784126281739,
43
+ "stdev": 7.692813873291012e-05,
44
+ "p50": 0.014234784126281739,
45
+ "p90": 0.014296326637268066,
46
+ "p95": 0.014304019451141356,
47
+ "p99": 0.014310173702239989,
48
+ "values": [
49
+ 0.014157855987548828,
50
+ 0.014311712265014648
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 491.7531546597796
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2110.103552,
64
+ "max_global_vram": 6060.244992,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.1267911376953124,
73
+ "mean": 1.5633955688476562,
74
+ "stdev": 0.0011890258789062358,
75
+ "p50": 1.5633955688476562,
76
+ "p90": 1.5643467895507812,
77
+ "p95": 1.5644656921386717,
78
+ "p99": 1.5645608142089842,
79
+ "values": [
80
+ 1.5645845947265624,
81
+ 1.56220654296875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.23343991156943
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.126509575843812,
97
+ "mean": 0.012309092818282722,
98
+ "stdev": 5.2855735257225656e-05,
99
+ "p50": 0.012297216415405274,
100
+ "p90": 0.012330393695831299,
101
+ "p95": 0.012354559898376465,
102
+ "p99": 0.012616918916702271,
103
+ "values": [
104
+ 0.012327936172485352,
105
+ 0.01229312038421631,
106
+ 0.01235148811340332,
107
+ 0.012311552047729492,
108
+ 0.012292096138000488,
109
+ 0.012296192169189453,
110
+ 0.012299263954162597,
111
+ 0.012305407524108887,
112
+ 0.012311552047729492,
113
+ 0.012310527801513671,
114
+ 0.01245798397064209,
115
+ 0.012731391906738282,
116
+ 0.012596223831176758,
117
+ 0.012339200019836426,
118
+ 0.012354559898376465,
119
+ 0.012318719863891601,
120
+ 0.012310527801513671,
121
+ 0.012289024353027344,
122
+ 0.012297216415405274,
123
+ 0.012311552047729492,
124
+ 0.012314623832702636,
125
+ 0.01228492832183838,
126
+ 0.012299263954162597,
127
+ 0.012297216415405274,
128
+ 0.01229312038421631,
129
+ 0.012295167922973632,
130
+ 0.012296192169189453,
131
+ 0.012275712013244629,
132
+ 0.012326911926269531,
133
+ 0.012344320297241211,
134
+ 0.012304384231567383,
135
+ 0.012302335739135742,
136
+ 0.012296192169189453,
137
+ 0.012315648078918457,
138
+ 0.012285951614379884,
139
+ 0.012291071891784668,
140
+ 0.012297216415405274,
141
+ 0.012297216415405274,
142
+ 0.012299263954162597,
143
+ 0.012297216415405274,
144
+ 0.012306431770324706,
145
+ 0.012280832290649414,
146
+ 0.012279808044433594,
147
+ 0.012285951614379884,
148
+ 0.012301312446594239,
149
+ 0.012309503555297852,
150
+ 0.012298239707946777,
151
+ 0.012308480262756348,
152
+ 0.012305407524108887,
153
+ 0.012288000106811523,
154
+ 0.012304384231567383,
155
+ 0.012294143676757812,
156
+ 0.012302335739135742,
157
+ 0.012299263954162597,
158
+ 0.012289024353027344,
159
+ 0.012381183624267577,
160
+ 0.012296192169189453,
161
+ 0.012285951614379884,
162
+ 0.012313599586486817,
163
+ 0.012640255928039551,
164
+ 0.012662783622741699,
165
+ 0.012588031768798828,
166
+ 0.012317695617675782,
167
+ 0.012314623832702636,
168
+ 0.012304384231567383,
169
+ 0.012310527801513671,
170
+ 0.012274687767028808,
171
+ 0.012277759552001954,
172
+ 0.012331007957458496,
173
+ 0.012365823745727538,
174
+ 0.012307456016540527,
175
+ 0.012312576293945313,
176
+ 0.012298239707946777,
177
+ 0.012324864387512208,
178
+ 0.012300288200378418,
179
+ 0.012322815895080566,
180
+ 0.012274687767028808,
181
+ 0.012307456016540527,
182
+ 0.012297216415405274,
183
+ 0.012289024353027344,
184
+ 0.012332032203674317,
185
+ 0.012337151527404786,
186
+ 0.012300288200378418,
187
+ 0.012326911926269531,
188
+ 0.012291071891784668,
189
+ 0.012301312446594239,
190
+ 0.012291071891784668,
191
+ 0.012299263954162597,
192
+ 0.012307456016540527,
193
+ 0.012290047645568849,
194
+ 0.012285951614379884,
195
+ 0.012278783798217773,
196
+ 0.012283904075622559,
197
+ 0.012303359985351562,
198
+ 0.012378111839294433,
199
+ 0.012346367835998535,
200
+ 0.012311552047729492,
201
+ 0.012278783798217773,
202
+ 0.012291071891784668,
203
+ 0.012283904075622559,
204
+ 0.012285951614379884,
205
+ 0.012339200019836426,
206
+ 0.012295167922973632,
207
+ 0.012299263954162597,
208
+ 0.01229312038421631,
209
+ 0.012278783798217773,
210
+ 0.012313599586486817,
211
+ 0.012315648078918457,
212
+ 0.01232588768005371,
213
+ 0.012297216415405274,
214
+ 0.01228492832183838,
215
+ 0.012292096138000488,
216
+ 0.012295167922973632,
217
+ 0.012279808044433594,
218
+ 0.012290047645568849,
219
+ 0.012295167922973632,
220
+ 0.012294143676757812,
221
+ 0.01228492832183838,
222
+ 0.012305407524108887,
223
+ 0.012292096138000488,
224
+ 0.012295167922973632,
225
+ 0.01233510398864746,
226
+ 0.012318719863891601,
227
+ 0.01227673625946045,
228
+ 0.012277759552001954,
229
+ 0.012291071891784668,
230
+ 0.012302335739135742,
231
+ 0.012380160331726075,
232
+ 0.01229312038421631,
233
+ 0.012297216415405274,
234
+ 0.012400639533996583,
235
+ 0.012305407524108887,
236
+ 0.012290047645568849,
237
+ 0.012300288200378418,
238
+ 0.01228492832183838,
239
+ 0.012354559898376465,
240
+ 0.012301312446594239,
241
+ 0.012283904075622559,
242
+ 0.012300288200378418,
243
+ 0.012303359985351562,
244
+ 0.012297216415405274,
245
+ 0.012295167922973632,
246
+ 0.012299263954162597,
247
+ 0.01228492832183838,
248
+ 0.01229312038421631,
249
+ 0.012285951614379884,
250
+ 0.012273664474487305,
251
+ 0.012290047645568849,
252
+ 0.012290047645568849,
253
+ 0.01229312038421631,
254
+ 0.012292096138000488,
255
+ 0.012314623832702636,
256
+ 0.012311552047729492,
257
+ 0.012318719863891601,
258
+ 0.012296192169189453,
259
+ 0.012289024353027344,
260
+ 0.01229312038421631,
261
+ 0.012297280311584472,
262
+ 0.012277695655822754,
263
+ 0.012290047645568849,
264
+ 0.012304384231567383,
265
+ 0.012283904075622559,
266
+ 0.012283904075622559,
267
+ 0.012316672325134278,
268
+ 0.012301312446594239,
269
+ 0.012318719863891601,
270
+ 0.012324864387512208,
271
+ 0.01229312038421631,
272
+ 0.012308480262756348,
273
+ 0.012306431770324706,
274
+ 0.012278783798217773,
275
+ 0.012289024353027344,
276
+ 0.012297216415405274,
277
+ 0.01228492832183838,
278
+ 0.012310527801513671,
279
+ 0.01228492832183838,
280
+ 0.012348416328430176,
281
+ 0.012300288200378418,
282
+ 0.012267519950866699,
283
+ 0.012295167922973632,
284
+ 0.012340288162231445,
285
+ 0.012286911964416504,
286
+ 0.012304384231567383,
287
+ 0.012282879829406738,
288
+ 0.012292096138000488,
289
+ 0.012282879829406738,
290
+ 0.012289024353027344,
291
+ 0.012280832290649414,
292
+ 0.012278783798217773,
293
+ 0.012290047645568849,
294
+ 0.012299263954162597,
295
+ 0.012304384231567383,
296
+ 0.012312576293945313,
297
+ 0.012327936172485352,
298
+ 0.01229312038421631,
299
+ 0.012289024353027344,
300
+ 0.012288000106811523,
301
+ 0.012290047645568849,
302
+ 0.012300288200378418,
303
+ 0.012289024353027344,
304
+ 0.012288000106811523,
305
+ 0.01233510398864746,
306
+ 0.01229312038421631,
307
+ 0.012297216415405274,
308
+ 0.012288000106811523,
309
+ 0.012288000106811523,
310
+ 0.012323840141296387,
311
+ 0.012296192169189453,
312
+ 0.012286975860595703,
313
+ 0.012296192169189453,
314
+ 0.012298239707946777,
315
+ 0.012300288200378418,
316
+ 0.012316672325134278,
317
+ 0.012295167922973632,
318
+ 0.012291071891784668,
319
+ 0.012292096138000488,
320
+ 0.012363776206970215,
321
+ 0.012328960418701173,
322
+ 0.012289024353027344,
323
+ 0.012314623832702636,
324
+ 0.012299263954162597,
325
+ 0.012294143676757812,
326
+ 0.012280832290649414,
327
+ 0.012296192169189453,
328
+ 0.01228492832183838,
329
+ 0.012297216415405274,
330
+ 0.012304384231567383,
331
+ 0.012292096138000488,
332
+ 0.012294143676757812,
333
+ 0.012290047645568849,
334
+ 0.012280832290649414,
335
+ 0.012288000106811523,
336
+ 0.012304415702819824,
337
+ 0.012319711685180664,
338
+ 0.012296192169189453,
339
+ 0.012283904075622559,
340
+ 0.012278783798217773,
341
+ 0.01228492832183838,
342
+ 0.012275712013244629,
343
+ 0.012306431770324706,
344
+ 0.012301312446594239,
345
+ 0.012316672325134278,
346
+ 0.012304384231567383,
347
+ 0.012280832290649414,
348
+ 0.012291071891784668,
349
+ 0.012294143676757812,
350
+ 0.012313599586486817,
351
+ 0.012319744110107422,
352
+ 0.012296192169189453,
353
+ 0.012280832290649414,
354
+ 0.012323840141296387,
355
+ 0.012290047645568849,
356
+ 0.012295167922973632,
357
+ 0.012323840141296387
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.24075549375156
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.014234784126281739,
7
+ "prefill.throughput.value": 491.7531546597796,
8
+ "decode.latency.mean": 1.5633955688476562,
9
+ "decode.throughput.value": 81.23343991156943,
10
+ "per_token.latency.mean": 0.012309092818282722,
11
+ "per_token.throughput.value": 81.24075549375156
12
+ }
13
+ }
2024-09-16/commit=95e816f2bca48de32167ce6243e6770dee23923d/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpz5e_ov2f/commit=95e816f2bca48de32167ce6243e6770dee23923d
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d
88
+ - hydra.run.dir=_benchmark/tmpz5e_ov2f/commit\=95e816f2bca48de32167ce6243e6770dee23923d/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: 95e816f2bca48de32167ce6243e6770dee23923d
225
+ accelerate_version: 0.35.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.23.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-09-16/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.021094143867492678,
8
+ "prefill.throughput.value": 331.8456555512269,
9
+ "decode.latency.mean": 2.5467454833984373,
10
+ "decode.throughput.value": 49.867566597400305,
11
+ "per_token.latency.mean": 0.020051847052386426,
12
+ "per_token.throughput.value": 49.87071751482301
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.021761743545532227,
21
+ "prefill.throughput.value": 321.6654026527726,
22
+ "decode.latency.mean": 2.582259033203125,
23
+ "decode.throughput.value": 49.18174294949206,
24
+ "per_token.latency.mean": 0.020331761848269486,
25
+ "per_token.throughput.value": 49.18412912086681
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "95e816f2bca48de32167ce6243e6770dee23923d",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.014234784126281739,
34
+ "prefill.throughput.value": 491.7531546597796,
35
+ "decode.latency.mean": 1.5633955688476562,
36
+ "decode.throughput.value": 81.23343991156943,
37
+ "per_token.latency.mean": 0.012309092818282722,
38
+ "per_token.throughput.value": 81.24075549375156
39
+ }
40
+ }
41
+ ]
2024-09-16/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "95e816f2bca48de32167ce6243e6770dee23923d": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.021094143867492678,
7
+ "prefill.throughput.value": 331.8456555512269,
8
+ "decode.latency.mean": 2.5467454833984373,
9
+ "decode.throughput.value": 49.867566597400305,
10
+ "per_token.latency.mean": 0.020051847052386426,
11
+ "per_token.throughput.value": 49.87071751482301
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "95e816f2bca48de32167ce6243e6770dee23923d": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.021761743545532227,
19
+ "prefill.throughput.value": 321.6654026527726,
20
+ "decode.latency.mean": 2.582259033203125,
21
+ "decode.throughput.value": 49.18174294949206,
22
+ "per_token.latency.mean": 0.020331761848269486,
23
+ "per_token.throughput.value": 49.18412912086681
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "95e816f2bca48de32167ce6243e6770dee23923d": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.014234784126281739,
31
+ "prefill.throughput.value": 491.7531546597796,
32
+ "decode.latency.mean": 1.5633955688476562,
33
+ "decode.throughput.value": 81.23343991156943,
34
+ "per_token.latency.mean": 0.012309092818282722,
35
+ "per_token.throughput.value": 81.24075549375156
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }