hf-transformers-bot commited on
Commit
8cba6a4
1 Parent(s): e0b6f2d

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/multirun.yaml +234 -0
  31. 2024-09-03/summaries.json +21 -21
  32. 2024-09-03/summary.json +21 -21
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 7ed9789e210d8eca797fc21b9c783b1ce718ecb5
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
88
+ - hydra.run.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1320.235008,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.0903349609375,
120
+ "mean": 12.0903349609375,
121
+ "stdev": 0.0,
122
+ "p50": 12.0903349609375,
123
+ "p90": 12.0903349609375,
124
+ "p95": 12.0903349609375,
125
+ "p99": 12.0903349609375,
126
+ "values": [
127
+ 12.0903349609375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1717.223424,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.04036412811279297,
147
+ "mean": 0.020182064056396484,
148
+ "stdev": 7.204818725585829e-05,
149
+ "p50": 0.020182064056396484,
150
+ "p90": 0.02023970260620117,
151
+ "p95": 0.020246907424926755,
152
+ "p99": 0.020252671279907226,
153
+ "values": [
154
+ 0.020110015869140626,
155
+ 0.020254112243652343
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 346.8426212720015
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1771.278336,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 4.834597412109375,
178
+ "mean": 2.4172987060546873,
179
+ "stdev": 0.003500610351562461,
180
+ "p50": 2.4172987060546873,
181
+ "p90": 2.420099194335937,
182
+ "p95": 2.4204492553710937,
183
+ "p99": 2.4207293041992184,
184
+ "values": [
185
+ 2.42079931640625,
186
+ 2.413798095703125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 52.53798369307812
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 4.83428351974487,
202
+ "mean": 0.01903261228246013,
203
+ "stdev": 0.00035680424553614915,
204
+ "p50": 0.018873856544494628,
205
+ "p90": 0.01952471103668213,
206
+ "p95": 0.019621581172943117,
207
+ "p99": 0.019915550537109375,
208
+ "values": [
209
+ 0.01866035270690918,
210
+ 0.01866547203063965,
211
+ 0.018729984283447267,
212
+ 0.018747392654418944,
213
+ 0.018737152099609376,
214
+ 0.018769920349121092,
215
+ 0.018736127853393555,
216
+ 0.018718719482421875,
217
+ 0.018751487731933594,
218
+ 0.018728992462158204,
219
+ 0.018739168167114257,
220
+ 0.018715648651123046,
221
+ 0.018712575912475587,
222
+ 0.018874368667602538,
223
+ 0.018763776779174804,
224
+ 0.018708480834960937,
225
+ 0.018692096710205077,
226
+ 0.018766847610473633,
227
+ 0.019534847259521485,
228
+ 0.01940377616882324,
229
+ 0.019358720779418945,
230
+ 0.0194201602935791,
231
+ 0.01878118324279785,
232
+ 0.018701311111450195,
233
+ 0.018697216033935548,
234
+ 0.018745344161987306,
235
+ 0.018739200592041014,
236
+ 0.018741247177124023,
237
+ 0.0186060791015625,
238
+ 0.019017728805541992,
239
+ 0.01900441551208496,
240
+ 0.018782207489013672,
241
+ 0.018741247177124023,
242
+ 0.018739200592041014,
243
+ 0.018750463485717773,
244
+ 0.018739200592041014,
245
+ 0.018745344161987306,
246
+ 0.018840576171875,
247
+ 0.018724863052368163,
248
+ 0.018902015686035157,
249
+ 0.01880473518371582,
250
+ 0.018700288772583007,
251
+ 0.018774015426635742,
252
+ 0.018727935791015626,
253
+ 0.018685951232910156,
254
+ 0.018701343536376952,
255
+ 0.01872697639465332,
256
+ 0.01878518486022949,
257
+ 0.018700288772583007,
258
+ 0.018716672897338867,
259
+ 0.018735103607177735,
260
+ 0.018743295669555664,
261
+ 0.018803712844848632,
262
+ 0.018696191787719727,
263
+ 0.018735103607177735,
264
+ 0.018732032775878905,
265
+ 0.018742271423339844,
266
+ 0.01882828712463379,
267
+ 0.01882316780090332,
268
+ 0.018931711196899414,
269
+ 0.01888768005371094,
270
+ 0.02000486373901367,
271
+ 0.019955711364746095,
272
+ 0.01962598419189453,
273
+ 0.019491840362548828,
274
+ 0.019522560119628905,
275
+ 0.0194652156829834,
276
+ 0.019564544677734375,
277
+ 0.019526655197143555,
278
+ 0.01948057556152344,
279
+ 0.01961881637573242,
280
+ 0.019531776428222656,
281
+ 0.019474431991577147,
282
+ 0.019513343811035155,
283
+ 0.019499008178710937,
284
+ 0.019510271072387696,
285
+ 0.01987993621826172,
286
+ 0.02067251205444336,
287
+ 0.01971609687805176,
288
+ 0.019588096618652344,
289
+ 0.019620864868164063,
290
+ 0.01962291145324707,
291
+ 0.019494911193847657,
292
+ 0.019809280395507813,
293
+ 0.01982975959777832,
294
+ 0.019709951400756837,
295
+ 0.01955843162536621,
296
+ 0.0194815673828125,
297
+ 0.01944883155822754,
298
+ 0.01949388885498047,
299
+ 0.019503103256225587,
300
+ 0.019514368057250975,
301
+ 0.01947648048400879,
302
+ 0.018971647262573242,
303
+ 0.01884671974182129,
304
+ 0.01881804847717285,
305
+ 0.01883955192565918,
306
+ 0.01883647918701172,
307
+ 0.01886412811279297,
308
+ 0.019299392700195313,
309
+ 0.018917312622070314,
310
+ 0.018889728546142577,
311
+ 0.018900991439819336,
312
+ 0.01887129592895508,
313
+ 0.018916351318359375,
314
+ 0.018894847869873048,
315
+ 0.01884876823425293,
316
+ 0.01882828712463379,
317
+ 0.01886720085144043,
318
+ 0.01887948799133301,
319
+ 0.018865152359008788,
320
+ 0.01886617660522461,
321
+ 0.018874368667602538,
322
+ 0.018860031127929687,
323
+ 0.01889587211608887,
324
+ 0.018888704299926756,
325
+ 0.018982912063598634,
326
+ 0.018905088424682616,
327
+ 0.018869247436523438,
328
+ 0.018844671249389648,
329
+ 0.01884364891052246,
330
+ 0.01887539291381836,
331
+ 0.019363840103149413,
332
+ 0.019710975646972655,
333
+ 0.019525632858276368,
334
+ 0.019511295318603517,
335
+ 0.019547136306762695,
336
+ 0.018674720764160158,
337
+ 0.018649055480957032,
338
+ 0.018686975479125977,
339
+ 0.01874947166442871,
340
+ 0.018744287490844728,
341
+ 0.018729984283447267,
342
+ 0.018754560470581053,
343
+ 0.018724863052368163,
344
+ 0.018733055114746093,
345
+ 0.018539520263671876,
346
+ 0.019207168579101562,
347
+ 0.018821184158325194,
348
+ 0.018713632583618165,
349
+ 0.01877494430541992,
350
+ 0.018778112411499022,
351
+ 0.018761728286743166,
352
+ 0.018741247177124023,
353
+ 0.018767871856689454,
354
+ 0.018762752532958983,
355
+ 0.018693119049072265,
356
+ 0.018712575912475587,
357
+ 0.01876479911804199,
358
+ 0.018754560470581053,
359
+ 0.018759679794311524,
360
+ 0.019062784194946288,
361
+ 0.019414016723632813,
362
+ 0.019405824661254883,
363
+ 0.019364864349365234,
364
+ 0.019337215423583985,
365
+ 0.019516416549682617,
366
+ 0.01942937660217285,
367
+ 0.01939558410644531,
368
+ 0.019367935180664063,
369
+ 0.019347455978393553,
370
+ 0.01939148712158203,
371
+ 0.019361791610717775,
372
+ 0.019422208786010742,
373
+ 0.01938739204406738,
374
+ 0.01927577590942383,
375
+ 0.0194150390625,
376
+ 0.019393535614013673,
377
+ 0.019384319305419923,
378
+ 0.01944576072692871,
379
+ 0.01943756866455078,
380
+ 0.019409919738769533,
381
+ 0.018762752532958983,
382
+ 0.018767871856689454,
383
+ 0.018712575912475587,
384
+ 0.018708480834960937,
385
+ 0.018729984283447267,
386
+ 0.018717695236206054,
387
+ 0.018722816467285155,
388
+ 0.018740224838256835,
389
+ 0.018760704040527345,
390
+ 0.018762752532958983,
391
+ 0.018747392654418944,
392
+ 0.018739200592041014,
393
+ 0.018915327072143554,
394
+ 0.01882931137084961,
395
+ 0.01884774398803711,
396
+ 0.01888768005371094,
397
+ 0.018775039672851563,
398
+ 0.018851839065551757,
399
+ 0.018896896362304686,
400
+ 0.019219455718994142,
401
+ 0.019119104385375976,
402
+ 0.0189040641784668,
403
+ 0.018711551666259766,
404
+ 0.018894847869873048,
405
+ 0.01889587211608887,
406
+ 0.018880512237548826,
407
+ 0.018884607315063476,
408
+ 0.01884364891052246,
409
+ 0.018811904907226562,
410
+ 0.018912256240844725,
411
+ 0.018967552185058592,
412
+ 0.018906112670898437,
413
+ 0.018873344421386717,
414
+ 0.018877439498901367,
415
+ 0.018910207748413087,
416
+ 0.018897920608520507,
417
+ 0.018885631561279297,
418
+ 0.01885593605041504,
419
+ 0.018884607315063476,
420
+ 0.01886207962036133,
421
+ 0.018919424057006837,
422
+ 0.019499008178710937,
423
+ 0.019522560119628905,
424
+ 0.019599359512329103,
425
+ 0.0188723201751709,
426
+ 0.01880166435241699,
427
+ 0.01883955192565918,
428
+ 0.018881536483764647,
429
+ 0.01880575942993164,
430
+ 0.01883545684814453,
431
+ 0.018774015426635742,
432
+ 0.018749439239501953,
433
+ 0.018860031127929687,
434
+ 0.01882111930847168,
435
+ 0.018886655807495118,
436
+ 0.01987583923339844,
437
+ 0.019475456237792968,
438
+ 0.01945907211303711,
439
+ 0.019575807571411134,
440
+ 0.019530752182006835,
441
+ 0.0194969596862793,
442
+ 0.019494911193847657,
443
+ 0.019520511627197267,
444
+ 0.019629056930541993,
445
+ 0.019486719131469727,
446
+ 0.019487743377685548,
447
+ 0.01949286460876465,
448
+ 0.01885798454284668,
449
+ 0.018886655807495118,
450
+ 0.018824192047119142,
451
+ 0.01882316780090332,
452
+ 0.018902015686035157,
453
+ 0.018912256240844725,
454
+ 0.01884876823425293,
455
+ 0.019311616897583008,
456
+ 0.018989055633544923,
457
+ 0.018974720001220705,
458
+ 0.018874368667602538,
459
+ 0.018897920608520507,
460
+ 0.01884262466430664,
461
+ 0.01882828712463379,
462
+ 0.01883135986328125
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 52.54139500974174
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1320.235008,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.0903349609375,
15
+ "mean": 12.0903349609375,
16
+ "stdev": 0.0,
17
+ "p50": 12.0903349609375,
18
+ "p90": 12.0903349609375,
19
+ "p95": 12.0903349609375,
20
+ "p99": 12.0903349609375,
21
+ "values": [
22
+ 12.0903349609375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1717.223424,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.04036412811279297,
42
+ "mean": 0.020182064056396484,
43
+ "stdev": 7.204818725585829e-05,
44
+ "p50": 0.020182064056396484,
45
+ "p90": 0.02023970260620117,
46
+ "p95": 0.020246907424926755,
47
+ "p99": 0.020252671279907226,
48
+ "values": [
49
+ 0.020110015869140626,
50
+ 0.020254112243652343
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 346.8426212720015
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1771.278336,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 4.834597412109375,
73
+ "mean": 2.4172987060546873,
74
+ "stdev": 0.003500610351562461,
75
+ "p50": 2.4172987060546873,
76
+ "p90": 2.420099194335937,
77
+ "p95": 2.4204492553710937,
78
+ "p99": 2.4207293041992184,
79
+ "values": [
80
+ 2.42079931640625,
81
+ 2.413798095703125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 52.53798369307812
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 4.83428351974487,
97
+ "mean": 0.01903261228246013,
98
+ "stdev": 0.00035680424553614915,
99
+ "p50": 0.018873856544494628,
100
+ "p90": 0.01952471103668213,
101
+ "p95": 0.019621581172943117,
102
+ "p99": 0.019915550537109375,
103
+ "values": [
104
+ 0.01866035270690918,
105
+ 0.01866547203063965,
106
+ 0.018729984283447267,
107
+ 0.018747392654418944,
108
+ 0.018737152099609376,
109
+ 0.018769920349121092,
110
+ 0.018736127853393555,
111
+ 0.018718719482421875,
112
+ 0.018751487731933594,
113
+ 0.018728992462158204,
114
+ 0.018739168167114257,
115
+ 0.018715648651123046,
116
+ 0.018712575912475587,
117
+ 0.018874368667602538,
118
+ 0.018763776779174804,
119
+ 0.018708480834960937,
120
+ 0.018692096710205077,
121
+ 0.018766847610473633,
122
+ 0.019534847259521485,
123
+ 0.01940377616882324,
124
+ 0.019358720779418945,
125
+ 0.0194201602935791,
126
+ 0.01878118324279785,
127
+ 0.018701311111450195,
128
+ 0.018697216033935548,
129
+ 0.018745344161987306,
130
+ 0.018739200592041014,
131
+ 0.018741247177124023,
132
+ 0.0186060791015625,
133
+ 0.019017728805541992,
134
+ 0.01900441551208496,
135
+ 0.018782207489013672,
136
+ 0.018741247177124023,
137
+ 0.018739200592041014,
138
+ 0.018750463485717773,
139
+ 0.018739200592041014,
140
+ 0.018745344161987306,
141
+ 0.018840576171875,
142
+ 0.018724863052368163,
143
+ 0.018902015686035157,
144
+ 0.01880473518371582,
145
+ 0.018700288772583007,
146
+ 0.018774015426635742,
147
+ 0.018727935791015626,
148
+ 0.018685951232910156,
149
+ 0.018701343536376952,
150
+ 0.01872697639465332,
151
+ 0.01878518486022949,
152
+ 0.018700288772583007,
153
+ 0.018716672897338867,
154
+ 0.018735103607177735,
155
+ 0.018743295669555664,
156
+ 0.018803712844848632,
157
+ 0.018696191787719727,
158
+ 0.018735103607177735,
159
+ 0.018732032775878905,
160
+ 0.018742271423339844,
161
+ 0.01882828712463379,
162
+ 0.01882316780090332,
163
+ 0.018931711196899414,
164
+ 0.01888768005371094,
165
+ 0.02000486373901367,
166
+ 0.019955711364746095,
167
+ 0.01962598419189453,
168
+ 0.019491840362548828,
169
+ 0.019522560119628905,
170
+ 0.0194652156829834,
171
+ 0.019564544677734375,
172
+ 0.019526655197143555,
173
+ 0.01948057556152344,
174
+ 0.01961881637573242,
175
+ 0.019531776428222656,
176
+ 0.019474431991577147,
177
+ 0.019513343811035155,
178
+ 0.019499008178710937,
179
+ 0.019510271072387696,
180
+ 0.01987993621826172,
181
+ 0.02067251205444336,
182
+ 0.01971609687805176,
183
+ 0.019588096618652344,
184
+ 0.019620864868164063,
185
+ 0.01962291145324707,
186
+ 0.019494911193847657,
187
+ 0.019809280395507813,
188
+ 0.01982975959777832,
189
+ 0.019709951400756837,
190
+ 0.01955843162536621,
191
+ 0.0194815673828125,
192
+ 0.01944883155822754,
193
+ 0.01949388885498047,
194
+ 0.019503103256225587,
195
+ 0.019514368057250975,
196
+ 0.01947648048400879,
197
+ 0.018971647262573242,
198
+ 0.01884671974182129,
199
+ 0.01881804847717285,
200
+ 0.01883955192565918,
201
+ 0.01883647918701172,
202
+ 0.01886412811279297,
203
+ 0.019299392700195313,
204
+ 0.018917312622070314,
205
+ 0.018889728546142577,
206
+ 0.018900991439819336,
207
+ 0.01887129592895508,
208
+ 0.018916351318359375,
209
+ 0.018894847869873048,
210
+ 0.01884876823425293,
211
+ 0.01882828712463379,
212
+ 0.01886720085144043,
213
+ 0.01887948799133301,
214
+ 0.018865152359008788,
215
+ 0.01886617660522461,
216
+ 0.018874368667602538,
217
+ 0.018860031127929687,
218
+ 0.01889587211608887,
219
+ 0.018888704299926756,
220
+ 0.018982912063598634,
221
+ 0.018905088424682616,
222
+ 0.018869247436523438,
223
+ 0.018844671249389648,
224
+ 0.01884364891052246,
225
+ 0.01887539291381836,
226
+ 0.019363840103149413,
227
+ 0.019710975646972655,
228
+ 0.019525632858276368,
229
+ 0.019511295318603517,
230
+ 0.019547136306762695,
231
+ 0.018674720764160158,
232
+ 0.018649055480957032,
233
+ 0.018686975479125977,
234
+ 0.01874947166442871,
235
+ 0.018744287490844728,
236
+ 0.018729984283447267,
237
+ 0.018754560470581053,
238
+ 0.018724863052368163,
239
+ 0.018733055114746093,
240
+ 0.018539520263671876,
241
+ 0.019207168579101562,
242
+ 0.018821184158325194,
243
+ 0.018713632583618165,
244
+ 0.01877494430541992,
245
+ 0.018778112411499022,
246
+ 0.018761728286743166,
247
+ 0.018741247177124023,
248
+ 0.018767871856689454,
249
+ 0.018762752532958983,
250
+ 0.018693119049072265,
251
+ 0.018712575912475587,
252
+ 0.01876479911804199,
253
+ 0.018754560470581053,
254
+ 0.018759679794311524,
255
+ 0.019062784194946288,
256
+ 0.019414016723632813,
257
+ 0.019405824661254883,
258
+ 0.019364864349365234,
259
+ 0.019337215423583985,
260
+ 0.019516416549682617,
261
+ 0.01942937660217285,
262
+ 0.01939558410644531,
263
+ 0.019367935180664063,
264
+ 0.019347455978393553,
265
+ 0.01939148712158203,
266
+ 0.019361791610717775,
267
+ 0.019422208786010742,
268
+ 0.01938739204406738,
269
+ 0.01927577590942383,
270
+ 0.0194150390625,
271
+ 0.019393535614013673,
272
+ 0.019384319305419923,
273
+ 0.01944576072692871,
274
+ 0.01943756866455078,
275
+ 0.019409919738769533,
276
+ 0.018762752532958983,
277
+ 0.018767871856689454,
278
+ 0.018712575912475587,
279
+ 0.018708480834960937,
280
+ 0.018729984283447267,
281
+ 0.018717695236206054,
282
+ 0.018722816467285155,
283
+ 0.018740224838256835,
284
+ 0.018760704040527345,
285
+ 0.018762752532958983,
286
+ 0.018747392654418944,
287
+ 0.018739200592041014,
288
+ 0.018915327072143554,
289
+ 0.01882931137084961,
290
+ 0.01884774398803711,
291
+ 0.01888768005371094,
292
+ 0.018775039672851563,
293
+ 0.018851839065551757,
294
+ 0.018896896362304686,
295
+ 0.019219455718994142,
296
+ 0.019119104385375976,
297
+ 0.0189040641784668,
298
+ 0.018711551666259766,
299
+ 0.018894847869873048,
300
+ 0.01889587211608887,
301
+ 0.018880512237548826,
302
+ 0.018884607315063476,
303
+ 0.01884364891052246,
304
+ 0.018811904907226562,
305
+ 0.018912256240844725,
306
+ 0.018967552185058592,
307
+ 0.018906112670898437,
308
+ 0.018873344421386717,
309
+ 0.018877439498901367,
310
+ 0.018910207748413087,
311
+ 0.018897920608520507,
312
+ 0.018885631561279297,
313
+ 0.01885593605041504,
314
+ 0.018884607315063476,
315
+ 0.01886207962036133,
316
+ 0.018919424057006837,
317
+ 0.019499008178710937,
318
+ 0.019522560119628905,
319
+ 0.019599359512329103,
320
+ 0.0188723201751709,
321
+ 0.01880166435241699,
322
+ 0.01883955192565918,
323
+ 0.018881536483764647,
324
+ 0.01880575942993164,
325
+ 0.01883545684814453,
326
+ 0.018774015426635742,
327
+ 0.018749439239501953,
328
+ 0.018860031127929687,
329
+ 0.01882111930847168,
330
+ 0.018886655807495118,
331
+ 0.01987583923339844,
332
+ 0.019475456237792968,
333
+ 0.01945907211303711,
334
+ 0.019575807571411134,
335
+ 0.019530752182006835,
336
+ 0.0194969596862793,
337
+ 0.019494911193847657,
338
+ 0.019520511627197267,
339
+ 0.019629056930541993,
340
+ 0.019486719131469727,
341
+ 0.019487743377685548,
342
+ 0.01949286460876465,
343
+ 0.01885798454284668,
344
+ 0.018886655807495118,
345
+ 0.018824192047119142,
346
+ 0.01882316780090332,
347
+ 0.018902015686035157,
348
+ 0.018912256240844725,
349
+ 0.01884876823425293,
350
+ 0.019311616897583008,
351
+ 0.018989055633544923,
352
+ 0.018974720001220705,
353
+ 0.018874368667602538,
354
+ 0.018897920608520507,
355
+ 0.01884262466430664,
356
+ 0.01882828712463379,
357
+ 0.01883135986328125
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 52.54139500974174
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.020182064056396484,
7
+ "prefill.throughput.value": 346.8426212720015,
8
+ "decode.latency.mean": 2.4172987060546873,
9
+ "decode.throughput.value": 52.53798369307812,
10
+ "per_token.latency.mean": 0.01903261228246013,
11
+ "per_token.throughput.value": 52.54139500974174
12
+ }
13
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 7ed9789e210d8eca797fc21b9c783b1ce718ecb5
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
88
+ - hydra.run.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-09-03 08:58:02,531][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-09-03 08:58:04,895][process][ERROR] - + Received traceback from isolated process
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 7ed9789e210d8eca797fc21b9c783b1ce718ecb5
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
88
+ - hydra.run.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1319.092224,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.3471416015625,
120
+ "mean": 12.3471416015625,
121
+ "stdev": 0.0,
122
+ "p50": 12.3471416015625,
123
+ "p90": 12.3471416015625,
124
+ "p95": 12.3471416015625,
125
+ "p99": 12.3471416015625,
126
+ "values": [
127
+ 12.3471416015625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1777.262592,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.043444223403930664,
147
+ "mean": 0.021722111701965332,
148
+ "stdev": 3.3922195434576974e-06,
149
+ "p50": 0.021722111701965332,
150
+ "p90": 0.021724825477600098,
151
+ "p95": 0.021725164699554444,
152
+ "p99": 0.02172543607711792,
153
+ "values": [
154
+ 0.021718719482421874,
155
+ 0.02172550392150879
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 322.25227897003526
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1801.777152,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.2023725585937495,
178
+ "mean": 2.6011862792968747,
179
+ "stdev": 0.02596093750000006,
180
+ "p50": 2.6011862792968747,
181
+ "p90": 2.621955029296875,
182
+ "p95": 2.624551123046875,
183
+ "p99": 2.626627998046875,
184
+ "values": [
185
+ 2.575225341796875,
186
+ 2.627147216796875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 48.82387740194036
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.202118667602542,
202
+ "mean": 0.02048078215591551,
203
+ "stdev": 0.0005142246344054906,
204
+ "p50": 0.02075033664703369,
205
+ "p90": 0.02090813446044922,
206
+ "p95": 0.020953088760375976,
207
+ "p99": 0.02174282781600952,
208
+ "values": [
209
+ 0.020039680480957032,
210
+ 0.020017152786254884,
211
+ 0.020057088851928712,
212
+ 0.02002022361755371,
213
+ 0.020024320602416993,
214
+ 0.019985408782958985,
215
+ 0.02002841567993164,
216
+ 0.019984384536743165,
217
+ 0.020306943893432617,
218
+ 0.020801536560058592,
219
+ 0.02079539108276367,
220
+ 0.02008576011657715,
221
+ 0.02008064079284668,
222
+ 0.01998847961425781,
223
+ 0.02000588798522949,
224
+ 0.019998720169067383,
225
+ 0.019979263305664064,
226
+ 0.020007936477661133,
227
+ 0.02004172706604004,
228
+ 0.020006912231445313,
229
+ 0.020116479873657226,
230
+ 0.02003558349609375,
231
+ 0.020083711624145507,
232
+ 0.02003046417236328,
233
+ 0.020983808517456053,
234
+ 0.02087936019897461,
235
+ 0.020783103942871094,
236
+ 0.020787200927734374,
237
+ 0.02082815933227539,
238
+ 0.020731903076171874,
239
+ 0.021405696868896484,
240
+ 0.02087321662902832,
241
+ 0.020867071151733398,
242
+ 0.020822015762329102,
243
+ 0.02083020782470703,
244
+ 0.020847616195678712,
245
+ 0.020815872192382814,
246
+ 0.020776960372924806,
247
+ 0.02086809539794922,
248
+ 0.020805631637573242,
249
+ 0.02082713508605957,
250
+ 0.0208035831451416,
251
+ 0.02083839988708496,
252
+ 0.019967071533203123,
253
+ 0.020073375701904296,
254
+ 0.020033536911010744,
255
+ 0.020057088851928712,
256
+ 0.02005299186706543,
257
+ 0.02007244873046875,
258
+ 0.02003046417236328,
259
+ 0.02008780860900879,
260
+ 0.02011238479614258,
261
+ 0.020051967620849608,
262
+ 0.02002022361755371,
263
+ 0.02007244873046875,
264
+ 0.01999667167663574,
265
+ 0.020008960723876954,
266
+ 0.020008960723876954,
267
+ 0.02008883285522461,
268
+ 0.02002739143371582,
269
+ 0.02026393508911133,
270
+ 0.02004787254333496,
271
+ 0.02006118392944336,
272
+ 0.020031488418579102,
273
+ 0.02006425666809082,
274
+ 0.02003046417236328,
275
+ 0.020023296356201172,
276
+ 0.02005606460571289,
277
+ 0.02006118392944336,
278
+ 0.020057088851928712,
279
+ 0.020039680480957032,
280
+ 0.020802560806274413,
281
+ 0.02084966468811035,
282
+ 0.020934656143188478,
283
+ 0.02088243293762207,
284
+ 0.020793344497680662,
285
+ 0.020808704376220705,
286
+ 0.020775999069213867,
287
+ 0.02079532814025879,
288
+ 0.019956735610961913,
289
+ 0.019976192474365235,
290
+ 0.019971071243286134,
291
+ 0.019990528106689453,
292
+ 0.020015104293823242,
293
+ 0.02005606460571289,
294
+ 0.019976192474365235,
295
+ 0.020015104293823242,
296
+ 0.020009983062744142,
297
+ 0.02001817512512207,
298
+ 0.020000768661499024,
299
+ 0.020074495315551756,
300
+ 0.02000588798522949,
301
+ 0.02001919937133789,
302
+ 0.020008960723876954,
303
+ 0.020023296356201172,
304
+ 0.02108518409729004,
305
+ 0.020818944931030273,
306
+ 0.02084659194946289,
307
+ 0.020884479522705078,
308
+ 0.020853759765625,
309
+ 0.02088140869140625,
310
+ 0.020809728622436522,
311
+ 0.020785152435302736,
312
+ 0.020839424133300782,
313
+ 0.019971071243286134,
314
+ 0.020001792907714845,
315
+ 0.019970048904418947,
316
+ 0.019974143981933593,
317
+ 0.020122623443603514,
318
+ 0.019974143981933593,
319
+ 0.019968000411987305,
320
+ 0.019992576599121094,
321
+ 0.019986431121826173,
322
+ 0.020083711624145507,
323
+ 0.01998847961425781,
324
+ 0.020031488418579102,
325
+ 0.020033536911010744,
326
+ 0.02001919937133789,
327
+ 0.020016128540039063,
328
+ 0.019993600845336915,
329
+ 0.019982336044311523,
330
+ 0.019971071243286134,
331
+ 0.020173824310302735,
332
+ 0.02004172706604004,
333
+ 0.019989503860473632,
334
+ 0.019938304901123048,
335
+ 0.019914751052856446,
336
+ 0.02086297607421875,
337
+ 0.020913152694702147,
338
+ 0.02079641532897949,
339
+ 0.020764671325683593,
340
+ 0.020774911880493165,
341
+ 0.020753408432006838,
342
+ 0.02002124786376953,
343
+ 0.020015104293823242,
344
+ 0.02003455924987793,
345
+ 0.021910528182983398,
346
+ 0.020933631896972657,
347
+ 0.020754432678222655,
348
+ 0.020946943283081054,
349
+ 0.02082713508605957,
350
+ 0.02084249687194824,
351
+ 0.020809728622436522,
352
+ 0.020832256317138673,
353
+ 0.020805631637573242,
354
+ 0.02084864044189453,
355
+ 0.020799488067626954,
356
+ 0.020793344497680662,
357
+ 0.020916223526000977,
358
+ 0.02083020782470703,
359
+ 0.02088140869140625,
360
+ 0.020801536560058592,
361
+ 0.020781055450439453,
362
+ 0.02081279945373535,
363
+ 0.020788223266601562,
364
+ 0.02082918357849121,
365
+ 0.020755456924438476,
366
+ 0.02088140869140625,
367
+ 0.020816896438598635,
368
+ 0.020793344497680662,
369
+ 0.02005401611328125,
370
+ 0.020032512664794923,
371
+ 0.02006220817565918,
372
+ 0.01986764717102051,
373
+ 0.01983180809020996,
374
+ 0.02005401611328125,
375
+ 0.0200263671875,
376
+ 0.02006220817565918,
377
+ 0.02004582405090332,
378
+ 0.02009702491760254,
379
+ 0.02010419273376465,
380
+ 0.02008678436279297,
381
+ 0.020024320602416993,
382
+ 0.01998028755187988,
383
+ 0.019993600845336915,
384
+ 0.020405248641967775,
385
+ 0.020818944931030273,
386
+ 0.020747264862060546,
387
+ 0.020024320602416993,
388
+ 0.020025344848632814,
389
+ 0.019946495056152345,
390
+ 0.02003455924987793,
391
+ 0.020139007568359374,
392
+ 0.019982336044311523,
393
+ 0.020031488418579102,
394
+ 0.0200263671875,
395
+ 0.02008576011657715,
396
+ 0.019981311798095702,
397
+ 0.02000486373901367,
398
+ 0.020156415939331054,
399
+ 0.021012479782104493,
400
+ 0.021594112396240234,
401
+ 0.021129215240478515,
402
+ 0.020951040267944337,
403
+ 0.02092748832702637,
404
+ 0.02086809539794922,
405
+ 0.021117952346801756,
406
+ 0.020890623092651366,
407
+ 0.0209039363861084,
408
+ 0.020904960632324218,
409
+ 0.02085785675048828,
410
+ 0.020853759765625,
411
+ 0.02089574432373047,
412
+ 0.02090598487854004,
413
+ 0.024049663543701173,
414
+ 0.023151615142822265,
415
+ 0.020953088760375976,
416
+ 0.02085683250427246,
417
+ 0.020853759765625,
418
+ 0.020755456924438476,
419
+ 0.020788223266601562,
420
+ 0.02082815933227539,
421
+ 0.020797439575195312,
422
+ 0.02083635139465332,
423
+ 0.020805631637573242,
424
+ 0.020853792190551758,
425
+ 0.02006012725830078,
426
+ 0.0200898551940918,
427
+ 0.020158464431762696,
428
+ 0.020883455276489257,
429
+ 0.02085683250427246,
430
+ 0.020951040267944337,
431
+ 0.020154367446899413,
432
+ 0.02007756805419922,
433
+ 0.02127667236328125,
434
+ 0.02089369583129883,
435
+ 0.020822015762329102,
436
+ 0.02084659194946289,
437
+ 0.02081996726989746,
438
+ 0.020800512313842775,
439
+ 0.020774911880493165,
440
+ 0.020805631637573242,
441
+ 0.020788223266601562,
442
+ 0.020914176940917968,
443
+ 0.020793344497680662,
444
+ 0.020788223266601562,
445
+ 0.020793344497680662,
446
+ 0.020867071151733398,
447
+ 0.020814847946166993,
448
+ 0.020807680130004884,
449
+ 0.02084249687194824,
450
+ 0.02084864044189453,
451
+ 0.02082508850097656,
452
+ 0.020930559158325195,
453
+ 0.0208035831451416,
454
+ 0.020909055709838868,
455
+ 0.020953088760375976,
456
+ 0.02081996726989746,
457
+ 0.02085990333557129,
458
+ 0.020920320510864256,
459
+ 0.02081996726989746,
460
+ 0.020858879089355468,
461
+ 0.020979711532592774,
462
+ 0.02080460739135742
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 48.82626026619632
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1319.092224,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.3471416015625,
15
+ "mean": 12.3471416015625,
16
+ "stdev": 0.0,
17
+ "p50": 12.3471416015625,
18
+ "p90": 12.3471416015625,
19
+ "p95": 12.3471416015625,
20
+ "p99": 12.3471416015625,
21
+ "values": [
22
+ 12.3471416015625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1777.262592,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.043444223403930664,
42
+ "mean": 0.021722111701965332,
43
+ "stdev": 3.3922195434576974e-06,
44
+ "p50": 0.021722111701965332,
45
+ "p90": 0.021724825477600098,
46
+ "p95": 0.021725164699554444,
47
+ "p99": 0.02172543607711792,
48
+ "values": [
49
+ 0.021718719482421874,
50
+ 0.02172550392150879
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 322.25227897003526
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1801.777152,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.2023725585937495,
73
+ "mean": 2.6011862792968747,
74
+ "stdev": 0.02596093750000006,
75
+ "p50": 2.6011862792968747,
76
+ "p90": 2.621955029296875,
77
+ "p95": 2.624551123046875,
78
+ "p99": 2.626627998046875,
79
+ "values": [
80
+ 2.575225341796875,
81
+ 2.627147216796875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 48.82387740194036
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.202118667602542,
97
+ "mean": 0.02048078215591551,
98
+ "stdev": 0.0005142246344054906,
99
+ "p50": 0.02075033664703369,
100
+ "p90": 0.02090813446044922,
101
+ "p95": 0.020953088760375976,
102
+ "p99": 0.02174282781600952,
103
+ "values": [
104
+ 0.020039680480957032,
105
+ 0.020017152786254884,
106
+ 0.020057088851928712,
107
+ 0.02002022361755371,
108
+ 0.020024320602416993,
109
+ 0.019985408782958985,
110
+ 0.02002841567993164,
111
+ 0.019984384536743165,
112
+ 0.020306943893432617,
113
+ 0.020801536560058592,
114
+ 0.02079539108276367,
115
+ 0.02008576011657715,
116
+ 0.02008064079284668,
117
+ 0.01998847961425781,
118
+ 0.02000588798522949,
119
+ 0.019998720169067383,
120
+ 0.019979263305664064,
121
+ 0.020007936477661133,
122
+ 0.02004172706604004,
123
+ 0.020006912231445313,
124
+ 0.020116479873657226,
125
+ 0.02003558349609375,
126
+ 0.020083711624145507,
127
+ 0.02003046417236328,
128
+ 0.020983808517456053,
129
+ 0.02087936019897461,
130
+ 0.020783103942871094,
131
+ 0.020787200927734374,
132
+ 0.02082815933227539,
133
+ 0.020731903076171874,
134
+ 0.021405696868896484,
135
+ 0.02087321662902832,
136
+ 0.020867071151733398,
137
+ 0.020822015762329102,
138
+ 0.02083020782470703,
139
+ 0.020847616195678712,
140
+ 0.020815872192382814,
141
+ 0.020776960372924806,
142
+ 0.02086809539794922,
143
+ 0.020805631637573242,
144
+ 0.02082713508605957,
145
+ 0.0208035831451416,
146
+ 0.02083839988708496,
147
+ 0.019967071533203123,
148
+ 0.020073375701904296,
149
+ 0.020033536911010744,
150
+ 0.020057088851928712,
151
+ 0.02005299186706543,
152
+ 0.02007244873046875,
153
+ 0.02003046417236328,
154
+ 0.02008780860900879,
155
+ 0.02011238479614258,
156
+ 0.020051967620849608,
157
+ 0.02002022361755371,
158
+ 0.02007244873046875,
159
+ 0.01999667167663574,
160
+ 0.020008960723876954,
161
+ 0.020008960723876954,
162
+ 0.02008883285522461,
163
+ 0.02002739143371582,
164
+ 0.02026393508911133,
165
+ 0.02004787254333496,
166
+ 0.02006118392944336,
167
+ 0.020031488418579102,
168
+ 0.02006425666809082,
169
+ 0.02003046417236328,
170
+ 0.020023296356201172,
171
+ 0.02005606460571289,
172
+ 0.02006118392944336,
173
+ 0.020057088851928712,
174
+ 0.020039680480957032,
175
+ 0.020802560806274413,
176
+ 0.02084966468811035,
177
+ 0.020934656143188478,
178
+ 0.02088243293762207,
179
+ 0.020793344497680662,
180
+ 0.020808704376220705,
181
+ 0.020775999069213867,
182
+ 0.02079532814025879,
183
+ 0.019956735610961913,
184
+ 0.019976192474365235,
185
+ 0.019971071243286134,
186
+ 0.019990528106689453,
187
+ 0.020015104293823242,
188
+ 0.02005606460571289,
189
+ 0.019976192474365235,
190
+ 0.020015104293823242,
191
+ 0.020009983062744142,
192
+ 0.02001817512512207,
193
+ 0.020000768661499024,
194
+ 0.020074495315551756,
195
+ 0.02000588798522949,
196
+ 0.02001919937133789,
197
+ 0.020008960723876954,
198
+ 0.020023296356201172,
199
+ 0.02108518409729004,
200
+ 0.020818944931030273,
201
+ 0.02084659194946289,
202
+ 0.020884479522705078,
203
+ 0.020853759765625,
204
+ 0.02088140869140625,
205
+ 0.020809728622436522,
206
+ 0.020785152435302736,
207
+ 0.020839424133300782,
208
+ 0.019971071243286134,
209
+ 0.020001792907714845,
210
+ 0.019970048904418947,
211
+ 0.019974143981933593,
212
+ 0.020122623443603514,
213
+ 0.019974143981933593,
214
+ 0.019968000411987305,
215
+ 0.019992576599121094,
216
+ 0.019986431121826173,
217
+ 0.020083711624145507,
218
+ 0.01998847961425781,
219
+ 0.020031488418579102,
220
+ 0.020033536911010744,
221
+ 0.02001919937133789,
222
+ 0.020016128540039063,
223
+ 0.019993600845336915,
224
+ 0.019982336044311523,
225
+ 0.019971071243286134,
226
+ 0.020173824310302735,
227
+ 0.02004172706604004,
228
+ 0.019989503860473632,
229
+ 0.019938304901123048,
230
+ 0.019914751052856446,
231
+ 0.02086297607421875,
232
+ 0.020913152694702147,
233
+ 0.02079641532897949,
234
+ 0.020764671325683593,
235
+ 0.020774911880493165,
236
+ 0.020753408432006838,
237
+ 0.02002124786376953,
238
+ 0.020015104293823242,
239
+ 0.02003455924987793,
240
+ 0.021910528182983398,
241
+ 0.020933631896972657,
242
+ 0.020754432678222655,
243
+ 0.020946943283081054,
244
+ 0.02082713508605957,
245
+ 0.02084249687194824,
246
+ 0.020809728622436522,
247
+ 0.020832256317138673,
248
+ 0.020805631637573242,
249
+ 0.02084864044189453,
250
+ 0.020799488067626954,
251
+ 0.020793344497680662,
252
+ 0.020916223526000977,
253
+ 0.02083020782470703,
254
+ 0.02088140869140625,
255
+ 0.020801536560058592,
256
+ 0.020781055450439453,
257
+ 0.02081279945373535,
258
+ 0.020788223266601562,
259
+ 0.02082918357849121,
260
+ 0.020755456924438476,
261
+ 0.02088140869140625,
262
+ 0.020816896438598635,
263
+ 0.020793344497680662,
264
+ 0.02005401611328125,
265
+ 0.020032512664794923,
266
+ 0.02006220817565918,
267
+ 0.01986764717102051,
268
+ 0.01983180809020996,
269
+ 0.02005401611328125,
270
+ 0.0200263671875,
271
+ 0.02006220817565918,
272
+ 0.02004582405090332,
273
+ 0.02009702491760254,
274
+ 0.02010419273376465,
275
+ 0.02008678436279297,
276
+ 0.020024320602416993,
277
+ 0.01998028755187988,
278
+ 0.019993600845336915,
279
+ 0.020405248641967775,
280
+ 0.020818944931030273,
281
+ 0.020747264862060546,
282
+ 0.020024320602416993,
283
+ 0.020025344848632814,
284
+ 0.019946495056152345,
285
+ 0.02003455924987793,
286
+ 0.020139007568359374,
287
+ 0.019982336044311523,
288
+ 0.020031488418579102,
289
+ 0.0200263671875,
290
+ 0.02008576011657715,
291
+ 0.019981311798095702,
292
+ 0.02000486373901367,
293
+ 0.020156415939331054,
294
+ 0.021012479782104493,
295
+ 0.021594112396240234,
296
+ 0.021129215240478515,
297
+ 0.020951040267944337,
298
+ 0.02092748832702637,
299
+ 0.02086809539794922,
300
+ 0.021117952346801756,
301
+ 0.020890623092651366,
302
+ 0.0209039363861084,
303
+ 0.020904960632324218,
304
+ 0.02085785675048828,
305
+ 0.020853759765625,
306
+ 0.02089574432373047,
307
+ 0.02090598487854004,
308
+ 0.024049663543701173,
309
+ 0.023151615142822265,
310
+ 0.020953088760375976,
311
+ 0.02085683250427246,
312
+ 0.020853759765625,
313
+ 0.020755456924438476,
314
+ 0.020788223266601562,
315
+ 0.02082815933227539,
316
+ 0.020797439575195312,
317
+ 0.02083635139465332,
318
+ 0.020805631637573242,
319
+ 0.020853792190551758,
320
+ 0.02006012725830078,
321
+ 0.0200898551940918,
322
+ 0.020158464431762696,
323
+ 0.020883455276489257,
324
+ 0.02085683250427246,
325
+ 0.020951040267944337,
326
+ 0.020154367446899413,
327
+ 0.02007756805419922,
328
+ 0.02127667236328125,
329
+ 0.02089369583129883,
330
+ 0.020822015762329102,
331
+ 0.02084659194946289,
332
+ 0.02081996726989746,
333
+ 0.020800512313842775,
334
+ 0.020774911880493165,
335
+ 0.020805631637573242,
336
+ 0.020788223266601562,
337
+ 0.020914176940917968,
338
+ 0.020793344497680662,
339
+ 0.020788223266601562,
340
+ 0.020793344497680662,
341
+ 0.020867071151733398,
342
+ 0.020814847946166993,
343
+ 0.020807680130004884,
344
+ 0.02084249687194824,
345
+ 0.02084864044189453,
346
+ 0.02082508850097656,
347
+ 0.020930559158325195,
348
+ 0.0208035831451416,
349
+ 0.020909055709838868,
350
+ 0.020953088760375976,
351
+ 0.02081996726989746,
352
+ 0.02085990333557129,
353
+ 0.020920320510864256,
354
+ 0.02081996726989746,
355
+ 0.020858879089355468,
356
+ 0.020979711532592774,
357
+ 0.02080460739135742
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 48.82626026619632
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.021722111701965332,
7
+ "prefill.throughput.value": 322.25227897003526,
8
+ "decode.latency.mean": 2.6011862792968747,
9
+ "decode.throughput.value": 48.82387740194036,
10
+ "per_token.latency.mean": 0.02048078215591551,
11
+ "per_token.throughput.value": 48.82626026619632
12
+ }
13
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 7ed9789e210d8eca797fc21b9c783b1ce718ecb5
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
88
+ - hydra.run.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1338.1632,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.770521484375,
120
+ "mean": 12.770521484375,
121
+ "stdev": 0.0,
122
+ "p50": 12.770521484375,
123
+ "p90": 12.770521484375,
124
+ "p95": 12.770521484375,
125
+ "p99": 12.770521484375,
126
+ "values": [
127
+ 12.770521484375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1853.964288,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.028351807594299318,
147
+ "mean": 0.014175903797149659,
148
+ "stdev": 0.0002044157981872562,
149
+ "p50": 0.014175903797149659,
150
+ "p90": 0.014339436435699463,
151
+ "p95": 0.014359878015518189,
152
+ "p99": 0.014376231279373168,
153
+ "values": [
154
+ 0.013971487998962402,
155
+ 0.014380319595336914
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 493.7956761111405
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2024.12032,
169
+ "max_global_vram": 6071.779328,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.131478881835937,
178
+ "mean": 1.5657394409179686,
179
+ "stdev": 0.00030450439453122424,
180
+ "p50": 1.5657394409179686,
181
+ "p90": 1.5659830444335936,
182
+ "p95": 1.5660134948730469,
183
+ "p99": 1.5660378552246093,
184
+ "values": [
185
+ 1.5660439453125,
186
+ 1.5654349365234375
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.11183552069295
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.131193336486816,
202
+ "mean": 0.012327532820814238,
203
+ "stdev": 8.273771738078037e-05,
204
+ "p50": 0.012310527801513671,
205
+ "p90": 0.012340224266052247,
206
+ "p95": 0.01236689887046814,
207
+ "p99": 0.012884049797058105,
208
+ "values": [
209
+ 0.012322815895080566,
210
+ 0.012294143676757812,
211
+ 0.012314623832702636,
212
+ 0.012306431770324706,
213
+ 0.012297216415405274,
214
+ 0.012312576293945313,
215
+ 0.012320768356323243,
216
+ 0.012317695617675782,
217
+ 0.012315648078918457,
218
+ 0.012300288200378418,
219
+ 0.012302335739135742,
220
+ 0.012318719863891601,
221
+ 0.012321791648864745,
222
+ 0.012326911926269531,
223
+ 0.012309503555297852,
224
+ 0.012305407524108887,
225
+ 0.012308480262756348,
226
+ 0.012303359985351562,
227
+ 0.012304384231567383,
228
+ 0.012306431770324706,
229
+ 0.012300288200378418,
230
+ 0.012305407524108887,
231
+ 0.01285324764251709,
232
+ 0.012354559898376465,
233
+ 0.012323840141296387,
234
+ 0.01229312038421631,
235
+ 0.012308480262756348,
236
+ 0.012310527801513671,
237
+ 0.012309503555297852,
238
+ 0.012368895530700684,
239
+ 0.012354559898376465,
240
+ 0.012309568405151367,
241
+ 0.012327872276306152,
242
+ 0.0123504638671875,
243
+ 0.012331007957458496,
244
+ 0.01232588768005371,
245
+ 0.012314623832702636,
246
+ 0.012352512359619141,
247
+ 0.012328960418701173,
248
+ 0.012318719863891601,
249
+ 0.012926976203918457,
250
+ 0.012347392082214356,
251
+ 0.012314623832702636,
252
+ 0.012305407524108887,
253
+ 0.012307456016540527,
254
+ 0.012327936172485352,
255
+ 0.012332032203674317,
256
+ 0.012312576293945313,
257
+ 0.012301312446594239,
258
+ 0.012305407524108887,
259
+ 0.012304384231567383,
260
+ 0.012299263954162597,
261
+ 0.012314623832702636,
262
+ 0.012306431770324706,
263
+ 0.012300288200378418,
264
+ 0.012305407524108887,
265
+ 0.012302335739135742,
266
+ 0.012300288200378418,
267
+ 0.012324864387512208,
268
+ 0.012299263954162597,
269
+ 0.012308480262756348,
270
+ 0.012301312446594239,
271
+ 0.012305407524108887,
272
+ 0.012299263954162597,
273
+ 0.012304384231567383,
274
+ 0.012301312446594239,
275
+ 0.012306431770324706,
276
+ 0.012296192169189453,
277
+ 0.01233407974243164,
278
+ 0.012299263954162597,
279
+ 0.012306431770324706,
280
+ 0.012317695617675782,
281
+ 0.012316672325134278,
282
+ 0.012315648078918457,
283
+ 0.012315648078918457,
284
+ 0.012306431770324706,
285
+ 0.012295167922973632,
286
+ 0.012308480262756348,
287
+ 0.012312576293945313,
288
+ 0.012310527801513671,
289
+ 0.012304384231567383,
290
+ 0.012302335739135742,
291
+ 0.012305407524108887,
292
+ 0.012301312446594239,
293
+ 0.01232588768005371,
294
+ 0.012319744110107422,
295
+ 0.012309503555297852,
296
+ 0.012310527801513671,
297
+ 0.012305407524108887,
298
+ 0.01229312038421631,
299
+ 0.012305407524108887,
300
+ 0.012316672325134278,
301
+ 0.012310527801513671,
302
+ 0.012308480262756348,
303
+ 0.012301312446594239,
304
+ 0.012302335739135742,
305
+ 0.012298239707946777,
306
+ 0.012326911926269531,
307
+ 0.012314623832702636,
308
+ 0.012296192169189453,
309
+ 0.012303359985351562,
310
+ 0.012322815895080566,
311
+ 0.012302335739135742,
312
+ 0.012299263954162597,
313
+ 0.012310527801513671,
314
+ 0.012309503555297852,
315
+ 0.012360704421997071,
316
+ 0.012323840141296387,
317
+ 0.012307456016540527,
318
+ 0.012305407524108887,
319
+ 0.012416000366210938,
320
+ 0.012312576293945313,
321
+ 0.012305407524108887,
322
+ 0.012294143676757812,
323
+ 0.012314623832702636,
324
+ 0.012313599586486817,
325
+ 0.012314623832702636,
326
+ 0.012314623832702636,
327
+ 0.012311552047729492,
328
+ 0.012306431770324706,
329
+ 0.012305407524108887,
330
+ 0.012304384231567383,
331
+ 0.0123504638671875,
332
+ 0.01233510398864746,
333
+ 0.012918784141540527,
334
+ 0.012604415893554688,
335
+ 0.012327936172485352,
336
+ 0.012365823745727538,
337
+ 0.012323840141296387,
338
+ 0.012313599586486817,
339
+ 0.012303359985351562,
340
+ 0.012311552047729492,
341
+ 0.012294143676757812,
342
+ 0.012297216415405274,
343
+ 0.012309503555297852,
344
+ 0.012304384231567383,
345
+ 0.012323840141296387,
346
+ 0.012340224266052247,
347
+ 0.012314623832702636,
348
+ 0.012315648078918457,
349
+ 0.012304384231567383,
350
+ 0.012291071891784668,
351
+ 0.012302335739135742,
352
+ 0.012324864387512208,
353
+ 0.012312576293945313,
354
+ 0.012304384231567383,
355
+ 0.012296192169189453,
356
+ 0.012304384231567383,
357
+ 0.012304384231567383,
358
+ 0.012317695617675782,
359
+ 0.012326911926269531,
360
+ 0.012314623832702636,
361
+ 0.012308480262756348,
362
+ 0.012310527801513671,
363
+ 0.012307456016540527,
364
+ 0.012302335739135742,
365
+ 0.012308480262756348,
366
+ 0.012318719863891601,
367
+ 0.012303359985351562,
368
+ 0.012314623832702636,
369
+ 0.012302335739135742,
370
+ 0.012302335739135742,
371
+ 0.012321791648864745,
372
+ 0.012336128234863282,
373
+ 0.012311552047729492,
374
+ 0.012309503555297852,
375
+ 0.012295167922973632,
376
+ 0.012296192169189453,
377
+ 0.012308480262756348,
378
+ 0.012311552047729492,
379
+ 0.012315648078918457,
380
+ 0.012290047645568849,
381
+ 0.012317695617675782,
382
+ 0.012306431770324706,
383
+ 0.012310527801513671,
384
+ 0.012320768356323243,
385
+ 0.012315648078918457,
386
+ 0.012312576293945313,
387
+ 0.012313599586486817,
388
+ 0.012299263954162597,
389
+ 0.012310527801513671,
390
+ 0.012312576293945313,
391
+ 0.012308480262756348,
392
+ 0.01244979190826416,
393
+ 0.012340224266052247,
394
+ 0.012321791648864745,
395
+ 0.012312576293945313,
396
+ 0.012323840141296387,
397
+ 0.012319744110107422,
398
+ 0.012305407524108887,
399
+ 0.012430335998535156,
400
+ 0.012322815895080566,
401
+ 0.012311552047729492,
402
+ 0.012303359985351562,
403
+ 0.012305407524108887,
404
+ 0.012303359985351562,
405
+ 0.012299263954162597,
406
+ 0.012316672325134278,
407
+ 0.012389375686645507,
408
+ 0.012313599586486817,
409
+ 0.012328960418701173,
410
+ 0.012305407524108887,
411
+ 0.012317695617675782,
412
+ 0.012299263954162597,
413
+ 0.012307456016540527,
414
+ 0.012327936172485352,
415
+ 0.012313599586486817,
416
+ 0.01294438362121582,
417
+ 0.012360704421997071,
418
+ 0.012319744110107422,
419
+ 0.012340224266052247,
420
+ 0.012328960418701173,
421
+ 0.012296192169189453,
422
+ 0.012307456016540527,
423
+ 0.012298239707946777,
424
+ 0.012304384231567383,
425
+ 0.012319744110107422,
426
+ 0.012416000366210938,
427
+ 0.01263923168182373,
428
+ 0.012554240226745606,
429
+ 0.012354559898376465,
430
+ 0.012364800453186036,
431
+ 0.012304384231567383,
432
+ 0.012317695617675782,
433
+ 0.012313599586486817,
434
+ 0.012313599586486817,
435
+ 0.012314623832702636,
436
+ 0.012313599586486817,
437
+ 0.012300288200378418,
438
+ 0.012304384231567383,
439
+ 0.012310591697692871,
440
+ 0.012303296089172364,
441
+ 0.012291071891784668,
442
+ 0.012315648078918457,
443
+ 0.012322815895080566,
444
+ 0.012304384231567383,
445
+ 0.012302335739135742,
446
+ 0.012318719863891601,
447
+ 0.012301312446594239,
448
+ 0.012298239707946777,
449
+ 0.012304384231567383,
450
+ 0.012312576293945313,
451
+ 0.012306431770324706,
452
+ 0.012306431770324706,
453
+ 0.012297216415405274,
454
+ 0.012297216415405274,
455
+ 0.012331007957458496,
456
+ 0.012317695617675782,
457
+ 0.012303359985351562,
458
+ 0.012306431770324706,
459
+ 0.012307456016540527,
460
+ 0.012309503555297852,
461
+ 0.012295167922973632,
462
+ 0.012300288200378418
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.11923241539175
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1338.1632,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.770521484375,
15
+ "mean": 12.770521484375,
16
+ "stdev": 0.0,
17
+ "p50": 12.770521484375,
18
+ "p90": 12.770521484375,
19
+ "p95": 12.770521484375,
20
+ "p99": 12.770521484375,
21
+ "values": [
22
+ 12.770521484375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1853.964288,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.028351807594299318,
42
+ "mean": 0.014175903797149659,
43
+ "stdev": 0.0002044157981872562,
44
+ "p50": 0.014175903797149659,
45
+ "p90": 0.014339436435699463,
46
+ "p95": 0.014359878015518189,
47
+ "p99": 0.014376231279373168,
48
+ "values": [
49
+ 0.013971487998962402,
50
+ 0.014380319595336914
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 493.7956761111405
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2024.12032,
64
+ "max_global_vram": 6071.779328,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.131478881835937,
73
+ "mean": 1.5657394409179686,
74
+ "stdev": 0.00030450439453122424,
75
+ "p50": 1.5657394409179686,
76
+ "p90": 1.5659830444335936,
77
+ "p95": 1.5660134948730469,
78
+ "p99": 1.5660378552246093,
79
+ "values": [
80
+ 1.5660439453125,
81
+ 1.5654349365234375
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.11183552069295
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.131193336486816,
97
+ "mean": 0.012327532820814238,
98
+ "stdev": 8.273771738078037e-05,
99
+ "p50": 0.012310527801513671,
100
+ "p90": 0.012340224266052247,
101
+ "p95": 0.01236689887046814,
102
+ "p99": 0.012884049797058105,
103
+ "values": [
104
+ 0.012322815895080566,
105
+ 0.012294143676757812,
106
+ 0.012314623832702636,
107
+ 0.012306431770324706,
108
+ 0.012297216415405274,
109
+ 0.012312576293945313,
110
+ 0.012320768356323243,
111
+ 0.012317695617675782,
112
+ 0.012315648078918457,
113
+ 0.012300288200378418,
114
+ 0.012302335739135742,
115
+ 0.012318719863891601,
116
+ 0.012321791648864745,
117
+ 0.012326911926269531,
118
+ 0.012309503555297852,
119
+ 0.012305407524108887,
120
+ 0.012308480262756348,
121
+ 0.012303359985351562,
122
+ 0.012304384231567383,
123
+ 0.012306431770324706,
124
+ 0.012300288200378418,
125
+ 0.012305407524108887,
126
+ 0.01285324764251709,
127
+ 0.012354559898376465,
128
+ 0.012323840141296387,
129
+ 0.01229312038421631,
130
+ 0.012308480262756348,
131
+ 0.012310527801513671,
132
+ 0.012309503555297852,
133
+ 0.012368895530700684,
134
+ 0.012354559898376465,
135
+ 0.012309568405151367,
136
+ 0.012327872276306152,
137
+ 0.0123504638671875,
138
+ 0.012331007957458496,
139
+ 0.01232588768005371,
140
+ 0.012314623832702636,
141
+ 0.012352512359619141,
142
+ 0.012328960418701173,
143
+ 0.012318719863891601,
144
+ 0.012926976203918457,
145
+ 0.012347392082214356,
146
+ 0.012314623832702636,
147
+ 0.012305407524108887,
148
+ 0.012307456016540527,
149
+ 0.012327936172485352,
150
+ 0.012332032203674317,
151
+ 0.012312576293945313,
152
+ 0.012301312446594239,
153
+ 0.012305407524108887,
154
+ 0.012304384231567383,
155
+ 0.012299263954162597,
156
+ 0.012314623832702636,
157
+ 0.012306431770324706,
158
+ 0.012300288200378418,
159
+ 0.012305407524108887,
160
+ 0.012302335739135742,
161
+ 0.012300288200378418,
162
+ 0.012324864387512208,
163
+ 0.012299263954162597,
164
+ 0.012308480262756348,
165
+ 0.012301312446594239,
166
+ 0.012305407524108887,
167
+ 0.012299263954162597,
168
+ 0.012304384231567383,
169
+ 0.012301312446594239,
170
+ 0.012306431770324706,
171
+ 0.012296192169189453,
172
+ 0.01233407974243164,
173
+ 0.012299263954162597,
174
+ 0.012306431770324706,
175
+ 0.012317695617675782,
176
+ 0.012316672325134278,
177
+ 0.012315648078918457,
178
+ 0.012315648078918457,
179
+ 0.012306431770324706,
180
+ 0.012295167922973632,
181
+ 0.012308480262756348,
182
+ 0.012312576293945313,
183
+ 0.012310527801513671,
184
+ 0.012304384231567383,
185
+ 0.012302335739135742,
186
+ 0.012305407524108887,
187
+ 0.012301312446594239,
188
+ 0.01232588768005371,
189
+ 0.012319744110107422,
190
+ 0.012309503555297852,
191
+ 0.012310527801513671,
192
+ 0.012305407524108887,
193
+ 0.01229312038421631,
194
+ 0.012305407524108887,
195
+ 0.012316672325134278,
196
+ 0.012310527801513671,
197
+ 0.012308480262756348,
198
+ 0.012301312446594239,
199
+ 0.012302335739135742,
200
+ 0.012298239707946777,
201
+ 0.012326911926269531,
202
+ 0.012314623832702636,
203
+ 0.012296192169189453,
204
+ 0.012303359985351562,
205
+ 0.012322815895080566,
206
+ 0.012302335739135742,
207
+ 0.012299263954162597,
208
+ 0.012310527801513671,
209
+ 0.012309503555297852,
210
+ 0.012360704421997071,
211
+ 0.012323840141296387,
212
+ 0.012307456016540527,
213
+ 0.012305407524108887,
214
+ 0.012416000366210938,
215
+ 0.012312576293945313,
216
+ 0.012305407524108887,
217
+ 0.012294143676757812,
218
+ 0.012314623832702636,
219
+ 0.012313599586486817,
220
+ 0.012314623832702636,
221
+ 0.012314623832702636,
222
+ 0.012311552047729492,
223
+ 0.012306431770324706,
224
+ 0.012305407524108887,
225
+ 0.012304384231567383,
226
+ 0.0123504638671875,
227
+ 0.01233510398864746,
228
+ 0.012918784141540527,
229
+ 0.012604415893554688,
230
+ 0.012327936172485352,
231
+ 0.012365823745727538,
232
+ 0.012323840141296387,
233
+ 0.012313599586486817,
234
+ 0.012303359985351562,
235
+ 0.012311552047729492,
236
+ 0.012294143676757812,
237
+ 0.012297216415405274,
238
+ 0.012309503555297852,
239
+ 0.012304384231567383,
240
+ 0.012323840141296387,
241
+ 0.012340224266052247,
242
+ 0.012314623832702636,
243
+ 0.012315648078918457,
244
+ 0.012304384231567383,
245
+ 0.012291071891784668,
246
+ 0.012302335739135742,
247
+ 0.012324864387512208,
248
+ 0.012312576293945313,
249
+ 0.012304384231567383,
250
+ 0.012296192169189453,
251
+ 0.012304384231567383,
252
+ 0.012304384231567383,
253
+ 0.012317695617675782,
254
+ 0.012326911926269531,
255
+ 0.012314623832702636,
256
+ 0.012308480262756348,
257
+ 0.012310527801513671,
258
+ 0.012307456016540527,
259
+ 0.012302335739135742,
260
+ 0.012308480262756348,
261
+ 0.012318719863891601,
262
+ 0.012303359985351562,
263
+ 0.012314623832702636,
264
+ 0.012302335739135742,
265
+ 0.012302335739135742,
266
+ 0.012321791648864745,
267
+ 0.012336128234863282,
268
+ 0.012311552047729492,
269
+ 0.012309503555297852,
270
+ 0.012295167922973632,
271
+ 0.012296192169189453,
272
+ 0.012308480262756348,
273
+ 0.012311552047729492,
274
+ 0.012315648078918457,
275
+ 0.012290047645568849,
276
+ 0.012317695617675782,
277
+ 0.012306431770324706,
278
+ 0.012310527801513671,
279
+ 0.012320768356323243,
280
+ 0.012315648078918457,
281
+ 0.012312576293945313,
282
+ 0.012313599586486817,
283
+ 0.012299263954162597,
284
+ 0.012310527801513671,
285
+ 0.012312576293945313,
286
+ 0.012308480262756348,
287
+ 0.01244979190826416,
288
+ 0.012340224266052247,
289
+ 0.012321791648864745,
290
+ 0.012312576293945313,
291
+ 0.012323840141296387,
292
+ 0.012319744110107422,
293
+ 0.012305407524108887,
294
+ 0.012430335998535156,
295
+ 0.012322815895080566,
296
+ 0.012311552047729492,
297
+ 0.012303359985351562,
298
+ 0.012305407524108887,
299
+ 0.012303359985351562,
300
+ 0.012299263954162597,
301
+ 0.012316672325134278,
302
+ 0.012389375686645507,
303
+ 0.012313599586486817,
304
+ 0.012328960418701173,
305
+ 0.012305407524108887,
306
+ 0.012317695617675782,
307
+ 0.012299263954162597,
308
+ 0.012307456016540527,
309
+ 0.012327936172485352,
310
+ 0.012313599586486817,
311
+ 0.01294438362121582,
312
+ 0.012360704421997071,
313
+ 0.012319744110107422,
314
+ 0.012340224266052247,
315
+ 0.012328960418701173,
316
+ 0.012296192169189453,
317
+ 0.012307456016540527,
318
+ 0.012298239707946777,
319
+ 0.012304384231567383,
320
+ 0.012319744110107422,
321
+ 0.012416000366210938,
322
+ 0.01263923168182373,
323
+ 0.012554240226745606,
324
+ 0.012354559898376465,
325
+ 0.012364800453186036,
326
+ 0.012304384231567383,
327
+ 0.012317695617675782,
328
+ 0.012313599586486817,
329
+ 0.012313599586486817,
330
+ 0.012314623832702636,
331
+ 0.012313599586486817,
332
+ 0.012300288200378418,
333
+ 0.012304384231567383,
334
+ 0.012310591697692871,
335
+ 0.012303296089172364,
336
+ 0.012291071891784668,
337
+ 0.012315648078918457,
338
+ 0.012322815895080566,
339
+ 0.012304384231567383,
340
+ 0.012302335739135742,
341
+ 0.012318719863891601,
342
+ 0.012301312446594239,
343
+ 0.012298239707946777,
344
+ 0.012304384231567383,
345
+ 0.012312576293945313,
346
+ 0.012306431770324706,
347
+ 0.012306431770324706,
348
+ 0.012297216415405274,
349
+ 0.012297216415405274,
350
+ 0.012331007957458496,
351
+ 0.012317695617675782,
352
+ 0.012303359985351562,
353
+ 0.012306431770324706,
354
+ 0.012307456016540527,
355
+ 0.012309503555297852,
356
+ 0.012295167922973632,
357
+ 0.012300288200378418
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.11923241539175
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.014175903797149659,
7
+ "prefill.throughput.value": 493.7956761111405,
8
+ "decode.latency.mean": 1.5657394409179686,
9
+ "decode.throughput.value": 81.11183552069295,
10
+ "per_token.latency.mean": 0.012327532820814238,
11
+ "per_token.throughput.value": 81.11923241539175
12
+ }
13
+ }
2024-09-03/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpgfijev8v/commit=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5
88
+ - hydra.run.dir=_benchmark/tmpgfijev8v/commit\=7ed9789e210d8eca797fc21b9c783b1ce718ecb5/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: 7ed9789e210d8eca797fc21b9c783b1ce718ecb5
225
+ accelerate_version: 0.35.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.22.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-09-03/summaries.json CHANGED
@@ -1,41 +1,41 @@
1
  [
2
  {
3
  "model": "google/gemma-2b",
4
- "commit": "566302686a71de14125717dea9a6a45b24d42b37",
5
  "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
  "metrics": {
7
- "prefill.latency.mean": 0.02062836742401123,
8
- "prefill.throughput.value": 339.338535915938,
9
- "decode.latency.mean": 2.445818481445312,
10
- "decode.throughput.value": 51.92535789694076,
11
- "per_token.latency.mean": 0.019257146474883316,
12
- "per_token.throughput.value": 51.9287736271975
13
  }
14
  },
15
  {
16
  "model": "google/gemma-2b",
17
- "commit": "566302686a71de14125717dea9a6a45b24d42b37",
18
  "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
  "metrics": {
20
- "prefill.latency.mean": 0.02275548839569092,
21
- "prefill.throughput.value": 307.6180953921231,
22
- "decode.latency.mean": 2.6347849121093754,
23
- "decode.throughput.value": 48.201277992868654,
24
- "per_token.latency.mean": 0.020745304685878003,
25
- "per_token.throughput.value": 48.20367862231169
26
  }
27
  },
28
  {
29
  "model": "google/gemma-2b",
30
- "commit": "566302686a71de14125717dea9a6a45b24d42b37",
31
  "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
  "metrics": {
33
- "prefill.latency.mean": 0.014240432262420655,
34
- "prefill.throughput.value": 491.55811221211536,
35
- "decode.latency.mean": 1.5674649658203124,
36
- "decode.throughput.value": 81.02254453485422,
37
- "per_token.latency.mean": 0.01234111495280829,
38
- "per_token.throughput.value": 81.02995586897474
39
  }
40
  }
41
  ]
 
1
  [
2
  {
3
  "model": "google/gemma-2b",
4
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
5
  "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
  "metrics": {
7
+ "prefill.latency.mean": 0.020182064056396484,
8
+ "prefill.throughput.value": 346.8426212720015,
9
+ "decode.latency.mean": 2.4172987060546873,
10
+ "decode.throughput.value": 52.53798369307812,
11
+ "per_token.latency.mean": 0.01903261228246013,
12
+ "per_token.throughput.value": 52.54139500974174
13
  }
14
  },
15
  {
16
  "model": "google/gemma-2b",
17
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
18
  "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
  "metrics": {
20
+ "prefill.latency.mean": 0.021722111701965332,
21
+ "prefill.throughput.value": 322.25227897003526,
22
+ "decode.latency.mean": 2.6011862792968747,
23
+ "decode.throughput.value": 48.82387740194036,
24
+ "per_token.latency.mean": 0.02048078215591551,
25
+ "per_token.throughput.value": 48.82626026619632
26
  }
27
  },
28
  {
29
  "model": "google/gemma-2b",
30
+ "commit": "7ed9789e210d8eca797fc21b9c783b1ce718ecb5",
31
  "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
  "metrics": {
33
+ "prefill.latency.mean": 0.014175903797149659,
34
+ "prefill.throughput.value": 493.7956761111405,
35
+ "decode.latency.mean": 1.5657394409179686,
36
+ "decode.throughput.value": 81.11183552069295,
37
+ "per_token.latency.mean": 0.012327532820814238,
38
+ "per_token.throughput.value": 81.11923241539175
39
  }
40
  }
41
  ]
2024-09-03/summary.json CHANGED
@@ -1,38 +1,38 @@
1
  {
2
  "google/gemma-2b": {
3
  "backend.cache_implementation=null,backend.torch_compile=False": {
4
- "566302686a71de14125717dea9a6a45b24d42b37": {
5
  "metrics": {
6
- "prefill.latency.mean": 0.02062836742401123,
7
- "prefill.throughput.value": 339.338535915938,
8
- "decode.latency.mean": 2.445818481445312,
9
- "decode.throughput.value": 51.92535789694076,
10
- "per_token.latency.mean": 0.019257146474883316,
11
- "per_token.throughput.value": 51.9287736271975
12
  }
13
  }
14
  },
15
  "backend.cache_implementation=static,backend.torch_compile=False": {
16
- "566302686a71de14125717dea9a6a45b24d42b37": {
17
  "metrics": {
18
- "prefill.latency.mean": 0.02275548839569092,
19
- "prefill.throughput.value": 307.6180953921231,
20
- "decode.latency.mean": 2.6347849121093754,
21
- "decode.throughput.value": 48.201277992868654,
22
- "per_token.latency.mean": 0.020745304685878003,
23
- "per_token.throughput.value": 48.20367862231169
24
  }
25
  }
26
  },
27
  "backend.cache_implementation=static,backend.torch_compile=True": {
28
- "566302686a71de14125717dea9a6a45b24d42b37": {
29
  "metrics": {
30
- "prefill.latency.mean": 0.014240432262420655,
31
- "prefill.throughput.value": 491.55811221211536,
32
- "decode.latency.mean": 1.5674649658203124,
33
- "decode.throughput.value": 81.02254453485422,
34
- "per_token.latency.mean": 0.01234111495280829,
35
- "per_token.throughput.value": 81.02995586897474
36
  }
37
  }
38
  }
 
1
  {
2
  "google/gemma-2b": {
3
  "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "7ed9789e210d8eca797fc21b9c783b1ce718ecb5": {
5
  "metrics": {
6
+ "prefill.latency.mean": 0.020182064056396484,
7
+ "prefill.throughput.value": 346.8426212720015,
8
+ "decode.latency.mean": 2.4172987060546873,
9
+ "decode.throughput.value": 52.53798369307812,
10
+ "per_token.latency.mean": 0.01903261228246013,
11
+ "per_token.throughput.value": 52.54139500974174
12
  }
13
  }
14
  },
15
  "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "7ed9789e210d8eca797fc21b9c783b1ce718ecb5": {
17
  "metrics": {
18
+ "prefill.latency.mean": 0.021722111701965332,
19
+ "prefill.throughput.value": 322.25227897003526,
20
+ "decode.latency.mean": 2.6011862792968747,
21
+ "decode.throughput.value": 48.82387740194036,
22
+ "per_token.latency.mean": 0.02048078215591551,
23
+ "per_token.throughput.value": 48.82626026619632
24
  }
25
  }
26
  },
27
  "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "7ed9789e210d8eca797fc21b9c783b1ce718ecb5": {
29
  "metrics": {
30
+ "prefill.latency.mean": 0.014175903797149659,
31
+ "prefill.throughput.value": 493.7956761111405,
32
+ "decode.latency.mean": 1.5657394409179686,
33
+ "decode.throughput.value": 81.11183552069295,
34
+ "per_token.latency.mean": 0.012327532820814238,
35
+ "per_token.throughput.value": 81.11923241539175
36
  }
37
  }
38
  }