hf-transformers-bot commited on
Commit
62649ec
1 Parent(s): 4b893fb

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/multirun.yaml +234 -0
  31. 2024-08-27/summaries.json +41 -0
  32. 2024-08-27/summary.json +40 -0
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: c6b23fda65f9ae74f9a1026b340241f65aebe1a3
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
88
+ - hydra.run.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1319.563264,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.304673828125,
120
+ "mean": 12.304673828125,
121
+ "stdev": 0.0,
122
+ "p50": 12.304673828125,
123
+ "p90": 12.304673828125,
124
+ "p95": 12.304673828125,
125
+ "p99": 12.304673828125,
126
+ "values": [
127
+ 12.304673828125
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1715.408896,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.04151971244812012,
147
+ "mean": 0.02075985622406006,
148
+ "stdev": 0.0005112485885620105,
149
+ "p50": 0.02075985622406006,
150
+ "p90": 0.021168855094909667,
151
+ "p95": 0.02121997995376587,
152
+ "p99": 0.02126087984085083,
153
+ "values": [
154
+ 0.02024860763549805,
155
+ 0.02127110481262207
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 337.1892331261527
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1769.824256,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 4.99106298828125,
178
+ "mean": 2.495531494140625,
179
+ "stdev": 0.0010227050781250746,
180
+ "p50": 2.495531494140625,
181
+ "p90": 2.4963496582031253,
182
+ "p95": 2.4964519287109375,
183
+ "p99": 2.4965337451171874,
184
+ "values": [
185
+ 2.4945087890625,
186
+ 2.49655419921875
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 50.89096262587318
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 4.990741498947142,
202
+ "mean": 0.01964858857853206,
203
+ "stdev": 0.0002815677718437068,
204
+ "p50": 0.019663871765136717,
205
+ "p90": 0.01982832622528076,
206
+ "p95": 0.01989442481994629,
207
+ "p99": 0.02071226375579834,
208
+ "values": [
209
+ 0.018699264526367186,
210
+ 0.01927987289428711,
211
+ 0.019525632858276368,
212
+ 0.019517440795898438,
213
+ 0.019513376235961916,
214
+ 0.0196157112121582,
215
+ 0.019568639755249022,
216
+ 0.01901670455932617,
217
+ 0.01878835105895996,
218
+ 0.018720767974853517,
219
+ 0.018754560470581053,
220
+ 0.018782207489013672,
221
+ 0.018693119049072265,
222
+ 0.01965158462524414,
223
+ 0.019628032684326172,
224
+ 0.018737152099609376,
225
+ 0.019804159164428712,
226
+ 0.019615743637084963,
227
+ 0.018774015426635742,
228
+ 0.018734079360961914,
229
+ 0.019360767364501954,
230
+ 0.01962393569946289,
231
+ 0.019599359512329103,
232
+ 0.019524608612060547,
233
+ 0.01963212776184082,
234
+ 0.019604480743408204,
235
+ 0.019555328369140625,
236
+ 0.019581951141357423,
237
+ 0.019582975387573243,
238
+ 0.019545087814331053,
239
+ 0.019745792388916016,
240
+ 0.019730432510375977,
241
+ 0.019528703689575197,
242
+ 0.01965465545654297,
243
+ 0.019615743637084963,
244
+ 0.01959321594238281,
245
+ 0.019672063827514647,
246
+ 0.019603456497192383,
247
+ 0.019603456497192383,
248
+ 0.01961369514465332,
249
+ 0.019550207138061524,
250
+ 0.019520511627197267,
251
+ 0.019525632858276368,
252
+ 0.01988198471069336,
253
+ 0.01961881637573242,
254
+ 0.01963929557800293,
255
+ 0.01949388885498047,
256
+ 0.019559423446655275,
257
+ 0.019526655197143555,
258
+ 0.019528703689575197,
259
+ 0.019495935440063478,
260
+ 0.019489791870117186,
261
+ 0.019534847259521485,
262
+ 0.019487743377685548,
263
+ 0.019522560119628905,
264
+ 0.019537919998168944,
265
+ 0.01958502388000488,
266
+ 0.019664896011352538,
267
+ 0.01966592025756836,
268
+ 0.019740671157836915,
269
+ 0.019931135177612306,
270
+ 0.019936256408691407,
271
+ 0.019804159164428712,
272
+ 0.019700735092163087,
273
+ 0.01964134407043457,
274
+ 0.01967103958129883,
275
+ 0.01966796875,
276
+ 0.019751935958862304,
277
+ 0.019725311279296876,
278
+ 0.0196945915222168,
279
+ 0.01966592025756836,
280
+ 0.019752960205078125,
281
+ 0.01967103958129883,
282
+ 0.019772415161132813,
283
+ 0.020715520858764647,
284
+ 0.020633600234985353,
285
+ 0.019900415420532228,
286
+ 0.019733503341674806,
287
+ 0.019801088333129883,
288
+ 0.019749887466430666,
289
+ 0.019768320083618163,
290
+ 0.019917823791503905,
291
+ 0.01965977668762207,
292
+ 0.019687423706054686,
293
+ 0.019537919998168944,
294
+ 0.019715072631835938,
295
+ 0.019869760513305666,
296
+ 0.019715007781982423,
297
+ 0.01966592025756836,
298
+ 0.01983078384399414,
299
+ 0.019795967102050782,
300
+ 0.01966182327270508,
301
+ 0.01965158462524414,
302
+ 0.020290559768676757,
303
+ 0.02085990333557129,
304
+ 0.020060159683227538,
305
+ 0.019768320083618163,
306
+ 0.019737600326538086,
307
+ 0.019732479095458985,
308
+ 0.01981644821166992,
309
+ 0.019688447952270507,
310
+ 0.019711999893188475,
311
+ 0.019763200759887696,
312
+ 0.01965158462524414,
313
+ 0.019672063827514647,
314
+ 0.019695615768432616,
315
+ 0.01959321594238281,
316
+ 0.01966592025756836,
317
+ 0.019701759338378907,
318
+ 0.019746816635131836,
319
+ 0.019713024139404296,
320
+ 0.019750911712646483,
321
+ 0.01979084777832031,
322
+ 0.019603456497192383,
323
+ 0.019794944763183595,
324
+ 0.01983078384399414,
325
+ 0.019755008697509766,
326
+ 0.020586496353149415,
327
+ 0.019891199111938478,
328
+ 0.019746816635131836,
329
+ 0.019739648818969727,
330
+ 0.019703807830810546,
331
+ 0.019878911972045898,
332
+ 0.01967411231994629,
333
+ 0.01986355209350586,
334
+ 0.019693567276000978,
335
+ 0.019813375473022463,
336
+ 0.01948467254638672,
337
+ 0.019388416290283202,
338
+ 0.019513343811035155,
339
+ 0.019713024139404296,
340
+ 0.019211263656616212,
341
+ 0.01949286460876465,
342
+ 0.019594240188598632,
343
+ 0.01965158462524414,
344
+ 0.019536895751953123,
345
+ 0.01949388885498047,
346
+ 0.019619840621948242,
347
+ 0.019547136306762695,
348
+ 0.01963007926940918,
349
+ 0.019647487640380858,
350
+ 0.019583999633789064,
351
+ 0.019588096618652344,
352
+ 0.019518463134765626,
353
+ 0.01963417625427246,
354
+ 0.019727359771728514,
355
+ 0.019552255630493166,
356
+ 0.01952467155456543,
357
+ 0.019548095703125,
358
+ 0.019582975387573243,
359
+ 0.019565568923950196,
360
+ 0.01945907211303711,
361
+ 0.019529727935791014,
362
+ 0.019610624313354492,
363
+ 0.019591167449951173,
364
+ 0.019509248733520508,
365
+ 0.019522560119628905,
366
+ 0.019644416809082032,
367
+ 0.019548160552978516,
368
+ 0.019554304122924804,
369
+ 0.019478527069091797,
370
+ 0.019574783325195313,
371
+ 0.019726335525512697,
372
+ 0.019559423446655275,
373
+ 0.019544063568115236,
374
+ 0.019551231384277345,
375
+ 0.019586048126220702,
376
+ 0.019628032684326172,
377
+ 0.019620864868164063,
378
+ 0.019607551574707033,
379
+ 0.019525632858276368,
380
+ 0.019524608612060547,
381
+ 0.01957683181762695,
382
+ 0.019568639755249022,
383
+ 0.019581951141357423,
384
+ 0.019494911193847657,
385
+ 0.019505151748657225,
386
+ 0.019612672805786133,
387
+ 0.019518463134765626,
388
+ 0.019580928802490235,
389
+ 0.0196177921295166,
390
+ 0.019636224746704102,
391
+ 0.0196628475189209,
392
+ 0.019517440795898438,
393
+ 0.01963929557800293,
394
+ 0.019784704208374023,
395
+ 0.019690591812133788,
396
+ 0.019791776657104493,
397
+ 0.019700735092163087,
398
+ 0.019672063827514647,
399
+ 0.019808256149291992,
400
+ 0.019755008697509766,
401
+ 0.01963520050048828,
402
+ 0.019800064086914062,
403
+ 0.019731456756591798,
404
+ 0.019741695404052736,
405
+ 0.01987481689453125,
406
+ 0.019713024139404296,
407
+ 0.019742719650268553,
408
+ 0.01965670394897461,
409
+ 0.019643455505371093,
410
+ 0.019657663345336914,
411
+ 0.019704832077026366,
412
+ 0.01982259178161621,
413
+ 0.020791296005249024,
414
+ 0.020709375381469726,
415
+ 0.01985024070739746,
416
+ 0.019812351226806642,
417
+ 0.01985843276977539,
418
+ 0.019730432510375977,
419
+ 0.019787776947021486,
420
+ 0.019693567276000978,
421
+ 0.019800064086914062,
422
+ 0.01985024070739746,
423
+ 0.019761152267456054,
424
+ 0.019833856582641602,
425
+ 0.019720191955566405,
426
+ 0.019673088073730468,
427
+ 0.01979903984069824,
428
+ 0.019718143463134767,
429
+ 0.019749887466430666,
430
+ 0.019725311279296876,
431
+ 0.01943654441833496,
432
+ 0.01967513656616211,
433
+ 0.018931711196899414,
434
+ 0.018922496795654296,
435
+ 0.019726335525512697,
436
+ 0.019664896011352538,
437
+ 0.020075519561767577,
438
+ 0.019876863479614256,
439
+ 0.018984960556030273,
440
+ 0.019718143463134767,
441
+ 0.019730432510375977,
442
+ 0.019737600326538086,
443
+ 0.019751935958862304,
444
+ 0.019743743896484374,
445
+ 0.01969049644470215,
446
+ 0.019792896270751953,
447
+ 0.019766271591186522,
448
+ 0.019693567276000978,
449
+ 0.019697664260864257,
450
+ 0.019783679962158202,
451
+ 0.01966694450378418,
452
+ 0.01964543914794922,
453
+ 0.01968230438232422,
454
+ 0.0196628475189209,
455
+ 0.019720191955566405,
456
+ 0.019762176513671875,
457
+ 0.019785728454589844,
458
+ 0.019769344329833984,
459
+ 0.019736576080322265,
460
+ 0.019753984451293945,
461
+ 0.01965056037902832,
462
+ 0.01960550308227539
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 50.89424087654797
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1319.563264,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.304673828125,
15
+ "mean": 12.304673828125,
16
+ "stdev": 0.0,
17
+ "p50": 12.304673828125,
18
+ "p90": 12.304673828125,
19
+ "p95": 12.304673828125,
20
+ "p99": 12.304673828125,
21
+ "values": [
22
+ 12.304673828125
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1715.408896,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.04151971244812012,
42
+ "mean": 0.02075985622406006,
43
+ "stdev": 0.0005112485885620105,
44
+ "p50": 0.02075985622406006,
45
+ "p90": 0.021168855094909667,
46
+ "p95": 0.02121997995376587,
47
+ "p99": 0.02126087984085083,
48
+ "values": [
49
+ 0.02024860763549805,
50
+ 0.02127110481262207
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 337.1892331261527
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1769.824256,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 4.99106298828125,
73
+ "mean": 2.495531494140625,
74
+ "stdev": 0.0010227050781250746,
75
+ "p50": 2.495531494140625,
76
+ "p90": 2.4963496582031253,
77
+ "p95": 2.4964519287109375,
78
+ "p99": 2.4965337451171874,
79
+ "values": [
80
+ 2.4945087890625,
81
+ 2.49655419921875
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 50.89096262587318
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 4.990741498947142,
97
+ "mean": 0.01964858857853206,
98
+ "stdev": 0.0002815677718437068,
99
+ "p50": 0.019663871765136717,
100
+ "p90": 0.01982832622528076,
101
+ "p95": 0.01989442481994629,
102
+ "p99": 0.02071226375579834,
103
+ "values": [
104
+ 0.018699264526367186,
105
+ 0.01927987289428711,
106
+ 0.019525632858276368,
107
+ 0.019517440795898438,
108
+ 0.019513376235961916,
109
+ 0.0196157112121582,
110
+ 0.019568639755249022,
111
+ 0.01901670455932617,
112
+ 0.01878835105895996,
113
+ 0.018720767974853517,
114
+ 0.018754560470581053,
115
+ 0.018782207489013672,
116
+ 0.018693119049072265,
117
+ 0.01965158462524414,
118
+ 0.019628032684326172,
119
+ 0.018737152099609376,
120
+ 0.019804159164428712,
121
+ 0.019615743637084963,
122
+ 0.018774015426635742,
123
+ 0.018734079360961914,
124
+ 0.019360767364501954,
125
+ 0.01962393569946289,
126
+ 0.019599359512329103,
127
+ 0.019524608612060547,
128
+ 0.01963212776184082,
129
+ 0.019604480743408204,
130
+ 0.019555328369140625,
131
+ 0.019581951141357423,
132
+ 0.019582975387573243,
133
+ 0.019545087814331053,
134
+ 0.019745792388916016,
135
+ 0.019730432510375977,
136
+ 0.019528703689575197,
137
+ 0.01965465545654297,
138
+ 0.019615743637084963,
139
+ 0.01959321594238281,
140
+ 0.019672063827514647,
141
+ 0.019603456497192383,
142
+ 0.019603456497192383,
143
+ 0.01961369514465332,
144
+ 0.019550207138061524,
145
+ 0.019520511627197267,
146
+ 0.019525632858276368,
147
+ 0.01988198471069336,
148
+ 0.01961881637573242,
149
+ 0.01963929557800293,
150
+ 0.01949388885498047,
151
+ 0.019559423446655275,
152
+ 0.019526655197143555,
153
+ 0.019528703689575197,
154
+ 0.019495935440063478,
155
+ 0.019489791870117186,
156
+ 0.019534847259521485,
157
+ 0.019487743377685548,
158
+ 0.019522560119628905,
159
+ 0.019537919998168944,
160
+ 0.01958502388000488,
161
+ 0.019664896011352538,
162
+ 0.01966592025756836,
163
+ 0.019740671157836915,
164
+ 0.019931135177612306,
165
+ 0.019936256408691407,
166
+ 0.019804159164428712,
167
+ 0.019700735092163087,
168
+ 0.01964134407043457,
169
+ 0.01967103958129883,
170
+ 0.01966796875,
171
+ 0.019751935958862304,
172
+ 0.019725311279296876,
173
+ 0.0196945915222168,
174
+ 0.01966592025756836,
175
+ 0.019752960205078125,
176
+ 0.01967103958129883,
177
+ 0.019772415161132813,
178
+ 0.020715520858764647,
179
+ 0.020633600234985353,
180
+ 0.019900415420532228,
181
+ 0.019733503341674806,
182
+ 0.019801088333129883,
183
+ 0.019749887466430666,
184
+ 0.019768320083618163,
185
+ 0.019917823791503905,
186
+ 0.01965977668762207,
187
+ 0.019687423706054686,
188
+ 0.019537919998168944,
189
+ 0.019715072631835938,
190
+ 0.019869760513305666,
191
+ 0.019715007781982423,
192
+ 0.01966592025756836,
193
+ 0.01983078384399414,
194
+ 0.019795967102050782,
195
+ 0.01966182327270508,
196
+ 0.01965158462524414,
197
+ 0.020290559768676757,
198
+ 0.02085990333557129,
199
+ 0.020060159683227538,
200
+ 0.019768320083618163,
201
+ 0.019737600326538086,
202
+ 0.019732479095458985,
203
+ 0.01981644821166992,
204
+ 0.019688447952270507,
205
+ 0.019711999893188475,
206
+ 0.019763200759887696,
207
+ 0.01965158462524414,
208
+ 0.019672063827514647,
209
+ 0.019695615768432616,
210
+ 0.01959321594238281,
211
+ 0.01966592025756836,
212
+ 0.019701759338378907,
213
+ 0.019746816635131836,
214
+ 0.019713024139404296,
215
+ 0.019750911712646483,
216
+ 0.01979084777832031,
217
+ 0.019603456497192383,
218
+ 0.019794944763183595,
219
+ 0.01983078384399414,
220
+ 0.019755008697509766,
221
+ 0.020586496353149415,
222
+ 0.019891199111938478,
223
+ 0.019746816635131836,
224
+ 0.019739648818969727,
225
+ 0.019703807830810546,
226
+ 0.019878911972045898,
227
+ 0.01967411231994629,
228
+ 0.01986355209350586,
229
+ 0.019693567276000978,
230
+ 0.019813375473022463,
231
+ 0.01948467254638672,
232
+ 0.019388416290283202,
233
+ 0.019513343811035155,
234
+ 0.019713024139404296,
235
+ 0.019211263656616212,
236
+ 0.01949286460876465,
237
+ 0.019594240188598632,
238
+ 0.01965158462524414,
239
+ 0.019536895751953123,
240
+ 0.01949388885498047,
241
+ 0.019619840621948242,
242
+ 0.019547136306762695,
243
+ 0.01963007926940918,
244
+ 0.019647487640380858,
245
+ 0.019583999633789064,
246
+ 0.019588096618652344,
247
+ 0.019518463134765626,
248
+ 0.01963417625427246,
249
+ 0.019727359771728514,
250
+ 0.019552255630493166,
251
+ 0.01952467155456543,
252
+ 0.019548095703125,
253
+ 0.019582975387573243,
254
+ 0.019565568923950196,
255
+ 0.01945907211303711,
256
+ 0.019529727935791014,
257
+ 0.019610624313354492,
258
+ 0.019591167449951173,
259
+ 0.019509248733520508,
260
+ 0.019522560119628905,
261
+ 0.019644416809082032,
262
+ 0.019548160552978516,
263
+ 0.019554304122924804,
264
+ 0.019478527069091797,
265
+ 0.019574783325195313,
266
+ 0.019726335525512697,
267
+ 0.019559423446655275,
268
+ 0.019544063568115236,
269
+ 0.019551231384277345,
270
+ 0.019586048126220702,
271
+ 0.019628032684326172,
272
+ 0.019620864868164063,
273
+ 0.019607551574707033,
274
+ 0.019525632858276368,
275
+ 0.019524608612060547,
276
+ 0.01957683181762695,
277
+ 0.019568639755249022,
278
+ 0.019581951141357423,
279
+ 0.019494911193847657,
280
+ 0.019505151748657225,
281
+ 0.019612672805786133,
282
+ 0.019518463134765626,
283
+ 0.019580928802490235,
284
+ 0.0196177921295166,
285
+ 0.019636224746704102,
286
+ 0.0196628475189209,
287
+ 0.019517440795898438,
288
+ 0.01963929557800293,
289
+ 0.019784704208374023,
290
+ 0.019690591812133788,
291
+ 0.019791776657104493,
292
+ 0.019700735092163087,
293
+ 0.019672063827514647,
294
+ 0.019808256149291992,
295
+ 0.019755008697509766,
296
+ 0.01963520050048828,
297
+ 0.019800064086914062,
298
+ 0.019731456756591798,
299
+ 0.019741695404052736,
300
+ 0.01987481689453125,
301
+ 0.019713024139404296,
302
+ 0.019742719650268553,
303
+ 0.01965670394897461,
304
+ 0.019643455505371093,
305
+ 0.019657663345336914,
306
+ 0.019704832077026366,
307
+ 0.01982259178161621,
308
+ 0.020791296005249024,
309
+ 0.020709375381469726,
310
+ 0.01985024070739746,
311
+ 0.019812351226806642,
312
+ 0.01985843276977539,
313
+ 0.019730432510375977,
314
+ 0.019787776947021486,
315
+ 0.019693567276000978,
316
+ 0.019800064086914062,
317
+ 0.01985024070739746,
318
+ 0.019761152267456054,
319
+ 0.019833856582641602,
320
+ 0.019720191955566405,
321
+ 0.019673088073730468,
322
+ 0.01979903984069824,
323
+ 0.019718143463134767,
324
+ 0.019749887466430666,
325
+ 0.019725311279296876,
326
+ 0.01943654441833496,
327
+ 0.01967513656616211,
328
+ 0.018931711196899414,
329
+ 0.018922496795654296,
330
+ 0.019726335525512697,
331
+ 0.019664896011352538,
332
+ 0.020075519561767577,
333
+ 0.019876863479614256,
334
+ 0.018984960556030273,
335
+ 0.019718143463134767,
336
+ 0.019730432510375977,
337
+ 0.019737600326538086,
338
+ 0.019751935958862304,
339
+ 0.019743743896484374,
340
+ 0.01969049644470215,
341
+ 0.019792896270751953,
342
+ 0.019766271591186522,
343
+ 0.019693567276000978,
344
+ 0.019697664260864257,
345
+ 0.019783679962158202,
346
+ 0.01966694450378418,
347
+ 0.01964543914794922,
348
+ 0.01968230438232422,
349
+ 0.0196628475189209,
350
+ 0.019720191955566405,
351
+ 0.019762176513671875,
352
+ 0.019785728454589844,
353
+ 0.019769344329833984,
354
+ 0.019736576080322265,
355
+ 0.019753984451293945,
356
+ 0.01965056037902832,
357
+ 0.01960550308227539
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 50.89424087654797
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02075985622406006,
7
+ "prefill.throughput.value": 337.1892331261527,
8
+ "decode.latency.mean": 2.495531494140625,
9
+ "decode.throughput.value": 50.89096262587318,
10
+ "per_token.latency.mean": 0.01964858857853206,
11
+ "per_token.throughput.value": 50.89424087654797
12
+ }
13
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: c6b23fda65f9ae74f9a1026b340241f65aebe1a3
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
88
+ - hydra.run.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-08-27 13:53:23,647][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-08-27 13:53:26,048][process][ERROR] - + Received traceback from isolated process
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: c6b23fda65f9ae74f9a1026b340241f65aebe1a3
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
88
+ - hydra.run.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1319.30112,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.77019140625,
120
+ "mean": 12.77019140625,
121
+ "stdev": 0.0,
122
+ "p50": 12.77019140625,
123
+ "p90": 12.77019140625,
124
+ "p95": 12.77019140625,
125
+ "p99": 12.77019140625,
126
+ "values": [
127
+ 12.77019140625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1753.64096,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.0462674560546875,
147
+ "mean": 0.02313372802734375,
148
+ "stdev": 0.00034492874145507744,
149
+ "p50": 0.02313372802734375,
150
+ "p90": 0.02340967102050781,
151
+ "p95": 0.02344416389465332,
152
+ "p99": 0.023471758193969725,
153
+ "values": [
154
+ 0.022788799285888672,
155
+ 0.023478656768798827
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 302.5884972679758
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1778.089984,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.391643798828125,
178
+ "mean": 2.6958218994140624,
179
+ "stdev": 0.0068511962890625,
180
+ "p50": 2.6958218994140624,
181
+ "p90": 2.7013028564453125,
182
+ "p95": 2.7019879760742187,
183
+ "p99": 2.7025360717773435,
184
+ "values": [
185
+ 2.702673095703125,
186
+ 2.688970703125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 47.10993705763852
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.391390718460083,
202
+ "mean": 0.02122594771047277,
203
+ "stdev": 0.0002843550696369424,
204
+ "p50": 0.02112512016296387,
205
+ "p90": 0.021713613319396974,
206
+ "p95": 0.021857024002075197,
207
+ "p99": 0.02214663164138794,
208
+ "values": [
209
+ 0.020996095657348633,
210
+ 0.021032960891723632,
211
+ 0.02104729652404785,
212
+ 0.021072895050048827,
213
+ 0.02105548858642578,
214
+ 0.02106265640258789,
215
+ 0.02107494354248047,
216
+ 0.021089279174804687,
217
+ 0.02107187271118164,
218
+ 0.02107187271118164,
219
+ 0.0214783992767334,
220
+ 0.02165247917175293,
221
+ 0.021171199798583985,
222
+ 0.021032960891723632,
223
+ 0.02104115104675293,
224
+ 0.02102783966064453,
225
+ 0.02104319953918457,
226
+ 0.02105036735534668,
227
+ 0.02104217529296875,
228
+ 0.021015552520751952,
229
+ 0.02109132766723633,
230
+ 0.02106675148010254,
231
+ 0.0210513916015625,
232
+ 0.02103603172302246,
233
+ 0.021015552520751952,
234
+ 0.02103910446166992,
235
+ 0.02102374458312988,
236
+ 0.020997119903564454,
237
+ 0.021136383056640624,
238
+ 0.02109235191345215,
239
+ 0.02107084846496582,
240
+ 0.02107904052734375,
241
+ 0.021122047424316406,
242
+ 0.021537792205810546,
243
+ 0.02188390350341797,
244
+ 0.021167104721069335,
245
+ 0.02106265640258789,
246
+ 0.021024768829345702,
247
+ 0.02109542465209961,
248
+ 0.02104729652404785,
249
+ 0.0210565128326416,
250
+ 0.02107187271118164,
251
+ 0.02106572723388672,
252
+ 0.021000192642211913,
253
+ 0.021046272277832033,
254
+ 0.022131711959838866,
255
+ 0.022350847244262697,
256
+ 0.023023616790771483,
257
+ 0.021962751388549806,
258
+ 0.02168320083618164,
259
+ 0.021113855361938477,
260
+ 0.02103193664550781,
261
+ 0.02104422378540039,
262
+ 0.02105855941772461,
263
+ 0.021124095916748048,
264
+ 0.021046272277832033,
265
+ 0.0211015682220459,
266
+ 0.021129215240478515,
267
+ 0.021132287979125978,
268
+ 0.021135360717773437,
269
+ 0.021130239486694336,
270
+ 0.021124095916748048,
271
+ 0.02109644889831543,
272
+ 0.021128192901611328,
273
+ 0.022163455963134765,
274
+ 0.022012928009033202,
275
+ 0.021910528182983398,
276
+ 0.021110784530639647,
277
+ 0.021078016281127928,
278
+ 0.021124095916748048,
279
+ 0.021145599365234375,
280
+ 0.02106777572631836,
281
+ 0.021133312225341795,
282
+ 0.02171392059326172,
283
+ 0.021123071670532227,
284
+ 0.021082111358642578,
285
+ 0.021106687545776368,
286
+ 0.021127168655395507,
287
+ 0.021117952346801756,
288
+ 0.021644287109375,
289
+ 0.021780479431152345,
290
+ 0.02170675277709961,
291
+ 0.021139455795288087,
292
+ 0.02167398452758789,
293
+ 0.02189516830444336,
294
+ 0.021149696350097655,
295
+ 0.021980159759521483,
296
+ 0.021181440353393553,
297
+ 0.021139455795288087,
298
+ 0.021121023178100586,
299
+ 0.021214208602905273,
300
+ 0.021126144409179686,
301
+ 0.02122444725036621,
302
+ 0.0212674560546875,
303
+ 0.02184601593017578,
304
+ 0.02169241523742676,
305
+ 0.021123071670532227,
306
+ 0.0211015682220459,
307
+ 0.021158912658691405,
308
+ 0.021090303421020508,
309
+ 0.021130239486694336,
310
+ 0.02112512016296387,
311
+ 0.021142528533935546,
312
+ 0.021078016281127928,
313
+ 0.02150912094116211,
314
+ 0.02127359962463379,
315
+ 0.021128192901611328,
316
+ 0.021141504287719725,
317
+ 0.021168127059936523,
318
+ 0.021073919296264648,
319
+ 0.021129215240478515,
320
+ 0.021550079345703126,
321
+ 0.021140480041503908,
322
+ 0.0210831356048584,
323
+ 0.021133312225341795,
324
+ 0.02148659133911133,
325
+ 0.02182963180541992,
326
+ 0.021521408081054686,
327
+ 0.021554176330566405,
328
+ 0.021537792205810546,
329
+ 0.02185523223876953,
330
+ 0.021712896347045898,
331
+ 0.021760000228881835,
332
+ 0.021725183486938478,
333
+ 0.021719039916992186,
334
+ 0.02110771179199219,
335
+ 0.02109644889831543,
336
+ 0.021761024475097656,
337
+ 0.021719039916992186,
338
+ 0.021128192901611328,
339
+ 0.021126144409179686,
340
+ 0.021140480041503908,
341
+ 0.021134336471557616,
342
+ 0.021234687805175782,
343
+ 0.021345279693603517,
344
+ 0.02106572723388672,
345
+ 0.021131263732910157,
346
+ 0.021146623611450196,
347
+ 0.021037055969238282,
348
+ 0.021135360717773437,
349
+ 0.021106687545776368,
350
+ 0.021110784530639647,
351
+ 0.02105958366394043,
352
+ 0.02107494354248047,
353
+ 0.021082111358642578,
354
+ 0.02110361671447754,
355
+ 0.02106777572631836,
356
+ 0.02105036735534668,
357
+ 0.021090303421020508,
358
+ 0.021090303421020508,
359
+ 0.02110771179199219,
360
+ 0.021102592468261717,
361
+ 0.02106675148010254,
362
+ 0.02110054397583008,
363
+ 0.021134336471557616,
364
+ 0.02109337615966797,
365
+ 0.021073919296264648,
366
+ 0.021123071670532227,
367
+ 0.02109644889831543,
368
+ 0.02109951972961426,
369
+ 0.021110784530639647,
370
+ 0.021118976593017577,
371
+ 0.02112512016296387,
372
+ 0.02106675148010254,
373
+ 0.02106777572631836,
374
+ 0.021111808776855468,
375
+ 0.0210882568359375,
376
+ 0.02112512016296387,
377
+ 0.02108006477355957,
378
+ 0.021082111358642578,
379
+ 0.021082111358642578,
380
+ 0.0210513916015625,
381
+ 0.02106060791015625,
382
+ 0.021112831115722656,
383
+ 0.02107699203491211,
384
+ 0.021124095916748048,
385
+ 0.02162483215332031,
386
+ 0.0218603515625,
387
+ 0.021740543365478517,
388
+ 0.0212490234375,
389
+ 0.021737472534179687,
390
+ 0.021113855361938477,
391
+ 0.02109337615966797,
392
+ 0.02110771179199219,
393
+ 0.021209087371826172,
394
+ 0.021152767181396484,
395
+ 0.021141504287719725,
396
+ 0.021117952346801756,
397
+ 0.02120806312561035,
398
+ 0.021183488845825195,
399
+ 0.021113855361938477,
400
+ 0.02117740821838379,
401
+ 0.02114246368408203,
402
+ 0.021183488845825195,
403
+ 0.021197824478149413,
404
+ 0.021123071670532227,
405
+ 0.021102592468261717,
406
+ 0.021164031982421876,
407
+ 0.021161983489990235,
408
+ 0.021133312225341795,
409
+ 0.021149696350097655,
410
+ 0.021132287979125978,
411
+ 0.021150720596313476,
412
+ 0.021165056228637694,
413
+ 0.02109235191345215,
414
+ 0.021145599365234375,
415
+ 0.021127168655395507,
416
+ 0.021146623611450196,
417
+ 0.021177343368530274,
418
+ 0.02110054397583008,
419
+ 0.021189632415771483,
420
+ 0.021164031982421876,
421
+ 0.021123071670532227,
422
+ 0.021145599365234375,
423
+ 0.021102592468261717,
424
+ 0.021133312225341795,
425
+ 0.021118976593017577,
426
+ 0.021138431549072266,
427
+ 0.021104639053344726,
428
+ 0.021133312225341795,
429
+ 0.021139455795288087,
430
+ 0.021171199798583985,
431
+ 0.021112831115722656,
432
+ 0.021159936904907226,
433
+ 0.021061632156372072,
434
+ 0.021114879608154297,
435
+ 0.0210831356048584,
436
+ 0.021135360717773437,
437
+ 0.021128192901611328,
438
+ 0.02106470489501953,
439
+ 0.02086809539794922,
440
+ 0.020918272018432618,
441
+ 0.02109644889831543,
442
+ 0.021165056228637694,
443
+ 0.020933631896972657,
444
+ 0.02105958366394043,
445
+ 0.0211015682220459,
446
+ 0.021160959243774414,
447
+ 0.021137407302856445,
448
+ 0.021385215759277345,
449
+ 0.021207040786743164,
450
+ 0.021152767181396484,
451
+ 0.021148672103881837,
452
+ 0.02168422317504883,
453
+ 0.021148672103881837,
454
+ 0.021181440353393553,
455
+ 0.021936128616333008,
456
+ 0.02187571144104004,
457
+ 0.021766143798828123,
458
+ 0.021110784530639647,
459
+ 0.02112623977661133,
460
+ 0.021138336181640623,
461
+ 0.02107084846496582,
462
+ 0.02109337615966797
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 47.11214847224962
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1319.30112,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.77019140625,
15
+ "mean": 12.77019140625,
16
+ "stdev": 0.0,
17
+ "p50": 12.77019140625,
18
+ "p90": 12.77019140625,
19
+ "p95": 12.77019140625,
20
+ "p99": 12.77019140625,
21
+ "values": [
22
+ 12.77019140625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1753.64096,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.0462674560546875,
42
+ "mean": 0.02313372802734375,
43
+ "stdev": 0.00034492874145507744,
44
+ "p50": 0.02313372802734375,
45
+ "p90": 0.02340967102050781,
46
+ "p95": 0.02344416389465332,
47
+ "p99": 0.023471758193969725,
48
+ "values": [
49
+ 0.022788799285888672,
50
+ 0.023478656768798827
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 302.5884972679758
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1778.089984,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.391643798828125,
73
+ "mean": 2.6958218994140624,
74
+ "stdev": 0.0068511962890625,
75
+ "p50": 2.6958218994140624,
76
+ "p90": 2.7013028564453125,
77
+ "p95": 2.7019879760742187,
78
+ "p99": 2.7025360717773435,
79
+ "values": [
80
+ 2.702673095703125,
81
+ 2.688970703125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 47.10993705763852
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.391390718460083,
97
+ "mean": 0.02122594771047277,
98
+ "stdev": 0.0002843550696369424,
99
+ "p50": 0.02112512016296387,
100
+ "p90": 0.021713613319396974,
101
+ "p95": 0.021857024002075197,
102
+ "p99": 0.02214663164138794,
103
+ "values": [
104
+ 0.020996095657348633,
105
+ 0.021032960891723632,
106
+ 0.02104729652404785,
107
+ 0.021072895050048827,
108
+ 0.02105548858642578,
109
+ 0.02106265640258789,
110
+ 0.02107494354248047,
111
+ 0.021089279174804687,
112
+ 0.02107187271118164,
113
+ 0.02107187271118164,
114
+ 0.0214783992767334,
115
+ 0.02165247917175293,
116
+ 0.021171199798583985,
117
+ 0.021032960891723632,
118
+ 0.02104115104675293,
119
+ 0.02102783966064453,
120
+ 0.02104319953918457,
121
+ 0.02105036735534668,
122
+ 0.02104217529296875,
123
+ 0.021015552520751952,
124
+ 0.02109132766723633,
125
+ 0.02106675148010254,
126
+ 0.0210513916015625,
127
+ 0.02103603172302246,
128
+ 0.021015552520751952,
129
+ 0.02103910446166992,
130
+ 0.02102374458312988,
131
+ 0.020997119903564454,
132
+ 0.021136383056640624,
133
+ 0.02109235191345215,
134
+ 0.02107084846496582,
135
+ 0.02107904052734375,
136
+ 0.021122047424316406,
137
+ 0.021537792205810546,
138
+ 0.02188390350341797,
139
+ 0.021167104721069335,
140
+ 0.02106265640258789,
141
+ 0.021024768829345702,
142
+ 0.02109542465209961,
143
+ 0.02104729652404785,
144
+ 0.0210565128326416,
145
+ 0.02107187271118164,
146
+ 0.02106572723388672,
147
+ 0.021000192642211913,
148
+ 0.021046272277832033,
149
+ 0.022131711959838866,
150
+ 0.022350847244262697,
151
+ 0.023023616790771483,
152
+ 0.021962751388549806,
153
+ 0.02168320083618164,
154
+ 0.021113855361938477,
155
+ 0.02103193664550781,
156
+ 0.02104422378540039,
157
+ 0.02105855941772461,
158
+ 0.021124095916748048,
159
+ 0.021046272277832033,
160
+ 0.0211015682220459,
161
+ 0.021129215240478515,
162
+ 0.021132287979125978,
163
+ 0.021135360717773437,
164
+ 0.021130239486694336,
165
+ 0.021124095916748048,
166
+ 0.02109644889831543,
167
+ 0.021128192901611328,
168
+ 0.022163455963134765,
169
+ 0.022012928009033202,
170
+ 0.021910528182983398,
171
+ 0.021110784530639647,
172
+ 0.021078016281127928,
173
+ 0.021124095916748048,
174
+ 0.021145599365234375,
175
+ 0.02106777572631836,
176
+ 0.021133312225341795,
177
+ 0.02171392059326172,
178
+ 0.021123071670532227,
179
+ 0.021082111358642578,
180
+ 0.021106687545776368,
181
+ 0.021127168655395507,
182
+ 0.021117952346801756,
183
+ 0.021644287109375,
184
+ 0.021780479431152345,
185
+ 0.02170675277709961,
186
+ 0.021139455795288087,
187
+ 0.02167398452758789,
188
+ 0.02189516830444336,
189
+ 0.021149696350097655,
190
+ 0.021980159759521483,
191
+ 0.021181440353393553,
192
+ 0.021139455795288087,
193
+ 0.021121023178100586,
194
+ 0.021214208602905273,
195
+ 0.021126144409179686,
196
+ 0.02122444725036621,
197
+ 0.0212674560546875,
198
+ 0.02184601593017578,
199
+ 0.02169241523742676,
200
+ 0.021123071670532227,
201
+ 0.0211015682220459,
202
+ 0.021158912658691405,
203
+ 0.021090303421020508,
204
+ 0.021130239486694336,
205
+ 0.02112512016296387,
206
+ 0.021142528533935546,
207
+ 0.021078016281127928,
208
+ 0.02150912094116211,
209
+ 0.02127359962463379,
210
+ 0.021128192901611328,
211
+ 0.021141504287719725,
212
+ 0.021168127059936523,
213
+ 0.021073919296264648,
214
+ 0.021129215240478515,
215
+ 0.021550079345703126,
216
+ 0.021140480041503908,
217
+ 0.0210831356048584,
218
+ 0.021133312225341795,
219
+ 0.02148659133911133,
220
+ 0.02182963180541992,
221
+ 0.021521408081054686,
222
+ 0.021554176330566405,
223
+ 0.021537792205810546,
224
+ 0.02185523223876953,
225
+ 0.021712896347045898,
226
+ 0.021760000228881835,
227
+ 0.021725183486938478,
228
+ 0.021719039916992186,
229
+ 0.02110771179199219,
230
+ 0.02109644889831543,
231
+ 0.021761024475097656,
232
+ 0.021719039916992186,
233
+ 0.021128192901611328,
234
+ 0.021126144409179686,
235
+ 0.021140480041503908,
236
+ 0.021134336471557616,
237
+ 0.021234687805175782,
238
+ 0.021345279693603517,
239
+ 0.02106572723388672,
240
+ 0.021131263732910157,
241
+ 0.021146623611450196,
242
+ 0.021037055969238282,
243
+ 0.021135360717773437,
244
+ 0.021106687545776368,
245
+ 0.021110784530639647,
246
+ 0.02105958366394043,
247
+ 0.02107494354248047,
248
+ 0.021082111358642578,
249
+ 0.02110361671447754,
250
+ 0.02106777572631836,
251
+ 0.02105036735534668,
252
+ 0.021090303421020508,
253
+ 0.021090303421020508,
254
+ 0.02110771179199219,
255
+ 0.021102592468261717,
256
+ 0.02106675148010254,
257
+ 0.02110054397583008,
258
+ 0.021134336471557616,
259
+ 0.02109337615966797,
260
+ 0.021073919296264648,
261
+ 0.021123071670532227,
262
+ 0.02109644889831543,
263
+ 0.02109951972961426,
264
+ 0.021110784530639647,
265
+ 0.021118976593017577,
266
+ 0.02112512016296387,
267
+ 0.02106675148010254,
268
+ 0.02106777572631836,
269
+ 0.021111808776855468,
270
+ 0.0210882568359375,
271
+ 0.02112512016296387,
272
+ 0.02108006477355957,
273
+ 0.021082111358642578,
274
+ 0.021082111358642578,
275
+ 0.0210513916015625,
276
+ 0.02106060791015625,
277
+ 0.021112831115722656,
278
+ 0.02107699203491211,
279
+ 0.021124095916748048,
280
+ 0.02162483215332031,
281
+ 0.0218603515625,
282
+ 0.021740543365478517,
283
+ 0.0212490234375,
284
+ 0.021737472534179687,
285
+ 0.021113855361938477,
286
+ 0.02109337615966797,
287
+ 0.02110771179199219,
288
+ 0.021209087371826172,
289
+ 0.021152767181396484,
290
+ 0.021141504287719725,
291
+ 0.021117952346801756,
292
+ 0.02120806312561035,
293
+ 0.021183488845825195,
294
+ 0.021113855361938477,
295
+ 0.02117740821838379,
296
+ 0.02114246368408203,
297
+ 0.021183488845825195,
298
+ 0.021197824478149413,
299
+ 0.021123071670532227,
300
+ 0.021102592468261717,
301
+ 0.021164031982421876,
302
+ 0.021161983489990235,
303
+ 0.021133312225341795,
304
+ 0.021149696350097655,
305
+ 0.021132287979125978,
306
+ 0.021150720596313476,
307
+ 0.021165056228637694,
308
+ 0.02109235191345215,
309
+ 0.021145599365234375,
310
+ 0.021127168655395507,
311
+ 0.021146623611450196,
312
+ 0.021177343368530274,
313
+ 0.02110054397583008,
314
+ 0.021189632415771483,
315
+ 0.021164031982421876,
316
+ 0.021123071670532227,
317
+ 0.021145599365234375,
318
+ 0.021102592468261717,
319
+ 0.021133312225341795,
320
+ 0.021118976593017577,
321
+ 0.021138431549072266,
322
+ 0.021104639053344726,
323
+ 0.021133312225341795,
324
+ 0.021139455795288087,
325
+ 0.021171199798583985,
326
+ 0.021112831115722656,
327
+ 0.021159936904907226,
328
+ 0.021061632156372072,
329
+ 0.021114879608154297,
330
+ 0.0210831356048584,
331
+ 0.021135360717773437,
332
+ 0.021128192901611328,
333
+ 0.02106470489501953,
334
+ 0.02086809539794922,
335
+ 0.020918272018432618,
336
+ 0.02109644889831543,
337
+ 0.021165056228637694,
338
+ 0.020933631896972657,
339
+ 0.02105958366394043,
340
+ 0.0211015682220459,
341
+ 0.021160959243774414,
342
+ 0.021137407302856445,
343
+ 0.021385215759277345,
344
+ 0.021207040786743164,
345
+ 0.021152767181396484,
346
+ 0.021148672103881837,
347
+ 0.02168422317504883,
348
+ 0.021148672103881837,
349
+ 0.021181440353393553,
350
+ 0.021936128616333008,
351
+ 0.02187571144104004,
352
+ 0.021766143798828123,
353
+ 0.021110784530639647,
354
+ 0.02112623977661133,
355
+ 0.021138336181640623,
356
+ 0.02107084846496582,
357
+ 0.02109337615966797
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 47.11214847224962
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02313372802734375,
7
+ "prefill.throughput.value": 302.5884972679758,
8
+ "decode.latency.mean": 2.6958218994140624,
9
+ "decode.throughput.value": 47.10993705763852,
10
+ "per_token.latency.mean": 0.02122594771047277,
11
+ "per_token.throughput.value": 47.11214847224962
12
+ }
13
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: c6b23fda65f9ae74f9a1026b340241f65aebe1a3
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
88
+ - hydra.run.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1339.613184,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 13.4834521484375,
120
+ "mean": 13.4834521484375,
121
+ "stdev": 0.0,
122
+ "p50": 13.4834521484375,
123
+ "p90": 13.4834521484375,
124
+ "p95": 13.4834521484375,
125
+ "p99": 13.4834521484375,
126
+ "values": [
127
+ 13.4834521484375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1858.891776,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.028394752502441407,
147
+ "mean": 0.014197376251220703,
148
+ "stdev": 0.00017443180084228578,
149
+ "p50": 0.014197376251220703,
150
+ "p90": 0.014336921691894532,
151
+ "p95": 0.01435436487197876,
152
+ "p99": 0.014368319416046143,
153
+ "values": [
154
+ 0.014022944450378418,
155
+ 0.01437180805206299
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 493.0488476276124
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2028.4416,
169
+ "max_global_vram": 6071.779328,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.131483276367187,
178
+ "mean": 1.5657416381835936,
179
+ "stdev": 0.001327087402343774,
180
+ "p50": 1.5657416381835936,
181
+ "p90": 1.5668033081054686,
182
+ "p95": 1.5669360168457032,
183
+ "p99": 1.5670421838378905,
184
+ "values": [
185
+ 1.5670687255859375,
186
+ 1.56441455078125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.11172169332602
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.1311953868865996,
202
+ "mean": 0.012327540893254317,
203
+ "stdev": 5.417195347269009e-05,
204
+ "p50": 0.012315648078918457,
205
+ "p90": 0.012347084808349609,
206
+ "p95": 0.01236664342880249,
207
+ "p99": 0.012504289407730103,
208
+ "values": [
209
+ 0.012337151527404786,
210
+ 0.012328960418701173,
211
+ 0.012320768356323243,
212
+ 0.012327936172485352,
213
+ 0.012304384231567383,
214
+ 0.012311552047729492,
215
+ 0.012313599586486817,
216
+ 0.012312576293945313,
217
+ 0.012346367835998535,
218
+ 0.012337151527404786,
219
+ 0.01234124755859375,
220
+ 0.012312576293945313,
221
+ 0.012315648078918457,
222
+ 0.012301376342773437,
223
+ 0.012308416366577148,
224
+ 0.012337151527404786,
225
+ 0.012320768356323243,
226
+ 0.012327936172485352,
227
+ 0.012331007957458496,
228
+ 0.012305407524108887,
229
+ 0.012310527801513671,
230
+ 0.012305407524108887,
231
+ 0.012327936172485352,
232
+ 0.012331007957458496,
233
+ 0.012308480262756348,
234
+ 0.012320768356323243,
235
+ 0.012309568405151367,
236
+ 0.012329919815063477,
237
+ 0.012323840141296387,
238
+ 0.012304384231567383,
239
+ 0.012313599586486817,
240
+ 0.012315648078918457,
241
+ 0.012304384231567383,
242
+ 0.012404735565185548,
243
+ 0.01238425636291504,
244
+ 0.012363776206970215,
245
+ 0.012329983711242675,
246
+ 0.012305407524108887,
247
+ 0.012316672325134278,
248
+ 0.012319744110107422,
249
+ 0.012340224266052247,
250
+ 0.012331007957458496,
251
+ 0.012319744110107422,
252
+ 0.012320768356323243,
253
+ 0.012313599586486817,
254
+ 0.012300288200378418,
255
+ 0.01232588768005371,
256
+ 0.012328960418701173,
257
+ 0.012314623832702636,
258
+ 0.012412927627563476,
259
+ 0.01234124755859375,
260
+ 0.012324864387512208,
261
+ 0.012324864387512208,
262
+ 0.012307456016540527,
263
+ 0.012318719863891601,
264
+ 0.012371968269348145,
265
+ 0.012572671890258789,
266
+ 0.01276313591003418,
267
+ 0.012375040054321289,
268
+ 0.012312576293945313,
269
+ 0.012303359985351562,
270
+ 0.012323840141296387,
271
+ 0.012319744110107422,
272
+ 0.012324864387512208,
273
+ 0.012310527801513671,
274
+ 0.012318719863891601,
275
+ 0.01235865592956543,
276
+ 0.012314623832702636,
277
+ 0.012311552047729492,
278
+ 0.012332032203674317,
279
+ 0.012310527801513671,
280
+ 0.012314623832702636,
281
+ 0.012331007957458496,
282
+ 0.012302335739135742,
283
+ 0.012319744110107422,
284
+ 0.012309503555297852,
285
+ 0.012328960418701173,
286
+ 0.012307456016540527,
287
+ 0.012332032203674317,
288
+ 0.012323840141296387,
289
+ 0.012303359985351562,
290
+ 0.012320768356323243,
291
+ 0.012349439620971679,
292
+ 0.012313599586486817,
293
+ 0.012324864387512208,
294
+ 0.012313599586486817,
295
+ 0.012381183624267577,
296
+ 0.01293824005126953,
297
+ 0.012443648338317872,
298
+ 0.012347392082214356,
299
+ 0.012348416328430176,
300
+ 0.012353568077087403,
301
+ 0.012312543869018555,
302
+ 0.012311552047729492,
303
+ 0.012317695617675782,
304
+ 0.012311552047729492,
305
+ 0.012328960418701173,
306
+ 0.012313599586486817,
307
+ 0.012337151527404786,
308
+ 0.012315648078918457,
309
+ 0.012337151527404786,
310
+ 0.012305407524108887,
311
+ 0.012312576293945313,
312
+ 0.012354559898376465,
313
+ 0.012320768356323243,
314
+ 0.012312576293945313,
315
+ 0.012300288200378418,
316
+ 0.012309503555297852,
317
+ 0.012307456016540527,
318
+ 0.012329983711242675,
319
+ 0.012302335739135742,
320
+ 0.012317695617675782,
321
+ 0.012313599586486817,
322
+ 0.012319744110107422,
323
+ 0.012309503555297852,
324
+ 0.012314623832702636,
325
+ 0.012348416328430176,
326
+ 0.012318719863891601,
327
+ 0.012306431770324706,
328
+ 0.012329983711242675,
329
+ 0.012442624092102051,
330
+ 0.012329024314880371,
331
+ 0.012340160369873046,
332
+ 0.012340224266052247,
333
+ 0.012355584144592285,
334
+ 0.012420096397399903,
335
+ 0.012345343589782716,
336
+ 0.012353535652160644,
337
+ 0.012314623832702636,
338
+ 0.012383232116699219,
339
+ 0.012324864387512208,
340
+ 0.012310527801513671,
341
+ 0.01233510398864746,
342
+ 0.012316672325134278,
343
+ 0.012301312446594239,
344
+ 0.012301312446594239,
345
+ 0.012311552047729492,
346
+ 0.01234227180480957,
347
+ 0.012308480262756348,
348
+ 0.012314623832702636,
349
+ 0.012303359985351562,
350
+ 0.012322815895080566,
351
+ 0.012318719863891601,
352
+ 0.012309503555297852,
353
+ 0.012332032203674317,
354
+ 0.012323840141296387,
355
+ 0.012320768356323243,
356
+ 0.012317695617675782,
357
+ 0.012312576293945313,
358
+ 0.01232588768005371,
359
+ 0.012314623832702636,
360
+ 0.012310527801513671,
361
+ 0.012310527801513671,
362
+ 0.012344351768493652,
363
+ 0.012314592361450195,
364
+ 0.012300288200378418,
365
+ 0.012304384231567383,
366
+ 0.0123504638671875,
367
+ 0.012312576293945313,
368
+ 0.012307456016540527,
369
+ 0.012297216415405274,
370
+ 0.012303359985351562,
371
+ 0.012316672325134278,
372
+ 0.012314623832702636,
373
+ 0.012297216415405274,
374
+ 0.012308480262756348,
375
+ 0.012309503555297852,
376
+ 0.012310527801513671,
377
+ 0.012331007957458496,
378
+ 0.012299263954162597,
379
+ 0.012347392082214356,
380
+ 0.012317695617675782,
381
+ 0.012310527801513671,
382
+ 0.012316672325134278,
383
+ 0.012313599586486817,
384
+ 0.012315648078918457,
385
+ 0.012337151527404786,
386
+ 0.012314623832702636,
387
+ 0.012329983711242675,
388
+ 0.012307456016540527,
389
+ 0.012333056449890138,
390
+ 0.012312576293945313,
391
+ 0.012331007957458496,
392
+ 0.012324864387512208,
393
+ 0.012305407524108887,
394
+ 0.012320768356323243,
395
+ 0.012307456016540527,
396
+ 0.012307456016540527,
397
+ 0.01232588768005371,
398
+ 0.012309503555297852,
399
+ 0.012302335739135742,
400
+ 0.012313599586486817,
401
+ 0.012311552047729492,
402
+ 0.012308480262756348,
403
+ 0.012307456016540527,
404
+ 0.01234329605102539,
405
+ 0.012322815895080566,
406
+ 0.012319744110107422,
407
+ 0.012311552047729492,
408
+ 0.012312576293945313,
409
+ 0.012322815895080566,
410
+ 0.01232588768005371,
411
+ 0.012309503555297852,
412
+ 0.012313599586486817,
413
+ 0.012310527801513671,
414
+ 0.012310527801513671,
415
+ 0.012300288200378418,
416
+ 0.01233510398864746,
417
+ 0.012326911926269531,
418
+ 0.012311552047729492,
419
+ 0.012339200019836426,
420
+ 0.012311552047729492,
421
+ 0.012309503555297852,
422
+ 0.012313599586486817,
423
+ 0.012306431770324706,
424
+ 0.012321791648864745,
425
+ 0.012307456016540527,
426
+ 0.012311552047729492,
427
+ 0.012304384231567383,
428
+ 0.01233510398864746,
429
+ 0.01233407974243164,
430
+ 0.012336128234863282,
431
+ 0.012308480262756348,
432
+ 0.012309503555297852,
433
+ 0.012298239707946777,
434
+ 0.012313599586486817,
435
+ 0.012323840141296387,
436
+ 0.012311552047729492,
437
+ 0.012317695617675782,
438
+ 0.012306431770324706,
439
+ 0.012304384231567383,
440
+ 0.012311552047729492,
441
+ 0.012308480262756348,
442
+ 0.012347392082214356,
443
+ 0.012309503555297852,
444
+ 0.012309503555297852,
445
+ 0.012310527801513671,
446
+ 0.012333056449890138,
447
+ 0.012320768356323243,
448
+ 0.012328960418701173,
449
+ 0.012313599586486817,
450
+ 0.012312576293945313,
451
+ 0.012311552047729492,
452
+ 0.012315648078918457,
453
+ 0.012303359985351562,
454
+ 0.01234124755859375,
455
+ 0.012326911926269531,
456
+ 0.012306431770324706,
457
+ 0.012305407524108887,
458
+ 0.012312576293945313,
459
+ 0.012310527801513671,
460
+ 0.012313599586486817,
461
+ 0.012307456016540527,
462
+ 0.012316672325134278
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.11917929610797
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1339.613184,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 13.4834521484375,
15
+ "mean": 13.4834521484375,
16
+ "stdev": 0.0,
17
+ "p50": 13.4834521484375,
18
+ "p90": 13.4834521484375,
19
+ "p95": 13.4834521484375,
20
+ "p99": 13.4834521484375,
21
+ "values": [
22
+ 13.4834521484375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1858.891776,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.028394752502441407,
42
+ "mean": 0.014197376251220703,
43
+ "stdev": 0.00017443180084228578,
44
+ "p50": 0.014197376251220703,
45
+ "p90": 0.014336921691894532,
46
+ "p95": 0.01435436487197876,
47
+ "p99": 0.014368319416046143,
48
+ "values": [
49
+ 0.014022944450378418,
50
+ 0.01437180805206299
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 493.0488476276124
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2028.4416,
64
+ "max_global_vram": 6071.779328,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.131483276367187,
73
+ "mean": 1.5657416381835936,
74
+ "stdev": 0.001327087402343774,
75
+ "p50": 1.5657416381835936,
76
+ "p90": 1.5668033081054686,
77
+ "p95": 1.5669360168457032,
78
+ "p99": 1.5670421838378905,
79
+ "values": [
80
+ 1.5670687255859375,
81
+ 1.56441455078125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.11172169332602
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.1311953868865996,
97
+ "mean": 0.012327540893254317,
98
+ "stdev": 5.417195347269009e-05,
99
+ "p50": 0.012315648078918457,
100
+ "p90": 0.012347084808349609,
101
+ "p95": 0.01236664342880249,
102
+ "p99": 0.012504289407730103,
103
+ "values": [
104
+ 0.012337151527404786,
105
+ 0.012328960418701173,
106
+ 0.012320768356323243,
107
+ 0.012327936172485352,
108
+ 0.012304384231567383,
109
+ 0.012311552047729492,
110
+ 0.012313599586486817,
111
+ 0.012312576293945313,
112
+ 0.012346367835998535,
113
+ 0.012337151527404786,
114
+ 0.01234124755859375,
115
+ 0.012312576293945313,
116
+ 0.012315648078918457,
117
+ 0.012301376342773437,
118
+ 0.012308416366577148,
119
+ 0.012337151527404786,
120
+ 0.012320768356323243,
121
+ 0.012327936172485352,
122
+ 0.012331007957458496,
123
+ 0.012305407524108887,
124
+ 0.012310527801513671,
125
+ 0.012305407524108887,
126
+ 0.012327936172485352,
127
+ 0.012331007957458496,
128
+ 0.012308480262756348,
129
+ 0.012320768356323243,
130
+ 0.012309568405151367,
131
+ 0.012329919815063477,
132
+ 0.012323840141296387,
133
+ 0.012304384231567383,
134
+ 0.012313599586486817,
135
+ 0.012315648078918457,
136
+ 0.012304384231567383,
137
+ 0.012404735565185548,
138
+ 0.01238425636291504,
139
+ 0.012363776206970215,
140
+ 0.012329983711242675,
141
+ 0.012305407524108887,
142
+ 0.012316672325134278,
143
+ 0.012319744110107422,
144
+ 0.012340224266052247,
145
+ 0.012331007957458496,
146
+ 0.012319744110107422,
147
+ 0.012320768356323243,
148
+ 0.012313599586486817,
149
+ 0.012300288200378418,
150
+ 0.01232588768005371,
151
+ 0.012328960418701173,
152
+ 0.012314623832702636,
153
+ 0.012412927627563476,
154
+ 0.01234124755859375,
155
+ 0.012324864387512208,
156
+ 0.012324864387512208,
157
+ 0.012307456016540527,
158
+ 0.012318719863891601,
159
+ 0.012371968269348145,
160
+ 0.012572671890258789,
161
+ 0.01276313591003418,
162
+ 0.012375040054321289,
163
+ 0.012312576293945313,
164
+ 0.012303359985351562,
165
+ 0.012323840141296387,
166
+ 0.012319744110107422,
167
+ 0.012324864387512208,
168
+ 0.012310527801513671,
169
+ 0.012318719863891601,
170
+ 0.01235865592956543,
171
+ 0.012314623832702636,
172
+ 0.012311552047729492,
173
+ 0.012332032203674317,
174
+ 0.012310527801513671,
175
+ 0.012314623832702636,
176
+ 0.012331007957458496,
177
+ 0.012302335739135742,
178
+ 0.012319744110107422,
179
+ 0.012309503555297852,
180
+ 0.012328960418701173,
181
+ 0.012307456016540527,
182
+ 0.012332032203674317,
183
+ 0.012323840141296387,
184
+ 0.012303359985351562,
185
+ 0.012320768356323243,
186
+ 0.012349439620971679,
187
+ 0.012313599586486817,
188
+ 0.012324864387512208,
189
+ 0.012313599586486817,
190
+ 0.012381183624267577,
191
+ 0.01293824005126953,
192
+ 0.012443648338317872,
193
+ 0.012347392082214356,
194
+ 0.012348416328430176,
195
+ 0.012353568077087403,
196
+ 0.012312543869018555,
197
+ 0.012311552047729492,
198
+ 0.012317695617675782,
199
+ 0.012311552047729492,
200
+ 0.012328960418701173,
201
+ 0.012313599586486817,
202
+ 0.012337151527404786,
203
+ 0.012315648078918457,
204
+ 0.012337151527404786,
205
+ 0.012305407524108887,
206
+ 0.012312576293945313,
207
+ 0.012354559898376465,
208
+ 0.012320768356323243,
209
+ 0.012312576293945313,
210
+ 0.012300288200378418,
211
+ 0.012309503555297852,
212
+ 0.012307456016540527,
213
+ 0.012329983711242675,
214
+ 0.012302335739135742,
215
+ 0.012317695617675782,
216
+ 0.012313599586486817,
217
+ 0.012319744110107422,
218
+ 0.012309503555297852,
219
+ 0.012314623832702636,
220
+ 0.012348416328430176,
221
+ 0.012318719863891601,
222
+ 0.012306431770324706,
223
+ 0.012329983711242675,
224
+ 0.012442624092102051,
225
+ 0.012329024314880371,
226
+ 0.012340160369873046,
227
+ 0.012340224266052247,
228
+ 0.012355584144592285,
229
+ 0.012420096397399903,
230
+ 0.012345343589782716,
231
+ 0.012353535652160644,
232
+ 0.012314623832702636,
233
+ 0.012383232116699219,
234
+ 0.012324864387512208,
235
+ 0.012310527801513671,
236
+ 0.01233510398864746,
237
+ 0.012316672325134278,
238
+ 0.012301312446594239,
239
+ 0.012301312446594239,
240
+ 0.012311552047729492,
241
+ 0.01234227180480957,
242
+ 0.012308480262756348,
243
+ 0.012314623832702636,
244
+ 0.012303359985351562,
245
+ 0.012322815895080566,
246
+ 0.012318719863891601,
247
+ 0.012309503555297852,
248
+ 0.012332032203674317,
249
+ 0.012323840141296387,
250
+ 0.012320768356323243,
251
+ 0.012317695617675782,
252
+ 0.012312576293945313,
253
+ 0.01232588768005371,
254
+ 0.012314623832702636,
255
+ 0.012310527801513671,
256
+ 0.012310527801513671,
257
+ 0.012344351768493652,
258
+ 0.012314592361450195,
259
+ 0.012300288200378418,
260
+ 0.012304384231567383,
261
+ 0.0123504638671875,
262
+ 0.012312576293945313,
263
+ 0.012307456016540527,
264
+ 0.012297216415405274,
265
+ 0.012303359985351562,
266
+ 0.012316672325134278,
267
+ 0.012314623832702636,
268
+ 0.012297216415405274,
269
+ 0.012308480262756348,
270
+ 0.012309503555297852,
271
+ 0.012310527801513671,
272
+ 0.012331007957458496,
273
+ 0.012299263954162597,
274
+ 0.012347392082214356,
275
+ 0.012317695617675782,
276
+ 0.012310527801513671,
277
+ 0.012316672325134278,
278
+ 0.012313599586486817,
279
+ 0.012315648078918457,
280
+ 0.012337151527404786,
281
+ 0.012314623832702636,
282
+ 0.012329983711242675,
283
+ 0.012307456016540527,
284
+ 0.012333056449890138,
285
+ 0.012312576293945313,
286
+ 0.012331007957458496,
287
+ 0.012324864387512208,
288
+ 0.012305407524108887,
289
+ 0.012320768356323243,
290
+ 0.012307456016540527,
291
+ 0.012307456016540527,
292
+ 0.01232588768005371,
293
+ 0.012309503555297852,
294
+ 0.012302335739135742,
295
+ 0.012313599586486817,
296
+ 0.012311552047729492,
297
+ 0.012308480262756348,
298
+ 0.012307456016540527,
299
+ 0.01234329605102539,
300
+ 0.012322815895080566,
301
+ 0.012319744110107422,
302
+ 0.012311552047729492,
303
+ 0.012312576293945313,
304
+ 0.012322815895080566,
305
+ 0.01232588768005371,
306
+ 0.012309503555297852,
307
+ 0.012313599586486817,
308
+ 0.012310527801513671,
309
+ 0.012310527801513671,
310
+ 0.012300288200378418,
311
+ 0.01233510398864746,
312
+ 0.012326911926269531,
313
+ 0.012311552047729492,
314
+ 0.012339200019836426,
315
+ 0.012311552047729492,
316
+ 0.012309503555297852,
317
+ 0.012313599586486817,
318
+ 0.012306431770324706,
319
+ 0.012321791648864745,
320
+ 0.012307456016540527,
321
+ 0.012311552047729492,
322
+ 0.012304384231567383,
323
+ 0.01233510398864746,
324
+ 0.01233407974243164,
325
+ 0.012336128234863282,
326
+ 0.012308480262756348,
327
+ 0.012309503555297852,
328
+ 0.012298239707946777,
329
+ 0.012313599586486817,
330
+ 0.012323840141296387,
331
+ 0.012311552047729492,
332
+ 0.012317695617675782,
333
+ 0.012306431770324706,
334
+ 0.012304384231567383,
335
+ 0.012311552047729492,
336
+ 0.012308480262756348,
337
+ 0.012347392082214356,
338
+ 0.012309503555297852,
339
+ 0.012309503555297852,
340
+ 0.012310527801513671,
341
+ 0.012333056449890138,
342
+ 0.012320768356323243,
343
+ 0.012328960418701173,
344
+ 0.012313599586486817,
345
+ 0.012312576293945313,
346
+ 0.012311552047729492,
347
+ 0.012315648078918457,
348
+ 0.012303359985351562,
349
+ 0.01234124755859375,
350
+ 0.012326911926269531,
351
+ 0.012306431770324706,
352
+ 0.012305407524108887,
353
+ 0.012312576293945313,
354
+ 0.012310527801513671,
355
+ 0.012313599586486817,
356
+ 0.012307456016540527,
357
+ 0.012316672325134278
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.11917929610797
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.014197376251220703,
7
+ "prefill.throughput.value": 493.0488476276124,
8
+ "decode.latency.mean": 1.5657416381835936,
9
+ "decode.throughput.value": 81.11172169332602,
10
+ "per_token.latency.mean": 0.012327540893254317,
11
+ "per_token.throughput.value": 81.11917929610797
12
+ }
13
+ }
2024-08-27/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmppit1gmri/commit=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3
88
+ - hydra.run.dir=_benchmark/tmppit1gmri/commit\=c6b23fda65f9ae74f9a1026b340241f65aebe1a3/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: c6b23fda65f9ae74f9a1026b340241f65aebe1a3
225
+ accelerate_version: 0.34.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.22.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-08-27/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.02075985622406006,
8
+ "prefill.throughput.value": 337.1892331261527,
9
+ "decode.latency.mean": 2.495531494140625,
10
+ "decode.throughput.value": 50.89096262587318,
11
+ "per_token.latency.mean": 0.01964858857853206,
12
+ "per_token.throughput.value": 50.89424087654797
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.02313372802734375,
21
+ "prefill.throughput.value": 302.5884972679758,
22
+ "decode.latency.mean": 2.6958218994140624,
23
+ "decode.throughput.value": 47.10993705763852,
24
+ "per_token.latency.mean": 0.02122594771047277,
25
+ "per_token.throughput.value": 47.11214847224962
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "c6b23fda65f9ae74f9a1026b340241f65aebe1a3",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.014197376251220703,
34
+ "prefill.throughput.value": 493.0488476276124,
35
+ "decode.latency.mean": 1.5657416381835936,
36
+ "decode.throughput.value": 81.11172169332602,
37
+ "per_token.latency.mean": 0.012327540893254317,
38
+ "per_token.throughput.value": 81.11917929610797
39
+ }
40
+ }
41
+ ]
2024-08-27/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "c6b23fda65f9ae74f9a1026b340241f65aebe1a3": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02075985622406006,
7
+ "prefill.throughput.value": 337.1892331261527,
8
+ "decode.latency.mean": 2.495531494140625,
9
+ "decode.throughput.value": 50.89096262587318,
10
+ "per_token.latency.mean": 0.01964858857853206,
11
+ "per_token.throughput.value": 50.89424087654797
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "c6b23fda65f9ae74f9a1026b340241f65aebe1a3": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.02313372802734375,
19
+ "prefill.throughput.value": 302.5884972679758,
20
+ "decode.latency.mean": 2.6958218994140624,
21
+ "decode.throughput.value": 47.10993705763852,
22
+ "per_token.latency.mean": 0.02122594771047277,
23
+ "per_token.throughput.value": 47.11214847224962
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "c6b23fda65f9ae74f9a1026b340241f65aebe1a3": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.014197376251220703,
31
+ "prefill.throughput.value": 493.0488476276124,
32
+ "decode.latency.mean": 1.5657416381835936,
33
+ "decode.throughput.value": 81.11172169332602,
34
+ "per_token.latency.mean": 0.012327540893254317,
35
+ "per_token.throughput.value": 81.11917929610797
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }