hf-transformers-bot commited on
Commit
a0abdd7
1 Parent(s): 7395013

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/multirun.yaml +234 -0
  31. 2024-08-27/summaries.json +41 -0
  32. 2024-08-27/summary.json +40 -0
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 9578c2597e2d88b6f0b304b5a05864fd613ddcc1
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
88
+ - hydra.run.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1319.886848,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.6247099609375,
120
+ "mean": 11.6247099609375,
121
+ "stdev": 0.0,
122
+ "p50": 11.6247099609375,
123
+ "p90": 11.6247099609375,
124
+ "p95": 11.6247099609375,
125
+ "p99": 11.6247099609375,
126
+ "values": [
127
+ 11.6247099609375
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1714.958336,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.041375551223754886,
147
+ "mean": 0.020687775611877443,
148
+ "stdev": 3.196811676025278e-05,
149
+ "p50": 0.020687775611877443,
150
+ "p90": 0.020713350105285643,
151
+ "p95": 0.02071654691696167,
152
+ "p99": 0.02071910436630249,
153
+ "values": [
154
+ 0.02065580749511719,
155
+ 0.020719743728637694
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 338.3640721615861
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1769.074688,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.00892724609375,
178
+ "mean": 2.504463623046875,
179
+ "stdev": 0.01867016601562499,
180
+ "p50": 2.504463623046875,
181
+ "p90": 2.519399755859375,
182
+ "p95": 2.5212667724609372,
183
+ "p99": 2.5227603857421874,
184
+ "values": [
185
+ 2.5231337890625,
186
+ 2.48579345703125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 50.7094608327729
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.008598026275633,
202
+ "mean": 0.019718889867226908,
203
+ "stdev": 0.0004935351444897931,
204
+ "p50": 0.019902976036071777,
205
+ "p90": 0.020130918502807616,
206
+ "p95": 0.020281190395355225,
207
+ "p99": 0.02115054573059082,
208
+ "values": [
209
+ 0.01946828842163086,
210
+ 0.019903488159179687,
211
+ 0.019912704467773438,
212
+ 0.019943424224853516,
213
+ 0.019929088592529298,
214
+ 0.019900415420532228,
215
+ 0.019917823791503905,
216
+ 0.01988403129577637,
217
+ 0.0198604793548584,
218
+ 0.021011455535888672,
219
+ 0.020000768661499024,
220
+ 0.019810304641723633,
221
+ 0.01983180809020996,
222
+ 0.019939327239990236,
223
+ 0.019894271850585937,
224
+ 0.02008166313171387,
225
+ 0.019929088592529298,
226
+ 0.019902463912963866,
227
+ 0.019940351486206053,
228
+ 0.019895296096801757,
229
+ 0.020239360809326173,
230
+ 0.02001817512512207,
231
+ 0.019878911972045898,
232
+ 0.020015104293823242,
233
+ 0.019655679702758787,
234
+ 0.019917823791503905,
235
+ 0.019949567794799804,
236
+ 0.020158464431762696,
237
+ 0.019982336044311523,
238
+ 0.019992576599121094,
239
+ 0.019933183670043944,
240
+ 0.0199332160949707,
241
+ 0.019948511123657225,
242
+ 0.019928064346313477,
243
+ 0.01991372871398926,
244
+ 0.019843072891235353,
245
+ 0.019878911972045898,
246
+ 0.01985638427734375,
247
+ 0.019943424224853516,
248
+ 0.019910655975341796,
249
+ 0.019903488159179687,
250
+ 0.01988096046447754,
251
+ 0.02005504035949707,
252
+ 0.019952640533447266,
253
+ 0.01985740852355957,
254
+ 0.019899391174316407,
255
+ 0.019990528106689453,
256
+ 0.01985024070739746,
257
+ 0.02027724838256836,
258
+ 0.019951616287231445,
259
+ 0.020001792907714845,
260
+ 0.019412992477416992,
261
+ 0.021355520248413085,
262
+ 0.020316160202026368,
263
+ 0.02003558349609375,
264
+ 0.01990553665161133,
265
+ 0.0198973445892334,
266
+ 0.01999667167663574,
267
+ 0.019946495056152345,
268
+ 0.02005606460571289,
269
+ 0.019997695922851562,
270
+ 0.01983078384399414,
271
+ 0.02000383949279785,
272
+ 0.02001203155517578,
273
+ 0.02011238479614258,
274
+ 0.020107263565063475,
275
+ 0.020017152786254884,
276
+ 0.020067327499389647,
277
+ 0.02004684829711914,
278
+ 0.020051967620849608,
279
+ 0.020017152786254884,
280
+ 0.020025344848632814,
281
+ 0.02003558349609375,
282
+ 0.020106239318847655,
283
+ 0.020076543807983398,
284
+ 0.02002739143371582,
285
+ 0.02004172706604004,
286
+ 0.02010419273376465,
287
+ 0.020076543807983398,
288
+ 0.020059135437011717,
289
+ 0.02011136054992676,
290
+ 0.020091903686523437,
291
+ 0.0200130558013916,
292
+ 0.020023296356201172,
293
+ 0.02002841567993164,
294
+ 0.020025344848632814,
295
+ 0.02005504035949707,
296
+ 0.02004991912841797,
297
+ 0.019994623184204103,
298
+ 0.019919872283935547,
299
+ 0.019747840881347657,
300
+ 0.01987174415588379,
301
+ 0.019992576599121094,
302
+ 0.02000383949279785,
303
+ 0.02002841567993164,
304
+ 0.020025344848632814,
305
+ 0.02002739143371582,
306
+ 0.02003455924987793,
307
+ 0.02007756805419922,
308
+ 0.020180992126464844,
309
+ 0.020048896789550782,
310
+ 0.0194201602935791,
311
+ 0.019305471420288087,
312
+ 0.019350528717041016,
313
+ 0.019315711975097655,
314
+ 0.019284992218017577,
315
+ 0.01926553535461426,
316
+ 0.019311616897583008,
317
+ 0.019318784713745117,
318
+ 0.019296255111694336,
319
+ 0.01923993682861328,
320
+ 0.01921023941040039,
321
+ 0.019088384628295898,
322
+ 0.019005439758300782,
323
+ 0.0191016960144043,
324
+ 0.01908531188964844,
325
+ 0.019152896881103516,
326
+ 0.019357696533203125,
327
+ 0.019092479705810548,
328
+ 0.01906073570251465,
329
+ 0.019087360382080077,
330
+ 0.020152320861816408,
331
+ 0.019787776947021486,
332
+ 0.019668991088867188,
333
+ 0.019771392822265626,
334
+ 0.019773439407348634,
335
+ 0.019896320343017578,
336
+ 0.01942425537109375,
337
+ 0.019555328369140625,
338
+ 0.019851264953613282,
339
+ 0.01984921646118164,
340
+ 0.019587072372436523,
341
+ 0.019708927154541016,
342
+ 0.01985740852355957,
343
+ 0.019817472457885742,
344
+ 0.019574783325195313,
345
+ 0.018920448303222655,
346
+ 0.01889587211608887,
347
+ 0.018911231994628908,
348
+ 0.018966527938842775,
349
+ 0.018961408615112304,
350
+ 0.018926591873168946,
351
+ 0.018950143814086915,
352
+ 0.018905088424682616,
353
+ 0.018953216552734374,
354
+ 0.01902592086791992,
355
+ 0.018976768493652343,
356
+ 0.01887129592895508,
357
+ 0.019133440017700197,
358
+ 0.018968576431274413,
359
+ 0.018979839324951172,
360
+ 0.018924543380737305,
361
+ 0.01897881507873535,
362
+ 0.01898700714111328,
363
+ 0.01908531188964844,
364
+ 0.019112960815429687,
365
+ 0.018961408615112304,
366
+ 0.018962432861328125,
367
+ 0.018997247695922852,
368
+ 0.018935808181762694,
369
+ 0.018917375564575196,
370
+ 0.018898944854736328,
371
+ 0.01906892776489258,
372
+ 0.019134464263916014,
373
+ 0.019145727157592773,
374
+ 0.019058687210083008,
375
+ 0.018951168060302736,
376
+ 0.019153919219970703,
377
+ 0.019117055892944337,
378
+ 0.019191808700561523,
379
+ 0.019164159774780275,
380
+ 0.01903206443786621,
381
+ 0.01902387237548828,
382
+ 0.018985984802246093,
383
+ 0.018924543380737305,
384
+ 0.019171327590942384,
385
+ 0.019144704818725586,
386
+ 0.01922150421142578,
387
+ 0.019136512756347656,
388
+ 0.019269632339477538,
389
+ 0.019142656326293944,
390
+ 0.019131391525268555,
391
+ 0.019183616638183593,
392
+ 0.01920204734802246,
393
+ 0.019283967971801756,
394
+ 0.019269632339477538,
395
+ 0.01926144027709961,
396
+ 0.019292160034179686,
397
+ 0.01924608039855957,
398
+ 0.01926348876953125,
399
+ 0.019277824401855468,
400
+ 0.019289087295532227,
401
+ 0.01927987289428711,
402
+ 0.019317760467529296,
403
+ 0.019338239669799806,
404
+ 0.01927884864807129,
405
+ 0.019283967971801756,
406
+ 0.019363840103149413,
407
+ 0.02165657615661621,
408
+ 0.02030899238586426,
409
+ 0.020183040618896485,
410
+ 0.01943961524963379,
411
+ 0.02012876892089844,
412
+ 0.020880384445190428,
413
+ 0.02012876892089844,
414
+ 0.02007961654663086,
415
+ 0.02003763198852539,
416
+ 0.020131839752197265,
417
+ 0.020060159683227538,
418
+ 0.020050943374633787,
419
+ 0.02002739143371582,
420
+ 0.020059135437011717,
421
+ 0.0200447998046875,
422
+ 0.020124671936035156,
423
+ 0.020140031814575195,
424
+ 0.019991552352905274,
425
+ 0.019320831298828126,
426
+ 0.019289087295532227,
427
+ 0.01924300765991211,
428
+ 0.019949567794799804,
429
+ 0.02001408004760742,
430
+ 0.02001408004760742,
431
+ 0.019983360290527344,
432
+ 0.020530176162719727,
433
+ 0.020137983322143553,
434
+ 0.02002022361755371,
435
+ 0.0200898551940918,
436
+ 0.0200263671875,
437
+ 0.02002022361755371,
438
+ 0.02007244873046875,
439
+ 0.02023526382446289,
440
+ 0.020131839752197265,
441
+ 0.02007756805419922,
442
+ 0.02001203155517578,
443
+ 0.02007040023803711,
444
+ 0.02005504035949707,
445
+ 0.02004172706604004,
446
+ 0.02004172706604004,
447
+ 0.02063155174255371,
448
+ 0.021307392120361326,
449
+ 0.020254720687866212,
450
+ 0.02046668815612793,
451
+ 0.020265983581542968,
452
+ 0.02003865623474121,
453
+ 0.020578304290771485,
454
+ 0.020495359420776366,
455
+ 0.020288511276245116,
456
+ 0.020114431381225584,
457
+ 0.02008678436279297,
458
+ 0.020107263565063475,
459
+ 0.01943756866455078,
460
+ 0.019289087295532227,
461
+ 0.01926553535461426,
462
+ 0.019216384887695313
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 50.71279401291322
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1319.886848,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.6247099609375,
15
+ "mean": 11.6247099609375,
16
+ "stdev": 0.0,
17
+ "p50": 11.6247099609375,
18
+ "p90": 11.6247099609375,
19
+ "p95": 11.6247099609375,
20
+ "p99": 11.6247099609375,
21
+ "values": [
22
+ 11.6247099609375
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1714.958336,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.041375551223754886,
42
+ "mean": 0.020687775611877443,
43
+ "stdev": 3.196811676025278e-05,
44
+ "p50": 0.020687775611877443,
45
+ "p90": 0.020713350105285643,
46
+ "p95": 0.02071654691696167,
47
+ "p99": 0.02071910436630249,
48
+ "values": [
49
+ 0.02065580749511719,
50
+ 0.020719743728637694
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 338.3640721615861
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1769.074688,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.00892724609375,
73
+ "mean": 2.504463623046875,
74
+ "stdev": 0.01867016601562499,
75
+ "p50": 2.504463623046875,
76
+ "p90": 2.519399755859375,
77
+ "p95": 2.5212667724609372,
78
+ "p99": 2.5227603857421874,
79
+ "values": [
80
+ 2.5231337890625,
81
+ 2.48579345703125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 50.7094608327729
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.008598026275633,
97
+ "mean": 0.019718889867226908,
98
+ "stdev": 0.0004935351444897931,
99
+ "p50": 0.019902976036071777,
100
+ "p90": 0.020130918502807616,
101
+ "p95": 0.020281190395355225,
102
+ "p99": 0.02115054573059082,
103
+ "values": [
104
+ 0.01946828842163086,
105
+ 0.019903488159179687,
106
+ 0.019912704467773438,
107
+ 0.019943424224853516,
108
+ 0.019929088592529298,
109
+ 0.019900415420532228,
110
+ 0.019917823791503905,
111
+ 0.01988403129577637,
112
+ 0.0198604793548584,
113
+ 0.021011455535888672,
114
+ 0.020000768661499024,
115
+ 0.019810304641723633,
116
+ 0.01983180809020996,
117
+ 0.019939327239990236,
118
+ 0.019894271850585937,
119
+ 0.02008166313171387,
120
+ 0.019929088592529298,
121
+ 0.019902463912963866,
122
+ 0.019940351486206053,
123
+ 0.019895296096801757,
124
+ 0.020239360809326173,
125
+ 0.02001817512512207,
126
+ 0.019878911972045898,
127
+ 0.020015104293823242,
128
+ 0.019655679702758787,
129
+ 0.019917823791503905,
130
+ 0.019949567794799804,
131
+ 0.020158464431762696,
132
+ 0.019982336044311523,
133
+ 0.019992576599121094,
134
+ 0.019933183670043944,
135
+ 0.0199332160949707,
136
+ 0.019948511123657225,
137
+ 0.019928064346313477,
138
+ 0.01991372871398926,
139
+ 0.019843072891235353,
140
+ 0.019878911972045898,
141
+ 0.01985638427734375,
142
+ 0.019943424224853516,
143
+ 0.019910655975341796,
144
+ 0.019903488159179687,
145
+ 0.01988096046447754,
146
+ 0.02005504035949707,
147
+ 0.019952640533447266,
148
+ 0.01985740852355957,
149
+ 0.019899391174316407,
150
+ 0.019990528106689453,
151
+ 0.01985024070739746,
152
+ 0.02027724838256836,
153
+ 0.019951616287231445,
154
+ 0.020001792907714845,
155
+ 0.019412992477416992,
156
+ 0.021355520248413085,
157
+ 0.020316160202026368,
158
+ 0.02003558349609375,
159
+ 0.01990553665161133,
160
+ 0.0198973445892334,
161
+ 0.01999667167663574,
162
+ 0.019946495056152345,
163
+ 0.02005606460571289,
164
+ 0.019997695922851562,
165
+ 0.01983078384399414,
166
+ 0.02000383949279785,
167
+ 0.02001203155517578,
168
+ 0.02011238479614258,
169
+ 0.020107263565063475,
170
+ 0.020017152786254884,
171
+ 0.020067327499389647,
172
+ 0.02004684829711914,
173
+ 0.020051967620849608,
174
+ 0.020017152786254884,
175
+ 0.020025344848632814,
176
+ 0.02003558349609375,
177
+ 0.020106239318847655,
178
+ 0.020076543807983398,
179
+ 0.02002739143371582,
180
+ 0.02004172706604004,
181
+ 0.02010419273376465,
182
+ 0.020076543807983398,
183
+ 0.020059135437011717,
184
+ 0.02011136054992676,
185
+ 0.020091903686523437,
186
+ 0.0200130558013916,
187
+ 0.020023296356201172,
188
+ 0.02002841567993164,
189
+ 0.020025344848632814,
190
+ 0.02005504035949707,
191
+ 0.02004991912841797,
192
+ 0.019994623184204103,
193
+ 0.019919872283935547,
194
+ 0.019747840881347657,
195
+ 0.01987174415588379,
196
+ 0.019992576599121094,
197
+ 0.02000383949279785,
198
+ 0.02002841567993164,
199
+ 0.020025344848632814,
200
+ 0.02002739143371582,
201
+ 0.02003455924987793,
202
+ 0.02007756805419922,
203
+ 0.020180992126464844,
204
+ 0.020048896789550782,
205
+ 0.0194201602935791,
206
+ 0.019305471420288087,
207
+ 0.019350528717041016,
208
+ 0.019315711975097655,
209
+ 0.019284992218017577,
210
+ 0.01926553535461426,
211
+ 0.019311616897583008,
212
+ 0.019318784713745117,
213
+ 0.019296255111694336,
214
+ 0.01923993682861328,
215
+ 0.01921023941040039,
216
+ 0.019088384628295898,
217
+ 0.019005439758300782,
218
+ 0.0191016960144043,
219
+ 0.01908531188964844,
220
+ 0.019152896881103516,
221
+ 0.019357696533203125,
222
+ 0.019092479705810548,
223
+ 0.01906073570251465,
224
+ 0.019087360382080077,
225
+ 0.020152320861816408,
226
+ 0.019787776947021486,
227
+ 0.019668991088867188,
228
+ 0.019771392822265626,
229
+ 0.019773439407348634,
230
+ 0.019896320343017578,
231
+ 0.01942425537109375,
232
+ 0.019555328369140625,
233
+ 0.019851264953613282,
234
+ 0.01984921646118164,
235
+ 0.019587072372436523,
236
+ 0.019708927154541016,
237
+ 0.01985740852355957,
238
+ 0.019817472457885742,
239
+ 0.019574783325195313,
240
+ 0.018920448303222655,
241
+ 0.01889587211608887,
242
+ 0.018911231994628908,
243
+ 0.018966527938842775,
244
+ 0.018961408615112304,
245
+ 0.018926591873168946,
246
+ 0.018950143814086915,
247
+ 0.018905088424682616,
248
+ 0.018953216552734374,
249
+ 0.01902592086791992,
250
+ 0.018976768493652343,
251
+ 0.01887129592895508,
252
+ 0.019133440017700197,
253
+ 0.018968576431274413,
254
+ 0.018979839324951172,
255
+ 0.018924543380737305,
256
+ 0.01897881507873535,
257
+ 0.01898700714111328,
258
+ 0.01908531188964844,
259
+ 0.019112960815429687,
260
+ 0.018961408615112304,
261
+ 0.018962432861328125,
262
+ 0.018997247695922852,
263
+ 0.018935808181762694,
264
+ 0.018917375564575196,
265
+ 0.018898944854736328,
266
+ 0.01906892776489258,
267
+ 0.019134464263916014,
268
+ 0.019145727157592773,
269
+ 0.019058687210083008,
270
+ 0.018951168060302736,
271
+ 0.019153919219970703,
272
+ 0.019117055892944337,
273
+ 0.019191808700561523,
274
+ 0.019164159774780275,
275
+ 0.01903206443786621,
276
+ 0.01902387237548828,
277
+ 0.018985984802246093,
278
+ 0.018924543380737305,
279
+ 0.019171327590942384,
280
+ 0.019144704818725586,
281
+ 0.01922150421142578,
282
+ 0.019136512756347656,
283
+ 0.019269632339477538,
284
+ 0.019142656326293944,
285
+ 0.019131391525268555,
286
+ 0.019183616638183593,
287
+ 0.01920204734802246,
288
+ 0.019283967971801756,
289
+ 0.019269632339477538,
290
+ 0.01926144027709961,
291
+ 0.019292160034179686,
292
+ 0.01924608039855957,
293
+ 0.01926348876953125,
294
+ 0.019277824401855468,
295
+ 0.019289087295532227,
296
+ 0.01927987289428711,
297
+ 0.019317760467529296,
298
+ 0.019338239669799806,
299
+ 0.01927884864807129,
300
+ 0.019283967971801756,
301
+ 0.019363840103149413,
302
+ 0.02165657615661621,
303
+ 0.02030899238586426,
304
+ 0.020183040618896485,
305
+ 0.01943961524963379,
306
+ 0.02012876892089844,
307
+ 0.020880384445190428,
308
+ 0.02012876892089844,
309
+ 0.02007961654663086,
310
+ 0.02003763198852539,
311
+ 0.020131839752197265,
312
+ 0.020060159683227538,
313
+ 0.020050943374633787,
314
+ 0.02002739143371582,
315
+ 0.020059135437011717,
316
+ 0.0200447998046875,
317
+ 0.020124671936035156,
318
+ 0.020140031814575195,
319
+ 0.019991552352905274,
320
+ 0.019320831298828126,
321
+ 0.019289087295532227,
322
+ 0.01924300765991211,
323
+ 0.019949567794799804,
324
+ 0.02001408004760742,
325
+ 0.02001408004760742,
326
+ 0.019983360290527344,
327
+ 0.020530176162719727,
328
+ 0.020137983322143553,
329
+ 0.02002022361755371,
330
+ 0.0200898551940918,
331
+ 0.0200263671875,
332
+ 0.02002022361755371,
333
+ 0.02007244873046875,
334
+ 0.02023526382446289,
335
+ 0.020131839752197265,
336
+ 0.02007756805419922,
337
+ 0.02001203155517578,
338
+ 0.02007040023803711,
339
+ 0.02005504035949707,
340
+ 0.02004172706604004,
341
+ 0.02004172706604004,
342
+ 0.02063155174255371,
343
+ 0.021307392120361326,
344
+ 0.020254720687866212,
345
+ 0.02046668815612793,
346
+ 0.020265983581542968,
347
+ 0.02003865623474121,
348
+ 0.020578304290771485,
349
+ 0.020495359420776366,
350
+ 0.020288511276245116,
351
+ 0.020114431381225584,
352
+ 0.02008678436279297,
353
+ 0.020107263565063475,
354
+ 0.01943756866455078,
355
+ 0.019289087295532227,
356
+ 0.01926553535461426,
357
+ 0.019216384887695313
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 50.71279401291322
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.020687775611877443,
7
+ "prefill.throughput.value": 338.3640721615861,
8
+ "decode.latency.mean": 2.504463623046875,
9
+ "decode.throughput.value": 50.7094608327729,
10
+ "per_token.latency.mean": 0.019718889867226908,
11
+ "per_token.throughput.value": 50.71279401291322
12
+ }
13
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 9578c2597e2d88b6f0b304b5a05864fd613ddcc1
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
88
+ - hydra.run.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-08-27 02:38:31,045][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-08-27 02:38:33,437][process][ERROR] - + Received traceback from isolated process
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 9578c2597e2d88b6f0b304b5a05864fd613ddcc1
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
88
+ - hydra.run.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1318.494208,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.6820244140625,
120
+ "mean": 11.6820244140625,
121
+ "stdev": 0.0,
122
+ "p50": 11.6820244140625,
123
+ "p90": 11.6820244140625,
124
+ "p95": 11.6820244140625,
125
+ "p99": 11.6820244140625,
126
+ "values": [
127
+ 11.6820244140625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1744.601088,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.04474979209899902,
147
+ "mean": 0.02237489604949951,
148
+ "stdev": 2.859210968017556e-05,
149
+ "p50": 0.02237489604949951,
150
+ "p90": 0.02239776973724365,
151
+ "p95": 0.022400628948211668,
152
+ "p99": 0.02240291631698608,
153
+ "values": [
154
+ 0.022403488159179686,
155
+ 0.022346303939819335
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 312.8506154627064
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1769.488384,
169
+ "max_global_vram": 6793.199616,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.278910156249999,
178
+ "mean": 2.6394550781249997,
179
+ "stdev": 0.011486083984374895,
180
+ "p50": 2.6394550781249997,
181
+ "p90": 2.6486439453124997,
182
+ "p95": 2.6497925537109372,
183
+ "p99": 2.650711440429687,
184
+ "values": [
185
+ 2.627968994140625,
186
+ 2.650941162109375
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 48.1159922184459
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.278648300170897,
202
+ "mean": 0.02078207992193267,
203
+ "stdev": 0.00035438229734022865,
204
+ "p50": 0.0205819034576416,
205
+ "p90": 0.02134026279449463,
206
+ "p95": 0.02140047378540039,
207
+ "p99": 0.021629153480529784,
208
+ "values": [
209
+ 0.020549631118774413,
210
+ 0.020557823181152343,
211
+ 0.021441535949707033,
212
+ 0.02145792007446289,
213
+ 0.021345279693603517,
214
+ 0.021355520248413085,
215
+ 0.020626432418823244,
216
+ 0.020567039489746093,
217
+ 0.020549631118774413,
218
+ 0.02058137512207031,
219
+ 0.0206059513092041,
220
+ 0.02060697555541992,
221
+ 0.020604000091552735,
222
+ 0.02055465507507324,
223
+ 0.020547584533691408,
224
+ 0.020547584533691408,
225
+ 0.020550655364990233,
226
+ 0.0206376953125,
227
+ 0.020603904724121092,
228
+ 0.020562944412231447,
229
+ 0.020541439056396483,
230
+ 0.020527103424072265,
231
+ 0.020525056838989256,
232
+ 0.020526079177856444,
233
+ 0.020538368225097657,
234
+ 0.020477951049804686,
235
+ 0.020361215591430663,
236
+ 0.020335615158081053,
237
+ 0.020574207305908202,
238
+ 0.020557823181152343,
239
+ 0.020566015243530272,
240
+ 0.020518911361694335,
241
+ 0.020575231552124023,
242
+ 0.020564992904663085,
243
+ 0.020570112228393556,
244
+ 0.020540416717529295,
245
+ 0.020599807739257812,
246
+ 0.020558847427368163,
247
+ 0.020596736907958983,
248
+ 0.020546560287475587,
249
+ 0.020586496353149415,
250
+ 0.020519935607910156,
251
+ 0.020537343978881836,
252
+ 0.02068992042541504,
253
+ 0.020545536041259766,
254
+ 0.020533248901367186,
255
+ 0.020550655364990233,
256
+ 0.020559871673583984,
257
+ 0.02061311912536621,
258
+ 0.020561920166015626,
259
+ 0.020609024047851563,
260
+ 0.020519935607910156,
261
+ 0.020726783752441406,
262
+ 0.020633600234985353,
263
+ 0.02062233543395996,
264
+ 0.020571136474609376,
265
+ 0.020534271240234374,
266
+ 0.02049228858947754,
267
+ 0.020553728103637696,
268
+ 0.020551679611206054,
269
+ 0.020574207305908202,
270
+ 0.02059775924682617,
271
+ 0.020531200408935548,
272
+ 0.020531200408935548,
273
+ 0.020534271240234374,
274
+ 0.020562944412231447,
275
+ 0.020557823181152343,
276
+ 0.020539392471313478,
277
+ 0.020555776596069338,
278
+ 0.02048102378845215,
279
+ 0.020488191604614257,
280
+ 0.020497407913208008,
281
+ 0.021561344146728514,
282
+ 0.021311487197875977,
283
+ 0.02127769660949707,
284
+ 0.020533248901367186,
285
+ 0.020552703857421875,
286
+ 0.020541439056396483,
287
+ 0.020577280044555665,
288
+ 0.020545568466186524,
289
+ 0.02055369567871094,
290
+ 0.020542463302612304,
291
+ 0.020537343978881836,
292
+ 0.020560895919799805,
293
+ 0.0206059513092041,
294
+ 0.02059878349304199,
295
+ 0.02060492706298828,
296
+ 0.020569087982177735,
297
+ 0.020550655364990233,
298
+ 0.020563968658447264,
299
+ 0.020559871673583984,
300
+ 0.020784128189086915,
301
+ 0.02061311912536621,
302
+ 0.020564992904663085,
303
+ 0.02052403259277344,
304
+ 0.020554752349853517,
305
+ 0.020528127670288086,
306
+ 0.020519935607910156,
307
+ 0.020600831985473633,
308
+ 0.02061311912536621,
309
+ 0.020590591430664062,
310
+ 0.02049228858947754,
311
+ 0.020571136474609376,
312
+ 0.020538368225097657,
313
+ 0.020547584533691408,
314
+ 0.020529151916503906,
315
+ 0.020538368225097657,
316
+ 0.020559871673583984,
317
+ 0.020519935607910156,
318
+ 0.020517887115478514,
319
+ 0.020552703857421875,
320
+ 0.02130227279663086,
321
+ 0.02131865692138672,
322
+ 0.021348352432250976,
323
+ 0.021334016799926758,
324
+ 0.021391359329223633,
325
+ 0.02128998374938965,
326
+ 0.02126643180847168,
327
+ 0.021327871322631836,
328
+ 0.021353471755981446,
329
+ 0.02128998374938965,
330
+ 0.021285888671875,
331
+ 0.02122137641906738,
332
+ 0.021271551132202148,
333
+ 0.02126540756225586,
334
+ 0.02125619125366211,
335
+ 0.020486143112182616,
336
+ 0.020510719299316405,
337
+ 0.020530176162719727,
338
+ 0.020516864776611327,
339
+ 0.020552703857421875,
340
+ 0.020572160720825194,
341
+ 0.020644863128662108,
342
+ 0.020559871673583984,
343
+ 0.020553728103637696,
344
+ 0.020576255798339844,
345
+ 0.020544511795043945,
346
+ 0.020518911361694335,
347
+ 0.02064896011352539,
348
+ 0.020557823181152343,
349
+ 0.020556800842285155,
350
+ 0.020519935607910156,
351
+ 0.020511743545532226,
352
+ 0.020496383666992187,
353
+ 0.020549631118774413,
354
+ 0.020503551483154296,
355
+ 0.020513792037963868,
356
+ 0.020548608779907225,
357
+ 0.020728832244873048,
358
+ 0.020544511795043945,
359
+ 0.02058243179321289,
360
+ 0.02050966453552246,
361
+ 0.020489215850830078,
362
+ 0.02049126434326172,
363
+ 0.020551679611206054,
364
+ 0.020548608779907225,
365
+ 0.020509695053100584,
366
+ 0.020566015243530272,
367
+ 0.020542463302612304,
368
+ 0.020544511795043945,
369
+ 0.020590591430664062,
370
+ 0.020551679611206054,
371
+ 0.02064076805114746,
372
+ 0.020540416717529295,
373
+ 0.020626432418823244,
374
+ 0.020488191604614257,
375
+ 0.02049126434326172,
376
+ 0.02107904052734375,
377
+ 0.021363712310791014,
378
+ 0.021246976852416992,
379
+ 0.02126848030090332,
380
+ 0.02126643180847168,
381
+ 0.02125619125366211,
382
+ 0.02125004768371582,
383
+ 0.02126335906982422,
384
+ 0.021324800491333007,
385
+ 0.02124595260620117,
386
+ 0.02127462387084961,
387
+ 0.02126848030090332,
388
+ 0.0212992000579834,
389
+ 0.021230592727661132,
390
+ 0.02126540756225586,
391
+ 0.020592639923095703,
392
+ 0.020580352783203124,
393
+ 0.02062335968017578,
394
+ 0.02061516761779785,
395
+ 0.020649984359741212,
396
+ 0.021497856140136717,
397
+ 0.021341184616088867,
398
+ 0.021324800491333007,
399
+ 0.021350400924682617,
400
+ 0.021413888931274414,
401
+ 0.021346303939819337,
402
+ 0.021335039138793945,
403
+ 0.02128895950317383,
404
+ 0.02131455993652344,
405
+ 0.02062950325012207,
406
+ 0.020576255798339844,
407
+ 0.020571136474609376,
408
+ 0.021362688064575194,
409
+ 0.021320703506469727,
410
+ 0.02063564872741699,
411
+ 0.020571136474609376,
412
+ 0.020585472106933594,
413
+ 0.020543487548828124,
414
+ 0.020616191864013672,
415
+ 0.02129715156555176,
416
+ 0.020577280044555665,
417
+ 0.020585472106933594,
418
+ 0.020578304290771485,
419
+ 0.02066022491455078,
420
+ 0.02066227149963379,
421
+ 0.020743167877197266,
422
+ 0.020583423614501953,
423
+ 0.020555776596069338,
424
+ 0.020649984359741212,
425
+ 0.020542463302612304,
426
+ 0.020611072540283205,
427
+ 0.020601856231689454,
428
+ 0.020783103942871094,
429
+ 0.020562944412231447,
430
+ 0.02063155174255371,
431
+ 0.020585472106933594,
432
+ 0.02061311912536621,
433
+ 0.020593664169311524,
434
+ 0.020580352783203124,
435
+ 0.020892671585083008,
436
+ 0.021307392120361326,
437
+ 0.021719039916992186,
438
+ 0.02067046356201172,
439
+ 0.02061516761779785,
440
+ 0.020580352783203124,
441
+ 0.02058956718444824,
442
+ 0.020669439315795898,
443
+ 0.020574207305908202,
444
+ 0.02169753646850586,
445
+ 0.021444608688354492,
446
+ 0.021353471755981446,
447
+ 0.021292032241821288,
448
+ 0.021568511962890623,
449
+ 0.021338111877441408,
450
+ 0.02129817581176758,
451
+ 0.021315584182739256,
452
+ 0.0212807674407959,
453
+ 0.021768192291259765,
454
+ 0.021414911270141602,
455
+ 0.021296127319335938,
456
+ 0.0213309440612793,
457
+ 0.02128998374938965,
458
+ 0.021409791946411134,
459
+ 0.021395456314086913,
460
+ 0.021357568740844726,
461
+ 0.021313535690307618,
462
+ 0.021410816192626952
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 48.118379091817246
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1318.494208,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.6820244140625,
15
+ "mean": 11.6820244140625,
16
+ "stdev": 0.0,
17
+ "p50": 11.6820244140625,
18
+ "p90": 11.6820244140625,
19
+ "p95": 11.6820244140625,
20
+ "p99": 11.6820244140625,
21
+ "values": [
22
+ 11.6820244140625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1744.601088,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.04474979209899902,
42
+ "mean": 0.02237489604949951,
43
+ "stdev": 2.859210968017556e-05,
44
+ "p50": 0.02237489604949951,
45
+ "p90": 0.02239776973724365,
46
+ "p95": 0.022400628948211668,
47
+ "p99": 0.02240291631698608,
48
+ "values": [
49
+ 0.022403488159179686,
50
+ 0.022346303939819335
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 312.8506154627064
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1769.488384,
64
+ "max_global_vram": 6793.199616,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.278910156249999,
73
+ "mean": 2.6394550781249997,
74
+ "stdev": 0.011486083984374895,
75
+ "p50": 2.6394550781249997,
76
+ "p90": 2.6486439453124997,
77
+ "p95": 2.6497925537109372,
78
+ "p99": 2.650711440429687,
79
+ "values": [
80
+ 2.627968994140625,
81
+ 2.650941162109375
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 48.1159922184459
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.278648300170897,
97
+ "mean": 0.02078207992193267,
98
+ "stdev": 0.00035438229734022865,
99
+ "p50": 0.0205819034576416,
100
+ "p90": 0.02134026279449463,
101
+ "p95": 0.02140047378540039,
102
+ "p99": 0.021629153480529784,
103
+ "values": [
104
+ 0.020549631118774413,
105
+ 0.020557823181152343,
106
+ 0.021441535949707033,
107
+ 0.02145792007446289,
108
+ 0.021345279693603517,
109
+ 0.021355520248413085,
110
+ 0.020626432418823244,
111
+ 0.020567039489746093,
112
+ 0.020549631118774413,
113
+ 0.02058137512207031,
114
+ 0.0206059513092041,
115
+ 0.02060697555541992,
116
+ 0.020604000091552735,
117
+ 0.02055465507507324,
118
+ 0.020547584533691408,
119
+ 0.020547584533691408,
120
+ 0.020550655364990233,
121
+ 0.0206376953125,
122
+ 0.020603904724121092,
123
+ 0.020562944412231447,
124
+ 0.020541439056396483,
125
+ 0.020527103424072265,
126
+ 0.020525056838989256,
127
+ 0.020526079177856444,
128
+ 0.020538368225097657,
129
+ 0.020477951049804686,
130
+ 0.020361215591430663,
131
+ 0.020335615158081053,
132
+ 0.020574207305908202,
133
+ 0.020557823181152343,
134
+ 0.020566015243530272,
135
+ 0.020518911361694335,
136
+ 0.020575231552124023,
137
+ 0.020564992904663085,
138
+ 0.020570112228393556,
139
+ 0.020540416717529295,
140
+ 0.020599807739257812,
141
+ 0.020558847427368163,
142
+ 0.020596736907958983,
143
+ 0.020546560287475587,
144
+ 0.020586496353149415,
145
+ 0.020519935607910156,
146
+ 0.020537343978881836,
147
+ 0.02068992042541504,
148
+ 0.020545536041259766,
149
+ 0.020533248901367186,
150
+ 0.020550655364990233,
151
+ 0.020559871673583984,
152
+ 0.02061311912536621,
153
+ 0.020561920166015626,
154
+ 0.020609024047851563,
155
+ 0.020519935607910156,
156
+ 0.020726783752441406,
157
+ 0.020633600234985353,
158
+ 0.02062233543395996,
159
+ 0.020571136474609376,
160
+ 0.020534271240234374,
161
+ 0.02049228858947754,
162
+ 0.020553728103637696,
163
+ 0.020551679611206054,
164
+ 0.020574207305908202,
165
+ 0.02059775924682617,
166
+ 0.020531200408935548,
167
+ 0.020531200408935548,
168
+ 0.020534271240234374,
169
+ 0.020562944412231447,
170
+ 0.020557823181152343,
171
+ 0.020539392471313478,
172
+ 0.020555776596069338,
173
+ 0.02048102378845215,
174
+ 0.020488191604614257,
175
+ 0.020497407913208008,
176
+ 0.021561344146728514,
177
+ 0.021311487197875977,
178
+ 0.02127769660949707,
179
+ 0.020533248901367186,
180
+ 0.020552703857421875,
181
+ 0.020541439056396483,
182
+ 0.020577280044555665,
183
+ 0.020545568466186524,
184
+ 0.02055369567871094,
185
+ 0.020542463302612304,
186
+ 0.020537343978881836,
187
+ 0.020560895919799805,
188
+ 0.0206059513092041,
189
+ 0.02059878349304199,
190
+ 0.02060492706298828,
191
+ 0.020569087982177735,
192
+ 0.020550655364990233,
193
+ 0.020563968658447264,
194
+ 0.020559871673583984,
195
+ 0.020784128189086915,
196
+ 0.02061311912536621,
197
+ 0.020564992904663085,
198
+ 0.02052403259277344,
199
+ 0.020554752349853517,
200
+ 0.020528127670288086,
201
+ 0.020519935607910156,
202
+ 0.020600831985473633,
203
+ 0.02061311912536621,
204
+ 0.020590591430664062,
205
+ 0.02049228858947754,
206
+ 0.020571136474609376,
207
+ 0.020538368225097657,
208
+ 0.020547584533691408,
209
+ 0.020529151916503906,
210
+ 0.020538368225097657,
211
+ 0.020559871673583984,
212
+ 0.020519935607910156,
213
+ 0.020517887115478514,
214
+ 0.020552703857421875,
215
+ 0.02130227279663086,
216
+ 0.02131865692138672,
217
+ 0.021348352432250976,
218
+ 0.021334016799926758,
219
+ 0.021391359329223633,
220
+ 0.02128998374938965,
221
+ 0.02126643180847168,
222
+ 0.021327871322631836,
223
+ 0.021353471755981446,
224
+ 0.02128998374938965,
225
+ 0.021285888671875,
226
+ 0.02122137641906738,
227
+ 0.021271551132202148,
228
+ 0.02126540756225586,
229
+ 0.02125619125366211,
230
+ 0.020486143112182616,
231
+ 0.020510719299316405,
232
+ 0.020530176162719727,
233
+ 0.020516864776611327,
234
+ 0.020552703857421875,
235
+ 0.020572160720825194,
236
+ 0.020644863128662108,
237
+ 0.020559871673583984,
238
+ 0.020553728103637696,
239
+ 0.020576255798339844,
240
+ 0.020544511795043945,
241
+ 0.020518911361694335,
242
+ 0.02064896011352539,
243
+ 0.020557823181152343,
244
+ 0.020556800842285155,
245
+ 0.020519935607910156,
246
+ 0.020511743545532226,
247
+ 0.020496383666992187,
248
+ 0.020549631118774413,
249
+ 0.020503551483154296,
250
+ 0.020513792037963868,
251
+ 0.020548608779907225,
252
+ 0.020728832244873048,
253
+ 0.020544511795043945,
254
+ 0.02058243179321289,
255
+ 0.02050966453552246,
256
+ 0.020489215850830078,
257
+ 0.02049126434326172,
258
+ 0.020551679611206054,
259
+ 0.020548608779907225,
260
+ 0.020509695053100584,
261
+ 0.020566015243530272,
262
+ 0.020542463302612304,
263
+ 0.020544511795043945,
264
+ 0.020590591430664062,
265
+ 0.020551679611206054,
266
+ 0.02064076805114746,
267
+ 0.020540416717529295,
268
+ 0.020626432418823244,
269
+ 0.020488191604614257,
270
+ 0.02049126434326172,
271
+ 0.02107904052734375,
272
+ 0.021363712310791014,
273
+ 0.021246976852416992,
274
+ 0.02126848030090332,
275
+ 0.02126643180847168,
276
+ 0.02125619125366211,
277
+ 0.02125004768371582,
278
+ 0.02126335906982422,
279
+ 0.021324800491333007,
280
+ 0.02124595260620117,
281
+ 0.02127462387084961,
282
+ 0.02126848030090332,
283
+ 0.0212992000579834,
284
+ 0.021230592727661132,
285
+ 0.02126540756225586,
286
+ 0.020592639923095703,
287
+ 0.020580352783203124,
288
+ 0.02062335968017578,
289
+ 0.02061516761779785,
290
+ 0.020649984359741212,
291
+ 0.021497856140136717,
292
+ 0.021341184616088867,
293
+ 0.021324800491333007,
294
+ 0.021350400924682617,
295
+ 0.021413888931274414,
296
+ 0.021346303939819337,
297
+ 0.021335039138793945,
298
+ 0.02128895950317383,
299
+ 0.02131455993652344,
300
+ 0.02062950325012207,
301
+ 0.020576255798339844,
302
+ 0.020571136474609376,
303
+ 0.021362688064575194,
304
+ 0.021320703506469727,
305
+ 0.02063564872741699,
306
+ 0.020571136474609376,
307
+ 0.020585472106933594,
308
+ 0.020543487548828124,
309
+ 0.020616191864013672,
310
+ 0.02129715156555176,
311
+ 0.020577280044555665,
312
+ 0.020585472106933594,
313
+ 0.020578304290771485,
314
+ 0.02066022491455078,
315
+ 0.02066227149963379,
316
+ 0.020743167877197266,
317
+ 0.020583423614501953,
318
+ 0.020555776596069338,
319
+ 0.020649984359741212,
320
+ 0.020542463302612304,
321
+ 0.020611072540283205,
322
+ 0.020601856231689454,
323
+ 0.020783103942871094,
324
+ 0.020562944412231447,
325
+ 0.02063155174255371,
326
+ 0.020585472106933594,
327
+ 0.02061311912536621,
328
+ 0.020593664169311524,
329
+ 0.020580352783203124,
330
+ 0.020892671585083008,
331
+ 0.021307392120361326,
332
+ 0.021719039916992186,
333
+ 0.02067046356201172,
334
+ 0.02061516761779785,
335
+ 0.020580352783203124,
336
+ 0.02058956718444824,
337
+ 0.020669439315795898,
338
+ 0.020574207305908202,
339
+ 0.02169753646850586,
340
+ 0.021444608688354492,
341
+ 0.021353471755981446,
342
+ 0.021292032241821288,
343
+ 0.021568511962890623,
344
+ 0.021338111877441408,
345
+ 0.02129817581176758,
346
+ 0.021315584182739256,
347
+ 0.0212807674407959,
348
+ 0.021768192291259765,
349
+ 0.021414911270141602,
350
+ 0.021296127319335938,
351
+ 0.0213309440612793,
352
+ 0.02128998374938965,
353
+ 0.021409791946411134,
354
+ 0.021395456314086913,
355
+ 0.021357568740844726,
356
+ 0.021313535690307618,
357
+ 0.021410816192626952
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 48.118379091817246
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02237489604949951,
7
+ "prefill.throughput.value": 312.8506154627064,
8
+ "decode.latency.mean": 2.6394550781249997,
9
+ "decode.throughput.value": 48.1159922184459,
10
+ "per_token.latency.mean": 0.02078207992193267,
11
+ "per_token.throughput.value": 48.118379091817246
12
+ }
13
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 9578c2597e2d88b6f0b304b5a05864fd613ddcc1
84
+ accelerate_version: 0.34.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.22.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
88
+ - hydra.run.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
94
+ "accelerate_version": "0.34.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.22.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1337.798656,
111
+ "max_global_vram": 6768.033792,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 14.0260244140625,
120
+ "mean": 14.0260244140625,
121
+ "stdev": 0.0,
122
+ "p50": 14.0260244140625,
123
+ "p90": 14.0260244140625,
124
+ "p95": 14.0260244140625,
125
+ "p99": 14.0260244140625,
126
+ "values": [
127
+ 14.0260244140625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1855.827968,
138
+ "max_global_vram": 6789.005312,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.0282490873336792,
147
+ "mean": 0.0141245436668396,
148
+ "stdev": 0.00015136003494262747,
149
+ "p50": 0.0141245436668396,
150
+ "p90": 0.014245631694793702,
151
+ "p95": 0.014260767698287964,
152
+ "p99": 0.014272876501083375,
153
+ "values": [
154
+ 0.013973183631896972,
155
+ 0.014275903701782227
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 495.5912321920887
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2025.803776,
169
+ "max_global_vram": 6071.779328,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.126360595703125,
178
+ "mean": 1.5631802978515625,
179
+ "stdev": 0.00015649414062501865,
180
+ "p50": 1.5631802978515625,
181
+ "p90": 1.5633054931640624,
182
+ "p95": 1.563321142578125,
183
+ "p99": 1.563333662109375,
184
+ "values": [
185
+ 1.5630238037109374,
186
+ 1.5633367919921874
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.24462685113738
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.1260805072784414,
202
+ "mean": 0.012307403571962371,
203
+ "stdev": 2.1436560582032447e-05,
204
+ "p50": 0.012302335739135742,
205
+ "p90": 0.012324557113647461,
206
+ "p95": 0.012333056449890138,
207
+ "p99": 0.012404592456817626,
208
+ "values": [
209
+ 0.012333056449890138,
210
+ 0.012291071891784668,
211
+ 0.012294143676757812,
212
+ 0.012338175773620605,
213
+ 0.012303359985351562,
214
+ 0.012309503555297852,
215
+ 0.012306431770324706,
216
+ 0.012295167922973632,
217
+ 0.012299263954162597,
218
+ 0.012294143676757812,
219
+ 0.012301312446594239,
220
+ 0.012295167922973632,
221
+ 0.012295167922973632,
222
+ 0.012295167922973632,
223
+ 0.012302335739135742,
224
+ 0.012308480262756348,
225
+ 0.012317695617675782,
226
+ 0.012327936172485352,
227
+ 0.012303359985351562,
228
+ 0.012307456016540527,
229
+ 0.012356608390808106,
230
+ 0.012317695617675782,
231
+ 0.01232588768005371,
232
+ 0.012297216415405274,
233
+ 0.012291071891784668,
234
+ 0.01229312038421631,
235
+ 0.012302335739135742,
236
+ 0.012318719863891601,
237
+ 0.012340224266052247,
238
+ 0.012314623832702636,
239
+ 0.012301312446594239,
240
+ 0.012304384231567383,
241
+ 0.012312576293945313,
242
+ 0.012306431770324706,
243
+ 0.012295167922973632,
244
+ 0.012299263954162597,
245
+ 0.012296192169189453,
246
+ 0.012310527801513671,
247
+ 0.012298239707946777,
248
+ 0.012322815895080566,
249
+ 0.012294143676757812,
250
+ 0.012305407524108887,
251
+ 0.012322815895080566,
252
+ 0.012303359985351562,
253
+ 0.012300288200378418,
254
+ 0.012311552047729492,
255
+ 0.012302335739135742,
256
+ 0.012298239707946777,
257
+ 0.012290047645568849,
258
+ 0.012297216415405274,
259
+ 0.012294207572937012,
260
+ 0.012303296089172364,
261
+ 0.012311552047729492,
262
+ 0.012310527801513671,
263
+ 0.01229312038421631,
264
+ 0.012324864387512208,
265
+ 0.012307456016540527,
266
+ 0.012302335739135742,
267
+ 0.012302335739135742,
268
+ 0.012305407524108887,
269
+ 0.012292096138000488,
270
+ 0.012300288200378418,
271
+ 0.012299263954162597,
272
+ 0.012294143676757812,
273
+ 0.012295167922973632,
274
+ 0.012301312446594239,
275
+ 0.012301312446594239,
276
+ 0.012298239707946777,
277
+ 0.012311552047729492,
278
+ 0.012324864387512208,
279
+ 0.012365823745727538,
280
+ 0.012319744110107422,
281
+ 0.012294143676757812,
282
+ 0.012307456016540527,
283
+ 0.012302335739135742,
284
+ 0.012300288200378418,
285
+ 0.012311552047729492,
286
+ 0.012297216415405274,
287
+ 0.012302335739135742,
288
+ 0.012294143676757812,
289
+ 0.012321791648864745,
290
+ 0.01232588768005371,
291
+ 0.012315648078918457,
292
+ 0.012304384231567383,
293
+ 0.012307456016540527,
294
+ 0.012295167922973632,
295
+ 0.012298239707946777,
296
+ 0.012305407524108887,
297
+ 0.012295167922973632,
298
+ 0.012294143676757812,
299
+ 0.012285951614379884,
300
+ 0.012300288200378418,
301
+ 0.012294143676757812,
302
+ 0.01229312038421631,
303
+ 0.012346367835998535,
304
+ 0.012319744110107422,
305
+ 0.012313599586486817,
306
+ 0.012317695617675782,
307
+ 0.012298239707946777,
308
+ 0.012296192169189453,
309
+ 0.012291071891784668,
310
+ 0.012296192169189453,
311
+ 0.012298239707946777,
312
+ 0.012309503555297852,
313
+ 0.012296192169189453,
314
+ 0.012307456016540527,
315
+ 0.012292096138000488,
316
+ 0.012322848320007324,
317
+ 0.0123156156539917,
318
+ 0.012300288200378418,
319
+ 0.012309503555297852,
320
+ 0.012299263954162597,
321
+ 0.012300288200378418,
322
+ 0.012295167922973632,
323
+ 0.012302335739135742,
324
+ 0.012301312446594239,
325
+ 0.01229312038421631,
326
+ 0.012303359985351562,
327
+ 0.012300288200378418,
328
+ 0.012294143676757812,
329
+ 0.012332032203674317,
330
+ 0.012319744110107422,
331
+ 0.012297216415405274,
332
+ 0.012312576293945313,
333
+ 0.01233407974243164,
334
+ 0.012315648078918457,
335
+ 0.012301312446594239,
336
+ 0.01236684799194336,
337
+ 0.012317695617675782,
338
+ 0.012302335739135742,
339
+ 0.012302335739135742,
340
+ 0.012301312446594239,
341
+ 0.012295167922973632,
342
+ 0.012300288200378418,
343
+ 0.012294143676757812,
344
+ 0.012288000106811523,
345
+ 0.012300288200378418,
346
+ 0.012296192169189453,
347
+ 0.012309503555297852,
348
+ 0.012312576293945313,
349
+ 0.012295167922973632,
350
+ 0.012304384231567383,
351
+ 0.01229312038421631,
352
+ 0.012302335739135742,
353
+ 0.012290047645568849,
354
+ 0.012492799758911132,
355
+ 0.012331007957458496,
356
+ 0.012305407524108887,
357
+ 0.012303359985351562,
358
+ 0.012386303901672363,
359
+ 0.012357631683349609,
360
+ 0.012327936172485352,
361
+ 0.012326911926269531,
362
+ 0.012308480262756348,
363
+ 0.012311552047729492,
364
+ 0.012301312446594239,
365
+ 0.012305407524108887,
366
+ 0.01229312038421631,
367
+ 0.012307456016540527,
368
+ 0.012298239707946777,
369
+ 0.012303359985351562,
370
+ 0.012297216415405274,
371
+ 0.012295167922973632,
372
+ 0.012294143676757812,
373
+ 0.012320768356323243,
374
+ 0.012300288200378418,
375
+ 0.012307456016540527,
376
+ 0.012302335739135742,
377
+ 0.012299263954162597,
378
+ 0.012295167922973632,
379
+ 0.012303359985351562,
380
+ 0.012304384231567383,
381
+ 0.012300288200378418,
382
+ 0.012299263954162597,
383
+ 0.01229520034790039,
384
+ 0.0123023042678833,
385
+ 0.012298239707946777,
386
+ 0.012446720123291016,
387
+ 0.012328960418701173,
388
+ 0.012297216415405274,
389
+ 0.012312576293945313,
390
+ 0.012313599586486817,
391
+ 0.012305407524108887,
392
+ 0.012332032203674317,
393
+ 0.012295167922973632,
394
+ 0.012309503555297852,
395
+ 0.012302335739135742,
396
+ 0.012307456016540527,
397
+ 0.012297216415405274,
398
+ 0.012312576293945313,
399
+ 0.012318719863891601,
400
+ 0.012292096138000488,
401
+ 0.012299263954162597,
402
+ 0.012300288200378418,
403
+ 0.012302335739135742,
404
+ 0.012300288200378418,
405
+ 0.012302335739135742,
406
+ 0.012303359985351562,
407
+ 0.01228700828552246,
408
+ 0.012296159744262696,
409
+ 0.012304384231567383,
410
+ 0.012308480262756348,
411
+ 0.012299263954162597,
412
+ 0.01232588768005371,
413
+ 0.012303359985351562,
414
+ 0.012314623832702636,
415
+ 0.012310527801513671,
416
+ 0.012300288200378418,
417
+ 0.012310527801513671,
418
+ 0.012297216415405274,
419
+ 0.012290047645568849,
420
+ 0.012300288200378418,
421
+ 0.012303359985351562,
422
+ 0.012304384231567383,
423
+ 0.012299263954162597,
424
+ 0.012289024353027344,
425
+ 0.012316672325134278,
426
+ 0.012311552047729492,
427
+ 0.012306431770324706,
428
+ 0.012302335739135742,
429
+ 0.012302335739135742,
430
+ 0.012302335739135742,
431
+ 0.012305407524108887,
432
+ 0.012297216415405274,
433
+ 0.012305407524108887,
434
+ 0.012296192169189453,
435
+ 0.012303359985351562,
436
+ 0.012301312446594239,
437
+ 0.012298239707946777,
438
+ 0.012314623832702636,
439
+ 0.012303359985351562,
440
+ 0.01242521572113037,
441
+ 0.012333056449890138,
442
+ 0.012302335739135742,
443
+ 0.012292096138000488,
444
+ 0.012296192169189453,
445
+ 0.012304384231567383,
446
+ 0.012292096138000488,
447
+ 0.012292096138000488,
448
+ 0.012306431770324706,
449
+ 0.012305407524108887,
450
+ 0.012301312446594239,
451
+ 0.012323840141296387,
452
+ 0.012309503555297852,
453
+ 0.012308480262756348,
454
+ 0.012298239707946777,
455
+ 0.012299263954162597,
456
+ 0.012295167922973632,
457
+ 0.012298239707946777,
458
+ 0.012291071891784668,
459
+ 0.012297216415405274,
460
+ 0.01228492832183838,
461
+ 0.012295167922973632,
462
+ 0.012305407524108887
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.25190615168505
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
93
+ "accelerate_version": "0.34.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.22.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1337.798656,
6
+ "max_global_vram": 6768.033792,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 14.0260244140625,
15
+ "mean": 14.0260244140625,
16
+ "stdev": 0.0,
17
+ "p50": 14.0260244140625,
18
+ "p90": 14.0260244140625,
19
+ "p95": 14.0260244140625,
20
+ "p99": 14.0260244140625,
21
+ "values": [
22
+ 14.0260244140625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1855.827968,
33
+ "max_global_vram": 6789.005312,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.0282490873336792,
42
+ "mean": 0.0141245436668396,
43
+ "stdev": 0.00015136003494262747,
44
+ "p50": 0.0141245436668396,
45
+ "p90": 0.014245631694793702,
46
+ "p95": 0.014260767698287964,
47
+ "p99": 0.014272876501083375,
48
+ "values": [
49
+ 0.013973183631896972,
50
+ 0.014275903701782227
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 495.5912321920887
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2025.803776,
64
+ "max_global_vram": 6071.779328,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.126360595703125,
73
+ "mean": 1.5631802978515625,
74
+ "stdev": 0.00015649414062501865,
75
+ "p50": 1.5631802978515625,
76
+ "p90": 1.5633054931640624,
77
+ "p95": 1.563321142578125,
78
+ "p99": 1.563333662109375,
79
+ "values": [
80
+ 1.5630238037109374,
81
+ 1.5633367919921874
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.24462685113738
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.1260805072784414,
97
+ "mean": 0.012307403571962371,
98
+ "stdev": 2.1436560582032447e-05,
99
+ "p50": 0.012302335739135742,
100
+ "p90": 0.012324557113647461,
101
+ "p95": 0.012333056449890138,
102
+ "p99": 0.012404592456817626,
103
+ "values": [
104
+ 0.012333056449890138,
105
+ 0.012291071891784668,
106
+ 0.012294143676757812,
107
+ 0.012338175773620605,
108
+ 0.012303359985351562,
109
+ 0.012309503555297852,
110
+ 0.012306431770324706,
111
+ 0.012295167922973632,
112
+ 0.012299263954162597,
113
+ 0.012294143676757812,
114
+ 0.012301312446594239,
115
+ 0.012295167922973632,
116
+ 0.012295167922973632,
117
+ 0.012295167922973632,
118
+ 0.012302335739135742,
119
+ 0.012308480262756348,
120
+ 0.012317695617675782,
121
+ 0.012327936172485352,
122
+ 0.012303359985351562,
123
+ 0.012307456016540527,
124
+ 0.012356608390808106,
125
+ 0.012317695617675782,
126
+ 0.01232588768005371,
127
+ 0.012297216415405274,
128
+ 0.012291071891784668,
129
+ 0.01229312038421631,
130
+ 0.012302335739135742,
131
+ 0.012318719863891601,
132
+ 0.012340224266052247,
133
+ 0.012314623832702636,
134
+ 0.012301312446594239,
135
+ 0.012304384231567383,
136
+ 0.012312576293945313,
137
+ 0.012306431770324706,
138
+ 0.012295167922973632,
139
+ 0.012299263954162597,
140
+ 0.012296192169189453,
141
+ 0.012310527801513671,
142
+ 0.012298239707946777,
143
+ 0.012322815895080566,
144
+ 0.012294143676757812,
145
+ 0.012305407524108887,
146
+ 0.012322815895080566,
147
+ 0.012303359985351562,
148
+ 0.012300288200378418,
149
+ 0.012311552047729492,
150
+ 0.012302335739135742,
151
+ 0.012298239707946777,
152
+ 0.012290047645568849,
153
+ 0.012297216415405274,
154
+ 0.012294207572937012,
155
+ 0.012303296089172364,
156
+ 0.012311552047729492,
157
+ 0.012310527801513671,
158
+ 0.01229312038421631,
159
+ 0.012324864387512208,
160
+ 0.012307456016540527,
161
+ 0.012302335739135742,
162
+ 0.012302335739135742,
163
+ 0.012305407524108887,
164
+ 0.012292096138000488,
165
+ 0.012300288200378418,
166
+ 0.012299263954162597,
167
+ 0.012294143676757812,
168
+ 0.012295167922973632,
169
+ 0.012301312446594239,
170
+ 0.012301312446594239,
171
+ 0.012298239707946777,
172
+ 0.012311552047729492,
173
+ 0.012324864387512208,
174
+ 0.012365823745727538,
175
+ 0.012319744110107422,
176
+ 0.012294143676757812,
177
+ 0.012307456016540527,
178
+ 0.012302335739135742,
179
+ 0.012300288200378418,
180
+ 0.012311552047729492,
181
+ 0.012297216415405274,
182
+ 0.012302335739135742,
183
+ 0.012294143676757812,
184
+ 0.012321791648864745,
185
+ 0.01232588768005371,
186
+ 0.012315648078918457,
187
+ 0.012304384231567383,
188
+ 0.012307456016540527,
189
+ 0.012295167922973632,
190
+ 0.012298239707946777,
191
+ 0.012305407524108887,
192
+ 0.012295167922973632,
193
+ 0.012294143676757812,
194
+ 0.012285951614379884,
195
+ 0.012300288200378418,
196
+ 0.012294143676757812,
197
+ 0.01229312038421631,
198
+ 0.012346367835998535,
199
+ 0.012319744110107422,
200
+ 0.012313599586486817,
201
+ 0.012317695617675782,
202
+ 0.012298239707946777,
203
+ 0.012296192169189453,
204
+ 0.012291071891784668,
205
+ 0.012296192169189453,
206
+ 0.012298239707946777,
207
+ 0.012309503555297852,
208
+ 0.012296192169189453,
209
+ 0.012307456016540527,
210
+ 0.012292096138000488,
211
+ 0.012322848320007324,
212
+ 0.0123156156539917,
213
+ 0.012300288200378418,
214
+ 0.012309503555297852,
215
+ 0.012299263954162597,
216
+ 0.012300288200378418,
217
+ 0.012295167922973632,
218
+ 0.012302335739135742,
219
+ 0.012301312446594239,
220
+ 0.01229312038421631,
221
+ 0.012303359985351562,
222
+ 0.012300288200378418,
223
+ 0.012294143676757812,
224
+ 0.012332032203674317,
225
+ 0.012319744110107422,
226
+ 0.012297216415405274,
227
+ 0.012312576293945313,
228
+ 0.01233407974243164,
229
+ 0.012315648078918457,
230
+ 0.012301312446594239,
231
+ 0.01236684799194336,
232
+ 0.012317695617675782,
233
+ 0.012302335739135742,
234
+ 0.012302335739135742,
235
+ 0.012301312446594239,
236
+ 0.012295167922973632,
237
+ 0.012300288200378418,
238
+ 0.012294143676757812,
239
+ 0.012288000106811523,
240
+ 0.012300288200378418,
241
+ 0.012296192169189453,
242
+ 0.012309503555297852,
243
+ 0.012312576293945313,
244
+ 0.012295167922973632,
245
+ 0.012304384231567383,
246
+ 0.01229312038421631,
247
+ 0.012302335739135742,
248
+ 0.012290047645568849,
249
+ 0.012492799758911132,
250
+ 0.012331007957458496,
251
+ 0.012305407524108887,
252
+ 0.012303359985351562,
253
+ 0.012386303901672363,
254
+ 0.012357631683349609,
255
+ 0.012327936172485352,
256
+ 0.012326911926269531,
257
+ 0.012308480262756348,
258
+ 0.012311552047729492,
259
+ 0.012301312446594239,
260
+ 0.012305407524108887,
261
+ 0.01229312038421631,
262
+ 0.012307456016540527,
263
+ 0.012298239707946777,
264
+ 0.012303359985351562,
265
+ 0.012297216415405274,
266
+ 0.012295167922973632,
267
+ 0.012294143676757812,
268
+ 0.012320768356323243,
269
+ 0.012300288200378418,
270
+ 0.012307456016540527,
271
+ 0.012302335739135742,
272
+ 0.012299263954162597,
273
+ 0.012295167922973632,
274
+ 0.012303359985351562,
275
+ 0.012304384231567383,
276
+ 0.012300288200378418,
277
+ 0.012299263954162597,
278
+ 0.01229520034790039,
279
+ 0.0123023042678833,
280
+ 0.012298239707946777,
281
+ 0.012446720123291016,
282
+ 0.012328960418701173,
283
+ 0.012297216415405274,
284
+ 0.012312576293945313,
285
+ 0.012313599586486817,
286
+ 0.012305407524108887,
287
+ 0.012332032203674317,
288
+ 0.012295167922973632,
289
+ 0.012309503555297852,
290
+ 0.012302335739135742,
291
+ 0.012307456016540527,
292
+ 0.012297216415405274,
293
+ 0.012312576293945313,
294
+ 0.012318719863891601,
295
+ 0.012292096138000488,
296
+ 0.012299263954162597,
297
+ 0.012300288200378418,
298
+ 0.012302335739135742,
299
+ 0.012300288200378418,
300
+ 0.012302335739135742,
301
+ 0.012303359985351562,
302
+ 0.01228700828552246,
303
+ 0.012296159744262696,
304
+ 0.012304384231567383,
305
+ 0.012308480262756348,
306
+ 0.012299263954162597,
307
+ 0.01232588768005371,
308
+ 0.012303359985351562,
309
+ 0.012314623832702636,
310
+ 0.012310527801513671,
311
+ 0.012300288200378418,
312
+ 0.012310527801513671,
313
+ 0.012297216415405274,
314
+ 0.012290047645568849,
315
+ 0.012300288200378418,
316
+ 0.012303359985351562,
317
+ 0.012304384231567383,
318
+ 0.012299263954162597,
319
+ 0.012289024353027344,
320
+ 0.012316672325134278,
321
+ 0.012311552047729492,
322
+ 0.012306431770324706,
323
+ 0.012302335739135742,
324
+ 0.012302335739135742,
325
+ 0.012302335739135742,
326
+ 0.012305407524108887,
327
+ 0.012297216415405274,
328
+ 0.012305407524108887,
329
+ 0.012296192169189453,
330
+ 0.012303359985351562,
331
+ 0.012301312446594239,
332
+ 0.012298239707946777,
333
+ 0.012314623832702636,
334
+ 0.012303359985351562,
335
+ 0.01242521572113037,
336
+ 0.012333056449890138,
337
+ 0.012302335739135742,
338
+ 0.012292096138000488,
339
+ 0.012296192169189453,
340
+ 0.012304384231567383,
341
+ 0.012292096138000488,
342
+ 0.012292096138000488,
343
+ 0.012306431770324706,
344
+ 0.012305407524108887,
345
+ 0.012301312446594239,
346
+ 0.012323840141296387,
347
+ 0.012309503555297852,
348
+ 0.012308480262756348,
349
+ 0.012298239707946777,
350
+ 0.012299263954162597,
351
+ 0.012295167922973632,
352
+ 0.012298239707946777,
353
+ 0.012291071891784668,
354
+ 0.012297216415405274,
355
+ 0.01228492832183838,
356
+ 0.012295167922973632,
357
+ 0.012305407524108887
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.25190615168505
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.0141245436668396,
7
+ "prefill.throughput.value": 495.5912321920887,
8
+ "decode.latency.mean": 1.5631802978515625,
9
+ "decode.throughput.value": 81.24462685113738,
10
+ "per_token.latency.mean": 0.012307403571962371,
11
+ "per_token.throughput.value": 81.25190615168505
12
+ }
13
+ }
2024-08-27/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpx0uk9jq5/commit=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1
88
+ - hydra.run.dir=_benchmark/tmpx0uk9jq5/commit\=9578c2597e2d88b6f0b304b5a05864fd613ddcc1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-211.872.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: 9578c2597e2d88b6f0b304b5a05864fd613ddcc1
225
+ accelerate_version: 0.34.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.22.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-08-27/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.020687775611877443,
8
+ "prefill.throughput.value": 338.3640721615861,
9
+ "decode.latency.mean": 2.504463623046875,
10
+ "decode.throughput.value": 50.7094608327729,
11
+ "per_token.latency.mean": 0.019718889867226908,
12
+ "per_token.throughput.value": 50.71279401291322
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.02237489604949951,
21
+ "prefill.throughput.value": 312.8506154627064,
22
+ "decode.latency.mean": 2.6394550781249997,
23
+ "decode.throughput.value": 48.1159922184459,
24
+ "per_token.latency.mean": 0.02078207992193267,
25
+ "per_token.throughput.value": 48.118379091817246
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "9578c2597e2d88b6f0b304b5a05864fd613ddcc1",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.0141245436668396,
34
+ "prefill.throughput.value": 495.5912321920887,
35
+ "decode.latency.mean": 1.5631802978515625,
36
+ "decode.throughput.value": 81.24462685113738,
37
+ "per_token.latency.mean": 0.012307403571962371,
38
+ "per_token.throughput.value": 81.25190615168505
39
+ }
40
+ }
41
+ ]
2024-08-27/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "9578c2597e2d88b6f0b304b5a05864fd613ddcc1": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.020687775611877443,
7
+ "prefill.throughput.value": 338.3640721615861,
8
+ "decode.latency.mean": 2.504463623046875,
9
+ "decode.throughput.value": 50.7094608327729,
10
+ "per_token.latency.mean": 0.019718889867226908,
11
+ "per_token.throughput.value": 50.71279401291322
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "9578c2597e2d88b6f0b304b5a05864fd613ddcc1": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.02237489604949951,
19
+ "prefill.throughput.value": 312.8506154627064,
20
+ "decode.latency.mean": 2.6394550781249997,
21
+ "decode.throughput.value": 48.1159922184459,
22
+ "per_token.latency.mean": 0.02078207992193267,
23
+ "per_token.throughput.value": 48.118379091817246
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "9578c2597e2d88b6f0b304b5a05864fd613ddcc1": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.0141245436668396,
31
+ "prefill.throughput.value": 495.5912321920887,
32
+ "decode.latency.mean": 1.5631802978515625,
33
+ "decode.throughput.value": 81.24462685113738,
34
+ "per_token.latency.mean": 0.012307403571962371,
35
+ "per_token.throughput.value": 81.25190615168505
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }