hf-transformers-bot commited on
Commit
ccffce3
1 Parent(s): a7788e8

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +91 -0
  2. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +443 -0
  5. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +102 -0
  7. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +339 -0
  8. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +91 -0
  10. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +1 -0
  13. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +102 -0
  14. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +91 -0
  15. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +443 -0
  18. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +102 -0
  20. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +339 -0
  21. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +91 -0
  23. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +443 -0
  26. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +102 -0
  28. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +339 -0
  29. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/multirun.yaml +232 -0
  31. 2024-06-16/summaries.json +41 -0
  32. 2024-06-16/summary.json +40 -0
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: eed9ed679878ada2f6d2eefccdbda368cabc88b1
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1
88
+ - hydra.run.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1738.227712,
109
+ "max_global_vram": 6793.199616,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 6146.752512,
112
+ "max_allocated": 5034.576896
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.045458688735961914,
118
+ "mean": 0.022729344367980957,
119
+ "stdev": 0.002710240364074707,
120
+ "p50": 0.022729344367980957,
121
+ "p90": 0.024897536659240722,
122
+ "p95": 0.02516856069564819,
123
+ "p99": 0.02538537992477417,
124
+ "values": [
125
+ 0.025439584732055664,
126
+ 0.02001910400390625
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 307.9719276839752
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 1738.32192,
140
+ "max_global_vram": 6795.296768,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 6148.849664,
143
+ "max_allocated": 5034.577408
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 4.775644775390625,
149
+ "mean": 2.3878223876953126,
150
+ "stdev": 0.013817749023437464,
151
+ "p50": 2.3878223876953126,
152
+ "p90": 2.3988765869140627,
153
+ "p95": 2.400258361816406,
154
+ "p99": 2.401363781738281,
155
+ "values": [
156
+ 2.40164013671875,
157
+ 2.374004638671875
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 53.186535420072985
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 4.776072187423707,
173
+ "mean": 0.018877755681516624,
174
+ "stdev": 0.0013918402423089098,
175
+ "p50": 0.018550783157348632,
176
+ "p90": 0.019393766403198242,
177
+ "p95": 0.019474841690063475,
178
+ "p99": 0.019828408737182616,
179
+ "values": [
180
+ 0.019301376342773437,
181
+ 0.01923686408996582,
182
+ 0.01925017547607422,
183
+ 0.021184511184692383,
184
+ 0.019475456237792968,
185
+ 0.019316736221313476,
186
+ 0.01929430389404297,
187
+ 0.01925315284729004,
188
+ 0.01922047996520996,
189
+ 0.019180543899536134,
190
+ 0.019198976516723632,
191
+ 0.019284992218017577,
192
+ 0.019174400329589843,
193
+ 0.019170303344726563,
194
+ 0.019153919219970703,
195
+ 0.019397632598876953,
196
+ 0.01929523277282715,
197
+ 0.019167232513427734,
198
+ 0.019268608093261717,
199
+ 0.019173376083374022,
200
+ 0.01922969627380371,
201
+ 0.019238943099975585,
202
+ 0.019377119064331056,
203
+ 0.019252223968505858,
204
+ 0.01922969627380371,
205
+ 0.019163135528564454,
206
+ 0.018884607315063476,
207
+ 0.018998271942138673,
208
+ 0.019182592391967773,
209
+ 0.01945292854309082,
210
+ 0.018923519134521484,
211
+ 0.018922496795654296,
212
+ 0.01905971145629883,
213
+ 0.0192491512298584,
214
+ 0.019371007919311522,
215
+ 0.019307519912719725,
216
+ 0.019194879531860352,
217
+ 0.019191808700561523,
218
+ 0.019338239669799806,
219
+ 0.01924095916748047,
220
+ 0.019377151489257814,
221
+ 0.019173376083374022,
222
+ 0.01925119972229004,
223
+ 0.01918976020812988,
224
+ 0.019183616638183593,
225
+ 0.01902694320678711,
226
+ 0.01904844856262207,
227
+ 0.019186687469482423,
228
+ 0.019594240188598632,
229
+ 0.01940787124633789,
230
+ 0.019204095840454103,
231
+ 0.019792896270751953,
232
+ 0.019451904296875,
233
+ 0.019539968490600586,
234
+ 0.019530752182006835,
235
+ 0.019350528717041016,
236
+ 0.019370016098022462,
237
+ 0.0193525447845459,
238
+ 0.019298303604125978,
239
+ 0.019366912841796875,
240
+ 0.019389440536499023,
241
+ 0.019362815856933592,
242
+ 0.01879347229003906,
243
+ 0.01847603225708008,
244
+ 0.018449407577514648,
245
+ 0.018388992309570314,
246
+ 0.018379776000976563,
247
+ 0.01840742492675781,
248
+ 0.01838489532470703,
249
+ 0.01862553596496582,
250
+ 0.018440191268920898,
251
+ 0.018365440368652345,
252
+ 0.01840947151184082,
253
+ 0.018480127334594726,
254
+ 0.018357248306274415,
255
+ 0.01846272087097168,
256
+ 0.018655231475830078,
257
+ 0.018363391876220703,
258
+ 0.01838591957092285,
259
+ 0.018364479064941406,
260
+ 0.01834796714782715,
261
+ 0.018747392654418944,
262
+ 0.01855897521972656,
263
+ 0.018431999206542968,
264
+ 0.01841766357421875,
265
+ 0.018370559692382812,
266
+ 0.018566144943237304,
267
+ 0.018396160125732423,
268
+ 0.018457599639892578,
269
+ 0.018388992309570314,
270
+ 0.01839308738708496,
271
+ 0.01841459274291992,
272
+ 0.018473983764648438,
273
+ 0.01839411163330078,
274
+ 0.018568191528320312,
275
+ 0.018387968063354493,
276
+ 0.018367488861083983,
277
+ 0.018498559951782227,
278
+ 0.018359296798706053,
279
+ 0.01863680076599121,
280
+ 0.018550783157348632,
281
+ 0.018405376434326173,
282
+ 0.018397184371948243,
283
+ 0.018358272552490236,
284
+ 0.018214912414550782,
285
+ 0.018267135620117187,
286
+ 0.01847500801086426,
287
+ 0.018592767715454102,
288
+ 0.018502656936645507,
289
+ 0.018395135879516602,
290
+ 0.01847091293334961,
291
+ 0.018511871337890624,
292
+ 0.018420736312866212,
293
+ 0.018447359085083007,
294
+ 0.01843507194519043,
295
+ 0.01840640068054199,
296
+ 0.018421760559082033,
297
+ 0.018355199813842774,
298
+ 0.018381824493408205,
299
+ 0.018340864181518556,
300
+ 0.019243104934692383,
301
+ 0.0194169921875,
302
+ 0.01925836753845215,
303
+ 0.01939455986022949,
304
+ 0.019327999114990235,
305
+ 0.01939059257507324,
306
+ 0.03963584136962891,
307
+ 0.018341888427734376,
308
+ 0.01838387107849121,
309
+ 0.018378751754760742,
310
+ 0.01824460792541504,
311
+ 0.01845145606994629,
312
+ 0.018344959259033202,
313
+ 0.018326528549194337,
314
+ 0.0182794246673584,
315
+ 0.01888768005371094,
316
+ 0.01842585563659668,
317
+ 0.01846067237854004,
318
+ 0.018307071685791015,
319
+ 0.018363391876220703,
320
+ 0.018397184371948243,
321
+ 0.018353151321411132,
322
+ 0.01835212707519531,
323
+ 0.019808256149291992,
324
+ 0.019335168838500977,
325
+ 0.019187711715698243,
326
+ 0.019314687728881837,
327
+ 0.018373632431030275,
328
+ 0.018276351928710938,
329
+ 0.018357248306274415,
330
+ 0.018283519744873047,
331
+ 0.018300928115844727,
332
+ 0.018329599380493163,
333
+ 0.018327743530273437,
334
+ 0.018350912094116212,
335
+ 0.018306047439575195,
336
+ 0.018303071975708008,
337
+ 0.018447263717651367,
338
+ 0.018693119049072265,
339
+ 0.019611648559570313,
340
+ 0.019537919998168944,
341
+ 0.019283967971801756,
342
+ 0.019288063049316406,
343
+ 0.0192675838470459,
344
+ 0.019344383239746094,
345
+ 0.01920204734802246,
346
+ 0.019191808700561523,
347
+ 0.018300928115844727,
348
+ 0.018276384353637695,
349
+ 0.01827734375,
350
+ 0.018249727249145507,
351
+ 0.018265087127685545,
352
+ 0.01829478454589844,
353
+ 0.01926144027709961,
354
+ 0.019195903778076173,
355
+ 0.01841459274291992,
356
+ 0.019406848907470704,
357
+ 0.019317760467529296,
358
+ 0.01943142318725586,
359
+ 0.019363840103149413,
360
+ 0.01924812889099121,
361
+ 0.019401727676391603,
362
+ 0.01927884864807129,
363
+ 0.019342336654663086,
364
+ 0.01943142318725586,
365
+ 0.018487295150756835,
366
+ 0.01846784019470215,
367
+ 0.018463743209838866,
368
+ 0.01985024070739746,
369
+ 0.018431999206542968,
370
+ 0.01843404769897461,
371
+ 0.018489343643188477,
372
+ 0.0186429443359375,
373
+ 0.018515968322753908,
374
+ 0.01839411163330078,
375
+ 0.018501632690429686,
376
+ 0.018431999206542968,
377
+ 0.01841766357421875,
378
+ 0.018515968322753908,
379
+ 0.018423807144165038,
380
+ 0.0184453125,
381
+ 0.018361343383789062,
382
+ 0.01882111930847168,
383
+ 0.01923891258239746,
384
+ 0.01923993682861328,
385
+ 0.019495935440063478,
386
+ 0.019330047607421876,
387
+ 0.01843916893005371,
388
+ 0.01843507194519043,
389
+ 0.01846272087097168,
390
+ 0.01835212707519531,
391
+ 0.01823027229309082,
392
+ 0.019567615509033204,
393
+ 0.019449855804443358,
394
+ 0.019401727676391603,
395
+ 0.019474431991577147,
396
+ 0.019312639236450196,
397
+ 0.018611200332641603,
398
+ 0.01844633674621582,
399
+ 0.018563072204589845,
400
+ 0.018440191268920898,
401
+ 0.018503679275512695,
402
+ 0.01846681594848633,
403
+ 0.018573312759399413,
404
+ 0.018572288513183592,
405
+ 0.018191360473632814,
406
+ 0.018242559432983398,
407
+ 0.01829478454589844,
408
+ 0.01825279998779297,
409
+ 0.01840742492675781,
410
+ 0.01824870491027832,
411
+ 0.018250751495361327,
412
+ 0.018267135620117187,
413
+ 0.018421760559082033,
414
+ 0.01846886444091797,
415
+ 0.019308544158935546,
416
+ 0.01842585563659668,
417
+ 0.01844131278991699,
418
+ 0.018512800216674806,
419
+ 0.018544639587402344,
420
+ 0.01845145606994629,
421
+ 0.018776063919067384,
422
+ 0.018562047958374024,
423
+ 0.018267135620117187,
424
+ 0.018158592224121094,
425
+ 0.01924300765991211,
426
+ 0.01923174476623535,
427
+ 0.01859993553161621,
428
+ 0.018481151580810547,
429
+ 0.01840230369567871,
430
+ 0.0185743350982666,
431
+ 0.018527231216430663,
432
+ 0.01839206314086914
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 52.97239867232251
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1738.227712,
6
+ "max_global_vram": 6793.199616,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6146.752512,
9
+ "max_allocated": 5034.576896
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.045458688735961914,
15
+ "mean": 0.022729344367980957,
16
+ "stdev": 0.002710240364074707,
17
+ "p50": 0.022729344367980957,
18
+ "p90": 0.024897536659240722,
19
+ "p95": 0.02516856069564819,
20
+ "p99": 0.02538537992477417,
21
+ "values": [
22
+ 0.025439584732055664,
23
+ 0.02001910400390625
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 307.9719276839752
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 1738.32192,
37
+ "max_global_vram": 6795.296768,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 6148.849664,
40
+ "max_allocated": 5034.577408
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 4.775644775390625,
46
+ "mean": 2.3878223876953126,
47
+ "stdev": 0.013817749023437464,
48
+ "p50": 2.3878223876953126,
49
+ "p90": 2.3988765869140627,
50
+ "p95": 2.400258361816406,
51
+ "p99": 2.401363781738281,
52
+ "values": [
53
+ 2.40164013671875,
54
+ 2.374004638671875
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 53.186535420072985
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 4.776072187423707,
70
+ "mean": 0.018877755681516624,
71
+ "stdev": 0.0013918402423089098,
72
+ "p50": 0.018550783157348632,
73
+ "p90": 0.019393766403198242,
74
+ "p95": 0.019474841690063475,
75
+ "p99": 0.019828408737182616,
76
+ "values": [
77
+ 0.019301376342773437,
78
+ 0.01923686408996582,
79
+ 0.01925017547607422,
80
+ 0.021184511184692383,
81
+ 0.019475456237792968,
82
+ 0.019316736221313476,
83
+ 0.01929430389404297,
84
+ 0.01925315284729004,
85
+ 0.01922047996520996,
86
+ 0.019180543899536134,
87
+ 0.019198976516723632,
88
+ 0.019284992218017577,
89
+ 0.019174400329589843,
90
+ 0.019170303344726563,
91
+ 0.019153919219970703,
92
+ 0.019397632598876953,
93
+ 0.01929523277282715,
94
+ 0.019167232513427734,
95
+ 0.019268608093261717,
96
+ 0.019173376083374022,
97
+ 0.01922969627380371,
98
+ 0.019238943099975585,
99
+ 0.019377119064331056,
100
+ 0.019252223968505858,
101
+ 0.01922969627380371,
102
+ 0.019163135528564454,
103
+ 0.018884607315063476,
104
+ 0.018998271942138673,
105
+ 0.019182592391967773,
106
+ 0.01945292854309082,
107
+ 0.018923519134521484,
108
+ 0.018922496795654296,
109
+ 0.01905971145629883,
110
+ 0.0192491512298584,
111
+ 0.019371007919311522,
112
+ 0.019307519912719725,
113
+ 0.019194879531860352,
114
+ 0.019191808700561523,
115
+ 0.019338239669799806,
116
+ 0.01924095916748047,
117
+ 0.019377151489257814,
118
+ 0.019173376083374022,
119
+ 0.01925119972229004,
120
+ 0.01918976020812988,
121
+ 0.019183616638183593,
122
+ 0.01902694320678711,
123
+ 0.01904844856262207,
124
+ 0.019186687469482423,
125
+ 0.019594240188598632,
126
+ 0.01940787124633789,
127
+ 0.019204095840454103,
128
+ 0.019792896270751953,
129
+ 0.019451904296875,
130
+ 0.019539968490600586,
131
+ 0.019530752182006835,
132
+ 0.019350528717041016,
133
+ 0.019370016098022462,
134
+ 0.0193525447845459,
135
+ 0.019298303604125978,
136
+ 0.019366912841796875,
137
+ 0.019389440536499023,
138
+ 0.019362815856933592,
139
+ 0.01879347229003906,
140
+ 0.01847603225708008,
141
+ 0.018449407577514648,
142
+ 0.018388992309570314,
143
+ 0.018379776000976563,
144
+ 0.01840742492675781,
145
+ 0.01838489532470703,
146
+ 0.01862553596496582,
147
+ 0.018440191268920898,
148
+ 0.018365440368652345,
149
+ 0.01840947151184082,
150
+ 0.018480127334594726,
151
+ 0.018357248306274415,
152
+ 0.01846272087097168,
153
+ 0.018655231475830078,
154
+ 0.018363391876220703,
155
+ 0.01838591957092285,
156
+ 0.018364479064941406,
157
+ 0.01834796714782715,
158
+ 0.018747392654418944,
159
+ 0.01855897521972656,
160
+ 0.018431999206542968,
161
+ 0.01841766357421875,
162
+ 0.018370559692382812,
163
+ 0.018566144943237304,
164
+ 0.018396160125732423,
165
+ 0.018457599639892578,
166
+ 0.018388992309570314,
167
+ 0.01839308738708496,
168
+ 0.01841459274291992,
169
+ 0.018473983764648438,
170
+ 0.01839411163330078,
171
+ 0.018568191528320312,
172
+ 0.018387968063354493,
173
+ 0.018367488861083983,
174
+ 0.018498559951782227,
175
+ 0.018359296798706053,
176
+ 0.01863680076599121,
177
+ 0.018550783157348632,
178
+ 0.018405376434326173,
179
+ 0.018397184371948243,
180
+ 0.018358272552490236,
181
+ 0.018214912414550782,
182
+ 0.018267135620117187,
183
+ 0.01847500801086426,
184
+ 0.018592767715454102,
185
+ 0.018502656936645507,
186
+ 0.018395135879516602,
187
+ 0.01847091293334961,
188
+ 0.018511871337890624,
189
+ 0.018420736312866212,
190
+ 0.018447359085083007,
191
+ 0.01843507194519043,
192
+ 0.01840640068054199,
193
+ 0.018421760559082033,
194
+ 0.018355199813842774,
195
+ 0.018381824493408205,
196
+ 0.018340864181518556,
197
+ 0.019243104934692383,
198
+ 0.0194169921875,
199
+ 0.01925836753845215,
200
+ 0.01939455986022949,
201
+ 0.019327999114990235,
202
+ 0.01939059257507324,
203
+ 0.03963584136962891,
204
+ 0.018341888427734376,
205
+ 0.01838387107849121,
206
+ 0.018378751754760742,
207
+ 0.01824460792541504,
208
+ 0.01845145606994629,
209
+ 0.018344959259033202,
210
+ 0.018326528549194337,
211
+ 0.0182794246673584,
212
+ 0.01888768005371094,
213
+ 0.01842585563659668,
214
+ 0.01846067237854004,
215
+ 0.018307071685791015,
216
+ 0.018363391876220703,
217
+ 0.018397184371948243,
218
+ 0.018353151321411132,
219
+ 0.01835212707519531,
220
+ 0.019808256149291992,
221
+ 0.019335168838500977,
222
+ 0.019187711715698243,
223
+ 0.019314687728881837,
224
+ 0.018373632431030275,
225
+ 0.018276351928710938,
226
+ 0.018357248306274415,
227
+ 0.018283519744873047,
228
+ 0.018300928115844727,
229
+ 0.018329599380493163,
230
+ 0.018327743530273437,
231
+ 0.018350912094116212,
232
+ 0.018306047439575195,
233
+ 0.018303071975708008,
234
+ 0.018447263717651367,
235
+ 0.018693119049072265,
236
+ 0.019611648559570313,
237
+ 0.019537919998168944,
238
+ 0.019283967971801756,
239
+ 0.019288063049316406,
240
+ 0.0192675838470459,
241
+ 0.019344383239746094,
242
+ 0.01920204734802246,
243
+ 0.019191808700561523,
244
+ 0.018300928115844727,
245
+ 0.018276384353637695,
246
+ 0.01827734375,
247
+ 0.018249727249145507,
248
+ 0.018265087127685545,
249
+ 0.01829478454589844,
250
+ 0.01926144027709961,
251
+ 0.019195903778076173,
252
+ 0.01841459274291992,
253
+ 0.019406848907470704,
254
+ 0.019317760467529296,
255
+ 0.01943142318725586,
256
+ 0.019363840103149413,
257
+ 0.01924812889099121,
258
+ 0.019401727676391603,
259
+ 0.01927884864807129,
260
+ 0.019342336654663086,
261
+ 0.01943142318725586,
262
+ 0.018487295150756835,
263
+ 0.01846784019470215,
264
+ 0.018463743209838866,
265
+ 0.01985024070739746,
266
+ 0.018431999206542968,
267
+ 0.01843404769897461,
268
+ 0.018489343643188477,
269
+ 0.0186429443359375,
270
+ 0.018515968322753908,
271
+ 0.01839411163330078,
272
+ 0.018501632690429686,
273
+ 0.018431999206542968,
274
+ 0.01841766357421875,
275
+ 0.018515968322753908,
276
+ 0.018423807144165038,
277
+ 0.0184453125,
278
+ 0.018361343383789062,
279
+ 0.01882111930847168,
280
+ 0.01923891258239746,
281
+ 0.01923993682861328,
282
+ 0.019495935440063478,
283
+ 0.019330047607421876,
284
+ 0.01843916893005371,
285
+ 0.01843507194519043,
286
+ 0.01846272087097168,
287
+ 0.01835212707519531,
288
+ 0.01823027229309082,
289
+ 0.019567615509033204,
290
+ 0.019449855804443358,
291
+ 0.019401727676391603,
292
+ 0.019474431991577147,
293
+ 0.019312639236450196,
294
+ 0.018611200332641603,
295
+ 0.01844633674621582,
296
+ 0.018563072204589845,
297
+ 0.018440191268920898,
298
+ 0.018503679275512695,
299
+ 0.01846681594848633,
300
+ 0.018573312759399413,
301
+ 0.018572288513183592,
302
+ 0.018191360473632814,
303
+ 0.018242559432983398,
304
+ 0.01829478454589844,
305
+ 0.01825279998779297,
306
+ 0.01840742492675781,
307
+ 0.01824870491027832,
308
+ 0.018250751495361327,
309
+ 0.018267135620117187,
310
+ 0.018421760559082033,
311
+ 0.01846886444091797,
312
+ 0.019308544158935546,
313
+ 0.01842585563659668,
314
+ 0.01844131278991699,
315
+ 0.018512800216674806,
316
+ 0.018544639587402344,
317
+ 0.01845145606994629,
318
+ 0.018776063919067384,
319
+ 0.018562047958374024,
320
+ 0.018267135620117187,
321
+ 0.018158592224121094,
322
+ 0.01924300765991211,
323
+ 0.01923174476623535,
324
+ 0.01859993553161621,
325
+ 0.018481151580810547,
326
+ 0.01840230369567871,
327
+ 0.0185743350982666,
328
+ 0.018527231216430663,
329
+ 0.01839206314086914
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 52.97239867232251
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022729344367980957,
7
+ "prefill.throughput.value": 307.9719276839752,
8
+ "decode.latency.mean": 2.3878223876953126,
9
+ "decode.throughput.value": 53.186535420072985,
10
+ "per_token.latency.mean": 0.018877755681516624,
11
+ "per_token.throughput.value": 52.97239867232251
12
+ }
13
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: eed9ed679878ada2f6d2eefccdbda368cabc88b1
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1
88
+ - hydra.run.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [ISOLATED-PROCESS][2024-06-16 02:35:22,006][process][ERROR] - + Sending traceback to main process
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: eed9ed679878ada2f6d2eefccdbda368cabc88b1
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1
88
+ - hydra.run.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1754.312704,
109
+ "max_global_vram": 6793.199616,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 6146.752512,
112
+ "max_allocated": 5036.936192
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.04457942390441895,
118
+ "mean": 0.022289711952209475,
119
+ "stdev": 0.0002221593856811515,
120
+ "p50": 0.022289711952209475,
121
+ "p90": 0.022467439460754395,
122
+ "p95": 0.02248965539932251,
123
+ "p99": 0.022507428150177002,
124
+ "values": [
125
+ 0.02206755256652832,
126
+ 0.022511871337890624
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 314.04622971388926
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 1754.329088,
140
+ "max_global_vram": 6793.199616,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 6146.752512,
143
+ "max_allocated": 5036.936704
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 5.197515380859375,
149
+ "mean": 2.5987576904296876,
150
+ "stdev": 0.01779040527343745,
151
+ "p50": 2.5987576904296876,
152
+ "p90": 2.6129900146484375,
153
+ "p95": 2.614769055175781,
154
+ "p99": 2.616192287597656,
155
+ "values": [
156
+ 2.616548095703125,
157
+ 2.58096728515625
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 48.86950425108752
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 5.199338497161862,
173
+ "mean": 0.020550745048070616,
174
+ "stdev": 0.001485052481425994,
175
+ "p50": 0.020602752685546875,
176
+ "p90": 0.020843519592285156,
177
+ "p95": 0.020971519470214844,
178
+ "p99": 0.021587354431152343,
179
+ "values": [
180
+ 0.022055936813354493,
181
+ 0.020694015502929687,
182
+ 0.019979263305664064,
183
+ 0.01990553665161133,
184
+ 0.02059775924682617,
185
+ 0.02082099151611328,
186
+ 0.02102579116821289,
187
+ 0.020137983322143553,
188
+ 0.019955711364746095,
189
+ 0.020600831985473633,
190
+ 0.020780031204223632,
191
+ 0.020686847686767578,
192
+ 0.02051584053039551,
193
+ 0.02040012741088867,
194
+ 0.020332544326782227,
195
+ 0.02021990394592285,
196
+ 0.020479999542236327,
197
+ 0.020752384185791017,
198
+ 0.02001919937133789,
199
+ 0.020717567443847656,
200
+ 0.021572608947753907,
201
+ 0.02086604881286621,
202
+ 0.019985408782958985,
203
+ 0.020007936477661133,
204
+ 0.019961856842041017,
205
+ 0.019961952209472656,
206
+ 0.019948448181152344,
207
+ 0.02004991912841797,
208
+ 0.019994623184204103,
209
+ 0.019919872283935547,
210
+ 0.02004275131225586,
211
+ 0.019975168228149414,
212
+ 0.020024320602416993,
213
+ 0.02005401611328125,
214
+ 0.02003865623474121,
215
+ 0.020445184707641603,
216
+ 0.021414911270141602,
217
+ 0.02069606399536133,
218
+ 0.020756479263305663,
219
+ 0.020145151138305666,
220
+ 0.02066022491455078,
221
+ 0.019916799545288084,
222
+ 0.020051967620849608,
223
+ 0.020001792907714845,
224
+ 0.019931135177612306,
225
+ 0.02084556770324707,
226
+ 0.020786176681518553,
227
+ 0.02085785675048828,
228
+ 0.021021696090698243,
229
+ 0.020971519470214844,
230
+ 0.02101043128967285,
231
+ 0.020937728881835937,
232
+ 0.02066329574584961,
233
+ 0.020626432418823244,
234
+ 0.020711423873901368,
235
+ 0.02061516761779785,
236
+ 0.020723712921142577,
237
+ 0.020716543197631835,
238
+ 0.02078108787536621,
239
+ 0.020603872299194335,
240
+ 0.020928512573242186,
241
+ 0.020752384185791017,
242
+ 0.020799488067626954,
243
+ 0.02058956718444824,
244
+ 0.02064896011352539,
245
+ 0.020700159072875975,
246
+ 0.020785152435302736,
247
+ 0.020585472106933594,
248
+ 0.02057318305969238,
249
+ 0.021061632156372072,
250
+ 0.02064793586730957,
251
+ 0.020445184707641603,
252
+ 0.0206878719329834,
253
+ 0.020585472106933594,
254
+ 0.020750335693359375,
255
+ 0.02064384078979492,
256
+ 0.02071244812011719,
257
+ 0.02065510368347168,
258
+ 0.0208353271484375,
259
+ 0.020971519470214844,
260
+ 0.02082508850097656,
261
+ 0.02084556770324707,
262
+ 0.020787200927734374,
263
+ 0.02068070411682129,
264
+ 0.020718591690063477,
265
+ 0.020758527755737305,
266
+ 0.02064384078979492,
267
+ 0.02066431999206543,
268
+ 0.020720640182495118,
269
+ 0.02058457565307617,
270
+ 0.020568960189819335,
271
+ 0.02066739273071289,
272
+ 0.020582399368286132,
273
+ 0.020653055191040038,
274
+ 0.020580352783203124,
275
+ 0.02030899238586426,
276
+ 0.020462591171264647,
277
+ 0.02067558479309082,
278
+ 0.020900863647460938,
279
+ 0.020586496353149415,
280
+ 0.02063871955871582,
281
+ 0.020588544845581053,
282
+ 0.020694015502929687,
283
+ 0.020793344497680662,
284
+ 0.020633600234985353,
285
+ 0.02063564872741699,
286
+ 0.020730880737304686,
287
+ 0.020717567443847656,
288
+ 0.020810752868652343,
289
+ 0.02064896011352539,
290
+ 0.020759552001953126,
291
+ 0.02069708824157715,
292
+ 0.020747264862060546,
293
+ 0.02064691162109375,
294
+ 0.020892704010009765,
295
+ 0.020643808364868163,
296
+ 0.020619264602661135,
297
+ 0.020603904724121092,
298
+ 0.020737024307250978,
299
+ 0.020754432678222655,
300
+ 0.021540864944458008,
301
+ 0.020923391342163086,
302
+ 0.020731903076171874,
303
+ 0.020741119384765624,
304
+ 0.020702207565307617,
305
+ 0.020647104263305665,
306
+ 0.04330271911621094,
307
+ 0.020743167877197266,
308
+ 0.02065715217590332,
309
+ 0.020685823440551757,
310
+ 0.02063974380493164,
311
+ 0.020798463821411133,
312
+ 0.02063667106628418,
313
+ 0.02064384078979492,
314
+ 0.020649984359741212,
315
+ 0.020937728881835937,
316
+ 0.020641792297363282,
317
+ 0.02062131118774414,
318
+ 0.020677631378173827,
319
+ 0.02062233543395996,
320
+ 0.020694015502929687,
321
+ 0.02067046356201172,
322
+ 0.020164608001708984,
323
+ 0.020331520080566406,
324
+ 0.020706304550170897,
325
+ 0.020956159591674805,
326
+ 0.020792320251464845,
327
+ 0.020684799194335936,
328
+ 0.0206561279296875,
329
+ 0.02088960075378418,
330
+ 0.021603328704833984,
331
+ 0.021173248291015623,
332
+ 0.02069811248779297,
333
+ 0.02063155174255371,
334
+ 0.020591615676879883,
335
+ 0.020577280044555665,
336
+ 0.020642816543579103,
337
+ 0.02067251205444336,
338
+ 0.020607999801635742,
339
+ 0.020642816543579103,
340
+ 0.02062950325012207,
341
+ 0.020586496353149415,
342
+ 0.020593664169311524,
343
+ 0.020596736907958983,
344
+ 0.020556800842285155,
345
+ 0.020550655364990233,
346
+ 0.02002022361755371,
347
+ 0.01993744087219238,
348
+ 0.019947359085083008,
349
+ 0.020017152786254884,
350
+ 0.019968191146850587,
351
+ 0.020066112518310548,
352
+ 0.02004172706604004,
353
+ 0.019969024658203126,
354
+ 0.019978239059448243,
355
+ 0.019927040100097656,
356
+ 0.019987455368041994,
357
+ 0.020668415069580077,
358
+ 0.02059775924682617,
359
+ 0.02081996726989746,
360
+ 0.020690944671630858,
361
+ 0.02069708824157715,
362
+ 0.020674560546875,
363
+ 0.020831232070922853,
364
+ 0.02067148780822754,
365
+ 0.02065407943725586,
366
+ 0.020797439575195312,
367
+ 0.020032512664794923,
368
+ 0.020057088851928712,
369
+ 0.020000768661499024,
370
+ 0.019999744415283204,
371
+ 0.020031648635864256,
372
+ 0.020076383590698244,
373
+ 0.020190208435058594,
374
+ 0.019983360290527344,
375
+ 0.020067327499389647,
376
+ 0.019992576599121094,
377
+ 0.020040704727172853,
378
+ 0.02002124786376953,
379
+ 0.020009983062744142,
380
+ 0.020000768661499024,
381
+ 0.02001203155517578,
382
+ 0.019979263305664064,
383
+ 0.019984384536743165,
384
+ 0.02002022361755371,
385
+ 0.019984384536743165,
386
+ 0.0200263671875,
387
+ 0.02006118392944336,
388
+ 0.020057088851928712,
389
+ 0.019961856842041017,
390
+ 0.020031488418579102,
391
+ 0.01999772834777832,
392
+ 0.020003807067871093,
393
+ 0.020040704727172853,
394
+ 0.0200130558013916,
395
+ 0.02006220817565918,
396
+ 0.019941375732421874,
397
+ 0.020133888244628906,
398
+ 0.020068351745605468,
399
+ 0.020110336303710938,
400
+ 0.019969024658203126,
401
+ 0.02001728057861328,
402
+ 0.019996639251708984,
403
+ 0.01997916793823242,
404
+ 0.019981439590454102,
405
+ 0.020602752685546875,
406
+ 0.020566015243530272,
407
+ 0.02006630325317383,
408
+ 0.019968000411987305,
409
+ 0.019971103668212892,
410
+ 0.019940319061279296,
411
+ 0.020173824310302735,
412
+ 0.02000486373901367,
413
+ 0.020015104293823242,
414
+ 0.019986431121826173,
415
+ 0.019958784103393554,
416
+ 0.019992576599121094,
417
+ 0.019978239059448243,
418
+ 0.019976192474365235,
419
+ 0.020053087234497072,
420
+ 0.02015635108947754,
421
+ 0.019811296463012697,
422
+ 0.019877887725830077,
423
+ 0.019956735610961913,
424
+ 0.019968000411987305,
425
+ 0.019945472717285157,
426
+ 0.02008883285522461,
427
+ 0.02126950454711914,
428
+ 0.02070425605773926,
429
+ 0.020592639923095703,
430
+ 0.02065510368347168,
431
+ 0.020610048294067384,
432
+ 0.020558847427368163
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 48.66003629848368
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1754.312704,
6
+ "max_global_vram": 6793.199616,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6146.752512,
9
+ "max_allocated": 5036.936192
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.04457942390441895,
15
+ "mean": 0.022289711952209475,
16
+ "stdev": 0.0002221593856811515,
17
+ "p50": 0.022289711952209475,
18
+ "p90": 0.022467439460754395,
19
+ "p95": 0.02248965539932251,
20
+ "p99": 0.022507428150177002,
21
+ "values": [
22
+ 0.02206755256652832,
23
+ 0.022511871337890624
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 314.04622971388926
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 1754.329088,
37
+ "max_global_vram": 6793.199616,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 6146.752512,
40
+ "max_allocated": 5036.936704
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 5.197515380859375,
46
+ "mean": 2.5987576904296876,
47
+ "stdev": 0.01779040527343745,
48
+ "p50": 2.5987576904296876,
49
+ "p90": 2.6129900146484375,
50
+ "p95": 2.614769055175781,
51
+ "p99": 2.616192287597656,
52
+ "values": [
53
+ 2.616548095703125,
54
+ 2.58096728515625
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 48.86950425108752
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 5.199338497161862,
70
+ "mean": 0.020550745048070616,
71
+ "stdev": 0.001485052481425994,
72
+ "p50": 0.020602752685546875,
73
+ "p90": 0.020843519592285156,
74
+ "p95": 0.020971519470214844,
75
+ "p99": 0.021587354431152343,
76
+ "values": [
77
+ 0.022055936813354493,
78
+ 0.020694015502929687,
79
+ 0.019979263305664064,
80
+ 0.01990553665161133,
81
+ 0.02059775924682617,
82
+ 0.02082099151611328,
83
+ 0.02102579116821289,
84
+ 0.020137983322143553,
85
+ 0.019955711364746095,
86
+ 0.020600831985473633,
87
+ 0.020780031204223632,
88
+ 0.020686847686767578,
89
+ 0.02051584053039551,
90
+ 0.02040012741088867,
91
+ 0.020332544326782227,
92
+ 0.02021990394592285,
93
+ 0.020479999542236327,
94
+ 0.020752384185791017,
95
+ 0.02001919937133789,
96
+ 0.020717567443847656,
97
+ 0.021572608947753907,
98
+ 0.02086604881286621,
99
+ 0.019985408782958985,
100
+ 0.020007936477661133,
101
+ 0.019961856842041017,
102
+ 0.019961952209472656,
103
+ 0.019948448181152344,
104
+ 0.02004991912841797,
105
+ 0.019994623184204103,
106
+ 0.019919872283935547,
107
+ 0.02004275131225586,
108
+ 0.019975168228149414,
109
+ 0.020024320602416993,
110
+ 0.02005401611328125,
111
+ 0.02003865623474121,
112
+ 0.020445184707641603,
113
+ 0.021414911270141602,
114
+ 0.02069606399536133,
115
+ 0.020756479263305663,
116
+ 0.020145151138305666,
117
+ 0.02066022491455078,
118
+ 0.019916799545288084,
119
+ 0.020051967620849608,
120
+ 0.020001792907714845,
121
+ 0.019931135177612306,
122
+ 0.02084556770324707,
123
+ 0.020786176681518553,
124
+ 0.02085785675048828,
125
+ 0.021021696090698243,
126
+ 0.020971519470214844,
127
+ 0.02101043128967285,
128
+ 0.020937728881835937,
129
+ 0.02066329574584961,
130
+ 0.020626432418823244,
131
+ 0.020711423873901368,
132
+ 0.02061516761779785,
133
+ 0.020723712921142577,
134
+ 0.020716543197631835,
135
+ 0.02078108787536621,
136
+ 0.020603872299194335,
137
+ 0.020928512573242186,
138
+ 0.020752384185791017,
139
+ 0.020799488067626954,
140
+ 0.02058956718444824,
141
+ 0.02064896011352539,
142
+ 0.020700159072875975,
143
+ 0.020785152435302736,
144
+ 0.020585472106933594,
145
+ 0.02057318305969238,
146
+ 0.021061632156372072,
147
+ 0.02064793586730957,
148
+ 0.020445184707641603,
149
+ 0.0206878719329834,
150
+ 0.020585472106933594,
151
+ 0.020750335693359375,
152
+ 0.02064384078979492,
153
+ 0.02071244812011719,
154
+ 0.02065510368347168,
155
+ 0.0208353271484375,
156
+ 0.020971519470214844,
157
+ 0.02082508850097656,
158
+ 0.02084556770324707,
159
+ 0.020787200927734374,
160
+ 0.02068070411682129,
161
+ 0.020718591690063477,
162
+ 0.020758527755737305,
163
+ 0.02064384078979492,
164
+ 0.02066431999206543,
165
+ 0.020720640182495118,
166
+ 0.02058457565307617,
167
+ 0.020568960189819335,
168
+ 0.02066739273071289,
169
+ 0.020582399368286132,
170
+ 0.020653055191040038,
171
+ 0.020580352783203124,
172
+ 0.02030899238586426,
173
+ 0.020462591171264647,
174
+ 0.02067558479309082,
175
+ 0.020900863647460938,
176
+ 0.020586496353149415,
177
+ 0.02063871955871582,
178
+ 0.020588544845581053,
179
+ 0.020694015502929687,
180
+ 0.020793344497680662,
181
+ 0.020633600234985353,
182
+ 0.02063564872741699,
183
+ 0.020730880737304686,
184
+ 0.020717567443847656,
185
+ 0.020810752868652343,
186
+ 0.02064896011352539,
187
+ 0.020759552001953126,
188
+ 0.02069708824157715,
189
+ 0.020747264862060546,
190
+ 0.02064691162109375,
191
+ 0.020892704010009765,
192
+ 0.020643808364868163,
193
+ 0.020619264602661135,
194
+ 0.020603904724121092,
195
+ 0.020737024307250978,
196
+ 0.020754432678222655,
197
+ 0.021540864944458008,
198
+ 0.020923391342163086,
199
+ 0.020731903076171874,
200
+ 0.020741119384765624,
201
+ 0.020702207565307617,
202
+ 0.020647104263305665,
203
+ 0.04330271911621094,
204
+ 0.020743167877197266,
205
+ 0.02065715217590332,
206
+ 0.020685823440551757,
207
+ 0.02063974380493164,
208
+ 0.020798463821411133,
209
+ 0.02063667106628418,
210
+ 0.02064384078979492,
211
+ 0.020649984359741212,
212
+ 0.020937728881835937,
213
+ 0.020641792297363282,
214
+ 0.02062131118774414,
215
+ 0.020677631378173827,
216
+ 0.02062233543395996,
217
+ 0.020694015502929687,
218
+ 0.02067046356201172,
219
+ 0.020164608001708984,
220
+ 0.020331520080566406,
221
+ 0.020706304550170897,
222
+ 0.020956159591674805,
223
+ 0.020792320251464845,
224
+ 0.020684799194335936,
225
+ 0.0206561279296875,
226
+ 0.02088960075378418,
227
+ 0.021603328704833984,
228
+ 0.021173248291015623,
229
+ 0.02069811248779297,
230
+ 0.02063155174255371,
231
+ 0.020591615676879883,
232
+ 0.020577280044555665,
233
+ 0.020642816543579103,
234
+ 0.02067251205444336,
235
+ 0.020607999801635742,
236
+ 0.020642816543579103,
237
+ 0.02062950325012207,
238
+ 0.020586496353149415,
239
+ 0.020593664169311524,
240
+ 0.020596736907958983,
241
+ 0.020556800842285155,
242
+ 0.020550655364990233,
243
+ 0.02002022361755371,
244
+ 0.01993744087219238,
245
+ 0.019947359085083008,
246
+ 0.020017152786254884,
247
+ 0.019968191146850587,
248
+ 0.020066112518310548,
249
+ 0.02004172706604004,
250
+ 0.019969024658203126,
251
+ 0.019978239059448243,
252
+ 0.019927040100097656,
253
+ 0.019987455368041994,
254
+ 0.020668415069580077,
255
+ 0.02059775924682617,
256
+ 0.02081996726989746,
257
+ 0.020690944671630858,
258
+ 0.02069708824157715,
259
+ 0.020674560546875,
260
+ 0.020831232070922853,
261
+ 0.02067148780822754,
262
+ 0.02065407943725586,
263
+ 0.020797439575195312,
264
+ 0.020032512664794923,
265
+ 0.020057088851928712,
266
+ 0.020000768661499024,
267
+ 0.019999744415283204,
268
+ 0.020031648635864256,
269
+ 0.020076383590698244,
270
+ 0.020190208435058594,
271
+ 0.019983360290527344,
272
+ 0.020067327499389647,
273
+ 0.019992576599121094,
274
+ 0.020040704727172853,
275
+ 0.02002124786376953,
276
+ 0.020009983062744142,
277
+ 0.020000768661499024,
278
+ 0.02001203155517578,
279
+ 0.019979263305664064,
280
+ 0.019984384536743165,
281
+ 0.02002022361755371,
282
+ 0.019984384536743165,
283
+ 0.0200263671875,
284
+ 0.02006118392944336,
285
+ 0.020057088851928712,
286
+ 0.019961856842041017,
287
+ 0.020031488418579102,
288
+ 0.01999772834777832,
289
+ 0.020003807067871093,
290
+ 0.020040704727172853,
291
+ 0.0200130558013916,
292
+ 0.02006220817565918,
293
+ 0.019941375732421874,
294
+ 0.020133888244628906,
295
+ 0.020068351745605468,
296
+ 0.020110336303710938,
297
+ 0.019969024658203126,
298
+ 0.02001728057861328,
299
+ 0.019996639251708984,
300
+ 0.01997916793823242,
301
+ 0.019981439590454102,
302
+ 0.020602752685546875,
303
+ 0.020566015243530272,
304
+ 0.02006630325317383,
305
+ 0.019968000411987305,
306
+ 0.019971103668212892,
307
+ 0.019940319061279296,
308
+ 0.020173824310302735,
309
+ 0.02000486373901367,
310
+ 0.020015104293823242,
311
+ 0.019986431121826173,
312
+ 0.019958784103393554,
313
+ 0.019992576599121094,
314
+ 0.019978239059448243,
315
+ 0.019976192474365235,
316
+ 0.020053087234497072,
317
+ 0.02015635108947754,
318
+ 0.019811296463012697,
319
+ 0.019877887725830077,
320
+ 0.019956735610961913,
321
+ 0.019968000411987305,
322
+ 0.019945472717285157,
323
+ 0.02008883285522461,
324
+ 0.02126950454711914,
325
+ 0.02070425605773926,
326
+ 0.020592639923095703,
327
+ 0.02065510368347168,
328
+ 0.020610048294067384,
329
+ 0.020558847427368163
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 48.66003629848368
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022289711952209475,
7
+ "prefill.throughput.value": 314.04622971388926,
8
+ "decode.latency.mean": 2.5987576904296876,
9
+ "decode.throughput.value": 48.86950425108752,
10
+ "per_token.latency.mean": 0.020550745048070616,
11
+ "per_token.throughput.value": 48.66003629848368
12
+ }
13
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.3.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model: google/gemma-2b
9
+ processor: null
10
+ device: cuda
11
+ device_ids: '0'
12
+ seed: 42
13
+ inter_op_num_threads: null
14
+ intra_op_num_threads: null
15
+ model_kwargs: {}
16
+ processor_kwargs: {}
17
+ hub_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ latency: true
51
+ memory: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ start_method: spawn
65
+ environment:
66
+ cpu: ' AMD EPYC 7R32'
67
+ cpu_count: 16
68
+ cpu_ram_mb: 66697.29792
69
+ system: Linux
70
+ machine: x86_64
71
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
72
+ processor: x86_64
73
+ python_version: 3.8.10
74
+ gpu:
75
+ - NVIDIA A10G
76
+ gpu_count: 1
77
+ gpu_vram_mb: 24146608128
78
+ optimum_benchmark_version: 0.2.1
79
+ optimum_benchmark_commit: null
80
+ transformers_version: 4.42.0.dev0
81
+ transformers_commit: eed9ed679878ada2f6d2eefccdbda368cabc88b1
82
+ accelerate_version: 0.32.0.dev0
83
+ accelerate_commit: null
84
+ diffusers_version: null
85
+ diffusers_commit: null
86
+ optimum_version: 1.21.0.dev0
87
+ optimum_commit: null
88
+ timm_version: 0.9.16
89
+ timm_commit: null
90
+ peft_version: 0.11.2.dev0
91
+ peft_commit: null
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1
88
+ - hydra.run.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,443 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "latency": true,
57
+ "memory": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "start_method": "spawn"
73
+ },
74
+ "environment": {
75
+ "cpu": " AMD EPYC 7R32",
76
+ "cpu_count": 16,
77
+ "cpu_ram_mb": 66697.29792,
78
+ "system": "Linux",
79
+ "machine": "x86_64",
80
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
81
+ "processor": "x86_64",
82
+ "python_version": "3.8.10",
83
+ "gpu": [
84
+ "NVIDIA A10G"
85
+ ],
86
+ "gpu_count": 1,
87
+ "gpu_vram_mb": 24146608128,
88
+ "optimum_benchmark_version": "0.2.1",
89
+ "optimum_benchmark_commit": null,
90
+ "transformers_version": "4.42.0.dev0",
91
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
92
+ "accelerate_version": "0.32.0.dev0",
93
+ "accelerate_commit": null,
94
+ "diffusers_version": null,
95
+ "diffusers_commit": null,
96
+ "optimum_version": "1.21.0.dev0",
97
+ "optimum_commit": null,
98
+ "timm_version": "0.9.16",
99
+ "timm_commit": null,
100
+ "peft_version": "0.11.2.dev0",
101
+ "peft_commit": null
102
+ }
103
+ },
104
+ "report": {
105
+ "prefill": {
106
+ "memory": {
107
+ "unit": "MB",
108
+ "max_ram": 1928.871936,
109
+ "max_global_vram": 6094.848,
110
+ "max_process_vram": 0.0,
111
+ "max_reserved": 5379.19488,
112
+ "max_allocated": 5021.251584
113
+ },
114
+ "latency": {
115
+ "unit": "s",
116
+ "count": 2,
117
+ "total": 0.028069279670715333,
118
+ "mean": 0.014034639835357667,
119
+ "stdev": 0.00011380815505981478,
120
+ "p50": 0.014034639835357667,
121
+ "p90": 0.014125686359405518,
122
+ "p95": 0.014137067174911498,
123
+ "p99": 0.014146171827316284,
124
+ "values": [
125
+ 0.01414844799041748,
126
+ 0.013920831680297851
127
+ ]
128
+ },
129
+ "throughput": {
130
+ "unit": "tokens/s",
131
+ "value": 498.76591648364223
132
+ },
133
+ "energy": null,
134
+ "efficiency": null
135
+ },
136
+ "decode": {
137
+ "memory": {
138
+ "unit": "MB",
139
+ "max_ram": 2272.821248,
140
+ "max_global_vram": 8907.128832,
141
+ "max_process_vram": 0.0,
142
+ "max_reserved": 5379.19488,
143
+ "max_allocated": 5027.089408
144
+ },
145
+ "latency": {
146
+ "unit": "s",
147
+ "count": 2,
148
+ "total": 3.1459737548828124,
149
+ "mean": 1.5729868774414062,
150
+ "stdev": 0.017520202636718718,
151
+ "p50": 1.5729868774414062,
152
+ "p90": 1.5870030395507813,
153
+ "p95": 1.588755059814453,
154
+ "p99": 1.5901566760253905,
155
+ "values": [
156
+ 1.590507080078125,
157
+ 1.5554666748046875
158
+ ]
159
+ },
160
+ "throughput": {
161
+ "unit": "tokens/s",
162
+ "value": 80.73811792160406
163
+ },
164
+ "energy": null,
165
+ "efficiency": null
166
+ },
167
+ "per_token": {
168
+ "memory": null,
169
+ "latency": {
170
+ "unit": "s",
171
+ "count": 253,
172
+ "total": 3.146998783111571,
173
+ "mean": 0.012438730368029931,
174
+ "stdev": 0.0009024142000934347,
175
+ "p50": 0.01226035213470459,
176
+ "p90": 0.012727705764770508,
177
+ "p95": 0.012750642967224122,
178
+ "p99": 0.012817571563720704,
179
+ "values": [
180
+ 0.012702848434448242,
181
+ 0.01272000026702881,
182
+ 0.01273855972290039,
183
+ 0.012786687850952149,
184
+ 0.01268019199371338,
185
+ 0.01277337646484375,
186
+ 0.012697600364685058,
187
+ 0.01276518440246582,
188
+ 0.012718079566955566,
189
+ 0.012714079856872559,
190
+ 0.012702624320983886,
191
+ 0.012748991966247559,
192
+ 0.012663616180419921,
193
+ 0.012716032028198243,
194
+ 0.012667903900146485,
195
+ 0.012751872062683106,
196
+ 0.012678144454956054,
197
+ 0.012725248336791992,
198
+ 0.012710911750793457,
199
+ 0.012755999565124512,
200
+ 0.012680159568786622,
201
+ 0.012722175598144531,
202
+ 0.012695551872253418,
203
+ 0.012761088371276855,
204
+ 0.01266585636138916,
205
+ 0.012809215545654297,
206
+ 0.012719103813171387,
207
+ 0.01274060821533203,
208
+ 0.012686528205871583,
209
+ 0.012731200218200683,
210
+ 0.01267404842376709,
211
+ 0.012793855667114258,
212
+ 0.012689408302307128,
213
+ 0.012774399757385254,
214
+ 0.012735487937927246,
215
+ 0.012826623916625977,
216
+ 0.012563455581665038,
217
+ 0.012733440399169921,
218
+ 0.012717056274414062,
219
+ 0.012746848106384278,
220
+ 0.012685215950012207,
221
+ 0.012676095962524414,
222
+ 0.012676095962524414,
223
+ 0.012728320121765137,
224
+ 0.012695551872253418,
225
+ 0.012716032028198243,
226
+ 0.012670975685119629,
227
+ 0.01273855972290039,
228
+ 0.01268019199371338,
229
+ 0.012930047988891602,
230
+ 0.012720128059387208,
231
+ 0.012720128059387208,
232
+ 0.012700672149658204,
233
+ 0.012706815719604492,
234
+ 0.01265561580657959,
235
+ 0.01273036766052246,
236
+ 0.012693504333496093,
237
+ 0.012698623657226562,
238
+ 0.012654591560363769,
239
+ 0.012670975685119629,
240
+ 0.012711935997009278,
241
+ 0.012749823570251465,
242
+ 0.012804096221923827,
243
+ 0.012733440399169921,
244
+ 0.012725248336791992,
245
+ 0.01273036766052246,
246
+ 0.012526592254638673,
247
+ 0.012608511924743653,
248
+ 0.012646400451660156,
249
+ 0.012578816413879394,
250
+ 0.012279808044433594,
251
+ 0.012463104248046876,
252
+ 0.012502016067504883,
253
+ 0.012577792167663575,
254
+ 0.01263923168182373,
255
+ 0.012444671630859374,
256
+ 0.012281951904296875,
257
+ 0.012396448135375977,
258
+ 0.012320768356323243,
259
+ 0.012432383537292481,
260
+ 0.012286975860595703,
261
+ 0.012257280349731446,
262
+ 0.012223487854003906,
263
+ 0.012266688346862793,
264
+ 0.012239680290222169,
265
+ 0.012272640228271485,
266
+ 0.012215295791625976,
267
+ 0.012274687767028808,
268
+ 0.012215295791625976,
269
+ 0.01226035213470459,
270
+ 0.012224512100219726,
271
+ 0.012255231857299804,
272
+ 0.012220416069030762,
273
+ 0.012255231857299804,
274
+ 0.012224512100219726,
275
+ 0.012290047645568849,
276
+ 0.012247039794921874,
277
+ 0.012259327888488769,
278
+ 0.012212224006652832,
279
+ 0.012254207611083985,
280
+ 0.012223487854003906,
281
+ 0.01226035213470459,
282
+ 0.012217344284057617,
283
+ 0.012273664474487305,
284
+ 0.012238847732543945,
285
+ 0.012257280349731446,
286
+ 0.012231679916381836,
287
+ 0.012339200019836426,
288
+ 0.01223475170135498,
289
+ 0.012254207611083985,
290
+ 0.012225536346435547,
291
+ 0.01227673625946045,
292
+ 0.012232704162597656,
293
+ 0.012256256103515625,
294
+ 0.012223487854003906,
295
+ 0.012259327888488769,
296
+ 0.012220416069030762,
297
+ 0.012266495704650878,
298
+ 0.012229632377624512,
299
+ 0.012267519950866699,
300
+ 0.01222758388519287,
301
+ 0.01225011157989502,
302
+ 0.01221939182281494,
303
+ 0.012266495704650878,
304
+ 0.012230655670166016,
305
+ 0.01225830364227295,
306
+ 0.02636595153808594,
307
+ 0.012233728408813477,
308
+ 0.012281855583190919,
309
+ 0.012228608131408691,
310
+ 0.012257280349731446,
311
+ 0.012213248252868653,
312
+ 0.01226137638092041,
313
+ 0.012241920471191407,
314
+ 0.01226035213470459,
315
+ 0.01224396800994873,
316
+ 0.012295167922973632,
317
+ 0.012223487854003906,
318
+ 0.012262399673461915,
319
+ 0.012223487854003906,
320
+ 0.012273664474487305,
321
+ 0.012224512100219726,
322
+ 0.012255231857299804,
323
+ 0.012224512100219726,
324
+ 0.01232588768005371,
325
+ 0.012223487854003906,
326
+ 0.012257280349731446,
327
+ 0.012220416069030762,
328
+ 0.012271615982055664,
329
+ 0.012215295791625976,
330
+ 0.012345343589782716,
331
+ 0.012232704162597656,
332
+ 0.012265472412109376,
333
+ 0.012230655670166016,
334
+ 0.012257280349731446,
335
+ 0.012215295791625976,
336
+ 0.01225216007232666,
337
+ 0.012212224006652832,
338
+ 0.012275712013244629,
339
+ 0.012225536346435547,
340
+ 0.012267519950866699,
341
+ 0.012218367576599122,
342
+ 0.012272640228271485,
343
+ 0.012213248252868653,
344
+ 0.01225113582611084,
345
+ 0.012216320037841797,
346
+ 0.01225113582611084,
347
+ 0.012238847732543945,
348
+ 0.012271615982055664,
349
+ 0.012209152221679688,
350
+ 0.01226854419708252,
351
+ 0.012216320037841797,
352
+ 0.012274687767028808,
353
+ 0.012232704162597656,
354
+ 0.012247039794921874,
355
+ 0.012209152221679688,
356
+ 0.01225830364227295,
357
+ 0.012216320037841797,
358
+ 0.012275712013244629,
359
+ 0.012215295791625976,
360
+ 0.01225113582611084,
361
+ 0.012225536346435547,
362
+ 0.012255231857299804,
363
+ 0.012211199760437011,
364
+ 0.012281855583190919,
365
+ 0.012246015548706055,
366
+ 0.012257280349731446,
367
+ 0.01222758388519287,
368
+ 0.012254207611083985,
369
+ 0.012231679916381836,
370
+ 0.012247039794921874,
371
+ 0.012204031944274902,
372
+ 0.012264448165893555,
373
+ 0.012353535652160644,
374
+ 0.012257280349731446,
375
+ 0.01221939182281494,
376
+ 0.012254207611083985,
377
+ 0.012217344284057617,
378
+ 0.01225113582611084,
379
+ 0.012236800193786621,
380
+ 0.012274687767028808,
381
+ 0.012229632377624512,
382
+ 0.012255231857299804,
383
+ 0.012215295791625976,
384
+ 0.012256256103515625,
385
+ 0.012208127975463867,
386
+ 0.012289024353027344,
387
+ 0.012213248252868653,
388
+ 0.012264448165893555,
389
+ 0.012249088287353516,
390
+ 0.01226137638092041,
391
+ 0.012240896224975586,
392
+ 0.01225113582611084,
393
+ 0.012221440315246582,
394
+ 0.01225113582611084,
395
+ 0.012217344284057617,
396
+ 0.012278783798217773,
397
+ 0.0122357759475708,
398
+ 0.01225113582611084,
399
+ 0.012229632377624512,
400
+ 0.012264448165893555,
401
+ 0.01222662353515625,
402
+ 0.012264384269714355,
403
+ 0.012211199760437011,
404
+ 0.012344320297241211,
405
+ 0.01224396800994873,
406
+ 0.012275712013244629,
407
+ 0.012233728408813477,
408
+ 0.012266495704650878,
409
+ 0.012211199760437011,
410
+ 0.01225011157989502,
411
+ 0.012208127975463867,
412
+ 0.012281855583190919,
413
+ 0.012225536346435547,
414
+ 0.012249088287353516,
415
+ 0.012204031944274902,
416
+ 0.012247039794921874,
417
+ 0.012239871978759765,
418
+ 0.01226956844329834,
419
+ 0.0122357759475708,
420
+ 0.012249088287353516,
421
+ 0.012239871978759765,
422
+ 0.012267519950866699,
423
+ 0.012215295791625976,
424
+ 0.012256256103515625,
425
+ 0.012211199760437011,
426
+ 0.01225216007232666,
427
+ 0.012231679916381836,
428
+ 0.01225830364227295,
429
+ 0.01223475170135498,
430
+ 0.012279808044433594,
431
+ 0.012216320037841797,
432
+ 0.012289024353027344
433
+ ]
434
+ },
435
+ "throughput": {
436
+ "unit": "tokens/s",
437
+ "value": 80.39405714350104
438
+ },
439
+ "energy": null,
440
+ "efficiency": null
441
+ }
442
+ }
443
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.3.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model": "google/gemma-2b",
10
+ "processor": "google/gemma-2b",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "model_kwargs": {},
17
+ "processor_kwargs": {},
18
+ "hub_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "latency": true,
56
+ "memory": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.29792,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29",
80
+ "processor": "x86_64",
81
+ "python_version": "3.8.10",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.0.dev0",
90
+ "transformers_commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
91
+ "accelerate_version": "0.32.0.dev0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": null,
94
+ "diffusers_commit": null,
95
+ "optimum_version": "1.21.0.dev0",
96
+ "optimum_commit": null,
97
+ "timm_version": "0.9.16",
98
+ "timm_commit": null,
99
+ "peft_version": "0.11.2.dev0",
100
+ "peft_commit": null
101
+ }
102
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,339 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1928.871936,
6
+ "max_global_vram": 6094.848,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 5379.19488,
9
+ "max_allocated": 5021.251584
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 2,
14
+ "total": 0.028069279670715333,
15
+ "mean": 0.014034639835357667,
16
+ "stdev": 0.00011380815505981478,
17
+ "p50": 0.014034639835357667,
18
+ "p90": 0.014125686359405518,
19
+ "p95": 0.014137067174911498,
20
+ "p99": 0.014146171827316284,
21
+ "values": [
22
+ 0.01414844799041748,
23
+ 0.013920831680297851
24
+ ]
25
+ },
26
+ "throughput": {
27
+ "unit": "tokens/s",
28
+ "value": 498.76591648364223
29
+ },
30
+ "energy": null,
31
+ "efficiency": null
32
+ },
33
+ "decode": {
34
+ "memory": {
35
+ "unit": "MB",
36
+ "max_ram": 2272.821248,
37
+ "max_global_vram": 8907.128832,
38
+ "max_process_vram": 0.0,
39
+ "max_reserved": 5379.19488,
40
+ "max_allocated": 5027.089408
41
+ },
42
+ "latency": {
43
+ "unit": "s",
44
+ "count": 2,
45
+ "total": 3.1459737548828124,
46
+ "mean": 1.5729868774414062,
47
+ "stdev": 0.017520202636718718,
48
+ "p50": 1.5729868774414062,
49
+ "p90": 1.5870030395507813,
50
+ "p95": 1.588755059814453,
51
+ "p99": 1.5901566760253905,
52
+ "values": [
53
+ 1.590507080078125,
54
+ 1.5554666748046875
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "tokens/s",
59
+ "value": 80.73811792160406
60
+ },
61
+ "energy": null,
62
+ "efficiency": null
63
+ },
64
+ "per_token": {
65
+ "memory": null,
66
+ "latency": {
67
+ "unit": "s",
68
+ "count": 253,
69
+ "total": 3.146998783111571,
70
+ "mean": 0.012438730368029931,
71
+ "stdev": 0.0009024142000934347,
72
+ "p50": 0.01226035213470459,
73
+ "p90": 0.012727705764770508,
74
+ "p95": 0.012750642967224122,
75
+ "p99": 0.012817571563720704,
76
+ "values": [
77
+ 0.012702848434448242,
78
+ 0.01272000026702881,
79
+ 0.01273855972290039,
80
+ 0.012786687850952149,
81
+ 0.01268019199371338,
82
+ 0.01277337646484375,
83
+ 0.012697600364685058,
84
+ 0.01276518440246582,
85
+ 0.012718079566955566,
86
+ 0.012714079856872559,
87
+ 0.012702624320983886,
88
+ 0.012748991966247559,
89
+ 0.012663616180419921,
90
+ 0.012716032028198243,
91
+ 0.012667903900146485,
92
+ 0.012751872062683106,
93
+ 0.012678144454956054,
94
+ 0.012725248336791992,
95
+ 0.012710911750793457,
96
+ 0.012755999565124512,
97
+ 0.012680159568786622,
98
+ 0.012722175598144531,
99
+ 0.012695551872253418,
100
+ 0.012761088371276855,
101
+ 0.01266585636138916,
102
+ 0.012809215545654297,
103
+ 0.012719103813171387,
104
+ 0.01274060821533203,
105
+ 0.012686528205871583,
106
+ 0.012731200218200683,
107
+ 0.01267404842376709,
108
+ 0.012793855667114258,
109
+ 0.012689408302307128,
110
+ 0.012774399757385254,
111
+ 0.012735487937927246,
112
+ 0.012826623916625977,
113
+ 0.012563455581665038,
114
+ 0.012733440399169921,
115
+ 0.012717056274414062,
116
+ 0.012746848106384278,
117
+ 0.012685215950012207,
118
+ 0.012676095962524414,
119
+ 0.012676095962524414,
120
+ 0.012728320121765137,
121
+ 0.012695551872253418,
122
+ 0.012716032028198243,
123
+ 0.012670975685119629,
124
+ 0.01273855972290039,
125
+ 0.01268019199371338,
126
+ 0.012930047988891602,
127
+ 0.012720128059387208,
128
+ 0.012720128059387208,
129
+ 0.012700672149658204,
130
+ 0.012706815719604492,
131
+ 0.01265561580657959,
132
+ 0.01273036766052246,
133
+ 0.012693504333496093,
134
+ 0.012698623657226562,
135
+ 0.012654591560363769,
136
+ 0.012670975685119629,
137
+ 0.012711935997009278,
138
+ 0.012749823570251465,
139
+ 0.012804096221923827,
140
+ 0.012733440399169921,
141
+ 0.012725248336791992,
142
+ 0.01273036766052246,
143
+ 0.012526592254638673,
144
+ 0.012608511924743653,
145
+ 0.012646400451660156,
146
+ 0.012578816413879394,
147
+ 0.012279808044433594,
148
+ 0.012463104248046876,
149
+ 0.012502016067504883,
150
+ 0.012577792167663575,
151
+ 0.01263923168182373,
152
+ 0.012444671630859374,
153
+ 0.012281951904296875,
154
+ 0.012396448135375977,
155
+ 0.012320768356323243,
156
+ 0.012432383537292481,
157
+ 0.012286975860595703,
158
+ 0.012257280349731446,
159
+ 0.012223487854003906,
160
+ 0.012266688346862793,
161
+ 0.012239680290222169,
162
+ 0.012272640228271485,
163
+ 0.012215295791625976,
164
+ 0.012274687767028808,
165
+ 0.012215295791625976,
166
+ 0.01226035213470459,
167
+ 0.012224512100219726,
168
+ 0.012255231857299804,
169
+ 0.012220416069030762,
170
+ 0.012255231857299804,
171
+ 0.012224512100219726,
172
+ 0.012290047645568849,
173
+ 0.012247039794921874,
174
+ 0.012259327888488769,
175
+ 0.012212224006652832,
176
+ 0.012254207611083985,
177
+ 0.012223487854003906,
178
+ 0.01226035213470459,
179
+ 0.012217344284057617,
180
+ 0.012273664474487305,
181
+ 0.012238847732543945,
182
+ 0.012257280349731446,
183
+ 0.012231679916381836,
184
+ 0.012339200019836426,
185
+ 0.01223475170135498,
186
+ 0.012254207611083985,
187
+ 0.012225536346435547,
188
+ 0.01227673625946045,
189
+ 0.012232704162597656,
190
+ 0.012256256103515625,
191
+ 0.012223487854003906,
192
+ 0.012259327888488769,
193
+ 0.012220416069030762,
194
+ 0.012266495704650878,
195
+ 0.012229632377624512,
196
+ 0.012267519950866699,
197
+ 0.01222758388519287,
198
+ 0.01225011157989502,
199
+ 0.01221939182281494,
200
+ 0.012266495704650878,
201
+ 0.012230655670166016,
202
+ 0.01225830364227295,
203
+ 0.02636595153808594,
204
+ 0.012233728408813477,
205
+ 0.012281855583190919,
206
+ 0.012228608131408691,
207
+ 0.012257280349731446,
208
+ 0.012213248252868653,
209
+ 0.01226137638092041,
210
+ 0.012241920471191407,
211
+ 0.01226035213470459,
212
+ 0.01224396800994873,
213
+ 0.012295167922973632,
214
+ 0.012223487854003906,
215
+ 0.012262399673461915,
216
+ 0.012223487854003906,
217
+ 0.012273664474487305,
218
+ 0.012224512100219726,
219
+ 0.012255231857299804,
220
+ 0.012224512100219726,
221
+ 0.01232588768005371,
222
+ 0.012223487854003906,
223
+ 0.012257280349731446,
224
+ 0.012220416069030762,
225
+ 0.012271615982055664,
226
+ 0.012215295791625976,
227
+ 0.012345343589782716,
228
+ 0.012232704162597656,
229
+ 0.012265472412109376,
230
+ 0.012230655670166016,
231
+ 0.012257280349731446,
232
+ 0.012215295791625976,
233
+ 0.01225216007232666,
234
+ 0.012212224006652832,
235
+ 0.012275712013244629,
236
+ 0.012225536346435547,
237
+ 0.012267519950866699,
238
+ 0.012218367576599122,
239
+ 0.012272640228271485,
240
+ 0.012213248252868653,
241
+ 0.01225113582611084,
242
+ 0.012216320037841797,
243
+ 0.01225113582611084,
244
+ 0.012238847732543945,
245
+ 0.012271615982055664,
246
+ 0.012209152221679688,
247
+ 0.01226854419708252,
248
+ 0.012216320037841797,
249
+ 0.012274687767028808,
250
+ 0.012232704162597656,
251
+ 0.012247039794921874,
252
+ 0.012209152221679688,
253
+ 0.01225830364227295,
254
+ 0.012216320037841797,
255
+ 0.012275712013244629,
256
+ 0.012215295791625976,
257
+ 0.01225113582611084,
258
+ 0.012225536346435547,
259
+ 0.012255231857299804,
260
+ 0.012211199760437011,
261
+ 0.012281855583190919,
262
+ 0.012246015548706055,
263
+ 0.012257280349731446,
264
+ 0.01222758388519287,
265
+ 0.012254207611083985,
266
+ 0.012231679916381836,
267
+ 0.012247039794921874,
268
+ 0.012204031944274902,
269
+ 0.012264448165893555,
270
+ 0.012353535652160644,
271
+ 0.012257280349731446,
272
+ 0.01221939182281494,
273
+ 0.012254207611083985,
274
+ 0.012217344284057617,
275
+ 0.01225113582611084,
276
+ 0.012236800193786621,
277
+ 0.012274687767028808,
278
+ 0.012229632377624512,
279
+ 0.012255231857299804,
280
+ 0.012215295791625976,
281
+ 0.012256256103515625,
282
+ 0.012208127975463867,
283
+ 0.012289024353027344,
284
+ 0.012213248252868653,
285
+ 0.012264448165893555,
286
+ 0.012249088287353516,
287
+ 0.01226137638092041,
288
+ 0.012240896224975586,
289
+ 0.01225113582611084,
290
+ 0.012221440315246582,
291
+ 0.01225113582611084,
292
+ 0.012217344284057617,
293
+ 0.012278783798217773,
294
+ 0.0122357759475708,
295
+ 0.01225113582611084,
296
+ 0.012229632377624512,
297
+ 0.012264448165893555,
298
+ 0.01222662353515625,
299
+ 0.012264384269714355,
300
+ 0.012211199760437011,
301
+ 0.012344320297241211,
302
+ 0.01224396800994873,
303
+ 0.012275712013244629,
304
+ 0.012233728408813477,
305
+ 0.012266495704650878,
306
+ 0.012211199760437011,
307
+ 0.01225011157989502,
308
+ 0.012208127975463867,
309
+ 0.012281855583190919,
310
+ 0.012225536346435547,
311
+ 0.012249088287353516,
312
+ 0.012204031944274902,
313
+ 0.012247039794921874,
314
+ 0.012239871978759765,
315
+ 0.01226956844329834,
316
+ 0.0122357759475708,
317
+ 0.012249088287353516,
318
+ 0.012239871978759765,
319
+ 0.012267519950866699,
320
+ 0.012215295791625976,
321
+ 0.012256256103515625,
322
+ 0.012211199760437011,
323
+ 0.01225216007232666,
324
+ 0.012231679916381836,
325
+ 0.01225830364227295,
326
+ 0.01223475170135498,
327
+ 0.012279808044433594,
328
+ 0.012216320037841797,
329
+ 0.012289024353027344
330
+ ]
331
+ },
332
+ "throughput": {
333
+ "unit": "tokens/s",
334
+ "value": 80.39405714350104
335
+ },
336
+ "energy": null,
337
+ "efficiency": null
338
+ }
339
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.014034639835357667,
7
+ "prefill.throughput.value": 498.76591648364223,
8
+ "decode.latency.mean": 1.5729868774414062,
9
+ "decode.throughput.value": 80.73811792160406,
10
+ "per_token.latency.mean": 0.012438730368029931,
11
+ "per_token.throughput.value": 80.39405714350104
12
+ }
13
+ }
2024-06-16/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/multirun.yaml ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpfrzfx6p6/commit=eed9ed679878ada2f6d2eefccdbda368cabc88b1
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1
88
+ - hydra.run.dir=_benchmark/tmpfrzfx6p6/commit\=eed9ed679878ada2f6d2eefccdbda368cabc88b1/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.3.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model: google/gemma-2b
150
+ processor: null
151
+ device: cuda
152
+ device_ids: '0'
153
+ seed: 42
154
+ inter_op_num_threads: null
155
+ intra_op_num_threads: null
156
+ model_kwargs: {}
157
+ processor_kwargs: {}
158
+ hub_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ latency: true
192
+ memory: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ start_method: spawn
206
+ environment:
207
+ cpu: ' AMD EPYC 7R32'
208
+ cpu_count: 16
209
+ cpu_ram_mb: 66697.29792
210
+ system: Linux
211
+ machine: x86_64
212
+ platform: Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29
213
+ processor: x86_64
214
+ python_version: 3.8.10
215
+ gpu:
216
+ - NVIDIA A10G
217
+ gpu_count: 1
218
+ gpu_vram_mb: 24146608128
219
+ optimum_benchmark_version: 0.2.1
220
+ optimum_benchmark_commit: null
221
+ transformers_version: 4.42.0.dev0
222
+ transformers_commit: eed9ed679878ada2f6d2eefccdbda368cabc88b1
223
+ accelerate_version: 0.32.0.dev0
224
+ accelerate_commit: null
225
+ diffusers_version: null
226
+ diffusers_commit: null
227
+ optimum_version: 1.21.0.dev0
228
+ optimum_commit: null
229
+ timm_version: 0.9.16
230
+ timm_commit: null
231
+ peft_version: 0.11.2.dev0
232
+ peft_commit: null
2024-06-16/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.022729344367980957,
8
+ "prefill.throughput.value": 307.9719276839752,
9
+ "decode.latency.mean": 2.3878223876953126,
10
+ "decode.throughput.value": 53.186535420072985,
11
+ "per_token.latency.mean": 0.018877755681516624,
12
+ "per_token.throughput.value": 52.97239867232251
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.022289711952209475,
21
+ "prefill.throughput.value": 314.04622971388926,
22
+ "decode.latency.mean": 2.5987576904296876,
23
+ "decode.throughput.value": 48.86950425108752,
24
+ "per_token.latency.mean": 0.020550745048070616,
25
+ "per_token.throughput.value": 48.66003629848368
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "eed9ed679878ada2f6d2eefccdbda368cabc88b1",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.014034639835357667,
34
+ "prefill.throughput.value": 498.76591648364223,
35
+ "decode.latency.mean": 1.5729868774414062,
36
+ "decode.throughput.value": 80.73811792160406,
37
+ "per_token.latency.mean": 0.012438730368029931,
38
+ "per_token.throughput.value": 80.39405714350104
39
+ }
40
+ }
41
+ ]
2024-06-16/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "eed9ed679878ada2f6d2eefccdbda368cabc88b1": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022729344367980957,
7
+ "prefill.throughput.value": 307.9719276839752,
8
+ "decode.latency.mean": 2.3878223876953126,
9
+ "decode.throughput.value": 53.186535420072985,
10
+ "per_token.latency.mean": 0.018877755681516624,
11
+ "per_token.throughput.value": 52.97239867232251
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "eed9ed679878ada2f6d2eefccdbda368cabc88b1": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.022289711952209475,
19
+ "prefill.throughput.value": 314.04622971388926,
20
+ "decode.latency.mean": 2.5987576904296876,
21
+ "decode.throughput.value": 48.86950425108752,
22
+ "per_token.latency.mean": 0.020550745048070616,
23
+ "per_token.throughput.value": 48.66003629848368
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "eed9ed679878ada2f6d2eefccdbda368cabc88b1": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.014034639835357667,
31
+ "prefill.throughput.value": 498.76591648364223,
32
+ "decode.latency.mean": 1.5729868774414062,
33
+ "decode.throughput.value": 80.73811792160406,
34
+ "per_token.latency.mean": 0.012438730368029931,
35
+ "per_token.throughput.value": 80.39405714350104
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }