IlyasMoutawwakil HF staff commited on
Commit
e3b701e
·
verified ·
1 Parent(s): 942affc

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -73,10 +73,10 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
- "cpu_ram_mb": 66697.29792,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -86,15 +86,15 @@
86
  "gpu_vram_mb": 24146608128,
87
  "optimum_benchmark_version": "0.2.1",
88
  "optimum_benchmark_commit": null,
89
- "transformers_version": "4.41.1",
90
  "transformers_commit": null,
91
- "accelerate_version": "0.30.1",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.27.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
- "timm_version": "1.0.3",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 869.208064,
108
  "max_global_vram": 1229.45536,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 591.396864,
@@ -112,208 +112,198 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 177,
116
- "total": 1.0042329287528995,
117
- "mean": 0.005673632365835588,
118
- "stdev": 0.0002255925016277513,
119
- "p50": 0.005618688106536865,
120
- "p90": 0.005950873470306396,
121
- "p95": 0.00601907205581665,
122
- "p99": 0.006297518177032472,
123
  "values": [
124
- 0.006268928050994873,
125
- 0.005730303764343261,
126
- 0.0057415680885314945,
127
- 0.0059770879745483394,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
128
  0.00587775993347168,
129
- 0.005884928226470947,
130
- 0.005765151977539062,
131
- 0.005815296173095703,
132
- 0.005819392204284668,
133
- 0.005750783920288086,
134
- 0.005757952213287353,
135
- 0.005699584007263184,
136
- 0.005578752040863037,
137
- 0.005544960021972656,
138
- 0.005615615844726562,
139
- 0.005710847854614258,
140
- 0.005573631763458252,
141
- 0.005548031806945801,
142
- 0.005526527881622314,
143
- 0.005489664077758789,
144
- 0.005575679779052735,
145
- 0.005425151824951172,
146
- 0.00561356782913208,
147
- 0.005610496044158936,
148
- 0.0056442880630493165,
149
- 0.005574656009674072,
150
- 0.005563392162322998,
151
- 0.005554175853729248,
152
- 0.0054876160621643065,
153
- 0.005517312049865723,
154
- 0.005661695957183838,
155
- 0.0056145920753479005,
156
- 0.005608448028564453,
157
- 0.005558400154113769,
158
- 0.00557260799407959,
159
- 0.005618688106536865,
160
- 0.006061056137084961,
161
- 0.005950463771820068,
162
- 0.005938176155090332,
163
- 0.00591871976852417,
164
- 0.0059361281394958495,
165
- 0.005983232021331787,
166
- 0.005787648200988769,
167
  0.005847040176391601,
168
- 0.005753856182098388,
169
- 0.005793791770935058,
170
- 0.005682176113128662,
171
- 0.005653503894805908,
172
- 0.005921792030334472,
173
- 0.00568832015991211,
174
- 0.005701632022857666,
175
- 0.005938176155090332,
176
- 0.005980160236358643,
177
- 0.00592793607711792,
178
- 0.005981184005737304,
179
- 0.005911551952362061,
180
- 0.005868544101715088,
181
- 0.005806079864501953,
182
- 0.005697535991668701,
183
- 0.005790719985961914,
184
- 0.005890048027038574,
185
  0.005848063945770264,
186
- 0.005776383876800537,
187
- 0.005787648200988769,
188
- 0.005718016147613526,
189
- 0.005642240047454834,
190
- 0.005690368175506591,
191
- 0.0058757119178771975,
 
 
 
 
 
 
 
 
 
 
 
 
 
192
  0.005832704067230224,
193
- 0.005928959846496582,
194
- 0.005891071796417236,
195
- 0.005788671970367432,
196
- 0.005888000011444092,
197
- 0.0059023361206054685,
198
- 0.006090752124786377,
199
- 0.00602726411819458,
200
- 0.005874688148498535,
201
- 0.006277120113372803,
202
- 0.006182911872863769,
203
- 0.006362112045288086,
204
- 0.005914624214172363,
205
- 0.005951488018035888,
206
- 0.006017024040222168,
207
- 0.005961728096008301,
208
- 0.005960703849792481,
209
- 0.005868544101715088,
210
- 0.005961728096008301,
211
- 0.005898240089416504,
212
- 0.0058716158866882326,
213
- 0.005925888061523437,
214
- 0.00586137580871582,
215
- 0.005681151866912842,
216
- 0.005708799839019775,
217
- 0.0056145920753479005,
218
- 0.005606400012969971,
219
- 0.005794816017150879,
220
- 0.005796864032745362,
221
- 0.005666816234588623,
222
- 0.0055623679161071774,
223
- 0.005464064121246338,
224
- 0.005438464164733887,
225
- 0.005618688106536865,
226
- 0.005593088150024414,
227
- 0.005764095783233642,
228
  0.005833727836608887,
229
- 0.0057825279235839844,
 
 
 
 
 
 
 
 
 
 
 
 
230
  0.005822463989257813,
231
- 0.0059023361206054685,
232
- 0.006187007904052734,
233
- 0.005640192031860352,
234
- 0.005433343887329102,
235
- 0.005434368133544922,
236
- 0.005442560195922852,
237
- 0.005402624130249023,
238
- 0.005451776027679443,
239
- 0.005401599884033203,
240
- 0.005500927925109863,
241
- 0.005420032024383545,
242
- 0.00541593599319458,
243
- 0.005458943843841553,
244
- 0.00541593599319458,
245
- 0.005392384052276611,
246
- 0.005432320117950439,
247
- 0.005404672145843506,
248
- 0.005388288021087646,
249
- 0.005447679996490478,
250
- 0.005465087890625,
251
- 0.005386240005493164,
252
- 0.005506048202514649,
253
- 0.005435391902923584,
254
- 0.005479423999786377,
255
- 0.005463039875030518,
256
- 0.005426176071166992,
257
- 0.005404672145843506,
258
- 0.005393407821655274,
259
- 0.005398528099060058,
260
- 0.005393407821655274,
261
- 0.0056258559226989744,
262
- 0.005436416149139404,
263
- 0.005434368133544922,
264
- 0.005393407821655274,
265
- 0.005368832111358642,
266
- 0.005469183921813964,
267
- 0.005430272102355957,
268
- 0.005419007778167725,
269
- 0.005409791946411133,
270
- 0.005399551868438721,
271
- 0.005421055793762207,
272
- 0.0054609918594360355,
273
- 0.005419072151184082,
274
- 0.005420032024383545,
275
- 0.005600255966186523,
276
- 0.005949440002441406,
277
- 0.005449728012084961,
278
- 0.006471680164337158,
279
- 0.005696512222290039,
280
- 0.00587775993347168,
281
- 0.005532671928405761,
282
- 0.005500927925109863,
283
- 0.0055203838348388675,
284
- 0.005508096218109131,
285
- 0.005485568046569824,
286
- 0.005436416149139404,
287
- 0.005458943843841553,
288
- 0.005465087890625,
289
- 0.00550707197189331,
290
- 0.005440512180328369,
291
- 0.005473279953002929,
292
- 0.005450751781463623,
293
- 0.005465087890625,
294
- 0.005575679779052735,
295
- 0.005543935775756836,
296
- 0.005563392162322998,
297
- 0.005566463947296142,
298
- 0.005536767959594726,
299
- 0.005550079822540284,
300
- 0.005560319900512695
301
  ]
302
  },
303
  "throughput": {
304
  "unit": "samples/s",
305
- "value": 176.25392967327454
306
  },
307
  "energy": {
308
  "unit": "kWh",
309
- "cpu": 6.333479676583802e-08,
310
- "ram": 3.461584478560219e-08,
311
- "gpu": 1.3496742823529406e-07,
312
- "total": 2.3291806978673424e-07
313
  },
314
  "efficiency": {
315
  "unit": "samples/kWh",
316
- "value": 4293355.173841281
317
  }
318
  }
319
  }
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.293824,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
86
  "gpu_vram_mb": 24146608128,
87
  "optimum_benchmark_version": "0.2.1",
88
  "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.31.0",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.29.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
+ "timm_version": "1.0.7",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 869.670912,
108
  "max_global_vram": 1229.45536,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 591.396864,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 167,
116
+ "total": 1.0031495051383974,
117
+ "mean": 0.0060068832643017794,
118
+ "stdev": 0.00020016329854263533,
119
+ "p50": 0.0059054079055786135,
120
+ "p90": 0.006282649612426757,
121
+ "p95": 0.006338355302810668,
122
+ "p99": 0.006516060104370117,
123
  "values": [
124
+ 0.006903808116912841,
125
+ 0.006348800182342529,
126
+ 0.006425600051879882,
127
+ 0.006407167911529541,
128
+ 0.006498303890228272,
129
+ 0.0063508481979370115,
130
+ 0.006366208076477051,
131
+ 0.006313983917236328,
132
+ 0.006308864116668702,
133
+ 0.006245376110076905,
134
+ 0.0062269439697265625,
135
+ 0.00628223991394043,
136
+ 0.006155263900756836,
137
+ 0.006157375812530517,
138
+ 0.006135807991027832,
139
+ 0.006262784004211426,
140
+ 0.006251520156860352,
141
+ 0.00637440013885498,
142
+ 0.006294528007507324,
143
+ 0.005955584049224853,
144
+ 0.005952511787414551,
145
+ 0.005909503936767578,
146
+ 0.00602726411819458,
147
+ 0.006086656093597412,
148
+ 0.005993472099304199,
149
+ 0.005967872142791748,
150
+ 0.006009856224060059,
151
+ 0.005914624214172363,
152
+ 0.005947391986846923,
153
+ 0.005933055877685547,
154
+ 0.005956607818603516,
155
+ 0.006550528049468994,
156
+ 0.006212607860565185,
157
+ 0.00611737585067749,
158
+ 0.006205440044403076,
159
+ 0.006109183788299561,
160
+ 0.006230016231536865,
161
+ 0.006164480209350586,
162
+ 0.006276095867156982,
163
+ 0.006239232063293457,
164
+ 0.006230016231536865,
165
+ 0.00624128007888794,
166
+ 0.005958655834197998,
167
+ 0.006087679862976075,
168
+ 0.006251520156860352,
169
+ 0.0062679038047790524,
170
+ 0.006207488059997559,
171
+ 0.00628326416015625,
172
+ 0.0061859841346740725,
173
+ 0.006141952037811279,
174
+ 0.006179840087890625,
175
+ 0.006160384178161621,
176
+ 0.0062975997924804685,
177
+ 0.005889023780822754,
178
+ 0.005827583789825439,
179
+ 0.00586137580871582,
180
+ 0.0059054079055786135,
181
+ 0.005852159976959229,
182
+ 0.006179840087890625,
183
+ 0.006154240131378174,
184
+ 0.006221824169158936,
185
+ 0.006253568172454834,
186
+ 0.006302720069885254,
187
+ 0.00623308801651001,
188
+ 0.006271999835968017,
189
+ 0.006171648025512695,
190
+ 0.006273024082183838,
191
+ 0.006153215885162353,
192
+ 0.006291456222534179,
193
+ 0.006135807991027832,
194
+ 0.006048768043518066,
195
+ 0.006095871925354004,
196
+ 0.005913599967956543,
197
+ 0.006023168087005615,
198
+ 0.0060631041526794435,
199
+ 0.005945343971252442,
200
+ 0.005849088191986084,
201
+ 0.005865471839904785,
202
+ 0.0058716158866882326,
203
+ 0.005913599967956543,
204
+ 0.005901311874389649,
205
+ 0.006119423866271972,
206
+ 0.006081535816192627,
207
+ 0.006091775894165039,
208
+ 0.006067200183868408,
209
+ 0.006199295997619629,
210
+ 0.0062873601913452145,
211
+ 0.006111231803894043,
212
+ 0.005944320201873779,
213
+ 0.005949440002441406,
214
+ 0.005855231761932373,
215
  0.00587775993347168,
216
+ 0.005853184223175049,
217
+ 0.0058787841796875,
218
+ 0.005836800098419189,
219
+ 0.0058787841796875,
220
+ 0.005834752082824707,
221
+ 0.00582041597366333,
222
+ 0.005984255790710449,
223
+ 0.005873663902282715,
224
+ 0.005881855964660645,
225
+ 0.005829631805419922,
226
+ 0.005851136207580566,
227
+ 0.005792768001556397,
228
+ 0.005881855964660645,
229
+ 0.005830656051635743,
230
+ 0.005937151908874512,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
231
  0.005847040176391601,
232
+ 0.005880832195281982,
233
+ 0.0058388481140136715,
234
+ 0.005881919860839844,
235
+ 0.005836800098419189,
236
+ 0.005832704067230224,
237
+ 0.005849088191986084,
238
+ 0.005858304023742676,
239
+ 0.005848063945770264,
240
+ 0.005827583789825439,
241
+ 0.005857279777526855,
242
+ 0.005830656051635743,
243
+ 0.00586240005493164,
244
+ 0.005860352039337159,
 
 
 
 
245
  0.005848063945770264,
246
+ 0.005876736164093017,
247
+ 0.005854207992553711,
248
+ 0.00582041597366333,
249
+ 0.0058091521263122555,
250
+ 0.00587775993347168,
251
+ 0.005817344188690185,
252
+ 0.005892096042633056,
253
+ 0.0058009600639343266,
254
+ 0.005880832195281982,
255
+ 0.005814271926879883,
256
+ 0.005843967914581299,
257
+ 0.005817344188690185,
258
+ 0.005846015930175781,
259
+ 0.005822463989257813,
260
+ 0.005850175857543946,
261
+ 0.005837823867797852,
262
+ 0.005830656051635743,
263
+ 0.005839871883392334,
264
+ 0.00582860803604126,
265
  0.005832704067230224,
266
+ 0.005817344188690185,
267
+ 0.005860352039337159,
268
+ 0.005798912048339844,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
269
  0.005833727836608887,
270
+ 0.005805056095123291,
271
+ 0.005818367958068848,
272
+ 0.005852159976959229,
273
+ 0.005859327793121338,
274
+ 0.0058388481140136715,
275
+ 0.00582860803604126,
276
+ 0.005886975765228271,
277
+ 0.005829631805419922,
278
+ 0.005873663902282715,
279
+ 0.005847040176391601,
280
+ 0.005901311874389649,
281
+ 0.00582041597366333,
282
+ 0.0058787841796875,
283
  0.005822463989257813,
284
+ 0.005848063945770264,
285
+ 0.00582144021987915,
286
+ 0.005840896129608154,
287
+ 0.005825535774230957,
288
+ 0.005825535774230957,
289
+ 0.005852159976959229,
290
+ 0.005827583789825439
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
291
  ]
292
  },
293
  "throughput": {
294
  "unit": "samples/s",
295
+ "value": 166.47568397789345
296
  },
297
  "energy": {
298
  "unit": "kWh",
299
+ "cpu": 6.822651719290113e-08,
300
+ "ram": 3.716695327299249e-08,
301
+ "gpu": 1.362064581714297e-07,
302
+ "total": 2.415999286373233e-07
303
  },
304
  "efficiency": {
305
  "unit": "samples/kWh",
306
+ "value": 4139074.0702624368
307
  }
308
  }
309
  }