IlyasMoutawwakil HF staff commited on
Commit
d4b8f22
·
verified ·
1 Parent(s): 89de04c

Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json CHANGED
@@ -84,7 +84,7 @@
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
- "optimum_benchmark_version": "0.3.0",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 876.412928,
108
  "max_global_vram": 1057.488896,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 406.847488,
@@ -112,235 +112,243 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 204,
116
- "total": 1.0073603105545041,
117
- "mean": 0.004938040738012277,
118
- "stdev": 0.00014371691996070864,
119
- "p50": 0.00485585618019104,
120
- "p90": 0.005174988698959351,
121
- "p95": 0.005221375942230225,
122
- "p99": 0.005323202576637268,
123
  "values": [
124
- 0.005561344146728516,
125
- 0.005169151782989502,
126
- 0.005210080146789551,
127
- 0.005253215789794922,
128
- 0.005228544235229492,
129
- 0.005179391860961914,
130
- 0.005245952129364013,
131
- 0.005149695873260498,
132
- 0.0049192957878112795,
133
- 0.004891647815704346,
134
- 0.004873216152191162,
135
- 0.005044223785400391,
136
- 0.00516812801361084,
137
- 0.005237728118896484,
138
- 0.005185535907745361,
139
- 0.005146624088287354,
140
- 0.005130239963531494,
141
- 0.005053440093994141,
142
- 0.005114880084991455,
143
- 0.005115903854370117,
144
- 0.005221375942230225,
145
- 0.005221375942230225,
146
- 0.005190656185150146,
147
- 0.005179391860961914,
148
- 0.005194752216339111,
149
- 0.0051404800415039064,
150
- 0.005138432025909424,
151
- 0.005262335777282715,
152
- 0.00515993595123291,
153
- 0.004928512096405029,
154
- 0.004917247772216797,
155
- 0.0048895998001098635,
156
- 0.004926464080810547,
157
- 0.005030911922454834,
158
- 0.005244927883148193,
159
- 0.0051968002319335935,
160
- 0.005128191947937012,
161
- 0.004908031940460205,
162
- 0.004984831809997559,
163
- 0.004851712226867676,
164
  0.004859903812408447,
165
- 0.004853759765625,
 
 
 
166
  0.004858880043029785,
167
- 0.0048558077812194825,
168
- 0.0048568320274353025,
 
 
 
 
 
 
 
 
 
 
169
  0.004860928058624267,
 
170
  0.004860928058624267,
171
- 0.0048568320274353025,
 
 
172
  0.004857855796813965,
173
- 0.005145599842071533,
174
- 0.00516099214553833,
175
- 0.0051773438453674315,
176
- 0.005174272060394287,
177
- 0.005137407779693603,
178
- 0.0051435518264770505,
179
- 0.005389311790466309,
180
- 0.005158912181854248,
181
- 0.005158912181854248,
182
- 0.005163008213043213,
183
- 0.005175295829772949,
184
- 0.005137407779693603,
185
- 0.0050206718444824215,
186
- 0.00510975980758667,
187
- 0.00515283203125,
188
- 0.0052715520858764645,
189
- 0.00532480001449585,
190
- 0.005157887935638428,
191
- 0.004853759765625,
192
- 0.004851712226867676,
193
- 0.004853759765625,
194
- 0.004853759765625,
195
- 0.004851712226867676,
196
- 0.004858880043029785,
197
- 0.004853759765625,
198
- 0.0049725441932678225,
199
- 0.005139455795288086,
200
- 0.0049500160217285155,
201
- 0.004914175987243652,
202
- 0.004969471931457519,
203
- 0.004930560111999512,
204
- 0.004877312183380127,
205
- 0.004853759765625,
206
- 0.004881408214569092,
207
- 0.004876287937164306,
208
- 0.004853759765625,
209
- 0.004848639965057373,
210
- 0.00485478401184082,
211
  0.00486195182800293,
212
- 0.0048558077812194825,
213
  0.004857855796813965,
214
- 0.00485478401184082,
215
- 0.0048527359962463375,
216
- 0.00485478401184082,
217
- 0.004853759765625,
218
- 0.0048496642112731934,
219
- 0.0048568320274353025,
220
- 0.004853759765625,
221
- 0.0048527359962463375,
222
- 0.00485478401184082,
223
- 0.004853759765625,
224
- 0.00485587215423584,
225
- 0.004858880043029785,
226
  0.004860928058624267,
227
- 0.0048558077812194825,
228
- 0.0048568320274353025,
229
- 0.004857855796813965,
230
- 0.0048527359962463375,
231
- 0.004850687980651855,
232
- 0.0048568320274353025,
233
- 0.004851712226867676,
234
- 0.0048558077812194825,
235
- 0.0048558077812194825,
 
 
 
 
 
 
236
  0.0048568320274353025,
 
 
 
 
 
 
 
 
237
  0.004858880043029785,
238
- 0.0048527359962463375,
239
- 0.004850687980651855,
 
 
 
240
  0.004850687980651855,
241
- 0.00485584020614624,
242
  0.004853759765625,
243
  0.0048527359962463375,
244
- 0.00485478401184082,
245
- 0.00485478401184082,
246
  0.004848639965057373,
 
 
247
  0.004850687980651855,
 
 
 
 
248
  0.004853759765625,
249
- 0.004851712226867676,
250
- 0.004853759765625,
251
- 0.0048558077812194825,
252
- 0.0048527359962463375,
253
- 0.0048527359962463375,
254
- 0.0048568320274353025,
255
- 0.0048527359962463375,
256
- 0.004851712226867676,
257
  0.004853759765625,
258
- 0.004851712226867676,
259
- 0.004857855796813965,
260
  0.004860928058624267,
261
- 0.00485478401184082,
262
- 0.004853856086730957,
263
- 0.00485478401184082,
264
- 0.004853759765625,
265
- 0.004851712226867676,
266
- 0.0048558077812194825,
267
- 0.0048568320274353025,
268
  0.004857855796813965,
269
- 0.004853759765625,
270
- 0.004853759765625,
271
- 0.0048527359962463375,
272
- 0.004850687980651855,
273
- 0.0048558077812194825,
274
- 0.0048527359962463375,
275
- 0.0048558077812194825,
276
- 0.0048527359962463375,
277
- 0.004850687980651855,
278
- 0.00485478401184082,
279
- 0.0048568320274353025,
280
- 0.00485478401184082,
281
- 0.00485478401184082,
282
- 0.0048558077812194825,
283
- 0.0048558077812194825,
284
- 0.0048527359962463375,
285
- 0.0048558077812194825,
286
- 0.0048527359962463375,
287
- 0.0048558077812194825,
288
- 0.004850687980651855,
289
  0.004858880043029785,
290
- 0.004850687980651855,
291
- 0.00485478401184082,
292
- 0.004853759765625,
293
- 0.004853759765625,
294
- 0.004853759765625,
295
- 0.0048558077812194825,
296
- 0.0048568320274353025,
297
- 0.0048527359962463375,
298
- 0.0048558077812194825,
299
- 0.0048527359962463375,
 
 
 
 
 
 
 
 
300
  0.0048558077812194825,
301
- 0.00485478401184082,
302
- 0.0048527359962463375,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
303
  0.004857855796813965,
304
- 0.0048568320274353025,
305
- 0.00485478401184082,
306
- 0.0048558077812194825,
307
- 0.004851712226867676,
308
- 0.004853759765625,
309
- 0.004853759765625,
310
  0.004857855796813965,
311
- 0.004850687980651855,
312
- 0.004853759765625,
 
 
 
313
  0.0048568320274353025,
314
- 0.0048527359962463375,
315
- 0.00485478401184082,
 
 
 
 
 
 
 
 
316
  0.004859903812408447,
 
317
  0.004858880043029785,
318
- 0.0048568320274353025,
319
- 0.0048527359962463375,
320
- 0.0048527359962463375,
 
 
321
  0.004857855796813965,
 
 
 
 
 
 
 
 
 
 
322
  0.004857855796813965,
323
- 0.0048558077812194825,
324
- 0.0048496642112731934,
325
- 0.0048558077812194825,
326
- 0.0048527359962463375,
327
- 0.004851712226867676
 
 
 
 
 
 
 
 
 
328
  ]
329
  },
330
  "throughput": {
331
  "unit": "samples/s",
332
- "value": 202.50946742949165
333
  },
334
  "energy": {
335
  "unit": "kWh",
336
- "cpu": 5.7579536950230413e-08,
337
- "ram": 3.148242451164379e-08,
338
- "gpu": 3.267314356872015e-07,
339
- "total": 4.1579339714907566e-07
340
  },
341
  "efficiency": {
342
  "unit": "samples/kWh",
343
- "value": 2405040.5967400847
344
  }
345
  }
346
  }
 
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.3.1",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 876.105728,
108
  "max_global_vram": 1057.488896,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 406.847488,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 212,
116
+ "total": 1.0311774311065671,
117
+ "mean": 0.004864044486351733,
118
+ "stdev": 4.300824366583922e-05,
119
+ "p50": 0.00486195182800293,
120
+ "p90": 0.004863999843597412,
121
+ "p95": 0.00486548478603363,
122
+ "p99": 0.004895068001747132,
123
  "values": [
124
+ 0.005483520030975342,
125
+ 0.004863999843597412,
126
+ 0.004863999843597412,
127
+ 0.004866047859191895,
128
+ 0.00486195182800293,
129
+ 0.004865024089813232,
130
+ 0.004867072105407715,
131
+ 0.004863999843597412,
132
+ 0.004865024089813232,
133
+ 0.004863999843597412,
134
+ 0.004860928058624267,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
135
  0.004859903812408447,
136
+ 0.00486297607421875,
137
+ 0.00486297607421875,
138
+ 0.00486195182800293,
139
+ 0.00486297607421875,
140
  0.004858880043029785,
141
+ 0.004867072105407715,
142
+ 0.004860960006713867,
143
+ 0.00486297607421875,
144
+ 0.00486195182800293,
145
+ 0.004857855796813965,
146
+ 0.004859903812408447,
147
+ 0.00486297607421875,
148
+ 0.00486297607421875,
149
+ 0.00486195182800293,
150
+ 0.00486297607421875,
151
+ 0.004858880043029785,
152
+ 0.004859903812408447,
153
  0.004860928058624267,
154
+ 0.00486297607421875,
155
  0.004860928058624267,
156
+ 0.00486297607421875,
157
+ 0.004859903812408447,
158
+ 0.00486297607421875,
159
  0.004857855796813965,
160
+ 0.004859903812408447,
161
+ 0.004860928058624267,
162
+ 0.004860928058624267,
163
+ 0.00486297607421875,
164
+ 0.004859903812408447,
165
+ 0.004863999843597412,
166
+ 0.00486195182800293,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
167
  0.00486195182800293,
 
168
  0.004857855796813965,
169
+ 0.00486297607421875,
 
 
 
 
 
 
 
 
 
 
 
170
  0.004860928058624267,
171
+ 0.004865024089813232,
172
+ 0.004863999843597412,
173
+ 0.004860991954803467,
174
+ 0.00486195182800293,
175
+ 0.00486297607421875,
176
+ 0.004859903812408447,
177
+ 0.00486297607421875,
178
+ 0.004865024089813232,
179
+ 0.00486297607421875,
180
+ 0.00486297607421875,
181
+ 0.00486195182800293,
182
+ 0.004867072105407715,
183
+ 0.0048895998001098635,
184
+ 0.004895743846893311,
185
+ 0.004895743846893311,
186
  0.0048568320274353025,
187
+ 0.004863999843597412,
188
+ 0.004859903812408447,
189
+ 0.00486195182800293,
190
+ 0.00486195182800293,
191
+ 0.00486297607421875,
192
+ 0.00486195182800293,
193
+ 0.004865024089813232,
194
+ 0.00486297607421875,
195
  0.004858880043029785,
196
+ 0.004858880043029785,
197
+ 0.0048496642112731934,
198
+ 0.004851712226867676,
199
+ 0.004848639965057373,
200
+ 0.004848639965057373,
201
  0.004850687980651855,
 
202
  0.004853759765625,
203
  0.0048527359962463375,
 
 
204
  0.004848639965057373,
205
+ 0.004848703861236572,
206
+ 0.0048568320274353025,
207
  0.004850687980651855,
208
+ 0.004850687980651855,
209
+ 0.004847616195678711,
210
+ 0.004848639965057373,
211
+ 0.0048496642112731934,
212
  0.004853759765625,
213
+ 0.004850687980651855,
214
+ 0.0048496642112731934,
215
+ 0.004847616195678711,
216
+ 0.004848639965057373,
 
 
 
 
217
  0.004853759765625,
218
+ 0.004859903812408447,
 
219
  0.004860928058624267,
220
+ 0.00486297607421875,
221
+ 0.004860928058624267,
222
+ 0.004860928058624267,
223
+ 0.004860928058624267,
224
+ 0.004858880043029785,
225
+ 0.004860928058624267,
226
+ 0.00486297607421875,
227
  0.004857855796813965,
228
+ 0.004868095874786377,
229
+ 0.004863999843597412,
230
+ 0.00486195182800293,
231
+ 0.00486195182800293,
232
+ 0.004859903812408447,
233
+ 0.004859903812408447,
234
+ 0.004860928058624267,
235
+ 0.004865024089813232,
 
 
 
 
 
 
 
 
 
 
 
 
236
  0.004858880043029785,
237
+ 0.004866047859191895,
238
+ 0.00486195182800293,
239
+ 0.004858880043029785,
240
+ 0.004865024089813232,
241
+ 0.004860928058624267,
242
+ 0.00486297607421875,
243
+ 0.004860928058624267,
244
+ 0.00486297607421875,
245
+ 0.004859903812408447,
246
+ 0.004860928058624267,
247
+ 0.00486297607421875,
248
+ 0.00486195182800293,
249
+ 0.00486297607421875,
250
+ 0.004860928058624267,
251
+ 0.00486297607421875,
252
+ 0.004858880043029785,
253
+ 0.00486195182800293,
254
+ 0.004859903812408447,
255
  0.0048558077812194825,
256
+ 0.004860928058624267,
257
+ 0.004859903812408447,
258
+ 0.00486195182800293,
259
+ 0.00486195182800293,
260
+ 0.004859903812408447,
261
+ 0.004859903812408447,
262
+ 0.00486195182800293,
263
+ 0.00486195182800293,
264
+ 0.00486195182800293,
265
+ 0.004859903812408447,
266
+ 0.004860928058624267,
267
+ 0.00486195182800293,
268
+ 0.00486297607421875,
269
+ 0.00486297607421875,
270
+ 0.004860928058624267,
271
+ 0.00486195182800293,
272
+ 0.004858880043029785,
273
+ 0.004866047859191895,
274
+ 0.004860928058624267,
275
+ 0.00486297607421875,
276
+ 0.004859903812408447,
277
+ 0.00486195182800293,
278
+ 0.004860928058624267,
279
+ 0.00486297607421875,
280
+ 0.00486297607421875,
281
+ 0.00486297607421875,
282
+ 0.00486297607421875,
283
  0.004857855796813965,
284
+ 0.00486297607421875,
 
 
 
 
 
285
  0.004857855796813965,
286
+ 0.004859903812408447,
287
+ 0.00486195182800293,
288
+ 0.00486406421661377,
289
+ 0.00486195182800293,
290
+ 0.00486195182800293,
291
  0.0048568320274353025,
292
+ 0.00486297607421875,
293
+ 0.00486297607421875,
294
+ 0.00486195182800293,
295
+ 0.004859903812408447,
296
+ 0.00486297607421875,
297
+ 0.00486297607421875,
298
+ 0.004859903812408447,
299
+ 0.004859903812408447,
300
+ 0.00486195182800293,
301
+ 0.00486195182800293,
302
  0.004859903812408447,
303
+ 0.004865024089813232,
304
  0.004858880043029785,
305
+ 0.004865024089813232,
306
+ 0.004860928058624267,
307
+ 0.00486195182800293,
308
+ 0.00486297607421875,
309
+ 0.00486195182800293,
310
  0.004857855796813965,
311
+ 0.00486297607421875,
312
+ 0.00486195182800293,
313
+ 0.00486297607421875,
314
+ 0.004863999843597412,
315
+ 0.00486297607421875,
316
+ 0.004858880043029785,
317
+ 0.004860928058624267,
318
+ 0.004860928058624267,
319
+ 0.004863999843597412,
320
+ 0.00486195182800293,
321
  0.004857855796813965,
322
+ 0.004860928058624267,
323
+ 0.004859903812408447,
324
+ 0.004863999843597412,
325
+ 0.00486297607421875,
326
+ 0.004860928058624267,
327
+ 0.004860928058624267,
328
+ 0.004863999843597412,
329
+ 0.00486195182800293,
330
+ 0.004859903812408447,
331
+ 0.00486297607421875,
332
+ 0.00486297607421875,
333
+ 0.004859903812408447,
334
+ 0.004860928058624267,
335
+ 0.00486195182800293
336
  ]
337
  },
338
  "throughput": {
339
  "unit": "samples/s",
340
+ "value": 205.59022492618033
341
  },
342
  "energy": {
343
  "unit": "kWh",
344
+ "cpu": 5.76926704266625e-08,
345
+ "ram": 3.1545104307606886e-08,
346
+ "gpu": 3.3605213824644606e-07,
347
+ "total": 4.252899129807155e-07
348
  },
349
  "efficiency": {
350
  "unit": "samples/kWh",
351
+ "value": 2351337.2160447743
352
  }
353
  }
354
  }