IlyasMoutawwakil HF staff commited on
Commit
a07835b
·
verified ·
1 Parent(s): 69fc3a6

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 765.485056,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.4303876953125,
117
- "mean": 7.4303876953125,
118
  "stdev": 0.0,
119
- "p50": 7.4303876953125,
120
- "p90": 7.4303876953125,
121
- "p95": 7.4303876953125,
122
- "p99": 7.4303876953125,
123
  "values": [
124
- 7.4303876953125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 1.0633139694438872e-06,
131
- "ram": 5.666658016994584e-07,
132
- "gpu": 1.6197235179994315e-06,
133
- "total": 3.2497032891427775e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 958.312448,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
@@ -145,181 +145,178 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 150,
149
- "total": 0.9997462720870974,
150
- "mean": 0.006664975147247314,
151
- "stdev": 0.0002994929534790365,
152
- "p50": 0.0068039679527282714,
153
- "p90": 0.006930943870544434,
154
- "p95": 0.007036518311500549,
155
- "p99": 0.007186268038749694,
156
  "values": [
157
- 0.007228415966033935,
158
- 0.007142399787902832,
159
- 0.007051263809204102,
160
- 0.0070563840866088865,
161
- 0.007060480117797851,
162
- 0.007053311824798584,
163
- 0.0069027838706970214,
164
- 0.006847487926483154,
165
- 0.006894591808319092,
166
- 0.006858751773834228,
167
- 0.006857791900634766,
168
- 0.006874112129211426,
169
- 0.006912000179290771,
170
- 0.006871039867401123,
171
- 0.0067717118263244626,
172
- 0.006814720153808594,
173
- 0.006825984001159668,
174
- 0.007060480117797851,
175
- 0.006949888229370117,
176
- 0.006908927917480469,
177
- 0.00689356803894043,
178
- 0.00684441614151001,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  0.006776832103729248,
180
- 0.006872064113616944,
181
- 0.006930431842803955,
182
- 0.006845439910888672,
183
- 0.006888383865356445,
184
- 0.006892543792724609,
185
- 0.006873087882995605,
186
- 0.006859776020050049,
187
- 0.0068577280044555666,
188
- 0.0068884482383728025,
189
- 0.006968319892883301,
190
- 0.006862847805023193,
191
- 0.006821887969970703,
192
  0.006778880119323731,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
193
  0.006754303932189941,
194
- 0.006807551860809326,
195
- 0.0068321280479431154,
196
- 0.006872064113616944,
197
- 0.006842368125915528,
198
- 0.00684441614151001,
199
- 0.00684441614151001,
200
- 0.006848512172698974,
201
- 0.006883327960968018,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
202
  0.0068853759765625,
203
- 0.006968319892883301,
204
- 0.006790143966674805,
205
- 0.00679423999786377,
206
- 0.0068249602317810056,
207
- 0.006850560188293457,
208
- 0.0068915200233459475,
209
- 0.007018496036529541,
210
- 0.006895616054534912,
211
- 0.006860799789428711,
212
- 0.006977536201477051,
213
  0.006807551860809326,
214
- 0.006854656219482422,
 
 
 
 
215
  0.006800384044647217,
216
- 0.0069027838706970214,
217
- 0.00687820816040039,
218
- 0.006792128086090088,
219
- 0.006845439910888672,
220
- 0.006833151817321777,
221
- 0.006814720153808594,
222
- 0.00677785587310791,
223
- 0.006831103801727295,
224
- 0.006860799789428711,
225
- 0.006916096210479736,
226
- 0.006945759773254395,
227
- 0.006897664070129395,
228
- 0.00693555212020874,
229
- 0.0069212160110473635,
230
- 0.006871039867401123,
231
- 0.006870016098022461,
232
- 0.006831039905548095,
233
- 0.006813695907592773,
234
- 0.006819839954376221,
235
- 0.006912992000579834,
236
- 0.006831103801727295,
237
- 0.006739967823028564,
238
- 0.00674508810043335,
239
- 0.006608895778656006,
240
- 0.006746111869812011,
241
- 0.00673689603805542,
242
- 0.006605823993682861,
243
- 0.006470655918121338,
244
- 0.006501376152038574,
245
- 0.006491136074066162,
246
- 0.006715392112731934,
247
- 0.00676966381072998,
248
- 0.006782976150512696,
249
- 0.006780928134918213,
250
- 0.006849535942077637,
251
- 0.00688643217086792,
252
- 0.0069079680442810055,
253
- 0.006590464115142822,
254
- 0.006594560146331787,
255
- 0.006508543968200684,
256
- 0.007353280067443848,
257
- 0.006542335987091064,
258
- 0.006490111827850342,
259
- 0.0064778242111206055,
260
- 0.0065136637687683106,
261
- 0.006434815883636475,
262
- 0.006484960079193115,
263
- 0.006501376152038574,
264
- 0.006427648067474365,
265
- 0.006499328136444092,
266
- 0.006437888145446777,
267
- 0.0064174079895019534,
268
- 0.0064143362045288085,
269
- 0.00639686393737793,
270
- 0.006432767868041992,
271
- 0.0064245758056640625,
272
- 0.006392831802368164,
273
- 0.006434815883636475,
274
- 0.006396927833557129,
275
- 0.006426623821258545,
276
- 0.00646560001373291,
277
- 0.006376448154449463,
278
- 0.006421504020690918,
279
- 0.006435872077941895,
280
- 0.006391808032989502,
281
- 0.006421504020690918,
282
- 0.006387712001800537,
283
- 0.006427648067474365,
284
- 0.0064245758056640625,
285
- 0.006135807991027832,
286
- 0.006124544143676758,
287
- 0.006132736206054688,
288
- 0.0061521921157836916,
289
- 0.006111231803894043,
290
- 0.00611737585067749,
291
- 0.006108160018920898,
292
- 0.006095871925354004,
293
- 0.006091775894165039,
294
- 0.006108160018920898,
295
- 0.0061265921592712404,
296
- 0.006103040218353272,
297
- 0.006103007793426513,
298
- 0.00608460807800293,
299
- 0.006118400096893311,
300
- 0.006106112003326416,
301
- 0.0061131839752197265,
302
- 0.006089727878570556,
303
- 0.006136832237243653,
304
- 0.0060631041526794435,
305
- 0.006096896171569824,
306
- 0.0060928001403808595
307
  ]
308
  },
309
  "throughput": {
310
  "unit": "samples/s",
311
- "value": 150.03806884606428
312
  },
313
  "energy": {
314
  "unit": "kWh",
315
- "cpu": 7.216292841293949e-08,
316
- "ram": 3.9456346913690235e-08,
317
- "gpu": 1.3254075643902343e-07,
318
- "total": 2.4416003176565314e-07
319
  },
320
  "efficiency": {
321
  "unit": "samples/kWh",
322
- "value": 4095674.4343800237
323
  }
324
  }
325
  }
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 787.894272,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.4492529296875,
117
+ "mean": 7.4492529296875,
118
  "stdev": 0.0,
119
+ "p50": 7.4492529296875,
120
+ "p90": 7.4492529296875,
121
+ "p95": 7.4492529296875,
122
+ "p99": 7.4492529296875,
123
  "values": [
124
+ 7.4492529296875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 1.0661731923607852e-06,
131
+ "ram": 5.675092877226037e-07,
132
+ "gpu": 1.602223504000317e-06,
133
+ "total": 3.235905984083706e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1077.583872,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 147,
149
+ "total": 0.9997752003669739,
150
+ "mean": 0.006801191839231115,
151
+ "stdev": 0.00035212494367062373,
152
+ "p50": 0.006779903888702392,
153
+ "p90": 0.006891110420227051,
154
+ "p95": 0.00700579833984375,
155
+ "p99": 0.0084855806159973,
156
  "values": [
157
+ 0.00971571159362793,
158
+ 0.009312255859375,
159
+ 0.007262207984924316,
160
+ 0.007009280204772949,
161
+ 0.007172095775604248,
162
+ 0.007515135765075683,
163
+ 0.006605823993682861,
164
+ 0.006435840129852295,
165
+ 0.006503424167633057,
166
+ 0.00652288007736206,
167
+ 0.006496255874633789,
168
+ 0.006512639999389648,
169
+ 0.00653715181350708,
170
+ 0.0065054078102111815,
171
+ 0.006524928092956543,
172
+ 0.0070767998695373535,
173
+ 0.006605855941772461,
174
+ 0.006492159843444824,
175
+ 0.006475808143615723,
176
+ 0.0067010560035705566,
177
+ 0.006508543968200684,
178
+ 0.006491136074066162,
179
+ 0.006520832061767578,
180
+ 0.006504447937011719,
181
+ 0.006464511871337891,
182
+ 0.006626304149627686,
183
+ 0.006738944053649902,
184
+ 0.006467584133148193,
185
+ 0.006468607902526856,
186
+ 0.006469632148742676,
187
+ 0.006429696083068847,
188
+ 0.006481919765472412,
189
+ 0.006937600135803222,
190
+ 0.0068249602317810056,
191
+ 0.00676966381072998,
192
+ 0.006779903888702392,
193
+ 0.0067645440101623535,
194
+ 0.00674508810043335,
195
+ 0.006782976150512696,
196
+ 0.006740960121154785,
197
+ 0.006945792198181152,
198
+ 0.006754303932189941,
199
+ 0.0067901120185852055,
200
+ 0.006713344097137451,
201
+ 0.00679423999786377,
202
+ 0.006791232109069824,
203
+ 0.006773759841918945,
204
+ 0.006774784088134766,
205
+ 0.006766592025756836,
206
+ 0.006760447978973389,
207
+ 0.006813695907592773,
208
+ 0.006755328178405762,
209
  0.006776832103729248,
210
+ 0.006748159885406494,
 
 
 
 
 
 
 
 
 
 
 
211
  0.006778880119323731,
212
+ 0.006782976150512696,
213
+ 0.006751232147216797,
214
+ 0.0067686400413513184,
215
+ 0.006800384044647217,
216
+ 0.006759391784667969,
217
+ 0.006785024166107178,
218
+ 0.006770688056945801,
219
+ 0.0067758078575134275,
220
+ 0.006743040084838867,
221
+ 0.006767615795135498,
222
+ 0.006776832103729248,
223
+ 0.006751232147216797,
224
+ 0.006910975933074951,
225
+ 0.006789120197296142,
226
+ 0.006763519763946534,
227
+ 0.006751232147216797,
228
+ 0.006747136116027832,
229
+ 0.006773759841918945,
230
+ 0.006740992069244385,
231
+ 0.00674508810043335,
232
+ 0.006758399963378906,
233
+ 0.006763519763946534,
234
+ 0.006779903888702392,
235
  0.006754303932189941,
236
+ 0.006754303932189941,
237
+ 0.006792191982269287,
238
+ 0.006759424209594727,
239
+ 0.006755328178405762,
240
+ 0.0067420158386230465,
241
+ 0.006770688056945801,
242
+ 0.00678604793548584,
243
+ 0.006754303932189941,
244
+ 0.006767615795135498,
245
+ 0.006793216228485107,
246
+ 0.006874112129211426,
247
+ 0.006781983852386475,
248
+ 0.0067675838470458985,
249
+ 0.006819839954376221,
250
+ 0.0067840638160705565,
251
+ 0.006788095951080322,
252
+ 0.00678604793548584,
253
+ 0.006772736072540283,
254
+ 0.006783999919891357,
255
+ 0.006782976150512696,
256
+ 0.006825984001159668,
257
+ 0.006880256175994873,
258
+ 0.007000063896179199,
259
+ 0.006913023948669434,
260
+ 0.00687820816040039,
261
+ 0.006899712085723877,
262
+ 0.006870016098022461,
263
+ 0.006881279945373535,
264
+ 0.006784031867980957,
265
+ 0.006820864200592041,
266
+ 0.006816768169403077,
267
+ 0.006792128086090088,
268
  0.0068853759765625,
269
+ 0.006821887969970703,
270
+ 0.006804512023925781,
271
+ 0.006803455829620361,
272
+ 0.006812672138214112,
 
 
 
 
 
 
273
  0.006807551860809326,
274
+ 0.006811647891998291,
275
+ 0.006823935985565186,
276
+ 0.0068249602317810056,
277
+ 0.006979584217071533,
278
+ 0.006766592025756836,
279
  0.006800384044647217,
280
+ 0.006796288013458252,
281
+ 0.006763519763946534,
282
+ 0.0067010560035705566,
283
+ 0.006680575847625733,
284
+ 0.006657023906707763,
285
+ 0.006687744140625,
286
+ 0.006779903888702392,
287
+ 0.006806528091430664,
288
+ 0.0068321280479431154,
289
+ 0.006803455829620361,
290
+ 0.006863935947418213,
291
+ 0.006781951904296875,
292
+ 0.007008255958557129,
293
+ 0.006789120197296142,
294
+ 0.006809599876403808,
295
+ 0.006808576107025147,
296
+ 0.0068055038452148435,
297
+ 0.006838272094726563,
298
+ 0.0068321280479431154,
299
+ 0.0068321280479431154,
300
+ 0.006818816184997558,
301
+ 0.006815743923187256,
302
+ 0.00679423999786377,
303
+ 0.0067983360290527345
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
304
  ]
305
  },
306
  "throughput": {
307
  "unit": "samples/s",
308
+ "value": 147.0330529763518
309
  },
310
  "energy": {
311
  "unit": "kWh",
312
+ "cpu": 7.700848735569603e-08,
313
+ "ram": 4.210524774954436e-08,
314
+ "gpu": 1.3937962088311895e-07,
315
+ "total": 2.584933559883593e-07
316
  },
317
  "efficiency": {
318
  "unit": "samples/kWh",
319
+ "value": 3868571.3842681237
320
  }
321
  }
322
  }