IlyasMoutawwakil HF staff commited on
Commit
e3db7fe
·
verified ·
1 Parent(s): 661d7cc

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 788.979712,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.34181884765625,
117
- "mean": 7.34181884765625,
118
  "stdev": 0.0,
119
- "p50": 7.34181884765625,
120
- "p90": 7.34181884765625,
121
- "p95": 7.34181884765625,
122
- "p99": 7.34181884765625,
123
  "values": [
124
- 7.34181884765625
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 9.00526207638682e-07,
131
- "ram": 4.695824427181267e-07,
132
- "gpu": 1.8308347980002318e-06,
133
- "total": 3.2009434483570403e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1084.071936,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
@@ -145,160 +145,169 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 129,
149
- "total": 1.0027385268211362,
150
- "mean": 0.007773166874582453,
151
- "stdev": 0.00026103824626296467,
152
- "p50": 0.007662623882293701,
153
- "p90": 0.008104959678649902,
154
- "p95": 0.008169862365722657,
155
- "p99": 0.00832499698638916,
156
  "values": [
157
- 0.008340479850769043,
158
- 0.008426495552062988,
159
- 0.008285183906555176,
160
- 0.008220671653747558,
161
- 0.008179679870605468,
162
- 0.00808448028564453,
163
- 0.008155136108398438,
164
- 0.008054783821105957,
165
- 0.008013824462890624,
166
- 0.008092703819274903,
167
- 0.008064000129699708,
168
- 0.007966720104217529,
169
- 0.007941120147705078,
170
- 0.008018943786621094,
171
- 0.008047616004943848,
172
- 0.00810086441040039,
173
- 0.00818073558807373,
174
- 0.008060928344726562,
175
- 0.007662623882293701,
176
- 0.007616511821746826,
177
- 0.007732223987579345,
178
- 0.007681024074554443,
179
- 0.007681024074554443,
180
- 0.00787660789489746,
181
- 0.008101887702941894,
182
- 0.008072192192077637,
183
- 0.008031231880187988,
184
- 0.008036352157592774,
185
- 0.008030207633972167,
186
- 0.008017919540405273,
187
- 0.00797388792037964,
188
- 0.007996416091918946,
189
- 0.008005632400512695,
190
- 0.008083456039428711,
191
- 0.008020992279052735,
192
- 0.00800556755065918,
193
- 0.007994368076324462,
194
- 0.008091648101806641,
195
- 0.008091648101806641,
196
- 0.00809881591796875,
197
- 0.008069120407104492,
198
- 0.007779327869415284,
199
- 0.00760319995880127,
200
- 0.0077199358940124516,
201
- 0.007696352005004883,
202
- 0.007658463954925537,
203
  0.007753727912902832,
204
- 0.007641088008880615,
205
- 0.007588863849639893,
206
- 0.0075857601165771485,
207
- 0.007631840229034424,
208
- 0.007600128173828125,
209
- 0.007636991977691651,
210
- 0.007588863849639893,
211
- 0.007575551986694336,
212
- 0.007623648166656494,
213
- 0.007633920192718506,
214
- 0.0076574721336364745,
215
- 0.008117247581481933,
216
- 0.007996416091918946,
217
- 0.008026111602783203,
218
- 0.008124416351318359,
219
- 0.008088607788085938,
220
- 0.008140800476074218,
221
- 0.008133631706237793,
222
- 0.008013824462890624,
223
- 0.00800972843170166,
224
- 0.007648255825042725,
225
- 0.00759500789642334,
226
- 0.007740352153778076,
227
- 0.00759603214263916,
228
- 0.007572480201721191,
229
- 0.007435264110565186,
230
- 0.007443456172943115,
231
- 0.0075970559120178225,
232
- 0.00742195177078247,
233
- 0.007271423816680909,
234
- 0.0073768959045410155,
235
- 0.007307263851165771,
236
- 0.00785100793838501,
237
- 0.007945216178894043,
238
- 0.007985151767730713,
239
- 0.008118271827697754,
240
- 0.008212479591369629,
241
- 0.007942143917083741,
242
- 0.00778547191619873,
243
- 0.00769536018371582,
244
- 0.00773529577255249,
245
- 0.007700543880462646,
246
- 0.007704576015472412,
247
- 0.007650303840637207,
248
- 0.007574560165405274,
249
- 0.0076574721336364745,
250
- 0.007618559837341309,
251
- 0.007600128173828125,
252
- 0.007596992015838623,
253
- 0.007561215877532959,
254
- 0.007628799915313721,
255
- 0.007566336154937744,
256
- 0.007550975799560547,
257
- 0.007562240123748779,
258
- 0.007609344005584716,
259
- 0.007532544136047363,
260
- 0.007557119846343994,
261
- 0.0075980801582336424,
262
- 0.007585792064666748,
263
- 0.007577663898468018,
264
- 0.007584767818450928,
265
- 0.007589888095855713,
266
- 0.0075632638931274416,
267
- 0.007562240123748779,
268
- 0.007543807983398438,
269
- 0.007521279811859131,
270
- 0.007561215877532959,
271
- 0.007579584121704102,
272
  0.007580671787261963,
273
- 0.007638016223907471,
274
- 0.007558144092559814,
275
- 0.007426047801971435,
276
- 0.007479296207427978,
277
- 0.007453695774078369,
278
- 0.007400447845458984,
279
- 0.007445536136627197,
280
- 0.007527423858642578,
281
- 0.007457791805267334,
282
- 0.0074332160949707035,
283
- 0.007422976016998291,
284
- 0.007456768035888672,
285
- 0.007411712169647216
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
286
  ]
287
  },
288
  "throughput": {
289
  "unit": "samples/s",
290
- "value": 128.64769483721093
291
  },
292
  "energy": {
293
  "unit": "kWh",
294
- "cpu": 8.65917404703981e-08,
295
- "ram": 4.734078302240591e-08,
296
- "gpu": 1.6241497178102032e-07,
297
- "total": 2.9634749527382436e-07
298
  },
299
  "efficiency": {
300
  "unit": "samples/kWh",
301
- "value": 3374416.912402119
302
  }
303
  }
304
  }
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 763.826176,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.408169921875,
117
+ "mean": 7.408169921875,
118
  "stdev": 0.0,
119
+ "p50": 7.408169921875,
120
+ "p90": 7.408169921875,
121
+ "p95": 7.408169921875,
122
+ "p99": 7.408169921875,
123
  "values": [
124
+ 7.408169921875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 8.859253916665151e-07,
131
+ "ram": 4.693504693417744e-07,
132
+ "gpu": 1.7433347279998887e-06,
133
+ "total": 3.098610589008178e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 964.153344,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 138,
149
+ "total": 1.0024035539627079,
150
+ "mean": 0.0072637938692949825,
151
+ "stdev": 0.00018445064897951318,
152
+ "p50": 0.007250944137573242,
153
+ "p90": 0.007534079933166504,
154
+ "p95": 0.0076551679134368895,
155
+ "p99": 0.007683154010772705,
156
  "values": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
  0.007753727912902832,
158
+ 0.007662591934204102,
159
+ 0.007689216136932373,
160
+ 0.007659520149230957,
161
+ 0.0076605439186096195,
162
+ 0.007646207809448242,
163
+ 0.007654399871826172,
164
+ 0.00765337610244751,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
165
  0.007580671787261963,
166
+ 0.007302080154418945,
167
+ 0.007328767776489258,
168
+ 0.007309311866760254,
169
+ 0.007282688140869141,
170
+ 0.007394303798675537,
171
+ 0.007451648235321045,
172
+ 0.007322624206542969,
173
+ 0.007362559795379638,
174
+ 0.007371776103973389,
175
+ 0.007358463764190673,
176
+ 0.007300096035003662,
177
+ 0.007258111953735351,
178
+ 0.007310336112976074,
179
+ 0.0073175039291381834,
180
+ 0.007330783843994141,
181
+ 0.007326720237731933,
182
+ 0.007353280067443848,
183
+ 0.0073400321006774905,
184
+ 0.0073471999168395995,
185
+ 0.007300096035003662,
186
+ 0.007269375801086426,
187
+ 0.007327744007110596,
188
+ 0.007329792022705078,
189
+ 0.007208960056304932,
190
+ 0.007173120021820068,
191
+ 0.007250944137573242,
192
+ 0.007265279769897461,
193
+ 0.007243775844573975,
194
+ 0.007288832187652588,
195
+ 0.007305215835571289,
196
+ 0.007326720237731933,
197
+ 0.007351295948028564,
198
+ 0.007358463764190673,
199
+ 0.007373824119567871,
200
+ 0.007514111995697022,
201
+ 0.007581696033477783,
202
+ 0.007592959880828858,
203
+ 0.007662591934204102,
204
+ 0.007629824161529541,
205
+ 0.007284736156463623,
206
+ 0.007314432144165039,
207
+ 0.007319551944732666,
208
+ 0.00724070405960083,
209
+ 0.007256063938140869,
210
+ 0.007279615879058838,
211
+ 0.0072375679016113284,
212
+ 0.007388160228729248,
213
+ 0.007229375839233399,
214
+ 0.007250944137573242,
215
+ 0.007247871875762939,
216
+ 0.007250944137573242,
217
+ 0.00724070405960083,
218
+ 0.007259136199951172,
219
+ 0.007268352031707764,
220
+ 0.007333888053894043,
221
+ 0.0074035201072692874,
222
+ 0.007371808052062988,
223
+ 0.007309311866760254,
224
+ 0.007372799873352051,
225
+ 0.00729804801940918,
226
+ 0.007320576190948487,
227
+ 0.007316480159759522,
228
+ 0.007234560012817383,
229
+ 0.007243775844573975,
230
+ 0.007248928070068359,
231
+ 0.0072540159225463864,
232
+ 0.007171072006225586,
233
+ 0.007054336071014404,
234
+ 0.0071485438346862796,
235
+ 0.007217152118682861,
236
+ 0.007001088142395019,
237
+ 0.00689356803894043,
238
+ 0.006892543792724609,
239
+ 0.006977536201477051,
240
+ 0.007357439994812012,
241
+ 0.00744652795791626,
242
+ 0.007385087966918945,
243
+ 0.007445504188537597,
244
+ 0.007672832012176513,
245
+ 0.007314432144165039,
246
+ 0.0071905279159545895,
247
+ 0.007236608028411865,
248
+ 0.00722431993484497,
249
+ 0.007237631797790528,
250
+ 0.007164927959442138,
251
+ 0.007205887794494629,
252
+ 0.0071905279159545895,
253
+ 0.007231488227844239,
254
+ 0.007205887794494629,
255
+ 0.0071823358535766605,
256
+ 0.007175168037414551,
257
+ 0.007197696208953858,
258
+ 0.007172063827514648,
259
+ 0.007163904190063477,
260
+ 0.007173120021820068,
261
+ 0.007145472049713135,
262
+ 0.007134208202362061,
263
+ 0.007135231971740722,
264
+ 0.007151616096496582,
265
+ 0.007150591850280762,
266
+ 0.007096288204193116,
267
+ 0.00719155216217041,
268
+ 0.007187456130981445,
269
+ 0.007223296165466309,
270
+ 0.007136256217956543,
271
+ 0.007178239822387696,
272
+ 0.0071833600997924804,
273
+ 0.0071792640686035155,
274
+ 0.007202816009521484,
275
+ 0.00714137601852417,
276
+ 0.007175168037414551,
277
+ 0.007204864025115967,
278
+ 0.007165952205657959,
279
+ 0.007158783912658692,
280
+ 0.0071157760620117185,
281
+ 0.0071495680809020995,
282
+ 0.007156735897064209,
283
+ 0.007170048236846924,
284
+ 0.0071792640686035155,
285
+ 0.007158783912658692,
286
+ 0.00689356803894043,
287
+ 0.006879231929779053,
288
+ 0.006897664070129395,
289
+ 0.006871039867401123,
290
+ 0.006884352207183838,
291
+ 0.0068884482383728025,
292
+ 0.006870016098022461,
293
+ 0.0068689918518066405,
294
+ 0.006889472007751465
295
  ]
296
  },
297
  "throughput": {
298
  "unit": "samples/s",
299
+ "value": 137.66910487742945
300
  },
301
  "energy": {
302
  "unit": "kWh",
303
+ "cpu": 8.719489930555627e-08,
304
+ "ram": 4.766417222569977e-08,
305
+ "gpu": 1.5942006217646838e-07,
306
+ "total": 2.9427913370772445e-07
307
  },
308
  "efficiency": {
309
  "unit": "samples/kWh",
310
+ "value": 3398134.2387435175
311
  }
312
  }
313
  }