IlyasMoutawwakil HF staff commited on
Commit
7d8be0a
·
verified ·
1 Parent(s): 6a988a5

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.6.0.dev20240917+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 811.147264,
108
  "max_global_vram": 1192.7552,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 8.4949970703125,
117
- "mean": 8.4949970703125,
118
  "stdev": 0.0,
119
- "p50": 8.4949970703125,
120
- "p90": 8.4949970703125,
121
- "p95": 8.4949970703125,
122
- "p99": 8.4949970703125,
123
  "values": [
124
- 8.4949970703125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 1.0381181749873274e-06,
131
- "ram": 5.496451973902203e-07,
132
- "gpu": 1.6905569080044591e-06,
133
- "total": 3.2783202803820067e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1098.928128,
141
  "max_global_vram": 1203.24096,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
@@ -146,180 +146,180 @@
146
  "latency": {
147
  "unit": "s",
148
  "count": 150,
149
- "total": 1.0000793929100034,
150
- "mean": 0.006667195952733357,
151
- "stdev": 0.0002283985679161313,
152
- "p50": 0.006562832117080688,
153
- "p90": 0.006862131214141846,
154
- "p95": 0.007115588784217832,
155
- "p99": 0.007373793277740478,
156
  "values": [
157
- 0.00739737606048584,
158
- 0.00734822416305542,
159
- 0.007349247932434082,
160
- 0.007325695991516113,
161
- 0.007342112064361572,
162
- 0.00729804801940918,
163
- 0.0077281279563903805,
164
- 0.00672051191329956,
165
- 0.006551551818847656,
166
- 0.006491136074066162,
167
- 0.006552544116973877,
168
- 0.0065177597999572755,
169
- 0.006472703933715821,
170
- 0.006487040042877197,
171
- 0.006466559886932373,
172
- 0.006631392002105713,
173
- 0.006508543968200684,
174
- 0.0064778242111206055,
175
- 0.006489088058471679,
176
- 0.00652288007736206,
177
- 0.006461440086364746,
178
- 0.006498367786407471,
179
- 0.006457344055175781,
 
 
 
 
 
 
 
 
 
 
 
 
180
  0.006479872226715088,
181
- 0.006446176052093506,
182
- 0.006465407848358154,
183
- 0.006481919765472412,
184
- 0.006493120193481445,
185
- 0.006486015796661377,
186
- 0.006529024124145508,
187
- 0.006497280120849609,
188
- 0.006471680164337158,
189
- 0.006529024124145508,
190
- 0.00648089599609375,
191
- 0.006486015796661377,
192
- 0.006505472183227539,
193
- 0.006505472183227539,
194
- 0.006505472183227539,
195
- 0.006511583805084229,
196
- 0.006452223777770996,
197
- 0.006497280120849609,
198
- 0.006453248023986816,
199
- 0.006473696231842041,
200
- 0.006495232105255127,
201
- 0.006467584133148193,
202
- 0.006505472183227539,
203
- 0.006470655918121338,
204
- 0.006436863899230957,
205
- 0.00648089599609375,
206
- 0.006457344055175781,
207
- 0.00648089599609375,
208
- 0.006499328136444092,
209
- 0.006463488101959228,
210
- 0.006472703933715821,
211
- 0.006519807815551758,
212
- 0.006482944011688232,
213
- 0.006515711784362793,
214
- 0.006475776195526123,
215
- 0.0064880638122558594,
216
- 0.006509568214416504,
217
- 0.006498303890228272,
218
- 0.006498303890228272,
 
 
 
 
 
 
 
 
 
 
 
219
  0.006651904106140137,
220
- 0.007050240039825439,
221
- 0.006662144184112549,
222
- 0.0065731201171875,
223
- 0.0065382399559021,
224
- 0.006533120155334473,
225
- 0.006536191940307618,
226
- 0.006504447937011719,
227
- 0.006533120155334473,
228
- 0.006497280120849609,
229
- 0.00648089599609375,
230
- 0.006497280120849609,
231
- 0.006486015796661377,
232
- 0.0066078720092773435,
233
- 0.006504479885101318,
234
- 0.006496255874633789,
235
- 0.006490111827850342,
236
- 0.006474751949310303,
237
- 0.006418464183807373,
238
- 0.006540287971496582,
239
- 0.006493184089660644,
240
- 0.006466559886932373,
241
- 0.006487040042877197,
242
- 0.0064839677810668945,
243
- 0.006500288009643555,
244
- 0.0065484800338745115,
245
- 0.006481919765472412,
246
  0.00667955207824707,
247
- 0.0067358717918396,
248
- 0.006667263984680176,
249
- 0.006686783790588379,
250
- 0.0067051520347595215,
251
- 0.006672383785247803,
252
- 0.006683648109436035,
253
- 0.006781951904296875,
 
 
 
 
 
 
 
 
 
 
 
254
  0.006685696125030518,
255
- 0.006750207901000976,
256
- 0.006729728221893311,
257
- 0.006724639892578125,
258
- 0.0067010560035705566,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
259
  0.0067420158386230465,
260
- 0.006711296081542969,
261
- 0.006677504062652588,
262
- 0.0069847040176391605,
263
- 0.007169055938720703,
264
- 0.00703385591506958,
265
- 0.006808576107025147,
266
- 0.006817791938781738,
267
- 0.0068249602317810056,
268
- 0.006773759841918945,
269
- 0.0068280320167541505,
270
- 0.006817791938781738,
271
- 0.006818816184997558,
272
- 0.006738880157470703,
273
- 0.006820864200592041,
274
- 0.006819839954376221,
275
- 0.006765567779541016,
276
- 0.0068055038452148435,
277
- 0.00693555212020874,
278
- 0.006838272094726563,
279
- 0.00682809591293335,
280
- 0.00687718391418457,
281
- 0.0068618240356445315,
282
- 0.00683622407913208,
283
- 0.006864895820617676,
284
- 0.006887423992156983,
285
- 0.006848512172698974,
286
- 0.006766592025756836,
287
- 0.006801407814025879,
288
- 0.006759424209594727,
289
- 0.006802432060241699,
290
- 0.006804480075836182,
291
- 0.0068310718536376954,
292
- 0.0067645440101623535,
293
- 0.006723584175109864,
294
- 0.006778880119323731,
295
- 0.006763519763946534,
296
- 0.006752255916595459,
297
- 0.006788095951080322,
298
- 0.006798367977142334,
299
- 0.006790143966674805,
300
- 0.006812672138214112,
301
- 0.006823935985565186,
302
- 0.006815711975097657,
303
- 0.006761472225189209,
304
- 0.006790143966674805,
305
- 0.0067983360290527345,
306
- 0.006808576107025147
307
  ]
308
  },
309
  "throughput": {
310
  "unit": "samples/s",
311
- "value": 149.98809200890952
312
  },
313
  "energy": {
314
  "unit": "kWh",
315
- "cpu": 7.823313306831999e-08,
316
- "ram": 4.269056784790297e-08,
317
- "gpu": 1.55718326328967e-07,
318
- "total": 2.7664202724518995e-07
319
  },
320
  "efficiency": {
321
  "unit": "samples/kWh",
322
- "value": 3614779.756922806
323
  }
324
  }
325
  }
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 799.162368,
108
  "max_global_vram": 1192.7552,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.9870517578125,
117
+ "mean": 7.9870517578125,
118
  "stdev": 0.0,
119
+ "p50": 7.9870517578125,
120
+ "p90": 7.9870517578125,
121
+ "p95": 7.9870517578125,
122
+ "p99": 7.9870517578125,
123
  "values": [
124
+ 7.9870517578125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 2.523608562500199e-06,
131
+ "ram": 1.3423439428583533e-06,
132
+ "gpu": 3.62166956399939e-06,
133
+ "total": 7.487622069357942e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1089.286144,
141
  "max_global_vram": 1203.24096,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
 
146
  "latency": {
147
  "unit": "s",
148
  "count": 150,
149
+ "total": 0.9992682905197142,
150
+ "mean": 0.006661788603464763,
151
+ "stdev": 0.00021736465453809132,
152
+ "p50": 0.006667263984680176,
153
+ "p90": 0.006838272094726563,
154
+ "p95": 0.006951833581924438,
155
+ "p99": 0.007703009381294247,
156
  "values": [
157
+ 0.007003136157989502,
158
+ 0.006979584217071533,
159
+ 0.006907904148101806,
160
+ 0.006960127830505371,
161
+ 0.006931456089019775,
162
+ 0.0069283838272094726,
163
+ 0.006832096099853515,
164
+ 0.006941696166992187,
165
+ 0.007547904014587403,
166
+ 0.007121920108795166,
167
+ 0.006694911956787109,
168
+ 0.006699007987976074,
169
+ 0.006686719894409179,
170
+ 0.00667955207824707,
171
+ 0.00664572811126709,
172
+ 0.006675456047058105,
173
+ 0.006659071922302246,
174
+ 0.006669312000274658,
175
+ 0.0066938881874084475,
176
+ 0.006556672096252441,
177
+ 0.0064471039772033695,
178
+ 0.006456352233886718,
179
+ 0.00642252779006958,
180
+ 0.006500351905822754,
181
+ 0.006432767868041992,
182
+ 0.0064174079895019534,
183
+ 0.0066344962120056155,
184
+ 0.0067051520347595215,
185
+ 0.006614016056060791,
186
+ 0.006680575847625733,
187
+ 0.0066713919639587406,
188
+ 0.006670335769653321,
189
+ 0.0066713600158691405,
190
+ 0.006670335769653321,
191
+ 0.0064245758056640625,
192
  0.006479872226715088,
193
+ 0.006460415840148926,
194
+ 0.00658841609954834,
195
+ 0.006749184131622315,
196
+ 0.00662937593460083,
197
+ 0.0066713600158691405,
198
+ 0.006686719894409179,
199
+ 0.006673408031463623,
200
+ 0.006651904106140137,
201
+ 0.0066499199867248535,
202
+ 0.006619135856628418,
203
+ 0.006856704235076904,
204
+ 0.006690815925598144,
205
+ 0.006636544227600098,
206
+ 0.006627327919006347,
207
+ 0.006659071922302246,
208
+ 0.006662112236022949,
209
+ 0.0066375679969787596,
210
+ 0.006710271835327148,
211
+ 0.006651904106140137,
212
+ 0.006603775978088379,
213
+ 0.0066406397819519045,
214
+ 0.006633471965789795,
215
+ 0.006594560146331787,
216
+ 0.0066344962120056155,
217
+ 0.0066119680404663084,
218
+ 0.006616000175476074,
219
+ 0.006623231887817383,
220
+ 0.006631423950195312,
221
+ 0.006433792114257812,
222
+ 0.0066119680404663084,
223
+ 0.0068618240356445315,
224
+ 0.006642687797546387,
225
+ 0.006721504211425781,
226
+ 0.006743040084838867,
227
+ 0.006396927833557129,
228
+ 0.006421504020690918,
229
+ 0.00638976001739502,
230
+ 0.006434815883636475,
231
+ 0.006435840129852295,
232
+ 0.0064174079895019534,
233
+ 0.006412288188934326,
234
+ 0.0064204797744750975,
235
+ 0.006399968147277832,
236
+ 0.006433792114257812,
237
+ 0.00638259220123291,
238
+ 0.00683622407913208,
239
+ 0.0066938881874084475,
240
+ 0.006696959972381592,
241
+ 0.006631423950195312,
242
  0.006651904106140137,
243
+ 0.006799424171447754,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
244
  0.00667955207824707,
245
+ 0.006668288230895996,
246
+ 0.0066007041931152345,
247
+ 0.006624192237854004,
248
+ 0.006552576065063476,
249
+ 0.0063907837867736815,
250
+ 0.006440959930419922,
251
+ 0.00642252779006958,
252
+ 0.0063836159706115725,
253
+ 0.00642252779006958,
254
+ 0.006401023864746094,
255
+ 0.006421504020690918,
256
+ 0.0064737281799316405,
257
+ 0.006402048110961914,
258
+ 0.006433792114257812,
259
+ 0.006378496170043945,
260
+ 0.006402048110961914,
261
+ 0.0064102401733398436,
262
+ 0.0067758078575134275,
263
  0.006685696125030518,
264
+ 0.006919167995452881,
265
+ 0.00785203218460083,
266
+ 0.007924736022949219,
267
+ 0.007020544052124023,
268
+ 0.0065771517753601075,
269
+ 0.006684671878814697,
270
+ 0.006716415882110595,
271
+ 0.006659071922302246,
272
+ 0.006713344097137451,
273
+ 0.006700032234191895,
274
+ 0.006638591766357422,
275
+ 0.0066938881874084475,
276
+ 0.00673689603805542,
277
+ 0.0066713600158691405,
278
+ 0.006726719856262207,
279
+ 0.006718463897705078,
280
+ 0.0066744318008422855,
281
+ 0.006740992069244385,
282
  0.0067420158386230465,
283
+ 0.006654975891113281,
284
+ 0.006713344097137451,
285
+ 0.006696959972381592,
286
+ 0.006651904106140137,
287
+ 0.006721536159515381,
288
+ 0.006672383785247803,
289
+ 0.006632448196411133,
290
+ 0.00672051191329956,
291
+ 0.006719488143920899,
292
+ 0.00667955207824707,
293
+ 0.00669593620300293,
294
+ 0.00673689603805542,
295
+ 0.006690815925598144,
296
+ 0.006667263984680176,
297
+ 0.0066877121925354,
298
+ 0.006669312000274658,
299
+ 0.006659071922302246,
300
+ 0.0066938881874084475,
301
+ 0.0066826238632202144,
302
+ 0.006661119937896728,
303
+ 0.0067010560035705566,
304
+ 0.0067041277885437015,
305
+ 0.006667263984680176,
306
+ 0.006722559928894043
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
307
  ]
308
  },
309
  "throughput": {
310
  "unit": "samples/s",
311
+ "value": 150.10983679066385
312
  },
313
  "energy": {
314
  "unit": "kWh",
315
+ "cpu": 7.672732945340256e-08,
316
+ "ram": 4.190022030068636e-08,
317
+ "gpu": 1.4601803796129068e-07,
318
+ "total": 2.6464558771537963e-07
319
  },
320
  "efficiency": {
321
  "unit": "samples/kWh",
322
+ "value": 3778638.4750743606
323
  }
324
  }
325
  }