IlyasMoutawwakil HF staff commited on
Commit
bf8a220
·
verified ·
1 Parent(s): 83b008c

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
- "model": "FacebookAI/roberta-base",
10
  "library": "transformers",
 
 
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -104,7 +102,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 889.99936,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,183 +110,189 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 152,
116
- "total": 0.9956670055389397,
117
- "mean": 0.006550440825914081,
118
- "stdev": 0.00020321217843591294,
119
- "p50": 0.0066206719875335695,
120
- "p90": 0.006772531270980835,
121
- "p95": 0.006805401682853699,
122
- "p99": 0.0069843455123901384,
123
  "values": [
124
- 0.007175168037414551,
125
- 0.006770688056945801,
126
- 0.006804480075836182,
127
- 0.00676966381072998,
128
- 0.006762495994567871,
129
- 0.00672051191329956,
130
- 0.00677785587310791,
131
- 0.0067983360290527345,
132
- 0.006725632190704346,
133
- 0.0066979842185974124,
134
- 0.0067358717918396,
135
- 0.006731808185577393,
136
- 0.006686719894409179,
137
- 0.006729728221893311,
138
- 0.006681600093841553,
139
- 0.006598656177520752,
140
- 0.006647808074951172,
141
- 0.0066826238632202144,
142
- 0.006649856090545654,
143
- 0.006793216228485107,
144
- 0.006772736072540283,
145
- 0.006803455829620361,
146
- 0.006750207901000976,
147
- 0.006721536159515381,
148
- 0.006661119937896728,
149
- 0.006603775978088379,
150
- 0.007080959796905518,
151
- 0.006809599876403808,
152
- 0.006762495994567871,
153
- 0.006879231929779053,
154
- 0.006731776237487793,
155
- 0.006702079772949219,
156
- 0.006599679946899414,
157
- 0.006721471786499024,
158
- 0.00672051191329956,
159
- 0.006696959972381592,
160
- 0.00667955207824707,
161
- 0.006551551818847656,
162
- 0.006511616230010986,
163
- 0.0066375679969787596,
164
- 0.0066447358131408694,
165
- 0.006590464115142822,
166
- 0.006604800224304199,
167
- 0.0067276802062988285,
168
- 0.006683648109436035,
169
- 0.006690815925598144,
170
- 0.006752255916595459,
171
- 0.006685696125030518,
172
- 0.006804480075836182,
173
- 0.006746111869812011,
174
- 0.006694911956787109,
175
- 0.006627327919006347,
176
- 0.006680575847625733,
177
- 0.006680575847625733,
178
- 0.006686719894409179,
179
- 0.0067040958404541015,
180
- 0.0068915200233459475,
181
- 0.006599679946899414,
182
- 0.006646783828735352,
183
- 0.006622208118438721,
184
- 0.006590496063232422,
185
- 0.006673408031463623,
186
- 0.006874112129211426,
187
- 0.0067358717918396,
188
- 0.006647808074951172,
189
- 0.006685696125030518,
190
- 0.006628352165222168,
191
- 0.006694911956787109,
192
- 0.006680575847625733,
193
- 0.006603775978088379,
194
- 0.006619135856628418,
195
- 0.006642687797546387,
196
- 0.006646783828735352,
197
- 0.006806528091430664,
198
- 0.006780928134918213,
199
- 0.006750207901000976,
200
- 0.006840320110321045,
201
- 0.0067338237762451176,
202
- 0.006661119937896728,
203
- 0.0067010560035705566,
204
- 0.00667852783203125,
205
  0.006638591766357422,
206
- 0.006707200050354004,
207
- 0.006676544189453125,
208
- 0.006625279903411865,
209
- 0.006465536117553711,
210
- 0.006516767978668213,
211
- 0.006635519981384277,
212
- 0.006512639999389648,
213
- 0.006341631889343262,
214
- 0.00628326416015625,
215
- 0.00639686393737793,
216
- 0.006387712001800537,
217
- 0.0064849920272827145,
218
- 0.0064245758056640625,
219
- 0.0064440321922302245,
220
- 0.006658048152923584,
221
- 0.006590464115142822,
222
- 0.006300672054290772,
223
- 0.006344704151153564,
224
- 0.006285312175750732,
225
- 0.0063836159706115725,
 
 
 
 
226
  0.0063907837867736815,
227
- 0.006364160060882569,
228
- 0.006388735771179199,
229
- 0.0063272957801818845,
230
- 0.006329343795776367,
231
- 0.006335487842559814,
232
- 0.006294528007507324,
233
- 0.0063836159706115725,
234
- 0.006318079948425293,
235
- 0.006377471923828125,
236
- 0.006335487842559814,
237
- 0.00626585578918457,
238
- 0.006303743839263916,
239
- 0.006280159950256348,
240
- 0.006305791854858398,
241
- 0.006290527820587158,
242
  0.006284287929534912,
243
- 0.006308864116668702,
244
- 0.00626585578918457,
245
- 0.006316031932830811,
246
- 0.006288383960723877,
247
- 0.006289408206939697,
248
- 0.006276095867156982,
249
- 0.006264832019805908,
250
- 0.006288383960723877,
251
- 0.00626585578918457,
252
- 0.006355967998504639,
253
- 0.006362112045288086,
254
- 0.006397952079772949,
255
- 0.00643891191482544,
256
- 0.006344704151153564,
257
- 0.006339583873748779,
258
- 0.006277120113372803,
259
- 0.006322175979614258,
260
- 0.0062904319763183595,
261
- 0.006276095867156982,
262
- 0.006298624038696289,
263
  0.00628326416015625,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
264
  0.006301695823669433,
265
- 0.006281216144561768,
266
- 0.006293504238128662,
267
- 0.006333439826965332,
268
- 0.006250495910644531,
269
- 0.0062873601913452145,
270
- 0.006246399879455566,
271
- 0.006301695823669433,
272
- 0.006306816101074219,
273
- 0.006334464073181153,
274
- 0.00637440013885498,
275
- 0.006345727920532227
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276
  ]
277
  },
278
  "throughput": {
279
  "unit": "samples/s",
280
- "value": 152.66148135312022
281
  },
282
  "energy": {
283
  "unit": "kWh",
284
- "cpu": 7.38967598266154e-08,
285
- "ram": 4.039728372173039e-08,
286
- "gpu": 1.3393730521772734e-07,
287
- "total": 2.4823134876607313e-07
288
  },
289
  "efficiency": {
290
  "unit": "samples/kWh",
291
- "value": 4028500.0463111307
292
  }
293
  }
294
  }
 
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
 
9
  "library": "transformers",
10
+ "model": "FacebookAI/roberta-base",
11
+ "processor": "FacebookAI/roberta-base",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 890.109952,
106
  "max_global_vram": 1195.900928,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 555.74528,
 
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 158,
114
+ "total": 0.9970523505210878,
115
+ "mean": 0.006310457914690428,
116
+ "stdev": 0.0002560881718122158,
117
+ "p50": 0.006318080186843872,
118
+ "p90": 0.006635724878311157,
119
+ "p95": 0.006662758588790894,
120
+ "p99": 0.0069552436161041285,
121
  "values": [
122
+ 0.0071823358535766605,
123
+ 0.006796288013458252,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
124
  0.006638591766357422,
125
+ 0.006532095909118653,
126
+ 0.006515711784362793,
127
+ 0.00657919979095459,
128
+ 0.006516736030578613,
129
+ 0.006594560146331787,
130
+ 0.006526976108551025,
131
+ 0.006546432018280029,
132
+ 0.006492159843444824,
133
+ 0.00643884801864624,
134
+ 0.006647776126861573,
135
+ 0.006475776195526123,
136
+ 0.0063569917678833006,
137
+ 0.006449151992797852,
138
+ 0.0066078720092773435,
139
+ 0.006359039783477783,
140
+ 0.006534143924713135,
141
+ 0.006412288188934326,
142
+ 0.006412288188934326,
143
+ 0.0064102401733398436,
144
+ 0.006450111865997314,
145
+ 0.006401023864746094,
146
+ 0.006342656135559082,
147
+ 0.006426623821258545,
148
+ 0.00643177604675293,
149
  0.0063907837867736815,
150
+ 0.006650879859924317,
151
+ 0.006445055961608887,
152
+ 0.00641539192199707,
153
+ 0.0065730881690979,
154
+ 0.006662144184112549,
155
+ 0.006586368083953857,
156
+ 0.006615039825439453,
157
+ 0.006528031826019287,
158
+ 0.006260735988616943,
 
 
 
 
 
 
159
  0.006284287929534912,
160
+ 0.0062975997924804685,
161
+ 0.006326272010803223,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
  0.00628326416015625,
163
+ 0.006395904064178467,
164
+ 0.006458367824554443,
165
+ 0.0066447358131408694,
166
+ 0.0063836159706115725,
167
+ 0.006434815883636475,
168
+ 0.006432735919952392,
169
+ 0.006545407772064209,
170
+ 0.00638976001739502,
171
+ 0.006407167911529541,
172
+ 0.006351871967315674,
173
+ 0.0064122557640075685,
174
+ 0.006401023864746094,
175
+ 0.006435840129852295,
176
+ 0.00638156795501709,
177
+ 0.006379519939422608,
178
+ 0.006363167762756348,
179
+ 0.006303743839263916,
180
+ 0.006319104194641113,
181
+ 0.006557695865631104,
182
+ 0.0065443840026855465,
183
+ 0.006644768238067627,
184
+ 0.006666240215301514,
185
+ 0.006563839912414551,
186
+ 0.006585343837738037,
187
+ 0.006558720111846924,
188
+ 0.0063170561790466305,
189
+ 0.006618144035339356,
190
+ 0.006561791896820069,
191
+ 0.006558784008026123,
192
+ 0.0065443840026855465,
193
+ 0.006658048152923584,
194
+ 0.006669312000274658,
195
+ 0.006576128005981445,
196
+ 0.006615039825439453,
197
+ 0.0066344962120056155,
198
+ 0.0067276802062988285,
199
+ 0.0066406397819519045,
200
+ 0.00659660816192627,
201
+ 0.00657203197479248,
202
+ 0.006612991809844971,
203
+ 0.006273024082183838,
204
+ 0.0062679038047790524,
205
+ 0.006136832237243653,
206
+ 0.006130688190460205,
207
  0.006301695823669433,
208
+ 0.006151167869567871,
209
+ 0.006024127960205078,
210
+ 0.00602623987197876,
211
+ 0.006073344230651856,
212
+ 0.00608460807800293,
213
+ 0.0061562881469726565,
214
+ 0.006109183788299561,
215
+ 0.006147039890289307,
216
+ 0.006329343795776367,
217
+ 0.006247424125671387,
218
+ 0.006077439785003662,
219
+ 0.006056960105895996,
220
+ 0.006089727878570556,
221
+ 0.006043680191040039,
222
+ 0.006047743797302246,
223
+ 0.006082560062408447,
224
+ 0.006108160018920898,
225
+ 0.006789120197296142,
226
+ 0.006312960147857666,
227
+ 0.006086656093597412,
228
+ 0.006107135772705078,
229
+ 0.006138912200927734,
230
+ 0.00608460807800293,
231
+ 0.006074368000030517,
232
+ 0.006060031890869141,
233
+ 0.006069248199462891,
234
+ 0.006032383918762207,
235
+ 0.006051839828491211,
236
+ 0.006033408164978027,
237
+ 0.006030303955078125,
238
+ 0.006017024040222168,
239
+ 0.006015999794006348,
240
+ 0.00601087999343872,
241
+ 0.00602623987197876,
242
+ 0.006001664161682129,
243
+ 0.006013887882232666,
244
+ 0.005984255790710449,
245
+ 0.006004735946655273,
246
+ 0.0060026879310607914,
247
+ 0.006115327835083008,
248
+ 0.006007808208465576,
249
+ 0.00601087999343872,
250
+ 0.006038527965545654,
251
+ 0.006055935859680176,
252
+ 0.006069248199462891,
253
+ 0.006056960105895996,
254
+ 0.006047743797302246,
255
+ 0.0060067839622497555,
256
+ 0.006005760192871094,
257
+ 0.006004735946655273,
258
+ 0.0059996161460876465,
259
+ 0.006009856224060059,
260
+ 0.006031360149383545,
261
+ 0.006037504196166992,
262
+ 0.006014976024627685,
263
+ 0.006039616107940674,
264
+ 0.006039552211761475,
265
+ 0.0060293121337890625,
266
+ 0.0067010560035705566,
267
+ 0.006269951820373535,
268
+ 0.00603436803817749,
269
+ 0.006000639915466309,
270
+ 0.0060661759376525876,
271
+ 0.006037504196166992,
272
+ 0.006052864074707031,
273
+ 0.006107135772705078,
274
+ 0.006077439785003662,
275
+ 0.0060702719688415525,
276
+ 0.006032383918762207,
277
+ 0.006038527965545654,
278
+ 0.006001664161682129,
279
+ 0.007165952205657959
280
  ]
281
  },
282
  "throughput": {
283
  "unit": "samples/s",
284
+ "value": 158.46710548089553
285
  },
286
  "energy": {
287
  "unit": "kWh",
288
+ "cpu": 7.606786632909053e-08,
289
+ "ram": 4.140192918182102e-08,
290
+ "gpu": 1.3334759429298747e-07,
291
+ "total": 2.5081738980389904e-07
292
  },
293
  "efficiency": {
294
  "unit": "samples/kWh",
295
+ "value": 3986964.3838565084
296
  }
297
  }
298
  }