IlyasMoutawwakil HF staff commited on
Commit
50782c1
·
verified ·
1 Parent(s): d0d5e0e

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "model": "FacebookAI/roberta-base",
@@ -78,7 +78,7 @@
78
  "machine": "x86_64",
79
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
- "python_version": "3.10.14",
82
  "gpu": [
83
  "NVIDIA A10G"
84
  ],
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 890.85952,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,172 +112,163 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 141,
116
- "total": 0.9968024640083314,
117
- "mean": 0.007069521021335682,
118
- "stdev": 0.00046649140372335165,
119
- "p50": 0.007013376235961914,
120
- "p90": 0.007370751857757568,
121
- "p95": 0.00754585599899292,
122
- "p99": 0.008827085113525385,
123
  "values": [
124
- 0.007755775928497314,
125
- 0.007445504188537597,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
126
  0.00754585599899292,
127
- 0.007284736156463623,
128
- 0.007223296165466309,
129
- 0.007176191806793213,
130
- 0.007321599960327148,
131
- 0.007341055870056152,
132
- 0.007320576190948487,
133
- 0.0072765440940856935,
134
- 0.007207935810089112,
135
- 0.007109632015228271,
136
- 0.007061503887176514,
137
- 0.00698367977142334,
138
- 0.006968319892883301,
139
- 0.007165952205657959,
140
- 0.00733081579208374,
141
- 0.007111680030822754,
142
- 0.007093247890472412,
143
- 0.007079936027526855,
144
- 0.007057407855987549,
145
- 0.007068672180175781,
146
- 0.007266304016113281,
147
- 0.0074035201072692874,
148
- 0.007377920150756836,
149
- 0.007290880203247071,
150
- 0.007351295948028564,
151
- 0.007370751857757568,
152
- 0.0070553598403930665,
153
- 0.0070563840866088865,
154
- 0.007038976192474365,
155
- 0.007293951988220215,
156
- 0.007296000003814697,
157
- 0.006973440170288086,
158
- 0.006961184024810791,
159
- 0.0069816322326660156,
160
- 0.006972415924072266,
161
- 0.0070860800743103025,
162
- 0.00729702377319336,
163
- 0.007310336112976074,
164
- 0.007064576148986816,
165
- 0.0070891518592834475,
166
- 0.007114751815795899,
167
- 0.007109632015228271,
168
- 0.007017471790313721,
169
- 0.007013376235961914,
170
- 0.007076863765716553,
171
- 0.00738099193572998,
172
- 0.007284736156463623,
173
- 0.007258111953735351,
174
- 0.007108607769012451,
175
- 0.006952960014343262,
176
- 0.006978559970855713,
177
- 0.0072724480628967286,
178
- 0.0072427520751953125,
179
- 0.007350240230560303,
180
- 0.007005184173583984,
181
- 0.006959136009216309,
182
- 0.00703385591506958,
183
- 0.0069959678649902345,
184
- 0.007014400005340577,
185
- 0.0070184640884399414,
186
- 0.007667712211608887,
187
- 0.008251392364501953,
188
- 0.010863615989685058,
189
- 0.009210880279541016,
190
- 0.007447519779205322,
191
  0.007490560054779053,
192
- 0.007250944137573242,
193
- 0.007345151901245117,
194
- 0.0069847040176391605,
195
- 0.007011328220367432,
196
- 0.007137279987335205,
197
- 0.007327744007110596,
198
- 0.007160831928253173,
199
- 0.007038976192474365,
200
- 0.007303167819976806,
201
- 0.007187456130981445,
202
- 0.006875135898590088,
203
- 0.00694374418258667,
204
- 0.007101439952850342,
205
- 0.007621632099151611,
206
- 0.007279615879058838,
207
- 0.007233535766601563,
208
- 0.007558144092559814,
209
- 0.007232480049133301,
210
- 0.0069918398857116695,
211
- 0.00695091199874878,
212
- 0.00687820816040039,
213
- 0.006854656219482422,
214
- 0.006894591808319092,
215
- 0.006859776020050049,
216
- 0.006871039867401123,
217
- 0.006906879901885986,
218
- 0.00690070390701294,
219
- 0.006692863941192627,
220
- 0.006648831844329834,
221
- 0.0066406397819519045,
222
- 0.006619135856628418,
223
- 0.006755328178405762,
224
- 0.006870016098022461,
225
- 0.006801407814025879,
226
- 0.0066078720092773435,
227
- 0.0066109437942504885,
228
- 0.006653952121734619,
229
- 0.006628352165222168,
230
- 0.006623231887817383,
231
- 0.006597631931304931,
232
- 0.006576128005981445,
233
- 0.0066406397819519045,
234
- 0.006632448196411133,
235
- 0.0065781760215759275,
236
- 0.006624256134033203,
237
- 0.006623199939727783,
238
- 0.006605823993682861,
239
- 0.006710271835327148,
240
- 0.006924287796020508,
241
- 0.006959104061126709,
242
- 0.006825984001159668,
243
- 0.006865920066833496,
244
- 0.006845439910888672,
245
- 0.006860799789428711,
246
- 0.0066406397819519045,
247
- 0.006621183872222901,
248
- 0.006795263767242431,
249
- 0.006859776020050049,
250
- 0.006894591808319092,
251
- 0.006845439910888672,
252
- 0.00683622407913208,
253
- 0.006880256175994873,
254
- 0.006870016098022461,
255
- 0.006840320110321045,
256
- 0.006830080032348633,
257
- 0.006865920066833496,
258
- 0.006823967933654785,
259
- 0.006841343879699707,
260
- 0.006658048152923584,
261
- 0.006604832172393799,
262
- 0.0066304001808166506,
263
- 0.006605823993682861,
264
- 0.006546432018280029
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
265
  ]
266
  },
267
  "throughput": {
268
  "unit": "samples/s",
269
- "value": 141.45229881657028
270
  },
271
  "energy": {
272
  "unit": "kWh",
273
- "cpu": 8.134306996743247e-08,
274
- "ram": 4.446755449874853e-08,
275
- "gpu": 1.5171435272607237e-07,
276
- "total": 2.775249771922534e-07
277
  },
278
  "efficiency": {
279
  "unit": "samples/kWh",
280
- "value": 3603279.280002453
281
  }
282
  }
283
  }
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "model": "FacebookAI/roberta-base",
 
78
  "machine": "x86_64",
79
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
+ "python_version": "3.10.12",
82
  "gpu": [
83
  "NVIDIA A10G"
84
  ],
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 909.471744,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 132,
116
+ "total": 0.9997109441757203,
117
+ "mean": 0.007573567758906972,
118
+ "stdev": 0.0002575650256898117,
119
+ "p50": 0.007501823902130127,
120
+ "p90": 0.00782069435119629,
121
+ "p95": 0.00798279707431793,
122
+ "p99": 0.008806525831222534,
123
  "values": [
124
+ 0.008065024375915527,
125
+ 0.007655424118041992,
126
+ 0.007904255867004394,
127
+ 0.00790835189819336,
128
+ 0.0076605439186096195,
129
+ 0.007823359966278077,
130
+ 0.0075939841270446775,
131
+ 0.007474080085754394,
132
+ 0.007570432186126709,
133
+ 0.007572480201721191,
134
+ 0.007549952030181885,
135
+ 0.007583744049072265,
136
+ 0.007524352073669433,
137
+ 0.007505919933319092,
138
+ 0.0075008001327514645,
139
+ 0.007463935852050781,
140
+ 0.007526400089263916,
141
+ 0.007556096076965332,
142
+ 0.007731359958648681,
143
+ 0.007527423858642578,
144
+ 0.007548927783966064,
145
+ 0.007510015964508057,
146
+ 0.007579648017883301,
147
+ 0.007601151943206787,
148
+ 0.007577600002288819,
149
+ 0.007796703815460205,
150
+ 0.007511040210723877,
151
+ 0.007501823902130127,
152
+ 0.007513088226318359,
153
+ 0.007546879768371582,
154
+ 0.007589888095855713,
155
  0.00754585599899292,
156
+ 0.007463935852050781,
157
+ 0.007461887836456299,
158
+ 0.007480319976806641,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
159
  0.007490560054779053,
160
+ 0.007432191848754883,
161
+ 0.00753772783279419,
162
+ 0.0075642881393432615,
163
+ 0.007710720062255859,
164
+ 0.007526400089263916,
165
+ 0.007555071830749512,
166
+ 0.007562240123748779,
167
+ 0.007506944179534912,
168
+ 0.007612415790557861,
169
+ 0.007458816051483155,
170
+ 0.007549952030181885,
171
+ 0.007488480091094971,
172
+ 0.007457791805267334,
173
+ 0.007451648235321045,
174
+ 0.007423999786376953,
175
+ 0.007398399829864502,
176
+ 0.007413760185241699,
177
+ 0.00744652795791626,
178
+ 0.007536640167236328,
179
+ 0.007559167861938477,
180
+ 0.007459839820861816,
181
+ 0.007611392021179199,
182
+ 0.007490560054779053,
183
+ 0.007477248191833496,
184
+ 0.0074579200744628905,
185
+ 0.007560192108154297,
186
+ 0.007498752117156982,
187
+ 0.007488512039184571,
188
+ 0.007650303840637207,
189
+ 0.007612415790557861,
190
+ 0.007590911865234375,
191
+ 0.00759603214263916,
192
+ 0.007516128063201904,
193
+ 0.007517183780670166,
194
+ 0.008433664321899414,
195
+ 0.007915520191192627,
196
+ 0.0077578239440917966,
197
+ 0.0076863040924072265,
198
+ 0.007741439819335938,
199
+ 0.007726079940795898,
200
+ 0.007745535850524903,
201
+ 0.00743833589553833,
202
+ 0.0073768959045410155,
203
+ 0.007451648235321045,
204
+ 0.007625728130340576,
205
+ 0.00768287992477417,
206
+ 0.007878655910491944,
207
+ 0.007863296031951903,
208
+ 0.007883808135986329,
209
+ 0.008824831962585449,
210
+ 0.008816703796386719,
211
+ 0.008783871650695801,
212
+ 0.00810905647277832,
213
+ 0.007448575973510742,
214
+ 0.007463935852050781,
215
+ 0.007460864067077637,
216
+ 0.007413760185241699,
217
+ 0.0074106879234313965,
218
+ 0.007342016220092774,
219
+ 0.007480319976806641,
220
+ 0.007384064197540283,
221
+ 0.007354368209838867,
222
+ 0.007333888053894043,
223
+ 0.007385087966918945,
224
+ 0.0073431038856506346,
225
+ 0.0073359360694885255,
226
+ 0.007367648124694824,
227
+ 0.00742195177078247,
228
+ 0.007475200176239013,
229
+ 0.007408639907836914,
230
+ 0.007418879985809326,
231
+ 0.007418879985809326,
232
+ 0.008253439903259278,
233
+ 0.007464960098266602,
234
+ 0.007479296207427978,
235
+ 0.0073994240760803225,
236
+ 0.007431168079376221,
237
+ 0.00742195177078247,
238
+ 0.007395328044891358,
239
+ 0.007423999786376953,
240
+ 0.007435264110565186,
241
+ 0.00743833589553833,
242
+ 0.007337984085083008,
243
+ 0.007394303798675537,
244
+ 0.0074609599113464355,
245
+ 0.007390207767486572,
246
+ 0.007494656085968018,
247
+ 0.007501823902130127,
248
+ 0.007458816051483155,
249
+ 0.007428095817565918,
250
+ 0.007367680072784424,
251
+ 0.007415808200836181,
252
+ 0.007434239864349365,
253
+ 0.007391232013702393,
254
+ 0.007400447845458984,
255
+ 0.007543807983398438
256
  ]
257
  },
258
  "throughput": {
259
  "unit": "samples/s",
260
+ "value": 132.03816640102542
261
  },
262
  "energy": {
263
  "unit": "kWh",
264
+ "cpu": 8.430648639128846e-08,
265
+ "ram": 4.602209843684017e-08,
266
+ "gpu": 1.5068175963120763e-07,
267
+ "total": 2.810103444593363e-07
268
  },
269
  "efficiency": {
270
  "unit": "samples/kWh",
271
+ "value": 3558587.858834874
272
  }
273
  }
274
  }