IlyasMoutawwakil HF staff commited on
Commit
6d8fd1a
·
verified ·
1 Parent(s): 3dd622c

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
- "model": "FacebookAI/roberta-base",
10
  "library": "transformers",
 
 
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -104,7 +102,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 891.0848,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,168 +110,177 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 137,
116
- "total": 0.9985012807846071,
117
- "mean": 0.007288330516675963,
118
- "stdev": 0.00039064226336823333,
119
- "p50": 0.007333888053894043,
120
- "p90": 0.007714816093444824,
121
- "p95": 0.007736524772644042,
122
- "p99": 0.008024514427185056,
123
  "values": [
124
- 0.008078335762023926,
125
- 0.007819263935089112,
126
- 0.007928832054138184,
127
- 0.007420928001403809,
128
- 0.007478271961212158,
129
- 0.007672832012176513,
130
- 0.007803904056549072,
131
- 0.007714816093444824,
132
- 0.007662591934204102,
133
- 0.007624703884124756,
134
- 0.007658495903015137,
135
  0.0076267518997192385,
136
- 0.00764518404006958,
137
- 0.007608320236206055,
138
- 0.007521279811859131,
139
- 0.00758681583404541,
140
- 0.007667712211608887,
141
- 0.007723008155822754,
142
- 0.0073471999168395995,
143
- 0.007333888053894043,
144
- 0.007615488052368164,
145
- 0.007778304100036621,
146
- 0.007634943962097168,
147
- 0.00764415979385376,
148
- 0.007649280071258545,
149
- 0.007675903797149658,
150
- 0.007658495903015137,
151
- 0.007714816093444824,
152
- 0.007659584045410156,
153
- 0.0076912641525268555,
154
- 0.007887872219085693,
155
- 0.007681024074554443,
156
- 0.007607295989990235,
157
- 0.007547904014587403,
158
- 0.007715839862823487,
159
- 0.007580671787261963,
160
- 0.007609344005584716,
161
- 0.007590911865234375,
162
- 0.007566336154937744,
163
- 0.0075857601165771485,
164
- 0.007656447887420655,
165
- 0.007716864109039307,
166
- 0.007726079940795898,
167
- 0.007718912124633789,
168
- 0.007688191890716553,
169
- 0.0076574721336364745,
170
- 0.007605216026306152,
171
- 0.007699456214904785,
172
- 0.007669760227203369,
173
- 0.0076943359375,
174
- 0.007685120105743408,
175
- 0.007606272220611572,
176
- 0.007539711952209473,
177
- 0.007684095859527588,
178
- 0.00760422420501709,
179
- 0.00760319995880127,
180
- 0.007721983909606934,
181
- 0.007548927783966064,
182
- 0.0075304961204528805,
183
- 0.00758784008026123,
184
- 0.0076605439186096195,
185
- 0.007640063762664795,
186
- 0.007623680114746094,
187
- 0.007616511821746826,
188
- 0.007616511821746826,
189
- 0.008318976402282715,
190
- 0.007374847888946533,
191
- 0.007301119804382325,
192
- 0.007356416225433349,
193
- 0.007263296127319336,
194
- 0.007266304016113281,
195
- 0.0073431038856506346,
196
- 0.007196671962738037,
197
- 0.00729702377319336,
198
- 0.007200767993927002,
199
- 0.007195648193359375,
200
- 0.007072768211364746,
201
- 0.007046144008636474,
202
- 0.0072468481063842774,
203
- 0.007031807899475098,
204
- 0.006882304191589355,
205
- 0.006863872051239014,
206
- 0.00703385591506958,
207
- 0.007143424034118652,
208
- 0.007136256217956543,
209
- 0.007207935810089112,
210
- 0.007207935810089112,
211
- 0.007264256000518799,
212
- 0.0069918718338012695,
213
- 0.006887423992156983,
214
- 0.006803455829620361,
215
- 0.006867968082427979,
216
- 0.006855743885040283,
217
- 0.006819839954376221,
218
- 0.006842368125915528,
219
- 0.006895616054534912,
220
- 0.006856704235076904,
221
  0.006823935985565186,
222
- 0.006842368125915528,
223
- 0.006850560188293457,
224
- 0.0068392958641052244,
225
- 0.006807551860809326,
226
- 0.006833151817321777,
227
- 0.006863872051239014,
228
  0.006806528091430664,
229
- 0.006850560188293457,
230
- 0.006842368125915528,
231
- 0.006862847805023193,
232
- 0.006789120197296142,
233
- 0.0068494720458984375,
234
- 0.00682700777053833,
235
- 0.006850560188293457,
236
- 0.006819839954376221,
237
- 0.006840320110321045,
238
- 0.006834176063537598,
239
- 0.006809599876403808,
240
- 0.006848512172698974,
241
- 0.0068392958641052244,
242
- 0.006788095951080322,
243
- 0.006873087882995605,
244
- 0.006825984001159668,
245
- 0.006854656219482422,
246
- 0.006789087772369385,
247
- 0.006832096099853515,
248
- 0.0068791680335998535,
249
- 0.006804480075836182,
250
- 0.006840320110321045,
251
- 0.006862847805023193,
252
- 0.006846464157104492,
253
- 0.006813695907592773,
254
  0.0068280320167541505,
255
- 0.006831103801727295,
256
- 0.006782976150512696,
257
- 0.00683516788482666,
 
 
 
 
 
 
 
 
 
258
  0.006846464157104492,
259
- 0.006848512172698974,
260
- 0.006795263767242431
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
261
  ]
262
  },
263
  "throughput": {
264
  "unit": "samples/s",
265
- "value": 137.20563271821496
266
  },
267
  "energy": {
268
  "unit": "kWh",
269
- "cpu": 8.069541493961512e-08,
270
- "ram": 4.410909769292778e-08,
271
- "gpu": 1.507616474419874e-07,
272
- "total": 2.7556616007453034e-07
273
  },
274
  "efficiency": {
275
  "unit": "samples/kWh",
276
- "value": 3628892.6032482996
277
  }
278
  }
279
  }
 
6
  "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
 
9
  "library": "transformers",
10
+ "model": "FacebookAI/roberta-base",
11
+ "processor": "FacebookAI/roberta-base",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 891.621376,
106
  "max_global_vram": 1195.900928,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 555.74528,
 
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 146,
114
+ "total": 0.9972845778465275,
115
+ "mean": 0.006830716286620049,
116
+ "stdev": 0.0001990599915854344,
117
+ "p50": 0.006783967971801758,
118
+ "p90": 0.006865440130233765,
119
+ "p95": 0.0073021438121795654,
120
+ "p99": 0.00782781438827515,
121
  "values": [
122
+ 0.008047616004943848,
 
 
 
 
 
 
 
 
 
 
123
  0.0076267518997192385,
124
+ 0.0074291200637817386,
125
+ 0.007386015892028809,
126
+ 0.007299071788787842,
127
+ 0.007346176147460938,
128
+ 0.007336959838867187,
129
+ 0.00799232006072998,
130
+ 0.006978559970855713,
131
+ 0.006864960193634033,
132
+ 0.0067900800704956054,
133
+ 0.0068269438743591305,
134
+ 0.006820799827575684,
135
+ 0.006787007808685303,
136
+ 0.006843391895294189,
137
+ 0.006874112129211426,
138
+ 0.006866943836212158,
139
+ 0.00678604793548584,
140
+ 0.0068269438743591305,
141
+ 0.006811647891998291,
142
+ 0.006774784088134766,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
143
  0.006823935985565186,
144
+ 0.006814720153808594,
145
+ 0.006822912216186523,
146
+ 0.006793216228485107,
147
+ 0.006838272094726563,
148
+ 0.006816768169403077,
149
+ 0.0067870721817016606,
150
  0.006806528091430664,
151
+ 0.0068197760581970215,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
152
  0.0068280320167541505,
153
+ 0.006812672138214112,
154
+ 0.006838304042816162,
155
+ 0.00678604793548584,
156
+ 0.0067604160308837894,
157
+ 0.006806528091430664,
158
+ 0.006788095951080322,
159
+ 0.006967296123504638,
160
+ 0.00678713607788086,
161
+ 0.006849535942077637,
162
+ 0.006841343879699707,
163
+ 0.006770688056945801,
164
+ 0.006865920066833496,
165
  0.006846464157104492,
166
+ 0.0067686400413513184,
167
+ 0.006779903888702392,
168
+ 0.006781951904296875,
169
+ 0.006807551860809326,
170
+ 0.0067573761940002445,
171
+ 0.006793216228485107,
172
+ 0.006767615795135498,
173
+ 0.006759456157684326,
174
+ 0.006793216228485107,
175
+ 0.006770688056945801,
176
+ 0.006762495994567871,
177
+ 0.006761472225189209,
178
+ 0.006776832103729248,
179
+ 0.006785024166107178,
180
+ 0.006767680168151855,
181
+ 0.006807551860809326,
182
+ 0.006817791938781738,
183
+ 0.00676966381072998,
184
+ 0.006779903888702392,
185
+ 0.006809599876403808,
186
+ 0.006812672138214112,
187
+ 0.007303167819976806,
188
+ 0.006812672138214112,
189
+ 0.0068055038452148435,
190
+ 0.006770688056945801,
191
+ 0.006782911777496338,
192
+ 0.0068055038452148435,
193
+ 0.006807551860809326,
194
+ 0.006802432060241699,
195
+ 0.006810656070709228,
196
+ 0.0067870721817016606,
197
+ 0.006762495994567871,
198
+ 0.006791168212890625,
199
+ 0.006796288013458252,
200
+ 0.006817791938781738,
201
+ 0.006759424209594727,
202
+ 0.0067870721817016606,
203
+ 0.00677785587310791,
204
+ 0.006739967823028564,
205
+ 0.006812672138214112,
206
+ 0.006806528091430664,
207
+ 0.0067573761940002445,
208
+ 0.0067717118263244626,
209
+ 0.0067983360290527345,
210
+ 0.006776832103729248,
211
+ 0.00673689603805542,
212
+ 0.006763519763946534,
213
+ 0.006789120197296142,
214
+ 0.006751232147216797,
215
+ 0.006800384044647217,
216
+ 0.0067645440101623535,
217
+ 0.006758399963378906,
218
+ 0.006780928134918213,
219
+ 0.006800352096557617,
220
+ 0.006767615795135498,
221
+ 0.006751232147216797,
222
+ 0.006737919807434082,
223
+ 0.006740992069244385,
224
+ 0.006727615833282471,
225
+ 0.006754303932189941,
226
+ 0.006719488143920899,
227
+ 0.006725632190704346,
228
+ 0.0068055038452148435,
229
+ 0.0067358717918396,
230
+ 0.006703104019165039,
231
+ 0.006749184131622315,
232
+ 0.0067348480224609375,
233
+ 0.006729728221893311,
234
+ 0.006715392112731934,
235
+ 0.0067358717918396,
236
+ 0.0067358717918396,
237
+ 0.0067276802062988285,
238
+ 0.006739967823028564,
239
+ 0.006750207901000976,
240
+ 0.006717440128326416,
241
+ 0.00679423999786377,
242
+ 0.006744063854217529,
243
+ 0.0067348480224609375,
244
+ 0.006790143966674805,
245
+ 0.006781951904296875,
246
+ 0.0067717118263244626,
247
+ 0.006740992069244385,
248
+ 0.006745120048522949,
249
+ 0.006756351947784424,
250
+ 0.006747136116027832,
251
+ 0.006763519763946534,
252
+ 0.006793216228485107,
253
+ 0.006718463897705078,
254
+ 0.006751232147216797,
255
+ 0.0067276802062988285,
256
+ 0.006703104019165039,
257
+ 0.006717472076416015,
258
+ 0.007165952205657959,
259
+ 0.006743040084838867,
260
+ 0.006721536159515381,
261
+ 0.006823935985565186,
262
+ 0.006763519763946534,
263
+ 0.006749184131622315,
264
+ 0.006754335880279541,
265
+ 0.006740992069244385,
266
+ 0.006762495994567871,
267
+ 0.006765567779541016
268
  ]
269
  },
270
  "throughput": {
271
  "unit": "samples/s",
272
+ "value": 146.3975310991604
273
  },
274
  "energy": {
275
  "unit": "kWh",
276
+ "cpu": 8.417799000292899e-08,
277
+ "ram": 4.6017656379572535e-08,
278
+ "gpu": 1.5290201357446433e-07,
279
+ "total": 2.830976599569658e-07
280
  },
281
  "efficiency": {
282
  "unit": "samples/kWh",
283
+ "value": 3532349.932359073
284
  }
285
  }
286
  }