IlyasMoutawwakil HF staff commited on
Commit
2e81069
·
verified ·
1 Parent(s): 743b321

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -73,10 +73,10 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
- "cpu_ram_mb": 66697.29792,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -86,15 +86,15 @@
86
  "gpu_vram_mb": 24146608128,
87
  "optimum_benchmark_version": "0.2.1",
88
  "optimum_benchmark_commit": null,
89
- "transformers_version": "4.41.1",
90
  "transformers_commit": null,
91
- "accelerate_version": "0.30.1",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.27.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
- "timm_version": "1.0.3",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 907.882496,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,162 +112,153 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 131,
116
- "total": 1.0012743368148795,
117
- "mean": 0.007643315548205194,
118
- "stdev": 0.0002595442981809403,
119
- "p50": 0.007656447887420655,
120
- "p90": 0.007930848121643066,
121
- "p95": 0.007986687898635865,
122
- "p99": 0.00813383674621582,
123
  "values": [
124
- 0.008762368202209473,
125
- 0.007656447887420655,
126
- 0.007848959922790527,
127
- 0.007888895988464355,
128
- 0.007984127998352051,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129
  0.008104960441589355,
130
- 0.007781375885009765,
131
- 0.007873536109924317,
132
- 0.007584767818450928,
133
- 0.00758681583404541,
134
- 0.007497727870941162,
135
- 0.007612415790557861,
136
- 0.007633920192718506,
137
- 0.00760319995880127,
138
- 0.007541728019714355,
139
- 0.007455743789672851,
140
- 0.00779366397857666,
141
- 0.007806975841522217,
142
- 0.008138751983642578,
143
- 0.007894015789031983,
144
- 0.008067071914672852,
145
- 0.007786496162414551,
146
- 0.007821343898773193,
147
- 0.007814144134521485,
148
- 0.007887872219085693,
149
- 0.007860223770141601,
150
- 0.007930848121643066,
151
- 0.00785920000076294,
152
- 0.007802879810333252,
153
- 0.007886847972869874,
154
- 0.007899136066436767,
155
- 0.007809023857116699,
156
- 0.007812096118927002,
157
- 0.007770143985748291,
158
- 0.00775980806350708,
159
- 0.007749504089355469,
160
- 0.007778304100036621,
161
- 0.007702527999877929,
162
- 0.007844863891601562,
163
- 0.007955455780029297,
164
- 0.007948287963867188,
165
- 0.008072192192077637,
166
- 0.007845888137817383,
167
  0.007989247798919678,
168
- 0.007827455997467042,
169
- 0.007878655910491944,
170
- 0.007539775848388672,
171
- 0.007635968208312988,
172
- 0.007764992237091065,
173
- 0.007847968101501466,
174
- 0.007877632141113282,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
175
  0.007974912166595459,
176
- 0.007889920234680176,
177
- 0.007831552028656007,
178
- 0.00788479995727539,
179
- 0.0077413759231567385,
180
- 0.007833439826965332,
181
- 0.007775231838226319,
182
- 0.007799776077270508,
183
- 0.00778223991394043,
184
- 0.007883776187896728,
185
- 0.008122367858886719,
186
- 0.007923711776733398,
187
  0.007964672088623047,
188
- 0.007879680156707763,
189
- 0.007812096118927002,
190
- 0.007817215919494629,
191
- 0.007824384212493896,
192
- 0.007748608112335205,
193
- 0.0076789441108703616,
194
- 0.007650303840637207,
195
- 0.007713791847229004,
196
- 0.007635968208312988,
197
- 0.007464960098266602,
198
- 0.007417856216430664,
199
- 0.007461887836456299,
200
- 0.007561215877532959,
201
- 0.007696383953094482,
202
- 0.007674880027770996,
203
- 0.007741439819335938,
204
- 0.007955455780029297,
205
- 0.0076912641525268555,
206
- 0.007515135765075683,
207
- 0.007458816051483155,
208
- 0.007415808200836181,
209
- 0.007456736087799072,
210
- 0.007443456172943115,
211
- 0.007386112213134765,
212
- 0.007372799873352051,
213
- 0.007477119922637939,
214
- 0.007482367992401123,
215
- 0.00743833589553833,
216
- 0.007428095817565918,
217
- 0.007425024032592774,
218
- 0.007425024032592774,
219
- 0.007458816051483155,
220
- 0.007412735939025879,
221
- 0.007368703842163086,
222
- 0.007325695991516113,
223
- 0.007341055870056152,
224
- 0.007279615879058838,
225
- 0.007262207984924316,
226
- 0.007274496078491211,
227
- 0.007322624206542969,
228
- 0.00733081579208374,
229
- 0.007257984161376953,
230
- 0.0072837119102478025,
231
- 0.007373824119567871,
232
- 0.007278592109680176,
233
- 0.007231488227844239,
234
- 0.007222271919250488,
235
- 0.007361536026000977,
236
- 0.007296000003814697,
237
- 0.007278592109680176,
238
- 0.007227392196655274,
239
- 0.007390207767486572,
240
- 0.00724889612197876,
241
- 0.007390016078948974,
242
- 0.007425024032592774,
243
- 0.007450623989105225,
244
- 0.00744652795791626,
245
- 0.007435264110565186,
246
- 0.0074321279525756834,
247
- 0.007428095817565918,
248
- 0.007458816051483155,
249
- 0.007449600219726562,
250
- 0.0074700798988342285,
251
- 0.007437312126159668,
252
- 0.007346176147460938,
253
- 0.0073134078979492185,
254
- 0.007328800201416016
255
  ]
256
  },
257
  "throughput": {
258
  "unit": "samples/s",
259
- "value": 130.83327434189476
260
  },
261
  "energy": {
262
  "unit": "kWh",
263
- "cpu": 8.526902428443294e-08,
264
- "ram": 4.6616730622504583e-08,
265
- "gpu": 1.6723034961151106e-07,
266
- "total": 2.991161045184486e-07
267
  },
268
  "efficiency": {
269
  "unit": "samples/kWh",
270
- "value": 3343183.415717167
271
  }
272
  }
273
  }
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.293824,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
86
  "gpu_vram_mb": 24146608128,
87
  "optimum_benchmark_version": "0.2.1",
88
  "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.31.0",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.29.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
+ "timm_version": "1.0.7",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 909.766656,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 122,
116
+ "total": 1.0027191996574405,
117
+ "mean": 0.008219009833257705,
118
+ "stdev": 0.0002897873745098747,
119
+ "p50": 0.008089087963104247,
120
+ "p90": 0.008587673854827881,
121
+ "p95": 0.008662733173370362,
122
+ "p99": 0.008938291273117065,
123
  "values": [
124
+ 0.009310208320617675,
125
+ 0.00877670383453369,
126
+ 0.008841216087341308,
127
+ 0.008688639640808106,
128
+ 0.008637439727783204,
129
+ 0.008624128341674805,
130
+ 0.008545280456542969,
131
+ 0.008532992362976074,
132
+ 0.008516575813293457,
133
+ 0.008443903923034669,
134
+ 0.008447967529296874,
135
+ 0.00850432014465332,
136
+ 0.00852070426940918,
137
+ 0.008606719970703124,
138
+ 0.008629247665405274,
139
+ 0.008444928169250488,
140
+ 0.008563712120056152,
141
+ 0.00848588752746582,
142
+ 0.008508416175842285,
143
+ 0.008570879936218261,
144
+ 0.008481792449951172,
145
+ 0.008477696418762207,
146
+ 0.008430591583251953,
147
+ 0.008401920318603515,
148
+ 0.00841318416595459,
149
+ 0.008421376228332519,
150
+ 0.008447999954223634,
151
+ 0.008514559745788575,
152
+ 0.008588288307189941,
153
+ 0.00860262393951416,
154
+ 0.008582143783569337,
155
+ 0.008455167770385743,
156
+ 0.008532992362976074,
157
+ 0.008475647926330567,
158
+ 0.008664064407348633,
159
+ 0.008454143524169922,
160
+ 0.00841318416595459,
161
+ 0.008506367683410645,
162
+ 0.008548352241516113,
163
+ 0.008335359573364258,
164
+ 0.008093695640563964,
165
+ 0.008089599609375,
166
+ 0.008099807739257812,
167
+ 0.00812335968017578,
168
+ 0.00819814395904541,
169
+ 0.008151040077209473,
170
+ 0.008349696159362792,
171
+ 0.008135680198669434,
172
+ 0.008088576316833495,
173
  0.008104960441589355,
174
+ 0.0080763521194458,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
175
  0.007989247798919678,
176
+ 0.007871488094329833,
177
+ 0.007967743873596191,
178
+ 0.00801689624786377,
179
+ 0.00781824016571045,
180
+ 0.0077547521591186525,
181
+ 0.0077619199752807615,
182
+ 0.007782400131225586,
183
+ 0.00820019245147705,
184
+ 0.008427519798278809,
185
+ 0.008468480110168456,
186
+ 0.008790016174316406,
187
+ 0.008434687614440918,
188
+ 0.008407039642333984,
189
+ 0.008964096069335938,
190
+ 0.008229887962341309,
191
+ 0.008078335762023926,
192
+ 0.00806604766845703,
193
+ 0.008072192192077637,
194
+ 0.008022015571594238,
195
+ 0.008033280372619628,
196
+ 0.008071167945861817,
197
+ 0.008257535934448243,
198
+ 0.008017919540405273,
199
+ 0.00797388792037964,
200
+ 0.008130559921264649,
201
+ 0.007985151767730713,
202
+ 0.007988224029541016,
203
+ 0.007994368076324462,
204
+ 0.007994368076324462,
205
+ 0.007998464107513427,
206
+ 0.007985151767730713,
207
+ 0.007977983951568603,
208
+ 0.007991360187530518,
209
+ 0.007982079982757568,
210
+ 0.007980031967163086,
211
+ 0.00800153636932373,
212
+ 0.00800153636932373,
213
+ 0.008004608154296875,
214
+ 0.008004608154296875,
215
+ 0.008026111602783203,
216
+ 0.007970816135406494,
217
+ 0.007995391845703125,
218
+ 0.007986176013946533,
219
+ 0.00799948787689209,
220
+ 0.007982079982757568,
221
+ 0.007985151767730713,
222
+ 0.007985151767730713,
223
+ 0.00800767993927002,
224
+ 0.007985151767730713,
225
+ 0.007970816135406494,
226
+ 0.007987199783325195,
227
+ 0.007980031967163086,
228
  0.007974912166595459,
229
+ 0.00799948787689209,
230
+ 0.00800767993927002,
231
+ 0.007986176013946533,
 
 
 
 
 
 
 
 
232
  0.007964672088623047,
233
+ 0.007953407764434815,
234
+ 0.007950335979461669,
235
+ 0.007956480026245117,
236
+ 0.007945184230804442,
237
+ 0.00796569585800171,
238
+ 0.007962624073028564,
239
+ 0.007981056213378907,
240
+ 0.008167424201965333,
241
+ 0.008003583908081055,
242
+ 0.00809779167175293,
243
+ 0.007988224029541016,
244
+ 0.008005632400512695,
245
+ 0.00799232006072998
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246
  ]
247
  },
248
  "throughput": {
249
  "unit": "samples/s",
250
+ "value": 121.66915726923249
251
  },
252
  "energy": {
253
  "unit": "kWh",
254
+ "cpu": 9.73193073316133e-08,
255
+ "ram": 5.306276868201506e-08,
256
+ "gpu": 1.7823329377049204e-07,
257
+ "total": 3.286153697841204e-07
258
  },
259
  "efficiency": {
260
  "unit": "samples/kWh",
261
+ "value": 3043071.2983903857
262
  }
263
  }
264
  }