Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
-
"version": "2.3.
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "multiple-choice",
|
9 |
"library": "transformers",
|
@@ -73,10 +73,10 @@
|
|
73 |
"environment": {
|
74 |
"cpu": " AMD EPYC 7R32",
|
75 |
"cpu_count": 16,
|
76 |
-
"cpu_ram_mb": 66697.
|
77 |
"system": "Linux",
|
78 |
"machine": "x86_64",
|
79 |
-
"platform": "Linux-5.10.
|
80 |
"processor": "x86_64",
|
81 |
"python_version": "3.10.12",
|
82 |
"gpu": [
|
@@ -86,15 +86,15 @@
|
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
"optimum_benchmark_version": "0.2.1",
|
88 |
"optimum_benchmark_commit": null,
|
89 |
-
"transformers_version": "4.
|
90 |
"transformers_commit": null,
|
91 |
-
"accelerate_version": "0.
|
92 |
"accelerate_commit": null,
|
93 |
-
"diffusers_version": "0.
|
94 |
"diffusers_commit": null,
|
95 |
"optimum_version": null,
|
96 |
"optimum_commit": null,
|
97 |
-
"timm_version": "1.0.
|
98 |
"timm_commit": null,
|
99 |
"peft_version": null,
|
100 |
"peft_commit": null
|
@@ -104,7 +104,7 @@
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
-
"max_ram":
|
108 |
"max_global_vram": 1195.900928,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 555.74528,
|
@@ -112,162 +112,153 @@
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
-
"count":
|
116 |
-
"total": 1.
|
117 |
-
"mean": 0.
|
118 |
-
"stdev": 0.
|
119 |
-
"p50": 0.
|
120 |
-
"p90": 0.
|
121 |
-
"p95": 0.
|
122 |
-
"p99": 0.
|
123 |
"values": [
|
124 |
-
0.
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
129 |
0.008104960441589355,
|
130 |
-
0.
|
131 |
-
0.007873536109924317,
|
132 |
-
0.007584767818450928,
|
133 |
-
0.00758681583404541,
|
134 |
-
0.007497727870941162,
|
135 |
-
0.007612415790557861,
|
136 |
-
0.007633920192718506,
|
137 |
-
0.00760319995880127,
|
138 |
-
0.007541728019714355,
|
139 |
-
0.007455743789672851,
|
140 |
-
0.00779366397857666,
|
141 |
-
0.007806975841522217,
|
142 |
-
0.008138751983642578,
|
143 |
-
0.007894015789031983,
|
144 |
-
0.008067071914672852,
|
145 |
-
0.007786496162414551,
|
146 |
-
0.007821343898773193,
|
147 |
-
0.007814144134521485,
|
148 |
-
0.007887872219085693,
|
149 |
-
0.007860223770141601,
|
150 |
-
0.007930848121643066,
|
151 |
-
0.00785920000076294,
|
152 |
-
0.007802879810333252,
|
153 |
-
0.007886847972869874,
|
154 |
-
0.007899136066436767,
|
155 |
-
0.007809023857116699,
|
156 |
-
0.007812096118927002,
|
157 |
-
0.007770143985748291,
|
158 |
-
0.00775980806350708,
|
159 |
-
0.007749504089355469,
|
160 |
-
0.007778304100036621,
|
161 |
-
0.007702527999877929,
|
162 |
-
0.007844863891601562,
|
163 |
-
0.007955455780029297,
|
164 |
-
0.007948287963867188,
|
165 |
-
0.008072192192077637,
|
166 |
-
0.007845888137817383,
|
167 |
0.007989247798919678,
|
168 |
-
0.
|
169 |
-
0.
|
170 |
-
0.
|
171 |
-
0.
|
172 |
-
0.
|
173 |
-
0.
|
174 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
175 |
0.007974912166595459,
|
176 |
-
0.
|
177 |
-
0.
|
178 |
-
0.
|
179 |
-
0.0077413759231567385,
|
180 |
-
0.007833439826965332,
|
181 |
-
0.007775231838226319,
|
182 |
-
0.007799776077270508,
|
183 |
-
0.00778223991394043,
|
184 |
-
0.007883776187896728,
|
185 |
-
0.008122367858886719,
|
186 |
-
0.007923711776733398,
|
187 |
0.007964672088623047,
|
188 |
-
0.
|
189 |
-
0.
|
190 |
-
0.
|
191 |
-
0.
|
192 |
-
0.
|
193 |
-
0.
|
194 |
-
0.
|
195 |
-
0.
|
196 |
-
0.
|
197 |
-
0.
|
198 |
-
0.
|
199 |
-
0.
|
200 |
-
0.
|
201 |
-
0.007696383953094482,
|
202 |
-
0.007674880027770996,
|
203 |
-
0.007741439819335938,
|
204 |
-
0.007955455780029297,
|
205 |
-
0.0076912641525268555,
|
206 |
-
0.007515135765075683,
|
207 |
-
0.007458816051483155,
|
208 |
-
0.007415808200836181,
|
209 |
-
0.007456736087799072,
|
210 |
-
0.007443456172943115,
|
211 |
-
0.007386112213134765,
|
212 |
-
0.007372799873352051,
|
213 |
-
0.007477119922637939,
|
214 |
-
0.007482367992401123,
|
215 |
-
0.00743833589553833,
|
216 |
-
0.007428095817565918,
|
217 |
-
0.007425024032592774,
|
218 |
-
0.007425024032592774,
|
219 |
-
0.007458816051483155,
|
220 |
-
0.007412735939025879,
|
221 |
-
0.007368703842163086,
|
222 |
-
0.007325695991516113,
|
223 |
-
0.007341055870056152,
|
224 |
-
0.007279615879058838,
|
225 |
-
0.007262207984924316,
|
226 |
-
0.007274496078491211,
|
227 |
-
0.007322624206542969,
|
228 |
-
0.00733081579208374,
|
229 |
-
0.007257984161376953,
|
230 |
-
0.0072837119102478025,
|
231 |
-
0.007373824119567871,
|
232 |
-
0.007278592109680176,
|
233 |
-
0.007231488227844239,
|
234 |
-
0.007222271919250488,
|
235 |
-
0.007361536026000977,
|
236 |
-
0.007296000003814697,
|
237 |
-
0.007278592109680176,
|
238 |
-
0.007227392196655274,
|
239 |
-
0.007390207767486572,
|
240 |
-
0.00724889612197876,
|
241 |
-
0.007390016078948974,
|
242 |
-
0.007425024032592774,
|
243 |
-
0.007450623989105225,
|
244 |
-
0.00744652795791626,
|
245 |
-
0.007435264110565186,
|
246 |
-
0.0074321279525756834,
|
247 |
-
0.007428095817565918,
|
248 |
-
0.007458816051483155,
|
249 |
-
0.007449600219726562,
|
250 |
-
0.0074700798988342285,
|
251 |
-
0.007437312126159668,
|
252 |
-
0.007346176147460938,
|
253 |
-
0.0073134078979492185,
|
254 |
-
0.007328800201416016
|
255 |
]
|
256 |
},
|
257 |
"throughput": {
|
258 |
"unit": "samples/s",
|
259 |
-
"value":
|
260 |
},
|
261 |
"energy": {
|
262 |
"unit": "kWh",
|
263 |
-
"cpu":
|
264 |
-
"ram":
|
265 |
-
"gpu": 1.
|
266 |
-
"total":
|
267 |
},
|
268 |
"efficiency": {
|
269 |
"unit": "samples/kWh",
|
270 |
-
"value":
|
271 |
}
|
272 |
}
|
273 |
}
|
|
|
3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
+
"version": "2.3.1+cu121",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "multiple-choice",
|
9 |
"library": "transformers",
|
|
|
73 |
"environment": {
|
74 |
"cpu": " AMD EPYC 7R32",
|
75 |
"cpu_count": 16,
|
76 |
+
"cpu_ram_mb": 66697.293824,
|
77 |
"system": "Linux",
|
78 |
"machine": "x86_64",
|
79 |
+
"platform": "Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35",
|
80 |
"processor": "x86_64",
|
81 |
"python_version": "3.10.12",
|
82 |
"gpu": [
|
|
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
"optimum_benchmark_version": "0.2.1",
|
88 |
"optimum_benchmark_commit": null,
|
89 |
+
"transformers_version": "4.42.3",
|
90 |
"transformers_commit": null,
|
91 |
+
"accelerate_version": "0.31.0",
|
92 |
"accelerate_commit": null,
|
93 |
+
"diffusers_version": "0.29.2",
|
94 |
"diffusers_commit": null,
|
95 |
"optimum_version": null,
|
96 |
"optimum_commit": null,
|
97 |
+
"timm_version": "1.0.7",
|
98 |
"timm_commit": null,
|
99 |
"peft_version": null,
|
100 |
"peft_commit": null
|
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
+
"max_ram": 909.766656,
|
108 |
"max_global_vram": 1195.900928,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 555.74528,
|
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
+
"count": 122,
|
116 |
+
"total": 1.0027191996574405,
|
117 |
+
"mean": 0.008219009833257705,
|
118 |
+
"stdev": 0.0002897873745098747,
|
119 |
+
"p50": 0.008089087963104247,
|
120 |
+
"p90": 0.008587673854827881,
|
121 |
+
"p95": 0.008662733173370362,
|
122 |
+
"p99": 0.008938291273117065,
|
123 |
"values": [
|
124 |
+
0.009310208320617675,
|
125 |
+
0.00877670383453369,
|
126 |
+
0.008841216087341308,
|
127 |
+
0.008688639640808106,
|
128 |
+
0.008637439727783204,
|
129 |
+
0.008624128341674805,
|
130 |
+
0.008545280456542969,
|
131 |
+
0.008532992362976074,
|
132 |
+
0.008516575813293457,
|
133 |
+
0.008443903923034669,
|
134 |
+
0.008447967529296874,
|
135 |
+
0.00850432014465332,
|
136 |
+
0.00852070426940918,
|
137 |
+
0.008606719970703124,
|
138 |
+
0.008629247665405274,
|
139 |
+
0.008444928169250488,
|
140 |
+
0.008563712120056152,
|
141 |
+
0.00848588752746582,
|
142 |
+
0.008508416175842285,
|
143 |
+
0.008570879936218261,
|
144 |
+
0.008481792449951172,
|
145 |
+
0.008477696418762207,
|
146 |
+
0.008430591583251953,
|
147 |
+
0.008401920318603515,
|
148 |
+
0.00841318416595459,
|
149 |
+
0.008421376228332519,
|
150 |
+
0.008447999954223634,
|
151 |
+
0.008514559745788575,
|
152 |
+
0.008588288307189941,
|
153 |
+
0.00860262393951416,
|
154 |
+
0.008582143783569337,
|
155 |
+
0.008455167770385743,
|
156 |
+
0.008532992362976074,
|
157 |
+
0.008475647926330567,
|
158 |
+
0.008664064407348633,
|
159 |
+
0.008454143524169922,
|
160 |
+
0.00841318416595459,
|
161 |
+
0.008506367683410645,
|
162 |
+
0.008548352241516113,
|
163 |
+
0.008335359573364258,
|
164 |
+
0.008093695640563964,
|
165 |
+
0.008089599609375,
|
166 |
+
0.008099807739257812,
|
167 |
+
0.00812335968017578,
|
168 |
+
0.00819814395904541,
|
169 |
+
0.008151040077209473,
|
170 |
+
0.008349696159362792,
|
171 |
+
0.008135680198669434,
|
172 |
+
0.008088576316833495,
|
173 |
0.008104960441589355,
|
174 |
+
0.0080763521194458,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
175 |
0.007989247798919678,
|
176 |
+
0.007871488094329833,
|
177 |
+
0.007967743873596191,
|
178 |
+
0.00801689624786377,
|
179 |
+
0.00781824016571045,
|
180 |
+
0.0077547521591186525,
|
181 |
+
0.0077619199752807615,
|
182 |
+
0.007782400131225586,
|
183 |
+
0.00820019245147705,
|
184 |
+
0.008427519798278809,
|
185 |
+
0.008468480110168456,
|
186 |
+
0.008790016174316406,
|
187 |
+
0.008434687614440918,
|
188 |
+
0.008407039642333984,
|
189 |
+
0.008964096069335938,
|
190 |
+
0.008229887962341309,
|
191 |
+
0.008078335762023926,
|
192 |
+
0.00806604766845703,
|
193 |
+
0.008072192192077637,
|
194 |
+
0.008022015571594238,
|
195 |
+
0.008033280372619628,
|
196 |
+
0.008071167945861817,
|
197 |
+
0.008257535934448243,
|
198 |
+
0.008017919540405273,
|
199 |
+
0.00797388792037964,
|
200 |
+
0.008130559921264649,
|
201 |
+
0.007985151767730713,
|
202 |
+
0.007988224029541016,
|
203 |
+
0.007994368076324462,
|
204 |
+
0.007994368076324462,
|
205 |
+
0.007998464107513427,
|
206 |
+
0.007985151767730713,
|
207 |
+
0.007977983951568603,
|
208 |
+
0.007991360187530518,
|
209 |
+
0.007982079982757568,
|
210 |
+
0.007980031967163086,
|
211 |
+
0.00800153636932373,
|
212 |
+
0.00800153636932373,
|
213 |
+
0.008004608154296875,
|
214 |
+
0.008004608154296875,
|
215 |
+
0.008026111602783203,
|
216 |
+
0.007970816135406494,
|
217 |
+
0.007995391845703125,
|
218 |
+
0.007986176013946533,
|
219 |
+
0.00799948787689209,
|
220 |
+
0.007982079982757568,
|
221 |
+
0.007985151767730713,
|
222 |
+
0.007985151767730713,
|
223 |
+
0.00800767993927002,
|
224 |
+
0.007985151767730713,
|
225 |
+
0.007970816135406494,
|
226 |
+
0.007987199783325195,
|
227 |
+
0.007980031967163086,
|
228 |
0.007974912166595459,
|
229 |
+
0.00799948787689209,
|
230 |
+
0.00800767993927002,
|
231 |
+
0.007986176013946533,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
232 |
0.007964672088623047,
|
233 |
+
0.007953407764434815,
|
234 |
+
0.007950335979461669,
|
235 |
+
0.007956480026245117,
|
236 |
+
0.007945184230804442,
|
237 |
+
0.00796569585800171,
|
238 |
+
0.007962624073028564,
|
239 |
+
0.007981056213378907,
|
240 |
+
0.008167424201965333,
|
241 |
+
0.008003583908081055,
|
242 |
+
0.00809779167175293,
|
243 |
+
0.007988224029541016,
|
244 |
+
0.008005632400512695,
|
245 |
+
0.00799232006072998
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
246 |
]
|
247 |
},
|
248 |
"throughput": {
|
249 |
"unit": "samples/s",
|
250 |
+
"value": 121.66915726923249
|
251 |
},
|
252 |
"energy": {
|
253 |
"unit": "kWh",
|
254 |
+
"cpu": 9.73193073316133e-08,
|
255 |
+
"ram": 5.306276868201506e-08,
|
256 |
+
"gpu": 1.7823329377049204e-07,
|
257 |
+
"total": 3.286153697841204e-07
|
258 |
},
|
259 |
"efficiency": {
|
260 |
"unit": "samples/kWh",
|
261 |
+
"value": 3043071.2983903857
|
262 |
}
|
263 |
}
|
264 |
}
|