Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json
CHANGED
@@ -84,7 +84,7 @@
|
|
84 |
],
|
85 |
"gpu_count": 1,
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
-
"optimum_benchmark_version": "0.3.
|
88 |
"optimum_benchmark_commit": null,
|
89 |
"transformers_version": "4.42.3",
|
90 |
"transformers_commit": null,
|
@@ -104,7 +104,7 @@
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
-
"max_ram": 976.
|
108 |
"max_global_vram": 1434.976256,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 794.820608,
|
@@ -112,102 +112,104 @@
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
-
"count":
|
116 |
-
"total": 1.
|
117 |
-
"mean": 0.
|
118 |
-
"stdev": 0.
|
119 |
-
"p50": 0.
|
120 |
-
"p90": 0.
|
121 |
-
"p95": 0.
|
122 |
-
"p99": 0.
|
123 |
"values": [
|
124 |
-
0.
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
129 |
-
0.
|
130 |
-
0.
|
131 |
-
0.
|
132 |
-
0.
|
133 |
-
0.
|
134 |
-
0.
|
135 |
-
0.
|
136 |
-
0.
|
137 |
-
0.
|
138 |
-
0.
|
139 |
-
0.
|
140 |
-
0.
|
141 |
-
0.
|
142 |
-
0.
|
143 |
-
0.
|
144 |
-
0.
|
145 |
-
0.
|
146 |
-
0.
|
147 |
-
0.
|
148 |
-
0.
|
149 |
-
0.
|
150 |
-
0.
|
151 |
-
0.
|
152 |
-
0.
|
153 |
-
0.
|
154 |
-
0.
|
155 |
-
0.014376992225646973,
|
156 |
-
0.014126079559326172,
|
157 |
-
0.013775903701782227,
|
158 |
-
0.01363865566253662,
|
159 |
-
0.014253055572509766,
|
160 |
-
0.014362624168395996,
|
161 |
-
0.014425087928771972,
|
162 |
-
0.014492671966552734,
|
163 |
-
0.014154751777648926,
|
164 |
-
0.015005696296691894,
|
165 |
-
0.014286848068237304,
|
166 |
-
0.013932543754577637,
|
167 |
-
0.014004223823547364,
|
168 |
-
0.01374617576599121,
|
169 |
-
0.013554656028747558,
|
170 |
-
0.013520895957946777,
|
171 |
-
0.013453311920166015,
|
172 |
-
0.013493247985839844,
|
173 |
-
0.013495295524597169,
|
174 |
-
0.0134901762008667,
|
175 |
-
0.013551615715026855,
|
176 |
-
0.013460479736328124,
|
177 |
-
0.013519871711730956,
|
178 |
-
0.013469696044921875,
|
179 |
-
0.013499391555786134,
|
180 |
-
0.013709312438964843,
|
181 |
-
0.013534208297729493,
|
182 |
-
0.01348198413848877,
|
183 |
-
0.013445119857788086,
|
184 |
-
0.013658080101013183,
|
185 |
-
0.013905983924865722,
|
186 |
-
0.013830143928527832,
|
187 |
0.01368883228302002,
|
188 |
-
0.
|
189 |
-
0.
|
190 |
-
0.
|
191 |
-
0.
|
192 |
-
0.
|
193 |
-
0.
|
194 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
195 |
]
|
196 |
},
|
197 |
"throughput": {
|
198 |
"unit": "samples/s",
|
199 |
-
"value":
|
200 |
},
|
201 |
"energy": {
|
202 |
"unit": "kWh",
|
203 |
-
"cpu": 1.
|
204 |
-
"ram": 8.
|
205 |
-
"gpu": 3.
|
206 |
-
"total": 5.
|
207 |
},
|
208 |
"efficiency": {
|
209 |
"unit": "samples/kWh",
|
210 |
-
"value":
|
211 |
}
|
212 |
}
|
213 |
}
|
|
|
84 |
],
|
85 |
"gpu_count": 1,
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
+
"optimum_benchmark_version": "0.3.1",
|
88 |
"optimum_benchmark_commit": null,
|
89 |
"transformers_version": "4.42.3",
|
90 |
"transformers_commit": null,
|
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
+
"max_ram": 976.080896,
|
108 |
"max_global_vram": 1434.976256,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 794.820608,
|
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
+
"count": 73,
|
116 |
+
"total": 1.0106326417922977,
|
117 |
+
"mean": 0.013844282764278045,
|
118 |
+
"stdev": 0.0010715488971016177,
|
119 |
+
"p50": 0.013612031936645508,
|
120 |
+
"p90": 0.014301798248291014,
|
121 |
+
"p95": 0.015009011077880857,
|
122 |
+
"p99": 0.01908179008483887,
|
123 |
"values": [
|
124 |
+
0.020207616806030275,
|
125 |
+
0.01864396858215332,
|
126 |
+
0.01460223960876465,
|
127 |
+
0.016536575317382812,
|
128 |
+
0.014818304061889649,
|
129 |
+
0.014376959800720214,
|
130 |
+
0.01487673568725586,
|
131 |
+
0.015207424163818359,
|
132 |
+
0.013744128227233888,
|
133 |
+
0.013275103569030762,
|
134 |
+
0.013041600227355957,
|
135 |
+
0.01306112003326416,
|
136 |
+
0.013022208213806152,
|
137 |
+
0.013090815544128418,
|
138 |
+
0.013020159721374512,
|
139 |
+
0.012989439964294434,
|
140 |
+
0.012991488456726074,
|
141 |
+
0.01349120044708252,
|
142 |
+
0.013637632369995116,
|
143 |
+
0.013553664207458497,
|
144 |
+
0.01367142391204834,
|
145 |
+
0.013667327880859375,
|
146 |
+
0.0138854398727417,
|
147 |
+
0.013824000358581542,
|
148 |
+
0.013752320289611816,
|
149 |
+
0.01377280044555664,
|
150 |
+
0.013650912284851074,
|
151 |
+
0.013801471710205078,
|
152 |
+
0.013702143669128418,
|
153 |
+
0.013888511657714844,
|
154 |
+
0.013637632369995116,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
155 |
0.01368883228302002,
|
156 |
+
0.013751296043395997,
|
157 |
+
0.013612031936645508,
|
158 |
+
0.013592608451843262,
|
159 |
+
0.013657088279724122,
|
160 |
+
0.01400115203857422,
|
161 |
+
0.01371241569519043,
|
162 |
+
0.01359769630432129,
|
163 |
+
0.01377280044555664,
|
164 |
+
0.013693951606750488,
|
165 |
+
0.01358233642578125,
|
166 |
+
0.01358028793334961,
|
167 |
+
0.01358950424194336,
|
168 |
+
0.013601792335510255,
|
169 |
+
0.013514752388000489,
|
170 |
+
0.013504511833190918,
|
171 |
+
0.013609984397888183,
|
172 |
+
0.013552639961242676,
|
173 |
+
0.013450271606445312,
|
174 |
+
0.013540351867675781,
|
175 |
+
0.013614080429077148,
|
176 |
+
0.013508607864379883,
|
177 |
+
0.013643775939941406,
|
178 |
+
0.013596672058105469,
|
179 |
+
0.01354751968383789,
|
180 |
+
0.01352086353302002,
|
181 |
+
0.0134717435836792,
|
182 |
+
0.013651968002319336,
|
183 |
+
0.013504511833190918,
|
184 |
+
0.013639679908752441,
|
185 |
+
0.013636608123779297,
|
186 |
+
0.01367142391204834,
|
187 |
+
0.013615103721618652,
|
188 |
+
0.013522944450378417,
|
189 |
+
0.013505536079406738,
|
190 |
+
0.013685759544372558,
|
191 |
+
0.013517791748046876,
|
192 |
+
0.013565952301025391,
|
193 |
+
0.013587455749511718,
|
194 |
+
0.013530112266540528,
|
195 |
+
0.013517791748046876,
|
196 |
+
0.013528063774108886
|
197 |
]
|
198 |
},
|
199 |
"throughput": {
|
200 |
"unit": "samples/s",
|
201 |
+
"value": 72.23198319672201
|
202 |
},
|
203 |
"energy": {
|
204 |
"unit": "kWh",
|
205 |
+
"cpu": 1.5782041240621498e-07,
|
206 |
+
"ram": 8.62766642443085e-08,
|
207 |
+
"gpu": 3.427410149333306e-07,
|
208 |
+
"total": 5.86838091583854e-07
|
209 |
},
|
210 |
"efficiency": {
|
211 |
"unit": "samples/kWh",
|
212 |
+
"value": 1704047.5291933375
|
213 |
}
|
214 |
}
|
215 |
}
|