IlyasMoutawwakil HF staff commited on
Commit
98fa85c
·
verified ·
1 Parent(s): d67c4db

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -84,7 +84,7 @@
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
- "optimum_benchmark_version": "0.3.0",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 976.142336,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
@@ -112,102 +112,104 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 71,
116
- "total": 1.000133632659912,
117
- "mean": 0.014086389192393129,
118
- "stdev": 0.0007870449991154138,
119
- "p50": 0.013996992111206055,
120
- "p90": 0.014700544357299805,
121
- "p95": 0.015127040386199952,
122
- "p99": 0.016504934406280508,
123
  "values": [
124
- 0.015121408462524414,
125
- 0.015286272048950195,
126
- 0.015096832275390625,
127
- 0.015338496208190918,
128
- 0.013991935729980469,
129
- 0.014082048416137695,
130
- 0.013932543754577637,
131
- 0.01420902442932129,
132
- 0.014161919593811035,
133
- 0.01417523193359375,
134
- 0.01459712028503418,
135
- 0.014153727531433105,
136
- 0.014060544013977052,
137
- 0.014073856353759765,
138
- 0.013963199615478516,
139
- 0.01397043228149414,
140
- 0.01409331226348877,
141
- 0.014700544357299805,
142
- 0.014232576370239258,
143
- 0.014162943840026856,
144
- 0.01409331226348877,
145
- 0.014060544013977052,
146
- 0.013998080253601074,
147
- 0.01428275203704834,
148
- 0.014340096473693848,
149
- 0.013996992111206055,
150
- 0.013932543754577637,
151
- 0.014541888236999511,
152
- 0.014573568344116212,
153
- 0.01922662353515625,
154
- 0.015132672309875488,
155
- 0.014376992225646973,
156
- 0.014126079559326172,
157
- 0.013775903701782227,
158
- 0.01363865566253662,
159
- 0.014253055572509766,
160
- 0.014362624168395996,
161
- 0.014425087928771972,
162
- 0.014492671966552734,
163
- 0.014154751777648926,
164
- 0.015005696296691894,
165
- 0.014286848068237304,
166
- 0.013932543754577637,
167
- 0.014004223823547364,
168
- 0.01374617576599121,
169
- 0.013554656028747558,
170
- 0.013520895957946777,
171
- 0.013453311920166015,
172
- 0.013493247985839844,
173
- 0.013495295524597169,
174
- 0.0134901762008667,
175
- 0.013551615715026855,
176
- 0.013460479736328124,
177
- 0.013519871711730956,
178
- 0.013469696044921875,
179
- 0.013499391555786134,
180
- 0.013709312438964843,
181
- 0.013534208297729493,
182
- 0.01348198413848877,
183
- 0.013445119857788086,
184
- 0.013658080101013183,
185
- 0.013905983924865722,
186
- 0.013830143928527832,
187
  0.01368883228302002,
188
- 0.01346560001373291,
189
- 0.013429759979248047,
190
- 0.013458432197570801,
191
- 0.013480959892272949,
192
- 0.01347379207611084,
193
- 0.013428735733032226,
194
- 0.013469696044921875
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 70.99051334887267
200
  },
201
  "energy": {
202
  "unit": "kWh",
203
- "cpu": 1.6127791445892497e-07,
204
- "ram": 8.779127442271151e-08,
205
- "gpu": 3.4143796083783817e-07,
206
- "total": 5.905071497194746e-07
207
  },
208
  "efficiency": {
209
  "unit": "samples/kWh",
210
- "value": 1693459.5973563713
211
  }
212
  }
213
  }
 
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.3.1",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 976.080896,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 73,
116
+ "total": 1.0106326417922977,
117
+ "mean": 0.013844282764278045,
118
+ "stdev": 0.0010715488971016177,
119
+ "p50": 0.013612031936645508,
120
+ "p90": 0.014301798248291014,
121
+ "p95": 0.015009011077880857,
122
+ "p99": 0.01908179008483887,
123
  "values": [
124
+ 0.020207616806030275,
125
+ 0.01864396858215332,
126
+ 0.01460223960876465,
127
+ 0.016536575317382812,
128
+ 0.014818304061889649,
129
+ 0.014376959800720214,
130
+ 0.01487673568725586,
131
+ 0.015207424163818359,
132
+ 0.013744128227233888,
133
+ 0.013275103569030762,
134
+ 0.013041600227355957,
135
+ 0.01306112003326416,
136
+ 0.013022208213806152,
137
+ 0.013090815544128418,
138
+ 0.013020159721374512,
139
+ 0.012989439964294434,
140
+ 0.012991488456726074,
141
+ 0.01349120044708252,
142
+ 0.013637632369995116,
143
+ 0.013553664207458497,
144
+ 0.01367142391204834,
145
+ 0.013667327880859375,
146
+ 0.0138854398727417,
147
+ 0.013824000358581542,
148
+ 0.013752320289611816,
149
+ 0.01377280044555664,
150
+ 0.013650912284851074,
151
+ 0.013801471710205078,
152
+ 0.013702143669128418,
153
+ 0.013888511657714844,
154
+ 0.013637632369995116,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
155
  0.01368883228302002,
156
+ 0.013751296043395997,
157
+ 0.013612031936645508,
158
+ 0.013592608451843262,
159
+ 0.013657088279724122,
160
+ 0.01400115203857422,
161
+ 0.01371241569519043,
162
+ 0.01359769630432129,
163
+ 0.01377280044555664,
164
+ 0.013693951606750488,
165
+ 0.01358233642578125,
166
+ 0.01358028793334961,
167
+ 0.01358950424194336,
168
+ 0.013601792335510255,
169
+ 0.013514752388000489,
170
+ 0.013504511833190918,
171
+ 0.013609984397888183,
172
+ 0.013552639961242676,
173
+ 0.013450271606445312,
174
+ 0.013540351867675781,
175
+ 0.013614080429077148,
176
+ 0.013508607864379883,
177
+ 0.013643775939941406,
178
+ 0.013596672058105469,
179
+ 0.01354751968383789,
180
+ 0.01352086353302002,
181
+ 0.0134717435836792,
182
+ 0.013651968002319336,
183
+ 0.013504511833190918,
184
+ 0.013639679908752441,
185
+ 0.013636608123779297,
186
+ 0.01367142391204834,
187
+ 0.013615103721618652,
188
+ 0.013522944450378417,
189
+ 0.013505536079406738,
190
+ 0.013685759544372558,
191
+ 0.013517791748046876,
192
+ 0.013565952301025391,
193
+ 0.013587455749511718,
194
+ 0.013530112266540528,
195
+ 0.013517791748046876,
196
+ 0.013528063774108886
197
  ]
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
+ "value": 72.23198319672201
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
+ "cpu": 1.5782041240621498e-07,
206
+ "ram": 8.62766642443085e-08,
207
+ "gpu": 3.427410149333306e-07,
208
+ "total": 5.86838091583854e-07
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
+ "value": 1704047.5291933375
213
  }
214
  }
215
  }