IlyasMoutawwakil HF staff commited on
Commit
0286c36
·
verified ·
1 Parent(s): 744b5a3

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -105,7 +105,7 @@
105
  "forward": {
106
  "memory": {
107
  "unit": "MB",
108
- "max_ram": 977.727488,
109
  "max_global_vram": 1434.976256,
110
  "max_process_vram": 0.0,
111
  "max_reserved": 794.820608,
@@ -114,102 +114,102 @@
114
  "latency": {
115
  "unit": "s",
116
  "count": 72,
117
- "total": 1.0054134082794188,
118
- "mean": 0.013964075114991932,
119
- "stdev": 0.00048452950488947103,
120
- "p50": 0.013945856094360352,
121
- "p90": 0.014506656169891358,
122
- "p95": 0.014775244665145874,
123
- "p99": 0.015397242612838747,
124
  "values": [
125
- 0.015508480072021484,
126
- 0.015351807594299317,
127
- 0.014829567909240723,
128
- 0.014742527961730957,
129
- 0.0140697603225708,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
130
  0.014285823822021485,
131
- 0.014215167999267577,
132
- 0.014203904151916504,
133
- 0.014248959541320801,
134
- 0.014112768173217773,
135
- 0.014070783615112305,
136
- 0.014263296127319336,
137
- 0.014172160148620605,
138
- 0.01477017593383789,
139
- 0.014510144233703613,
140
- 0.014246912002563476,
141
- 0.014369791984558105,
142
- 0.014346240043640136,
143
- 0.014115872383117676,
144
- 0.014164959907531739,
145
- 0.01447116756439209,
146
- 0.01417523193359375,
147
- 0.014225407600402832,
148
- 0.013937664031982423,
149
- 0.014053376197814941,
150
- 0.013881343841552735,
151
- 0.013858816146850587,
152
- 0.014052351951599122,
153
- 0.014575615882873535,
154
- 0.014781439781188965,
155
- 0.014459903717041016,
156
- 0.014475263595581055,
157
- 0.01429094409942627,
158
- 0.01399500846862793,
159
- 0.014006272315979004,
160
- 0.013950976371765136,
161
- 0.013911040306091308,
162
- 0.013899776458740234,
163
- 0.013953023910522461,
164
- 0.013915136337280273,
165
- 0.013876223564147949,
166
- 0.013430784225463867,
167
- 0.013430784225463867,
168
- 0.01346560001373291,
169
- 0.013398015975952148,
170
- 0.013431808471679688,
171
- 0.013451264381408692,
172
- 0.013424639701843261,
173
- 0.01342972755432129,
174
- 0.013460479736328124,
175
- 0.01343283176422119,
176
- 0.014407615661621093,
177
- 0.013944831848144532,
178
- 0.013946880340576171,
179
- 0.013883392333984374,
180
- 0.013922304153442382,
181
- 0.01388646411895752,
182
- 0.013930496215820312,
183
- 0.013429759979248047,
184
- 0.013411328315734864,
185
- 0.013492223739624023,
186
- 0.01347379207611084,
187
- 0.013392895698547362,
188
- 0.01338265609741211,
189
- 0.013392895698547362,
190
- 0.013420543670654296,
191
- 0.013371456146240234,
192
- 0.01335807991027832,
193
- 0.013455360412597657,
194
- 0.013389823913574218,
195
- 0.01338368034362793,
196
- 0.013435903549194337
197
  ]
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
- "value": 71.6123332025329
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
- "cpu": 1.611029108365377e-07,
206
- "ram": 8.807571214787856e-08,
207
- "gpu": 3.5670073581079516e-07,
208
- "total": 6.058793587952113e-07
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
- "value": 1650493.5932930543
213
  }
214
  }
215
  }
 
105
  "forward": {
106
  "memory": {
107
  "unit": "MB",
108
+ "max_ram": 982.491136,
109
  "max_global_vram": 1434.976256,
110
  "max_process_vram": 0.0,
111
  "max_reserved": 794.820608,
 
114
  "latency": {
115
  "unit": "s",
116
  "count": 72,
117
+ "total": 1.010008995056152,
118
+ "mean": 0.014027902709113225,
119
+ "stdev": 0.0009506995686880036,
120
+ "p50": 0.013721600055694581,
121
+ "p90": 0.014417202854156494,
122
+ "p95": 0.015691896533966066,
123
+ "p99": 0.01783655416488649,
124
  "values": [
125
+ 0.019886079788208007,
126
+ 0.016470016479492186,
127
+ 0.015333408355712891,
128
+ 0.015256575584411621,
129
+ 0.01359769630432129,
130
+ 0.01367347240447998,
131
+ 0.013515775680541992,
132
+ 0.013558783531188966,
133
+ 0.013570048332214356,
134
+ 0.013723648071289063,
135
+ 0.01358233642578125,
136
+ 0.013683712005615235,
137
+ 0.013699071884155273,
138
+ 0.013700096130371094,
139
+ 0.013634559631347656,
140
+ 0.013592576026916504,
141
+ 0.013674495697021484,
142
+ 0.013428735733032226,
143
+ 0.013601792335510255,
144
+ 0.013488127708435058,
145
+ 0.013592543601989747,
146
+ 0.013645824432373046,
147
+ 0.013745152473449707,
148
+ 0.013758463859558106,
149
+ 0.013535264015197754,
150
+ 0.013810688018798829,
151
+ 0.013616127967834473,
152
+ 0.013571071624755859,
153
+ 0.013503423690795898,
154
+ 0.013677568435668945,
155
+ 0.013578240394592284,
156
+ 0.013529088020324707,
157
+ 0.013629440307617188,
158
+ 0.01358028793334961,
159
+ 0.016130048751831053,
160
+ 0.01699942398071289,
161
+ 0.014419967651367188,
162
  0.014285823822021485,
163
+ 0.01459712028503418,
164
+ 0.014170111656188965,
165
+ 0.014166015625,
166
+ 0.014315487861633301,
167
+ 0.013908991813659668,
168
+ 0.014140512466430664,
169
+ 0.014145536422729492,
170
+ 0.014392319679260255,
171
+ 0.014256128311157227,
172
+ 0.014137408256530763,
173
+ 0.014092288017272948,
174
+ 0.014108672142028808,
175
+ 0.014024640083312988,
176
+ 0.013766655921936035,
177
+ 0.013807616233825683,
178
+ 0.013706175804138184,
179
+ 0.013651968002319336,
180
+ 0.013735936164855958,
181
+ 0.013514752388000489,
182
+ 0.01390182399749756,
183
+ 0.013734911918640137,
184
+ 0.013747200012207032,
185
+ 0.013619199752807617,
186
+ 0.014041088104248046,
187
+ 0.013784064292907714,
188
+ 0.013612031936645508,
189
+ 0.013742079734802246,
190
+ 0.013719552040100098,
191
+ 0.013739007949829102,
192
+ 0.013692928314208984,
193
+ 0.01366528034210205,
194
+ 0.0136878080368042,
195
+ 0.01368057632446289,
196
+ 0.013723648071289063
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
197
  ]
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
+ "value": 71.28649383562878
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
+ "cpu": 1.6237637269079597e-07,
206
+ "ram": 8.877343861993175e-08,
207
+ "gpu": 3.399013373698694e-07,
208
+ "total": 5.910511486805971e-07
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
+ "value": 1691900.9500824066
213
  }
214
  }
215
  }