IlyasMoutawwakil HF staff commited on
Commit
e9d7fee
·
verified ·
1 Parent(s): 4d1f789

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -105,7 +105,7 @@
105
  "forward": {
106
  "memory": {
107
  "unit": "MB",
108
- "max_ram": 982.491136,
109
  "max_global_vram": 1434.976256,
110
  "max_process_vram": 0.0,
111
  "max_reserved": 794.820608,
@@ -113,103 +113,105 @@
113
  },
114
  "latency": {
115
  "unit": "s",
116
- "count": 72,
117
- "total": 1.010008995056152,
118
- "mean": 0.014027902709113225,
119
- "stdev": 0.0009506995686880036,
120
- "p50": 0.013721600055694581,
121
- "p90": 0.014417202854156494,
122
- "p95": 0.015691896533966066,
123
- "p99": 0.01783655416488649,
124
  "values": [
125
- 0.019886079788208007,
126
- 0.016470016479492186,
127
- 0.015333408355712891,
128
- 0.015256575584411621,
129
- 0.01359769630432129,
130
- 0.01367347240447998,
131
- 0.013515775680541992,
132
- 0.013558783531188966,
133
- 0.013570048332214356,
134
- 0.013723648071289063,
135
- 0.01358233642578125,
136
- 0.013683712005615235,
137
- 0.013699071884155273,
138
- 0.013700096130371094,
139
- 0.013634559631347656,
140
- 0.013592576026916504,
141
- 0.013674495697021484,
142
- 0.013428735733032226,
143
- 0.013601792335510255,
144
- 0.013488127708435058,
145
- 0.013592543601989747,
146
- 0.013645824432373046,
147
- 0.013745152473449707,
148
- 0.013758463859558106,
149
- 0.013535264015197754,
150
- 0.013810688018798829,
151
- 0.013616127967834473,
152
- 0.013571071624755859,
153
- 0.013503423690795898,
154
- 0.013677568435668945,
155
- 0.013578240394592284,
156
- 0.013529088020324707,
157
- 0.013629440307617188,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
158
  0.01358028793334961,
159
- 0.016130048751831053,
160
- 0.01699942398071289,
161
- 0.014419967651367188,
162
- 0.014285823822021485,
163
- 0.01459712028503418,
164
- 0.014170111656188965,
165
- 0.014166015625,
166
- 0.014315487861633301,
167
- 0.013908991813659668,
168
- 0.014140512466430664,
169
- 0.014145536422729492,
170
- 0.014392319679260255,
171
- 0.014256128311157227,
172
- 0.014137408256530763,
173
- 0.014092288017272948,
174
- 0.014108672142028808,
175
- 0.014024640083312988,
176
- 0.013766655921936035,
177
- 0.013807616233825683,
178
- 0.013706175804138184,
179
- 0.013651968002319336,
180
- 0.013735936164855958,
181
- 0.013514752388000489,
182
- 0.01390182399749756,
183
- 0.013734911918640137,
184
- 0.013747200012207032,
185
- 0.013619199752807617,
186
- 0.014041088104248046,
187
- 0.013784064292907714,
188
- 0.013612031936645508,
189
- 0.013742079734802246,
190
- 0.013719552040100098,
191
- 0.013739007949829102,
192
- 0.013692928314208984,
193
- 0.01366528034210205,
194
- 0.0136878080368042,
195
- 0.01368057632446289,
196
- 0.013723648071289063
197
  ]
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
- "value": 71.28649383562878
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
- "cpu": 1.6237637269079597e-07,
206
- "ram": 8.877343861993175e-08,
207
- "gpu": 3.399013373698694e-07,
208
- "total": 5.910511486805971e-07
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
- "value": 1691900.9500824066
213
  }
214
  }
215
  }
 
105
  "forward": {
106
  "memory": {
107
  "unit": "MB",
108
+ "max_ram": 982.831104,
109
  "max_global_vram": 1434.976256,
110
  "max_process_vram": 0.0,
111
  "max_reserved": 794.820608,
 
113
  },
114
  "latency": {
115
  "unit": "s",
116
+ "count": 74,
117
+ "total": 0.9982882900238037,
118
+ "mean": 0.013490382297618968,
119
+ "stdev": 0.00035699389510758804,
120
+ "p50": 0.013436383724212646,
121
+ "p90": 0.013837785816192627,
122
+ "p95": 0.01429652452468872,
123
+ "p99": 0.014615294561386109,
124
  "values": [
125
+ 0.01467801570892334,
126
+ 0.014507871627807617,
127
+ 0.014420991897583007,
128
+ 0.014592096328735352,
129
+ 0.014006112098693848,
130
+ 0.01352188777923584,
131
+ 0.013643775939941406,
132
+ 0.01355673599243164,
133
+ 0.013428799629211426,
134
+ 0.013505536079406738,
135
+ 0.013422592163085938,
136
+ 0.013377599716186524,
137
+ 0.013624320030212403,
138
+ 0.013324288368225098,
139
+ 0.013297663688659669,
140
+ 0.013278240203857422,
141
+ 0.013336576461791993,
142
+ 0.01347379207611084,
143
+ 0.013532223701477051,
144
+ 0.013462528228759766,
145
+ 0.013550496101379395,
146
+ 0.013410304069519043,
147
+ 0.013301792144775391,
148
+ 0.013242400169372559,
149
+ 0.01326591968536377,
150
+ 0.013396991729736327,
151
+ 0.013179712295532227,
152
+ 0.01347481632232666,
153
+ 0.013765631675720215,
154
+ 0.013834112167358399,
155
+ 0.013546367645263672,
156
+ 0.013280256271362305,
157
+ 0.013222016334533692,
158
+ 0.013327360153198242,
159
+ 0.013271039962768554,
160
+ 0.013602815628051757,
161
+ 0.014229503631591797,
162
+ 0.014032896041870118,
163
+ 0.013546496391296388,
164
+ 0.013560832023620606,
165
+ 0.013429759979248047,
166
+ 0.013419520378112794,
167
+ 0.013345824241638184,
168
+ 0.012782591819763184,
169
+ 0.012999775886535645,
170
+ 0.013128704071044921,
171
+ 0.012957695960998536,
172
+ 0.012893183708190918,
173
+ 0.01291155242919922,
174
+ 0.012934240341186523,
175
+ 0.012900223731994629,
176
+ 0.013373439788818359,
177
+ 0.013839360237121581,
178
+ 0.01381987190246582,
179
+ 0.013644800186157227,
180
+ 0.013649920463562011,
181
+ 0.013461503982543945,
182
+ 0.01348300838470459,
183
+ 0.013395968437194825,
184
+ 0.013441056251525878,
185
+ 0.013372447967529297,
186
  0.01358028793334961,
187
+ 0.013527039527893067,
188
+ 0.013451264381408692,
189
+ 0.013373439788818359,
190
+ 0.013485024452209473,
191
+ 0.01344102382659912,
192
+ 0.013418496131896973,
193
+ 0.013413472175598145,
194
+ 0.013431743621826172,
195
+ 0.013547679901123048,
196
+ 0.013586496353149414,
197
+ 0.013417471885681152,
198
+ 0.013396991729736327
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
199
  ]
200
  },
201
  "throughput": {
202
  "unit": "samples/s",
203
+ "value": 74.1268837263788
204
  },
205
  "energy": {
206
  "unit": "kWh",
207
+ "cpu": 1.5601429115298165e-07,
208
+ "ram": 8.48876546085235e-08,
209
+ "gpu": 3.338458659220762e-07,
210
+ "total": 5.747478116835814e-07
211
  },
212
  "efficiency": {
213
  "unit": "samples/kWh",
214
+ "value": 1739893.5318618224
215
  }
216
  }
217
  }