IlyasMoutawwakil HF staff commited on
Commit
d1ef06a
·
verified ·
1 Parent(s): d5cc4d1

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -102,7 +102,7 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 975.491072,
106
  "max_global_vram": 1434.976256,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 794.820608,
@@ -110,102 +110,101 @@
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 71,
114
- "total": 1.0056614370346066,
115
- "mean": 0.014164245592036717,
116
- "stdev": 0.00044171859682510525,
117
- "p50": 0.014051360130310058,
118
- "p90": 0.014436351776123046,
119
- "p95": 0.01508351993560791,
120
- "p99": 0.015944806003570556,
121
  "values": [
122
- 0.01538764762878418,
123
- 0.01617919921875,
124
- 0.01579529571533203,
125
- 0.015844351768493654,
126
- 0.014779392242431641,
127
- 0.01420083236694336,
128
- 0.014054400444030762,
129
- 0.014000127792358399,
130
- 0.013830143928527832,
131
- 0.013831168174743653,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
132
  0.014029824256896972,
133
- 0.013916159629821777,
134
- 0.013868032455444336,
135
- 0.013856767654418945,
136
- 0.01397043228149414,
137
- 0.014118911743164063,
138
- 0.0141659517288208,
139
- 0.014126079559326172,
140
- 0.01397862434387207,
141
- 0.014105600357055664,
142
- 0.014051360130310058,
143
- 0.014134271621704102,
144
- 0.014071807861328126,
145
- 0.014046208381652832,
146
- 0.014112768173217773,
147
- 0.014109696388244629,
148
- 0.014051327705383301,
149
- 0.014015487670898438,
150
- 0.014034943580627441,
151
- 0.01425100803375244,
152
- 0.01390182399749756,
153
  0.01397555160522461,
154
- 0.013876223564147949,
 
 
 
 
 
 
 
 
 
155
  0.01386188793182373,
156
- 0.01386086368560791,
157
- 0.01384447956085205,
158
- 0.013927424430847168,
159
- 0.013837311744689941,
160
- 0.01386393642425537,
161
- 0.013969408035278321,
162
- 0.014004223823547364,
163
- 0.013921279907226563,
164
- 0.013874176025390626,
165
- 0.013881343841552735,
166
- 0.014483551979064941,
167
- 0.0140697603225708,
168
- 0.014072832107543945,
169
- 0.014585856437683106,
170
- 0.014299136161804199,
171
- 0.014231552124023437,
172
- 0.014375935554504395,
173
- 0.014306303977966308,
174
- 0.014115839958190919,
175
- 0.014170111656188965,
176
- 0.014113823890686035,
177
- 0.01386291217803955,
178
- 0.014436351776123046,
179
- 0.014094335556030273,
180
- 0.014057536125183106,
181
- 0.01407590389251709,
182
- 0.014056447982788087,
183
- 0.014050304412841797,
184
- 0.014042112350463867,
185
- 0.014034943580627441,
186
- 0.014164992332458496,
187
- 0.014210047721862793,
188
- 0.014139391899108887,
189
- 0.014013440132141113,
190
- 0.014020607948303223,
191
- 0.014040063858032227,
192
- 0.014019583702087402
193
  ]
194
  },
195
  "throughput": {
196
  "unit": "samples/s",
197
- "value": 70.60030084215782
198
  },
199
  "energy": {
200
  "unit": "kWh",
201
- "cpu": 1.6661593805492662e-07,
202
- "ram": 9.068229740023525e-08,
203
- "gpu": 3.232911536944357e-07,
204
- "total": 5.805893891495975e-07
205
  },
206
  "efficiency": {
207
  "unit": "samples/kWh",
208
- "value": 1722387.661036525
209
  }
210
  }
211
  }
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 975.515648,
106
  "max_global_vram": 1434.976256,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 794.820608,
 
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 70,
114
+ "total": 1.0092306232452397,
115
+ "mean": 0.01441758033207485,
116
+ "stdev": 0.00043281803130732914,
117
+ "p50": 0.014469631671905517,
118
+ "p90": 0.014797516822814941,
119
+ "p95": 0.015288847780227661,
120
+ "p99": 0.01584938042640686,
121
  "values": [
122
+ 0.015313920021057128,
123
+ 0.015992768287658693,
124
+ 0.01528927993774414,
125
+ 0.014533568382263183,
126
+ 0.014776288032531738,
127
+ 0.014534655570983887,
128
+ 0.014455807685852052,
129
+ 0.014361599922180175,
130
+ 0.014534655570983887,
131
+ 0.014725119590759277,
132
+ 0.014537728309631348,
133
+ 0.014548992156982422,
134
+ 0.014650367736816406,
135
+ 0.014739456176757813,
136
+ 0.014467071533203125,
137
+ 0.014560256004333496,
138
+ 0.014454784393310547,
139
+ 0.014499839782714843,
140
+ 0.01528831958770752,
141
+ 0.01448755168914795,
142
+ 0.01462377643585205,
143
+ 0.014896127700805664,
144
+ 0.01460223960876465,
145
+ 0.014525440216064453,
146
+ 0.014631936073303223,
147
+ 0.014531583786010742,
148
+ 0.014515199661254884,
149
+ 0.014474240303039551,
150
+ 0.014740480422973632,
151
+ 0.014414848327636719,
152
+ 0.014409728050231933,
153
+ 0.01447219181060791,
154
+ 0.014546879768371582,
155
+ 0.01578495979309082,
156
+ 0.01478656005859375,
157
+ 0.014486528396606446,
158
+ 0.01448960018157959,
159
+ 0.014337023735046387,
160
+ 0.014355456352233887,
161
+ 0.014321663856506347,
162
+ 0.014193663597106934,
163
+ 0.014403583526611329,
164
+ 0.014474240303039551,
165
+ 0.014929920196533204,
166
+ 0.014532608032226562,
167
+ 0.014476287841796874,
168
+ 0.01408512020111084,
169
+ 0.014196736335754395,
170
  0.014029824256896972,
171
+ 0.013910016059875489,
172
+ 0.013980671882629395,
173
+ 0.013893695831298828,
174
+ 0.014016511917114258,
175
+ 0.013957088470458984,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
176
  0.01397555160522461,
177
+ 0.013900799751281738,
178
+ 0.014395392417907715,
179
+ 0.014152704238891601,
180
+ 0.013963264465332031,
181
+ 0.013875200271606445,
182
+ 0.013956095695495606,
183
+ 0.013889535903930664,
184
+ 0.013908991813659668,
185
+ 0.013959168434143066,
186
+ 0.013917183876037598,
187
  0.01386188793182373,
188
+ 0.01396019172668457,
189
+ 0.013910016059875489,
190
+ 0.013896703720092773,
191
+ 0.013929471969604493
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
192
  ]
193
  },
194
  "throughput": {
195
  "unit": "samples/s",
196
+ "value": 69.3597661304717
197
  },
198
  "energy": {
199
  "unit": "kWh",
200
+ "cpu": 1.6292014259907392e-07,
201
+ "ram": 8.888658838707806e-08,
202
+ "gpu": 3.4823467736985594e-07,
203
+ "total": 6.000414083560079e-07
204
  },
205
  "efficiency": {
206
  "unit": "samples/kWh",
207
+ "value": 1666551.6513931893
208
  }
209
  }
210
  }