Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"forward": {
|
3 |
"memory": {
|
4 |
"unit": "MB",
|
5 |
-
"max_ram": 902.
|
6 |
"max_global_vram": 1195.900928,
|
7 |
"max_process_vram": 0.0,
|
8 |
"max_reserved": 555.74528,
|
@@ -10,183 +10,179 @@
|
|
10 |
},
|
11 |
"latency": {
|
12 |
"unit": "s",
|
13 |
-
"count":
|
14 |
-
"total":
|
15 |
-
"mean": 0.
|
16 |
-
"stdev": 0.
|
17 |
-
"p50": 0.
|
18 |
-
"p90": 0.
|
19 |
-
"p95": 0.
|
20 |
-
"p99": 0.
|
21 |
"values": [
|
22 |
-
0.
|
23 |
-
0.
|
24 |
-
0.
|
25 |
-
0.
|
26 |
-
0.
|
27 |
-
0.
|
28 |
-
0.
|
29 |
-
0.
|
30 |
-
0.
|
31 |
-
0.
|
32 |
-
0.
|
33 |
-
0.
|
34 |
-
0.
|
35 |
-
0.
|
36 |
-
0.
|
37 |
-
0.
|
38 |
-
0.
|
39 |
-
0.
|
40 |
-
0.
|
41 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
42 |
0.006364160060882569,
|
43 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
44 |
0.006323200225830078,
|
45 |
-
0.
|
46 |
-
0.
|
47 |
-
0.
|
48 |
0.006351871967315674,
|
49 |
-
0.
|
50 |
-
0.
|
51 |
-
0.
|
52 |
-
0.
|
53 |
-
0.
|
54 |
-
0.
|
55 |
-
0.
|
56 |
-
0.
|
57 |
-
0.
|
58 |
-
0.
|
59 |
-
0.
|
60 |
-
0.
|
61 |
-
0.
|
62 |
-
0.
|
63 |
-
0.
|
64 |
-
0.
|
65 |
-
0.
|
66 |
-
0.
|
67 |
-
0.
|
68 |
-
0.
|
69 |
-
0.
|
70 |
-
0.
|
71 |
-
0.
|
72 |
-
0.
|
73 |
-
0.
|
74 |
-
0.
|
75 |
-
0.006586368083953857,
|
76 |
-
0.00652185583114624,
|
77 |
-
0.006575039863586426,
|
78 |
-
0.006586368083953857,
|
79 |
-
0.006515711784362793,
|
80 |
-
0.006577087879180908,
|
81 |
-
0.0065792641639709475,
|
82 |
-
0.006536191940307618,
|
83 |
-
0.006560768127441406,
|
84 |
-
0.006555647850036621,
|
85 |
-
0.006533120155334473,
|
86 |
-
0.006554624080657959,
|
87 |
-
0.006621183872222901,
|
88 |
-
0.006549503803253174,
|
89 |
-
0.006576128005981445,
|
90 |
-
0.006553599834442139,
|
91 |
-
0.006550528049468994,
|
92 |
-
0.006594560146331787,
|
93 |
-
0.006563839912414551,
|
94 |
-
0.006545407772064209,
|
95 |
-
0.006560768127441406,
|
96 |
-
0.006521920204162598,
|
97 |
-
0.006527999877929688,
|
98 |
-
0.006557695865631104,
|
99 |
-
0.006525951862335205,
|
100 |
-
0.006563839912414551,
|
101 |
-
0.00658841609954834,
|
102 |
-
0.006512639999389648,
|
103 |
-
0.0065443840026855465,
|
104 |
-
0.006546432018280029,
|
105 |
-
0.006551583766937256,
|
106 |
-
0.0065669121742248536,
|
107 |
-
0.006818816184997558,
|
108 |
-
0.006546432018280029,
|
109 |
-
0.006582272052764892,
|
110 |
-
0.006553599834442139,
|
111 |
-
0.00653004789352417,
|
112 |
-
0.006558720111846924,
|
113 |
-
0.006504447937011719,
|
114 |
-
0.006559711933135986,
|
115 |
-
0.006559743881225586,
|
116 |
-
0.006520832061767578,
|
117 |
-
0.007136256217956543,
|
118 |
-
0.007790592193603516,
|
119 |
-
0.007093183994293213,
|
120 |
-
0.007525375843048096,
|
121 |
-
0.006603775978088379,
|
122 |
-
0.00658022403717041,
|
123 |
-
0.006584320068359375,
|
124 |
-
0.006603775978088379,
|
125 |
-
0.006813695907592773,
|
126 |
-
0.006600639820098877,
|
127 |
-
0.006502399921417237,
|
128 |
-
0.006582272052764892,
|
129 |
-
0.006584320068359375,
|
130 |
-
0.006524928092956543,
|
131 |
-
0.006555647850036621,
|
132 |
-
0.006639616012573242,
|
133 |
-
0.006553599834442139,
|
134 |
-
0.006602752208709717,
|
135 |
-
0.006549503803253174,
|
136 |
-
0.006565887928009034,
|
137 |
-
0.006563839912414551,
|
138 |
-
0.0065443840026855465,
|
139 |
-
0.006501376152038574,
|
140 |
-
0.006604800224304199,
|
141 |
-
0.006452223777770996,
|
142 |
-
0.006554624080657959,
|
143 |
-
0.006582272052764892,
|
144 |
-
0.006543360233306885,
|
145 |
-
0.006536191940307618,
|
146 |
-
0.006532063961029053,
|
147 |
-
0.006540287971496582,
|
148 |
-
0.0065413122177124024,
|
149 |
-
0.006552639961242676,
|
150 |
-
0.006552576065063476,
|
151 |
-
0.006561791896820069,
|
152 |
-
0.006552576065063476,
|
153 |
-
0.006533120155334473,
|
154 |
-
0.006648831844329834,
|
155 |
-
0.006490111827850342,
|
156 |
-
0.006635519981384277,
|
157 |
-
0.0065669121742248536,
|
158 |
-
0.0065146880149841305,
|
159 |
-
0.006562816143035889,
|
160 |
-
0.006591487884521485,
|
161 |
-
0.006535168170928955,
|
162 |
-
0.006567935943603515,
|
163 |
-
0.006554624080657959,
|
164 |
-
0.00653107213973999,
|
165 |
-
0.006568960189819336,
|
166 |
-
0.006964223861694336,
|
167 |
-
0.006465536117553711,
|
168 |
-
0.006472703933715821,
|
169 |
-
0.006445055961608887,
|
170 |
-
0.00643891191482544,
|
171 |
-
0.006520832061767578,
|
172 |
-
0.00652288007736206,
|
173 |
-
0.006534143924713135
|
174 |
]
|
175 |
},
|
176 |
"throughput": {
|
177 |
"unit": "samples/s",
|
178 |
-
"value":
|
179 |
},
|
180 |
"energy": {
|
181 |
"unit": "kWh",
|
182 |
-
"cpu": 7.
|
183 |
-
"ram": 4.
|
184 |
-
"gpu": 1.
|
185 |
-
"total": 2.
|
186 |
},
|
187 |
"efficiency": {
|
188 |
"unit": "samples/kWh",
|
189 |
-
"value":
|
190 |
}
|
191 |
}
|
192 |
}
|
|
|
2 |
"forward": {
|
3 |
"memory": {
|
4 |
"unit": "MB",
|
5 |
+
"max_ram": 902.28736,
|
6 |
"max_global_vram": 1195.900928,
|
7 |
"max_process_vram": 0.0,
|
8 |
"max_reserved": 555.74528,
|
|
|
10 |
},
|
11 |
"latency": {
|
12 |
"unit": "s",
|
13 |
+
"count": 148,
|
14 |
+
"total": 0.9989087686538696,
|
15 |
+
"mean": 0.006749383571985605,
|
16 |
+
"stdev": 0.0006199000503150585,
|
17 |
+
"p50": 0.0066923520565032955,
|
18 |
+
"p90": 0.006864076614379882,
|
19 |
+
"p95": 0.008348518562316899,
|
20 |
+
"p99": 0.009350686492919922,
|
21 |
"values": [
|
22 |
+
0.009878527641296387,
|
23 |
+
0.009399295806884766,
|
24 |
+
0.009274368286132812,
|
25 |
+
0.00929587173461914,
|
26 |
+
0.009267200469970703,
|
27 |
+
0.008657983779907227,
|
28 |
+
0.00859545612335205,
|
29 |
+
0.008596480369567871,
|
30 |
+
0.006862847805023193,
|
31 |
+
0.006762495994567871,
|
32 |
+
0.006724607944488525,
|
33 |
+
0.006703104019165039,
|
34 |
+
0.006758399963378906,
|
35 |
+
0.00673689603805542,
|
36 |
+
0.006625216007232666,
|
37 |
+
0.006660096168518067,
|
38 |
+
0.006668288230895996,
|
39 |
+
0.006676479816436768,
|
40 |
+
0.00679423999786377,
|
41 |
+
0.006791168212890625,
|
42 |
+
0.006763519763946534,
|
43 |
+
0.0068351998329162595,
|
44 |
+
0.006744063854217529,
|
45 |
+
0.006715392112731934,
|
46 |
+
0.0067338237762451176,
|
47 |
+
0.006805471897125244,
|
48 |
+
0.006761472225189209,
|
49 |
+
0.006737919807434082,
|
50 |
+
0.0067645440101623535,
|
51 |
+
0.006758399963378906,
|
52 |
+
0.00670201587677002,
|
53 |
+
0.006749184131622315,
|
54 |
+
0.006762495994567871,
|
55 |
+
0.00671123218536377,
|
56 |
+
0.00669593620300293,
|
57 |
+
0.006647808074951172,
|
58 |
+
0.006680575847625733,
|
59 |
+
0.006668288230895996,
|
60 |
+
0.0067123198509216305,
|
61 |
+
0.006713344097137451,
|
62 |
+
0.006668288230895996,
|
63 |
+
0.006713344097137451,
|
64 |
+
0.006740992069244385,
|
65 |
+
0.006738944053649902,
|
66 |
+
0.00677177619934082,
|
67 |
+
0.00675328016281128,
|
68 |
+
0.006754303932189941,
|
69 |
+
0.006789120197296142,
|
70 |
+
0.006726655960083008,
|
71 |
+
0.006652927875518798,
|
72 |
+
0.006708288192749023,
|
73 |
+
0.00679423999786377,
|
74 |
+
0.006740992069244385,
|
75 |
+
0.006699007987976074,
|
76 |
+
0.006744063854217529,
|
77 |
+
0.006663199901580811,
|
78 |
+
0.006631423950195312,
|
79 |
+
0.006711296081542969,
|
80 |
+
0.006662112236022949,
|
81 |
+
0.0066979842185974124,
|
82 |
+
0.0068280320167541505,
|
83 |
+
0.006656000137329102,
|
84 |
+
0.006585311889648437,
|
85 |
+
0.006691840171813965,
|
86 |
+
0.006692863941192627,
|
87 |
+
0.006721536159515381,
|
88 |
+
0.0067041277885437015,
|
89 |
+
0.006703104019165039,
|
90 |
+
0.0066938881874084475,
|
91 |
+
0.006762495994567871,
|
92 |
+
0.006841343879699707,
|
93 |
+
0.006866943836212158,
|
94 |
+
0.00679423999786377,
|
95 |
+
0.00682700777053833,
|
96 |
+
0.006744063854217529,
|
97 |
+
0.006730751991271973,
|
98 |
+
0.006732800006866455,
|
99 |
+
0.006663167953491211,
|
100 |
+
0.006756351947784424,
|
101 |
+
0.006658048152923584,
|
102 |
+
0.006669312000274658,
|
103 |
+
0.0064880638122558594,
|
104 |
+
0.006461440086364746,
|
105 |
+
0.0066979842185974124,
|
106 |
+
0.006568992137908935,
|
107 |
+
0.00638156795501709,
|
108 |
+
0.0065382399559021,
|
109 |
+
0.006713344097137451,
|
110 |
+
0.006862847805023193,
|
111 |
+
0.006855679988861084,
|
112 |
+
0.0068884482383728025,
|
113 |
+
0.006947840213775635,
|
114 |
+
0.006947807788848877,
|
115 |
+
0.006965248107910156,
|
116 |
+
0.007651328086853027,
|
117 |
+
0.007889920234680176,
|
118 |
+
0.006717440128326416,
|
119 |
+
0.0066416640281677245,
|
120 |
+
0.0066447358131408694,
|
121 |
+
0.006627327919006347,
|
122 |
+
0.006652927875518798,
|
123 |
+
0.006632415771484375,
|
124 |
+
0.006512639999389648,
|
125 |
+
0.006392831802368164,
|
126 |
+
0.0063539199829101565,
|
127 |
+
0.006343679904937744,
|
128 |
0.006364160060882569,
|
129 |
+
0.006333439826965332,
|
130 |
+
0.0063385281562805176,
|
131 |
+
0.006333439826965332,
|
132 |
+
0.006347775936126709,
|
133 |
+
0.006346752166748047,
|
134 |
+
0.006341631889343262,
|
135 |
+
0.006379519939422608,
|
136 |
+
0.006347775936126709,
|
137 |
+
0.006347775936126709,
|
138 |
+
0.0063610877990722655,
|
139 |
0.006323200225830078,
|
140 |
+
0.006354944229125976,
|
141 |
+
0.006309887886047363,
|
142 |
+
0.0063610877990722655,
|
143 |
0.006351871967315674,
|
144 |
+
0.006313983917236328,
|
145 |
+
0.0063784317970275875,
|
146 |
+
0.006330431938171387,
|
147 |
+
0.006371327877044678,
|
148 |
+
0.006351871967315674,
|
149 |
+
0.006340640068054199,
|
150 |
+
0.006354944229125976,
|
151 |
+
0.006308864116668702,
|
152 |
+
0.006354944229125976,
|
153 |
+
0.006341631889343262,
|
154 |
+
0.006307839870452881,
|
155 |
+
0.006351871967315674,
|
156 |
+
0.0063170561790466305,
|
157 |
+
0.006358016014099121,
|
158 |
+
0.006341631889343262,
|
159 |
+
0.006329343795776367,
|
160 |
+
0.006341631889343262,
|
161 |
+
0.006328320026397705,
|
162 |
+
0.006365119934082031,
|
163 |
+
0.006340608119964599,
|
164 |
+
0.006326272010803223,
|
165 |
+
0.006343679904937744,
|
166 |
+
0.0063201279640197755,
|
167 |
+
0.006352896213531494,
|
168 |
+
0.006354944229125976,
|
169 |
+
0.0063836159706115725
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
170 |
]
|
171 |
},
|
172 |
"throughput": {
|
173 |
"unit": "samples/s",
|
174 |
+
"value": 148.1616786680579
|
175 |
},
|
176 |
"energy": {
|
177 |
"unit": "kWh",
|
178 |
+
"cpu": 7.798429687469327e-08,
|
179 |
+
"ram": 4.263400563247973e-08,
|
180 |
+
"gpu": 1.3821648484210711e-07,
|
181 |
+
"total": 2.588347873492801e-07
|
182 |
},
|
183 |
"efficiency": {
|
184 |
"unit": "samples/kWh",
|
185 |
+
"value": 3863468.3159901816
|
186 |
}
|
187 |
}
|
188 |
}
|