IlyasMoutawwakil HF staff commited on
Commit
bba7297
·
verified ·
1 Parent(s): 9b77b0b

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 908.357632,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
@@ -10,179 +10,185 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 148,
14
- "total": 1.0002943978309633,
15
- "mean": 0.0067587459312902914,
16
- "stdev": 0.00043562922231685263,
17
- "p50": 0.0067645440101623535,
18
- "p90": 0.006954758262634277,
19
- "p95": 0.007115457773208619,
20
- "p99": 0.008922040643692017,
21
  "values": [
22
- 0.009449407577514649,
23
- 0.008961024284362793,
24
- 0.008878080368041993,
25
- 0.008804351806640624,
26
- 0.007475200176239013,
27
- 0.0070348801612854,
28
- 0.0069324798583984375,
29
- 0.006829055786132812,
30
- 0.007030784130096435,
31
- 0.006801439762115479,
32
- 0.00677785587310791,
33
- 0.006898655891418457,
34
- 0.006785024166107178,
35
- 0.0067758078575134275,
36
- 0.006725632190704346,
37
- 0.00672870397567749,
38
- 0.006795263767242431,
39
- 0.007176095962524414,
40
- 0.007268352031707764,
41
- 0.006990848064422607,
42
- 0.006916096210479736,
43
- 0.006845439910888672,
44
- 0.006927296161651612,
45
- 0.006785024166107178,
46
- 0.006858751773834228,
47
- 0.006834176063537598,
48
- 0.007046048164367676,
49
- 0.006907904148101806,
50
- 0.006841184139251709,
51
- 0.006781951904296875,
52
- 0.0067358717918396,
53
- 0.0068249602317810056,
54
  0.006803455829620361,
55
- 0.006797311782836914,
56
- 0.006841343879699707,
57
- 0.006785024166107178,
58
- 0.006633471965789795,
59
- 0.006694911956787109,
60
- 0.006792191982269287,
61
- 0.006729728221893311,
62
- 0.006706175804138184,
63
- 0.006989823818206787,
64
- 0.006750207901000976,
65
- 0.006968287944793701,
66
- 0.006822912216186523,
67
- 0.006846464157104492,
68
- 0.006806528091430664,
69
- 0.006806528091430664,
70
- 0.006858751773834228,
71
- 0.006761472225189209,
72
- 0.0067674560546875,
73
- 0.006814720153808594,
74
- 0.006831103801727295,
75
- 0.006842368125915528,
76
- 0.006788095951080322,
77
- 0.0067983360290527345,
78
- 0.006797311782836914,
79
- 0.006886367797851563,
80
- 0.006918144226074219,
81
- 0.006866943836212158,
82
- 0.006823935985565186,
83
- 0.006944896221160889,
84
- 0.006814720153808594,
85
- 0.006840320110321045,
86
- 0.006801407814025879,
87
- 0.0067645440101623535,
88
- 0.0067645440101623535,
89
- 0.006754303932189941,
90
- 0.00678604793548584,
91
- 0.0069489598274230955,
92
- 0.00688640022277832,
93
- 0.006818816184997558,
94
- 0.0070522880554199216,
95
- 0.007149472236633301,
96
- 0.006944767951965332,
97
- 0.006803455829620361,
98
- 0.006851583957672119,
99
- 0.006811647891998291,
100
- 0.006744063854217529,
101
- 0.006872064113616944,
102
- 0.006765567779541016,
103
- 0.0067645440101623535,
104
  0.006554624080657959,
105
- 0.0065762557983398435,
106
- 0.006717440128326416,
107
- 0.006623231887817383,
108
- 0.006464384078979492,
109
- 0.0064133119583129885,
110
  0.006617087841033936,
111
- 0.0065064959526062015,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112
  0.006474751949310303,
113
- 0.006890495777130127,
114
- 0.006595583915710449,
115
- 0.006638591766357422,
116
- 0.006752255916595459,
117
- 0.006903808116912841,
118
- 0.006608895778656006,
119
- 0.006455296039581298,
120
- 0.00652185583114624,
121
- 0.006504447937011719,
122
- 0.006531136035919189,
123
- 0.0065474557876586915,
 
 
 
 
 
 
124
  0.006500351905822754,
125
- 0.006524928092956543,
126
- 0.006478879928588867,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
127
  0.006398975849151611,
128
- 0.006651904106140137,
129
- 0.006661119937896728,
130
- 0.006478847980499268,
131
- 0.006461440086364746,
132
- 0.006449088096618653,
133
- 0.006452223777770996,
134
- 0.0064430079460144046,
135
- 0.006415359973907471,
136
- 0.006402048110961914,
137
- 0.006542143821716309,
138
  0.006397952079772949,
139
- 0.0064245758056640625,
140
- 0.006401023864746094,
141
- 0.006411263942718506,
142
- 0.006384640216827392,
143
- 0.006339583873748779,
144
- 0.006377471923828125,
145
- 0.006486911773681641,
146
- 0.006782976150512696,
147
- 0.006922207832336426,
148
- 0.00648089599609375,
149
- 0.0064440321922302245,
150
- 0.0064306240081787105,
151
- 0.006415487766265869,
 
 
 
 
 
152
  0.006392831802368164,
153
- 0.00647273588180542,
154
- 0.006411263942718506,
 
155
  0.006457344055175781,
156
- 0.0063937602043151855,
157
- 0.0064471039772033695,
158
- 0.0064644479751586915,
159
- 0.0063836159706115725,
160
- 0.006429696083068847,
161
- 0.0064542717933654785,
162
  0.006433792114257812,
163
- 0.006602752208709717,
164
- 0.0065924801826477055,
165
- 0.006446080207824707,
166
- 0.006411263942718506,
167
- 0.00637337589263916,
168
- 0.0064174079895019534,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
169
  0.006415359973907471
170
  ]
171
  },
172
  "throughput": {
173
  "unit": "samples/s",
174
- "value": 147.95644194441454
175
  },
176
  "energy": {
177
  "unit": "kWh",
178
- "cpu": 7.349243502558013e-08,
179
- "ram": 4.018104493988821e-08,
180
- "gpu": 1.3912219634568036e-07,
181
- "total": 2.527956763111487e-07
182
  },
183
  "efficiency": {
184
  "unit": "samples/kWh",
185
- "value": 3955763.858750374
186
  }
187
  }
188
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 908.005376,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 154,
14
+ "total": 0.9989794883728027,
15
+ "mean": 0.006486879794628589,
16
+ "stdev": 0.00014497112456265002,
17
+ "p50": 0.006457344055175781,
18
+ "p90": 0.006613094282150269,
19
+ "p95": 0.006732543921470642,
20
+ "p99": 0.007088906364440917,
21
  "values": [
22
+ 0.007197696208953858,
23
+ 0.006968319892883301,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  0.006803455829620361,
25
+ 0.006681600093841553,
26
+ 0.0068392958641052244,
27
+ 0.006658048152923584,
28
+ 0.006887423992156983,
29
+ 0.006586368083953857,
30
+ 0.007147520065307617,
31
+ 0.006624256134033203,
32
+ 0.006518784046173095,
33
+ 0.006469632148742676,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  0.006554624080657959,
35
+ 0.006474751949310303,
36
+ 0.006557695865631104,
 
 
 
37
  0.006617087841033936,
38
+ 0.006469632148742676,
39
+ 0.0065781760215759275,
40
+ 0.006565887928009034,
41
+ 0.006406144142150879,
42
+ 0.006555647850036621,
43
+ 0.006697951793670654,
44
+ 0.006412288188934326,
45
+ 0.006549503803253174,
46
+ 0.006479872226715088,
47
+ 0.006445055961608887,
48
+ 0.006497280120849609,
49
+ 0.006298624038696289,
50
+ 0.0064839677810668945,
51
+ 0.006542335987091064,
52
+ 0.006396927833557129,
53
+ 0.006540287971496582,
54
+ 0.00652288007736206,
55
+ 0.0064102401733398436,
56
+ 0.006558720111846924,
57
+ 0.006345727920532227,
58
+ 0.006322175979614258,
59
+ 0.006370304107666015,
60
+ 0.00638156795501709,
61
+ 0.006563839912414551,
62
+ 0.006494207859039307,
63
+ 0.006457344055175781,
64
+ 0.006512639999389648,
65
  0.006474751949310303,
66
+ 0.00643993616104126,
67
+ 0.006510591983795166,
68
+ 0.006449151992797852,
69
+ 0.006490111827850342,
70
+ 0.0065075201988220215,
71
+ 0.006395904064178467,
72
+ 0.0066007041931152345,
73
+ 0.0065710082054138185,
74
+ 0.006665215969085693,
75
+ 0.006576128005981445,
76
+ 0.006684671878814697,
77
+ 0.006429696083068847,
78
+ 0.0064143362045288085,
79
+ 0.006388768196105957,
80
+ 0.006354944229125976,
81
+ 0.006550528049468994,
82
+ 0.007036928176879883,
83
  0.006500351905822754,
84
+ 0.006492159843444824,
85
+ 0.006421504020690918,
86
+ 0.006487008094787598,
87
+ 0.0064880638122558594,
88
+ 0.006389791965484619,
89
+ 0.006472703933715821,
90
+ 0.006491136074066162,
91
+ 0.006419456005096436,
92
+ 0.006482944011688232,
93
+ 0.006504447937011719,
94
+ 0.006450175762176514,
95
+ 0.006512639999389648,
96
+ 0.006456319808959961,
97
+ 0.0064737281799316405,
98
+ 0.006405119895935059,
99
+ 0.0062975997924804685,
100
+ 0.006368256092071533,
101
+ 0.00642252779006958,
102
+ 0.006268928050994873,
103
+ 0.006421504020690918,
104
+ 0.006337535858154297,
105
+ 0.006333439826965332,
106
+ 0.006395904064178467,
107
+ 0.006269951820373535,
108
+ 0.00638156795501709,
109
+ 0.006400000095367431,
110
+ 0.006253568172454834,
111
+ 0.0063836159706115725,
112
+ 0.006296576023101807,
113
+ 0.006347775936126709,
114
+ 0.006387712001800537,
115
+ 0.006435840129852295,
116
+ 0.006561791896820069,
117
+ 0.00652185583114624,
118
  0.006398975849151611,
119
+ 0.006702079772949219,
120
+ 0.006519807815551758,
 
 
 
 
 
 
 
 
121
  0.006397952079772949,
122
+ 0.006556672096252441,
123
+ 0.006434815883636475,
124
+ 0.006553599834442139,
125
+ 0.006789120197296142,
126
+ 0.006589439868927002,
127
+ 0.00652185583114624,
128
+ 0.006552576065063476,
129
+ 0.006427648067474365,
130
+ 0.006542335987091064,
131
+ 0.0065382399559021,
132
+ 0.006494207859039307,
133
+ 0.006524928092956543,
134
+ 0.006495232105255127,
135
+ 0.006437888145446777,
136
+ 0.00652288007736206,
137
+ 0.006503424167633057,
138
+ 0.006487040042877197,
139
+ 0.006456319808959961,
140
  0.006392831802368164,
141
+ 0.0064737281799316405,
142
+ 0.006441952228546143,
143
+ 0.006384640216827392,
144
  0.006457344055175781,
145
+ 0.006412288188934326,
146
+ 0.006403071880340576,
147
+ 0.006481823921203614,
 
 
 
148
  0.006433792114257812,
149
+ 0.006449151992797852,
150
+ 0.006377471923828125,
151
+ 0.006597631931304931,
152
+ 0.00639081621170044,
153
+ 0.006375423908233643,
154
+ 0.006405119895935059,
155
+ 0.006366208076477051,
156
+ 0.006407167911529541,
157
+ 0.0063508481979370115,
158
+ 0.0064102401733398436,
159
+ 0.006386688232421875,
160
+ 0.006406144142150879,
161
+ 0.006474751949310303,
162
+ 0.006415359973907471,
163
+ 0.006375423908233643,
164
+ 0.006426623821258545,
165
+ 0.006385663986206055,
166
+ 0.006388800144195557,
167
+ 0.006603775978088379,
168
+ 0.00636518383026123,
169
+ 0.006421504020690918,
170
+ 0.006367231845855713,
171
+ 0.006377408027648926,
172
+ 0.0064204797744750975,
173
+ 0.006372352123260498,
174
+ 0.006397952079772949,
175
  0.006415359973907471
176
  ]
177
  },
178
  "throughput": {
179
  "unit": "samples/s",
180
+ "value": 154.15731933680078
181
  },
182
  "energy": {
183
  "unit": "kWh",
184
+ "cpu": 7.543949712191755e-08,
185
+ "ram": 4.124434678936419e-08,
186
+ "gpu": 1.3960733036943034e-07,
187
+ "total": 2.5629117428071206e-07
188
  },
189
  "efficiency": {
190
  "unit": "samples/kWh",
191
+ "value": 3901812.080757468
192
  }
193
  }
194
  }