IlyasMoutawwakil HF staff commited on
Commit
da759b3
·
verified ·
1 Parent(s): ca889ee

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 902.115328,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
@@ -10,183 +10,179 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 152,
14
- "total": 1.001573279380798,
15
- "mean": 0.006589297890663147,
16
- "stdev": 0.00022843924268888433,
17
- "p50": 0.006553599834442139,
18
- "p90": 0.006811033535003662,
19
- "p95": 0.007023871970176697,
20
- "p99": 0.0076116683959960955,
21
  "values": [
22
- 0.007549952030181885,
23
- 0.007012351989746094,
24
- 0.007113728046417236,
25
- 0.006969312191009522,
26
- 0.006897664070129395,
27
- 0.0067870721817016606,
28
- 0.007002111911773682,
29
- 0.007037951946258545,
30
- 0.007675903797149658,
31
- 0.006536255836486817,
32
- 0.006401023864746094,
33
- 0.006407167911529541,
34
- 0.006329343795776367,
35
- 0.006363135814666748,
36
- 0.006311935901641846,
37
- 0.006355967998504639,
38
- 0.006354911804199219,
39
- 0.006336512088775635,
40
- 0.006375423908233643,
41
- 0.006351967811584472,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  0.006364160060882569,
43
- 0.0063569917678833006,
 
 
 
 
 
 
 
 
 
44
  0.006323200225830078,
45
- 0.0064767999649047855,
46
- 0.006331391811370849,
47
- 0.0063508481979370115,
48
  0.006351871967315674,
49
- 0.006331391811370849,
50
- 0.006359039783477783,
51
- 0.0063211522102355954,
52
- 0.006336512088775635,
53
- 0.00636518383026123,
54
- 0.006425600051879882,
55
- 0.006599679946899414,
56
- 0.0065474557876586915,
57
- 0.006662144184112549,
58
- 0.006601727962493896,
59
- 0.006560768127441406,
60
- 0.006624288082122802,
61
- 0.006586304187774658,
62
- 0.0065413122177124024,
63
- 0.006584320068359375,
64
- 0.006576128005981445,
65
- 0.006542335987091064,
66
- 0.006605855941772461,
67
- 0.006552576065063476,
68
- 0.006553599834442139,
69
- 0.006552512168884277,
70
- 0.006567903995513916,
71
- 0.006550528049468994,
72
- 0.006560768127441406,
73
- 0.006874112129211426,
74
- 0.006550528049468994,
75
- 0.006586368083953857,
76
- 0.00652185583114624,
77
- 0.006575039863586426,
78
- 0.006586368083953857,
79
- 0.006515711784362793,
80
- 0.006577087879180908,
81
- 0.0065792641639709475,
82
- 0.006536191940307618,
83
- 0.006560768127441406,
84
- 0.006555647850036621,
85
- 0.006533120155334473,
86
- 0.006554624080657959,
87
- 0.006621183872222901,
88
- 0.006549503803253174,
89
- 0.006576128005981445,
90
- 0.006553599834442139,
91
- 0.006550528049468994,
92
- 0.006594560146331787,
93
- 0.006563839912414551,
94
- 0.006545407772064209,
95
- 0.006560768127441406,
96
- 0.006521920204162598,
97
- 0.006527999877929688,
98
- 0.006557695865631104,
99
- 0.006525951862335205,
100
- 0.006563839912414551,
101
- 0.00658841609954834,
102
- 0.006512639999389648,
103
- 0.0065443840026855465,
104
- 0.006546432018280029,
105
- 0.006551583766937256,
106
- 0.0065669121742248536,
107
- 0.006818816184997558,
108
- 0.006546432018280029,
109
- 0.006582272052764892,
110
- 0.006553599834442139,
111
- 0.00653004789352417,
112
- 0.006558720111846924,
113
- 0.006504447937011719,
114
- 0.006559711933135986,
115
- 0.006559743881225586,
116
- 0.006520832061767578,
117
- 0.007136256217956543,
118
- 0.007790592193603516,
119
- 0.007093183994293213,
120
- 0.007525375843048096,
121
- 0.006603775978088379,
122
- 0.00658022403717041,
123
- 0.006584320068359375,
124
- 0.006603775978088379,
125
- 0.006813695907592773,
126
- 0.006600639820098877,
127
- 0.006502399921417237,
128
- 0.006582272052764892,
129
- 0.006584320068359375,
130
- 0.006524928092956543,
131
- 0.006555647850036621,
132
- 0.006639616012573242,
133
- 0.006553599834442139,
134
- 0.006602752208709717,
135
- 0.006549503803253174,
136
- 0.006565887928009034,
137
- 0.006563839912414551,
138
- 0.0065443840026855465,
139
- 0.006501376152038574,
140
- 0.006604800224304199,
141
- 0.006452223777770996,
142
- 0.006554624080657959,
143
- 0.006582272052764892,
144
- 0.006543360233306885,
145
- 0.006536191940307618,
146
- 0.006532063961029053,
147
- 0.006540287971496582,
148
- 0.0065413122177124024,
149
- 0.006552639961242676,
150
- 0.006552576065063476,
151
- 0.006561791896820069,
152
- 0.006552576065063476,
153
- 0.006533120155334473,
154
- 0.006648831844329834,
155
- 0.006490111827850342,
156
- 0.006635519981384277,
157
- 0.0065669121742248536,
158
- 0.0065146880149841305,
159
- 0.006562816143035889,
160
- 0.006591487884521485,
161
- 0.006535168170928955,
162
- 0.006567935943603515,
163
- 0.006554624080657959,
164
- 0.00653107213973999,
165
- 0.006568960189819336,
166
- 0.006964223861694336,
167
- 0.006465536117553711,
168
- 0.006472703933715821,
169
- 0.006445055961608887,
170
- 0.00643891191482544,
171
- 0.006520832061767578,
172
- 0.00652288007736206,
173
- 0.006534143924713135
174
  ]
175
  },
176
  "throughput": {
177
  "unit": "samples/s",
178
- "value": 151.76123717474854
179
  },
180
  "energy": {
181
  "unit": "kWh",
182
- "cpu": 7.842443636711072e-08,
183
- "ram": 4.287034533756965e-08,
184
- "gpu": 1.4525581892715106e-07,
185
- "total": 2.6655060063183144e-07
186
  },
187
  "efficiency": {
188
  "unit": "samples/kWh",
189
- "value": 3751632.8893260807
190
  }
191
  }
192
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 902.28736,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 148,
14
+ "total": 0.9989087686538696,
15
+ "mean": 0.006749383571985605,
16
+ "stdev": 0.0006199000503150585,
17
+ "p50": 0.0066923520565032955,
18
+ "p90": 0.006864076614379882,
19
+ "p95": 0.008348518562316899,
20
+ "p99": 0.009350686492919922,
21
  "values": [
22
+ 0.009878527641296387,
23
+ 0.009399295806884766,
24
+ 0.009274368286132812,
25
+ 0.00929587173461914,
26
+ 0.009267200469970703,
27
+ 0.008657983779907227,
28
+ 0.00859545612335205,
29
+ 0.008596480369567871,
30
+ 0.006862847805023193,
31
+ 0.006762495994567871,
32
+ 0.006724607944488525,
33
+ 0.006703104019165039,
34
+ 0.006758399963378906,
35
+ 0.00673689603805542,
36
+ 0.006625216007232666,
37
+ 0.006660096168518067,
38
+ 0.006668288230895996,
39
+ 0.006676479816436768,
40
+ 0.00679423999786377,
41
+ 0.006791168212890625,
42
+ 0.006763519763946534,
43
+ 0.0068351998329162595,
44
+ 0.006744063854217529,
45
+ 0.006715392112731934,
46
+ 0.0067338237762451176,
47
+ 0.006805471897125244,
48
+ 0.006761472225189209,
49
+ 0.006737919807434082,
50
+ 0.0067645440101623535,
51
+ 0.006758399963378906,
52
+ 0.00670201587677002,
53
+ 0.006749184131622315,
54
+ 0.006762495994567871,
55
+ 0.00671123218536377,
56
+ 0.00669593620300293,
57
+ 0.006647808074951172,
58
+ 0.006680575847625733,
59
+ 0.006668288230895996,
60
+ 0.0067123198509216305,
61
+ 0.006713344097137451,
62
+ 0.006668288230895996,
63
+ 0.006713344097137451,
64
+ 0.006740992069244385,
65
+ 0.006738944053649902,
66
+ 0.00677177619934082,
67
+ 0.00675328016281128,
68
+ 0.006754303932189941,
69
+ 0.006789120197296142,
70
+ 0.006726655960083008,
71
+ 0.006652927875518798,
72
+ 0.006708288192749023,
73
+ 0.00679423999786377,
74
+ 0.006740992069244385,
75
+ 0.006699007987976074,
76
+ 0.006744063854217529,
77
+ 0.006663199901580811,
78
+ 0.006631423950195312,
79
+ 0.006711296081542969,
80
+ 0.006662112236022949,
81
+ 0.0066979842185974124,
82
+ 0.0068280320167541505,
83
+ 0.006656000137329102,
84
+ 0.006585311889648437,
85
+ 0.006691840171813965,
86
+ 0.006692863941192627,
87
+ 0.006721536159515381,
88
+ 0.0067041277885437015,
89
+ 0.006703104019165039,
90
+ 0.0066938881874084475,
91
+ 0.006762495994567871,
92
+ 0.006841343879699707,
93
+ 0.006866943836212158,
94
+ 0.00679423999786377,
95
+ 0.00682700777053833,
96
+ 0.006744063854217529,
97
+ 0.006730751991271973,
98
+ 0.006732800006866455,
99
+ 0.006663167953491211,
100
+ 0.006756351947784424,
101
+ 0.006658048152923584,
102
+ 0.006669312000274658,
103
+ 0.0064880638122558594,
104
+ 0.006461440086364746,
105
+ 0.0066979842185974124,
106
+ 0.006568992137908935,
107
+ 0.00638156795501709,
108
+ 0.0065382399559021,
109
+ 0.006713344097137451,
110
+ 0.006862847805023193,
111
+ 0.006855679988861084,
112
+ 0.0068884482383728025,
113
+ 0.006947840213775635,
114
+ 0.006947807788848877,
115
+ 0.006965248107910156,
116
+ 0.007651328086853027,
117
+ 0.007889920234680176,
118
+ 0.006717440128326416,
119
+ 0.0066416640281677245,
120
+ 0.0066447358131408694,
121
+ 0.006627327919006347,
122
+ 0.006652927875518798,
123
+ 0.006632415771484375,
124
+ 0.006512639999389648,
125
+ 0.006392831802368164,
126
+ 0.0063539199829101565,
127
+ 0.006343679904937744,
128
  0.006364160060882569,
129
+ 0.006333439826965332,
130
+ 0.0063385281562805176,
131
+ 0.006333439826965332,
132
+ 0.006347775936126709,
133
+ 0.006346752166748047,
134
+ 0.006341631889343262,
135
+ 0.006379519939422608,
136
+ 0.006347775936126709,
137
+ 0.006347775936126709,
138
+ 0.0063610877990722655,
139
  0.006323200225830078,
140
+ 0.006354944229125976,
141
+ 0.006309887886047363,
142
+ 0.0063610877990722655,
143
  0.006351871967315674,
144
+ 0.006313983917236328,
145
+ 0.0063784317970275875,
146
+ 0.006330431938171387,
147
+ 0.006371327877044678,
148
+ 0.006351871967315674,
149
+ 0.006340640068054199,
150
+ 0.006354944229125976,
151
+ 0.006308864116668702,
152
+ 0.006354944229125976,
153
+ 0.006341631889343262,
154
+ 0.006307839870452881,
155
+ 0.006351871967315674,
156
+ 0.0063170561790466305,
157
+ 0.006358016014099121,
158
+ 0.006341631889343262,
159
+ 0.006329343795776367,
160
+ 0.006341631889343262,
161
+ 0.006328320026397705,
162
+ 0.006365119934082031,
163
+ 0.006340608119964599,
164
+ 0.006326272010803223,
165
+ 0.006343679904937744,
166
+ 0.0063201279640197755,
167
+ 0.006352896213531494,
168
+ 0.006354944229125976,
169
+ 0.0063836159706115725
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
170
  ]
171
  },
172
  "throughput": {
173
  "unit": "samples/s",
174
+ "value": 148.1616786680579
175
  },
176
  "energy": {
177
  "unit": "kWh",
178
+ "cpu": 7.798429687469327e-08,
179
+ "ram": 4.263400563247973e-08,
180
+ "gpu": 1.3821648484210711e-07,
181
+ "total": 2.588347873492801e-07
182
  },
183
  "efficiency": {
184
  "unit": "samples/kWh",
185
+ "value": 3863468.3159901816
186
  }
187
  }
188
  }