Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark_report.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"forward": {
|
3 |
"memory": {
|
4 |
"unit": "MB",
|
5 |
-
"max_ram":
|
6 |
"max_global_vram": 1434.976256,
|
7 |
"max_process_vram": 0.0,
|
8 |
"max_reserved": 794.820608,
|
@@ -11,100 +11,100 @@
|
|
11 |
"latency": {
|
12 |
"unit": "s",
|
13 |
"count": 70,
|
14 |
-
"total": 1.
|
15 |
-
"mean": 0.
|
16 |
-
"stdev": 0.
|
17 |
-
"p50": 0.
|
18 |
-
"p90": 0.
|
19 |
-
"p95": 0.
|
20 |
-
"p99": 0.
|
21 |
"values": [
|
22 |
-
0.
|
23 |
-
0.
|
24 |
-
0.
|
25 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
0.014220288276672363,
|
27 |
-
0.
|
28 |
-
0.014218239784240723,
|
29 |
-
0.014090239524841308,
|
30 |
-
0.014104576110839843,
|
31 |
-
0.014255104064941406,
|
32 |
-
0.014347264289855957,
|
33 |
-
0.014137344360351562,
|
34 |
-
0.014187520027160644,
|
35 |
-
0.014226431846618653,
|
36 |
-
0.014338047981262207,
|
37 |
-
0.014148608207702636,
|
38 |
-
0.014213120460510254,
|
39 |
-
0.014131199836730958,
|
40 |
-
0.014117888450622559,
|
41 |
-
0.014134271621704102,
|
42 |
-
0.01409331226348877,
|
43 |
-
0.01438003158569336,
|
44 |
-
0.01448140811920166,
|
45 |
-
0.014326784133911133,
|
46 |
-
0.014186464309692382,
|
47 |
-
0.01436569595336914,
|
48 |
-
0.0142673921585083,
|
49 |
-
0.01415987205505371,
|
50 |
-
0.014188544273376465,
|
51 |
-
0.014384127616882325,
|
52 |
-
0.01406873607635498,
|
53 |
-
0.0140697603225708,
|
54 |
-
0.01407795238494873,
|
55 |
-
0.014150624275207519,
|
56 |
-
0.014461952209472656,
|
57 |
-
0.01448960018157959,
|
58 |
-
0.014296064376831055,
|
59 |
-
0.014125056266784668,
|
60 |
-
0.014151647567749023,
|
61 |
-
0.014020607948303223,
|
62 |
-
0.014054400444030762,
|
63 |
-
0.014038016319274902,
|
64 |
-
0.013894656181335448,
|
65 |
-
0.014030847549438476,
|
66 |
-
0.014228480339050293,
|
67 |
-
0.014315520286560059,
|
68 |
-
0.015007743835449219,
|
69 |
-
0.014940159797668457,
|
70 |
-
0.014850048065185547,
|
71 |
-
0.014645248413085938,
|
72 |
-
0.014537728309631348,
|
73 |
-
0.014452735900878906,
|
74 |
-
0.014501888275146485,
|
75 |
-
0.014515232086181641,
|
76 |
-
0.014433279991149902,
|
77 |
-
0.014564352035522461,
|
78 |
-
0.014943231582641601,
|
79 |
-
0.014523391723632812,
|
80 |
-
0.014533632278442383,
|
81 |
-
0.014446592330932618,
|
82 |
-
0.01448857593536377,
|
83 |
-
0.014561280250549317,
|
84 |
-
0.014525440216064453,
|
85 |
-
0.014503935813903808,
|
86 |
-
0.014445568084716797,
|
87 |
-
0.014445568084716797,
|
88 |
-
0.014988287925720215,
|
89 |
-
0.014632960319519044,
|
90 |
-
0.014562303543090821,
|
91 |
-
0.01449779224395752
|
92 |
]
|
93 |
},
|
94 |
"throughput": {
|
95 |
"unit": "samples/s",
|
96 |
-
"value": 69.
|
97 |
},
|
98 |
"energy": {
|
99 |
"unit": "kWh",
|
100 |
-
"cpu": 1.
|
101 |
-
"ram":
|
102 |
-
"gpu": 3.
|
103 |
-
"total":
|
104 |
},
|
105 |
"efficiency": {
|
106 |
"unit": "samples/kWh",
|
107 |
-
"value":
|
108 |
}
|
109 |
}
|
110 |
}
|
|
|
2 |
"forward": {
|
3 |
"memory": {
|
4 |
"unit": "MB",
|
5 |
+
"max_ram": 975.458304,
|
6 |
"max_global_vram": 1434.976256,
|
7 |
"max_process_vram": 0.0,
|
8 |
"max_reserved": 794.820608,
|
|
|
11 |
"latency": {
|
12 |
"unit": "s",
|
13 |
"count": 70,
|
14 |
+
"total": 1.007840227127075,
|
15 |
+
"mean": 0.01439771753038679,
|
16 |
+
"stdev": 0.00107056545905484,
|
17 |
+
"p50": 0.014158847808837891,
|
18 |
+
"p90": 0.014704332733154296,
|
19 |
+
"p95": 0.016207155609130858,
|
20 |
+
"p99": 0.01903002662658692,
|
21 |
"values": [
|
22 |
+
0.020493312835693358,
|
23 |
+
0.01741721534729004,
|
24 |
+
0.015558655738830567,
|
25 |
+
0.014436351776123046,
|
26 |
+
0.013902848243713378,
|
27 |
+
0.013955072402954101,
|
28 |
+
0.013876223564147949,
|
29 |
+
0.014328831672668458,
|
30 |
+
0.0140830717086792,
|
31 |
+
0.014141440391540527,
|
32 |
+
0.013847552299499511,
|
33 |
+
0.013931520462036133,
|
34 |
+
0.01419878387451172,
|
35 |
+
0.014423040390014649,
|
36 |
+
0.013848575592041015,
|
37 |
+
0.013941760063171387,
|
38 |
+
0.013741056442260742,
|
39 |
+
0.013829119682312012,
|
40 |
+
0.013815808296203613,
|
41 |
+
0.013864959716796875,
|
42 |
+
0.013930496215820312,
|
43 |
+
0.014217215538024902,
|
44 |
+
0.013934592247009277,
|
45 |
+
0.013890560150146485,
|
46 |
+
0.014021632194519042,
|
47 |
+
0.016486400604248046,
|
48 |
+
0.018372608184814454,
|
49 |
+
0.01467903995513916,
|
50 |
+
0.015865856170654297,
|
51 |
+
0.014231552124023437,
|
52 |
+
0.013827072143554688,
|
53 |
+
0.013798399925231934,
|
54 |
+
0.013827072143554688,
|
55 |
+
0.014258111953735352,
|
56 |
+
0.014212096214294433,
|
57 |
+
0.013934592247009277,
|
58 |
+
0.013874176025390626,
|
59 |
+
0.013867008209228515,
|
60 |
+
0.01375436782836914,
|
61 |
+
0.013774847984313965,
|
62 |
+
0.013730815887451172,
|
63 |
+
0.013551615715026855,
|
64 |
+
0.013505536079406738,
|
65 |
+
0.014270463943481445,
|
66 |
+
0.014343199729919433,
|
67 |
+
0.014414848327636719,
|
68 |
+
0.014224384307861328,
|
69 |
+
0.014664704322814942,
|
70 |
+
0.014590975761413574,
|
71 |
+
0.014334976196289062,
|
72 |
+
0.014664704322814942,
|
73 |
+
0.014379008293151856,
|
74 |
+
0.014199808120727539,
|
75 |
+
0.014310400009155273,
|
76 |
+
0.01426636791229248,
|
77 |
+
0.014155776023864745,
|
78 |
+
0.014589952468872071,
|
79 |
+
0.014139391899108887,
|
80 |
+
0.014286848068237304,
|
81 |
+
0.014116864204406738,
|
82 |
+
0.014089216232299804,
|
83 |
+
0.01425715160369873,
|
84 |
+
0.014542847633361817,
|
85 |
+
0.014931967735290527,
|
86 |
+
0.014094335556030273,
|
87 |
+
0.014294015884399413,
|
88 |
+
0.014161919593811035,
|
89 |
+
0.014045184135437011,
|
90 |
0.014220288276672363,
|
91 |
+
0.0140697603225708
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
92 |
]
|
93 |
},
|
94 |
"throughput": {
|
95 |
"unit": "samples/s",
|
96 |
+
"value": 69.45545346958446
|
97 |
},
|
98 |
"energy": {
|
99 |
"unit": "kWh",
|
100 |
+
"cpu": 1.651157345764532e-07,
|
101 |
+
"ram": 8.989148911296816e-08,
|
102 |
+
"gpu": 3.351144233424662e-07,
|
103 |
+
"total": 5.901216470318875e-07
|
104 |
},
|
105 |
"efficiency": {
|
106 |
"unit": "samples/kWh",
|
107 |
+
"value": 1694565.866257681
|
108 |
}
|
109 |
}
|
110 |
}
|