{ "best_metric": NaN, "best_model_checkpoint": "miner_id_24/checkpoint-50", "epoch": 0.38113387327298714, "eval_steps": 50, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0019056693663649356, "grad_norm": 1.625303864479065, "learning_rate": 1e-05, "loss": 2.0806, "step": 1 }, { "epoch": 0.0019056693663649356, "eval_loss": NaN, "eval_runtime": 121.9902, "eval_samples_per_second": 7.246, "eval_steps_per_second": 1.812, "step": 1 }, { "epoch": 0.003811338732729871, "grad_norm": 1.8919113874435425, "learning_rate": 2e-05, "loss": 2.4173, "step": 2 }, { "epoch": 0.005717008099094807, "grad_norm": 2.010105848312378, "learning_rate": 3e-05, "loss": 2.6527, "step": 3 }, { "epoch": 0.007622677465459742, "grad_norm": 1.7687952518463135, "learning_rate": 4e-05, "loss": 2.6151, "step": 4 }, { "epoch": 0.009528346831824679, "grad_norm": 1.6465333700180054, "learning_rate": 5e-05, "loss": 2.6334, "step": 5 }, { "epoch": 0.011434016198189614, "grad_norm": 1.404477834701538, "learning_rate": 6e-05, "loss": 2.547, "step": 6 }, { "epoch": 0.01333968556455455, "grad_norm": 1.258142352104187, "learning_rate": 7e-05, "loss": 2.5915, "step": 7 }, { "epoch": 0.015245354930919485, "grad_norm": 1.5234754085540771, "learning_rate": 8e-05, "loss": 2.5803, "step": 8 }, { "epoch": 0.01715102429728442, "grad_norm": 1.6583815813064575, "learning_rate": 9e-05, "loss": 2.5519, "step": 9 }, { "epoch": 0.019056693663649357, "grad_norm": 1.2219481468200684, "learning_rate": 0.0001, "loss": 2.3874, "step": 10 }, { "epoch": 0.02096236303001429, "grad_norm": 1.083174467086792, "learning_rate": 9.999316524962345e-05, "loss": 2.4246, "step": 11 }, { "epoch": 0.022868032396379228, "grad_norm": 1.1686445474624634, "learning_rate": 9.997266286704631e-05, "loss": 2.5171, "step": 12 }, { "epoch": 0.024773701762744165, "grad_norm": 1.1928153038024902, "learning_rate": 9.993849845741524e-05, "loss": 2.3594, "step": 13 }, { "epoch": 0.0266793711291091, "grad_norm": 1.1012721061706543, "learning_rate": 9.989068136093873e-05, "loss": 2.2893, "step": 14 }, { "epoch": 0.028585040495474036, "grad_norm": 1.0885003805160522, "learning_rate": 9.98292246503335e-05, "loss": 2.4461, "step": 15 }, { "epoch": 0.03049070986183897, "grad_norm": 1.0348674058914185, "learning_rate": 9.975414512725057e-05, "loss": 2.2247, "step": 16 }, { "epoch": 0.03239637922820391, "grad_norm": 1.0466992855072021, "learning_rate": 9.966546331768191e-05, "loss": 2.3983, "step": 17 }, { "epoch": 0.03430204859456884, "grad_norm": 1.0777167081832886, "learning_rate": 9.956320346634876e-05, "loss": 2.3311, "step": 18 }, { "epoch": 0.03620771796093378, "grad_norm": 1.0364409685134888, "learning_rate": 9.944739353007344e-05, "loss": 2.3972, "step": 19 }, { "epoch": 0.038113387327298714, "grad_norm": 1.2143179178237915, "learning_rate": 9.931806517013612e-05, "loss": 2.3061, "step": 20 }, { "epoch": 0.04001905669366365, "grad_norm": 1.1169112920761108, "learning_rate": 9.917525374361912e-05, "loss": 2.2864, "step": 21 }, { "epoch": 0.04192472606002858, "grad_norm": 2.8071861267089844, "learning_rate": 9.901899829374047e-05, "loss": 2.4627, "step": 22 }, { "epoch": 0.04383039542639352, "grad_norm": 1.6181660890579224, "learning_rate": 9.884934153917997e-05, "loss": 2.4329, "step": 23 }, { "epoch": 0.045736064792758456, "grad_norm": 7.056554317474365, "learning_rate": 9.86663298624003e-05, "loss": 2.3044, "step": 24 }, { "epoch": 0.04764173415912339, "grad_norm": 1.5961129665374756, "learning_rate": 9.847001329696653e-05, "loss": 2.311, "step": 25 }, { "epoch": 0.04954740352548833, "grad_norm": 1.412833571434021, "learning_rate": 9.826044551386744e-05, "loss": 2.3759, "step": 26 }, { "epoch": 0.05145307289185326, "grad_norm": 1.2436467409133911, "learning_rate": 9.803768380684242e-05, "loss": 2.2569, "step": 27 }, { "epoch": 0.0533587422582182, "grad_norm": 1.2787665128707886, "learning_rate": 9.780178907671789e-05, "loss": 2.415, "step": 28 }, { "epoch": 0.055264411624583135, "grad_norm": 1.1875362396240234, "learning_rate": 9.755282581475769e-05, "loss": 2.0744, "step": 29 }, { "epoch": 0.05717008099094807, "grad_norm": 1.468946933746338, "learning_rate": 9.729086208503174e-05, "loss": 2.3473, "step": 30 }, { "epoch": 0.05907575035731301, "grad_norm": 1.3427952527999878, "learning_rate": 9.701596950580806e-05, "loss": 2.2288, "step": 31 }, { "epoch": 0.06098141972367794, "grad_norm": 1.6122334003448486, "learning_rate": 9.672822322997305e-05, "loss": 2.427, "step": 32 }, { "epoch": 0.06288708909004288, "grad_norm": 2.7167866230010986, "learning_rate": 9.642770192448536e-05, "loss": 2.4146, "step": 33 }, { "epoch": 0.06479275845640782, "grad_norm": 1.385132074356079, "learning_rate": 9.611448774886924e-05, "loss": 2.1858, "step": 34 }, { "epoch": 0.06669842782277274, "grad_norm": 2.5712013244628906, "learning_rate": 9.578866633275288e-05, "loss": 2.3198, "step": 35 }, { "epoch": 0.06860409718913768, "grad_norm": 2.7239487171173096, "learning_rate": 9.545032675245813e-05, "loss": 2.302, "step": 36 }, { "epoch": 0.07050976655550262, "grad_norm": 1.5528560876846313, "learning_rate": 9.509956150664796e-05, "loss": 2.1923, "step": 37 }, { "epoch": 0.07241543592186755, "grad_norm": 1.9108004570007324, "learning_rate": 9.473646649103818e-05, "loss": 2.4346, "step": 38 }, { "epoch": 0.07432110528823249, "grad_norm": 1.7357029914855957, "learning_rate": 9.43611409721806e-05, "loss": 2.3665, "step": 39 }, { "epoch": 0.07622677465459743, "grad_norm": 1.6708160638809204, "learning_rate": 9.397368756032445e-05, "loss": 2.3946, "step": 40 }, { "epoch": 0.07813244402096237, "grad_norm": 1.7986937761306763, "learning_rate": 9.357421218136386e-05, "loss": 2.5051, "step": 41 }, { "epoch": 0.0800381133873273, "grad_norm": 1.7630176544189453, "learning_rate": 9.316282404787871e-05, "loss": 2.2678, "step": 42 }, { "epoch": 0.08194378275369224, "grad_norm": 1.8785440921783447, "learning_rate": 9.273963562927695e-05, "loss": 2.1948, "step": 43 }, { "epoch": 0.08384945212005716, "grad_norm": 2.319279193878174, "learning_rate": 9.230476262104677e-05, "loss": 2.3691, "step": 44 }, { "epoch": 0.0857551214864221, "grad_norm": 4.61505651473999, "learning_rate": 9.185832391312644e-05, "loss": 3.0313, "step": 45 }, { "epoch": 0.08766079085278704, "grad_norm": 0.0, "learning_rate": 9.140044155740101e-05, "loss": 0.0, "step": 46 }, { "epoch": 0.08956646021915197, "grad_norm": 0.0, "learning_rate": 9.093124073433463e-05, "loss": 0.0, "step": 47 }, { "epoch": 0.09147212958551691, "grad_norm": 0.0, "learning_rate": 9.045084971874738e-05, "loss": 0.0, "step": 48 }, { "epoch": 0.09337779895188185, "grad_norm": 0.0, "learning_rate": 8.995939984474624e-05, "loss": 0.0, "step": 49 }, { "epoch": 0.09528346831824679, "grad_norm": 3.32122540473938, "learning_rate": 8.945702546981969e-05, "loss": 2.6096, "step": 50 }, { "epoch": 0.09528346831824679, "eval_loss": NaN, "eval_runtime": 124.017, "eval_samples_per_second": 7.128, "eval_steps_per_second": 1.782, "step": 50 }, { "epoch": 0.09718913768461172, "grad_norm": 3.257054090499878, "learning_rate": 8.894386393810563e-05, "loss": 2.475, "step": 51 }, { "epoch": 0.09909480705097666, "grad_norm": 1.825034737586975, "learning_rate": 8.842005554284296e-05, "loss": 2.1688, "step": 52 }, { "epoch": 0.1010004764173416, "grad_norm": 1.2291990518569946, "learning_rate": 8.788574348801675e-05, "loss": 2.2075, "step": 53 }, { "epoch": 0.10290614578370652, "grad_norm": 0.8810988664627075, "learning_rate": 8.73410738492077e-05, "loss": 2.0334, "step": 54 }, { "epoch": 0.10481181515007146, "grad_norm": 0.8452022671699524, "learning_rate": 8.678619553365659e-05, "loss": 2.2784, "step": 55 }, { "epoch": 0.1067174845164364, "grad_norm": 0.8677942752838135, "learning_rate": 8.622126023955446e-05, "loss": 2.3311, "step": 56 }, { "epoch": 0.10862315388280133, "grad_norm": 0.8011043071746826, "learning_rate": 8.564642241456986e-05, "loss": 2.1818, "step": 57 }, { "epoch": 0.11052882324916627, "grad_norm": 0.8731439709663391, "learning_rate": 8.506183921362443e-05, "loss": 2.2584, "step": 58 }, { "epoch": 0.1124344926155312, "grad_norm": 0.903266966342926, "learning_rate": 8.44676704559283e-05, "loss": 2.1772, "step": 59 }, { "epoch": 0.11434016198189614, "grad_norm": 0.8754334449768066, "learning_rate": 8.386407858128706e-05, "loss": 2.2054, "step": 60 }, { "epoch": 0.11624583134826108, "grad_norm": 1.0421199798583984, "learning_rate": 8.32512286056924e-05, "loss": 2.1207, "step": 61 }, { "epoch": 0.11815150071462602, "grad_norm": 0.8233661651611328, "learning_rate": 8.262928807620843e-05, "loss": 2.0804, "step": 62 }, { "epoch": 0.12005717008099095, "grad_norm": 0.9691787362098694, "learning_rate": 8.199842702516583e-05, "loss": 2.264, "step": 63 }, { "epoch": 0.12196283944735588, "grad_norm": 0.9734489321708679, "learning_rate": 8.135881792367686e-05, "loss": 2.2164, "step": 64 }, { "epoch": 0.12386850881372082, "grad_norm": 1.0550222396850586, "learning_rate": 8.07106356344834e-05, "loss": 2.3509, "step": 65 }, { "epoch": 0.12577417818008577, "grad_norm": 0.9587940573692322, "learning_rate": 8.005405736415126e-05, "loss": 2.2228, "step": 66 }, { "epoch": 0.1276798475464507, "grad_norm": 0.966098427772522, "learning_rate": 7.938926261462366e-05, "loss": 2.2955, "step": 67 }, { "epoch": 0.12958551691281564, "grad_norm": 0.9006718993186951, "learning_rate": 7.871643313414718e-05, "loss": 2.2223, "step": 68 }, { "epoch": 0.13149118627918055, "grad_norm": 0.9012450575828552, "learning_rate": 7.803575286758364e-05, "loss": 2.2281, "step": 69 }, { "epoch": 0.1333968556455455, "grad_norm": 0.9210295677185059, "learning_rate": 7.734740790612136e-05, "loss": 2.1927, "step": 70 }, { "epoch": 0.13530252501191042, "grad_norm": 0.9985268712043762, "learning_rate": 7.66515864363997e-05, "loss": 2.1856, "step": 71 }, { "epoch": 0.13720819437827536, "grad_norm": 1.0568547248840332, "learning_rate": 7.594847868906076e-05, "loss": 2.2347, "step": 72 }, { "epoch": 0.1391138637446403, "grad_norm": 1.3930721282958984, "learning_rate": 7.52382768867422e-05, "loss": 2.2211, "step": 73 }, { "epoch": 0.14101953311100524, "grad_norm": 1.1122373342514038, "learning_rate": 7.452117519152542e-05, "loss": 2.2579, "step": 74 }, { "epoch": 0.14292520247737017, "grad_norm": 1.0553274154663086, "learning_rate": 7.379736965185368e-05, "loss": 2.2314, "step": 75 }, { "epoch": 0.1448308718437351, "grad_norm": 1.0698802471160889, "learning_rate": 7.30670581489344e-05, "loss": 2.1543, "step": 76 }, { "epoch": 0.14673654121010005, "grad_norm": 1.6581573486328125, "learning_rate": 7.233044034264034e-05, "loss": 2.2807, "step": 77 }, { "epoch": 0.14864221057646498, "grad_norm": 1.1972084045410156, "learning_rate": 7.158771761692464e-05, "loss": 2.2719, "step": 78 }, { "epoch": 0.15054787994282992, "grad_norm": 1.1362383365631104, "learning_rate": 7.083909302476453e-05, "loss": 2.1507, "step": 79 }, { "epoch": 0.15245354930919486, "grad_norm": 1.153376579284668, "learning_rate": 7.008477123264848e-05, "loss": 2.2695, "step": 80 }, { "epoch": 0.1543592186755598, "grad_norm": 1.2021276950836182, "learning_rate": 6.932495846462261e-05, "loss": 2.305, "step": 81 }, { "epoch": 0.15626488804192473, "grad_norm": 1.243454098701477, "learning_rate": 6.855986244591104e-05, "loss": 2.1861, "step": 82 }, { "epoch": 0.15817055740828967, "grad_norm": 1.4724884033203125, "learning_rate": 6.778969234612584e-05, "loss": 2.1153, "step": 83 }, { "epoch": 0.1600762267746546, "grad_norm": 1.2578407526016235, "learning_rate": 6.701465872208216e-05, "loss": 2.1895, "step": 84 }, { "epoch": 0.16198189614101954, "grad_norm": 1.3038272857666016, "learning_rate": 6.623497346023418e-05, "loss": 2.1578, "step": 85 }, { "epoch": 0.16388756550738448, "grad_norm": 1.8126806020736694, "learning_rate": 6.545084971874738e-05, "loss": 2.1949, "step": 86 }, { "epoch": 0.16579323487374942, "grad_norm": 1.3623523712158203, "learning_rate": 6.466250186922325e-05, "loss": 2.3411, "step": 87 }, { "epoch": 0.16769890424011433, "grad_norm": 1.370890498161316, "learning_rate": 6.387014543809223e-05, "loss": 2.2452, "step": 88 }, { "epoch": 0.16960457360647926, "grad_norm": 1.3815104961395264, "learning_rate": 6.307399704769099e-05, "loss": 2.1947, "step": 89 }, { "epoch": 0.1715102429728442, "grad_norm": 1.5048907995224, "learning_rate": 6.227427435703997e-05, "loss": 2.3206, "step": 90 }, { "epoch": 0.17341591233920914, "grad_norm": 1.8256431818008423, "learning_rate": 6.147119600233758e-05, "loss": 2.2274, "step": 91 }, { "epoch": 0.17532158170557408, "grad_norm": 1.8468575477600098, "learning_rate": 6.066498153718735e-05, "loss": 2.2922, "step": 92 }, { "epoch": 0.177227251071939, "grad_norm": 1.6818069219589233, "learning_rate": 5.985585137257401e-05, "loss": 2.2491, "step": 93 }, { "epoch": 0.17913292043830395, "grad_norm": 2.0240490436553955, "learning_rate": 5.90440267166055e-05, "loss": 2.575, "step": 94 }, { "epoch": 0.1810385898046689, "grad_norm": 2.463879346847534, "learning_rate": 5.8229729514036705e-05, "loss": 2.186, "step": 95 }, { "epoch": 0.18294425917103382, "grad_norm": 0.0, "learning_rate": 5.74131823855921e-05, "loss": 0.0, "step": 96 }, { "epoch": 0.18484992853739876, "grad_norm": 0.0, "learning_rate": 5.6594608567103456e-05, "loss": 0.0, "step": 97 }, { "epoch": 0.1867555979037637, "grad_norm": 0.0, "learning_rate": 5.577423184847932e-05, "loss": 0.0, "step": 98 }, { "epoch": 0.18866126727012864, "grad_norm": 0.0, "learning_rate": 5.495227651252315e-05, "loss": 0.0, "step": 99 }, { "epoch": 0.19056693663649357, "grad_norm": 7.715859413146973, "learning_rate": 5.4128967273616625e-05, "loss": 2.6799, "step": 100 }, { "epoch": 0.19056693663649357, "eval_loss": NaN, "eval_runtime": 123.8787, "eval_samples_per_second": 7.136, "eval_steps_per_second": 1.784, "step": 100 }, { "epoch": 0.1924726060028585, "grad_norm": 6.374715805053711, "learning_rate": 5.330452921628497e-05, "loss": 2.4025, "step": 101 }, { "epoch": 0.19437827536922345, "grad_norm": 3.087235927581787, "learning_rate": 5.247918773366112e-05, "loss": 2.3568, "step": 102 }, { "epoch": 0.19628394473558838, "grad_norm": 2.604992151260376, "learning_rate": 5.165316846586541e-05, "loss": 2.4599, "step": 103 }, { "epoch": 0.19818961410195332, "grad_norm": 2.30696964263916, "learning_rate": 5.0826697238317935e-05, "loss": 2.135, "step": 104 }, { "epoch": 0.20009528346831826, "grad_norm": 1.6723530292510986, "learning_rate": 5e-05, "loss": 2.1672, "step": 105 }, { "epoch": 0.2020009528346832, "grad_norm": 1.3099277019500732, "learning_rate": 4.917330276168208e-05, "loss": 2.1903, "step": 106 }, { "epoch": 0.2039066222010481, "grad_norm": 0.9453718662261963, "learning_rate": 4.834683153413459e-05, "loss": 2.2202, "step": 107 }, { "epoch": 0.20581229156741304, "grad_norm": 0.7894024848937988, "learning_rate": 4.7520812266338885e-05, "loss": 2.1519, "step": 108 }, { "epoch": 0.20771796093377798, "grad_norm": 0.7461780309677124, "learning_rate": 4.669547078371504e-05, "loss": 2.1794, "step": 109 }, { "epoch": 0.20962363030014292, "grad_norm": 0.8102741837501526, "learning_rate": 4.5871032726383386e-05, "loss": 2.1812, "step": 110 }, { "epoch": 0.21152929966650785, "grad_norm": 0.8100122213363647, "learning_rate": 4.504772348747687e-05, "loss": 2.1625, "step": 111 }, { "epoch": 0.2134349690328728, "grad_norm": 0.8052271604537964, "learning_rate": 4.4225768151520694e-05, "loss": 2.2185, "step": 112 }, { "epoch": 0.21534063839923773, "grad_norm": 0.8258956074714661, "learning_rate": 4.3405391432896555e-05, "loss": 2.228, "step": 113 }, { "epoch": 0.21724630776560266, "grad_norm": 0.8672061562538147, "learning_rate": 4.2586817614407895e-05, "loss": 2.2591, "step": 114 }, { "epoch": 0.2191519771319676, "grad_norm": 0.9598968625068665, "learning_rate": 4.17702704859633e-05, "loss": 2.266, "step": 115 }, { "epoch": 0.22105764649833254, "grad_norm": 0.9424980878829956, "learning_rate": 4.095597328339452e-05, "loss": 2.1901, "step": 116 }, { "epoch": 0.22296331586469748, "grad_norm": 0.8500011563301086, "learning_rate": 4.0144148627425993e-05, "loss": 2.1516, "step": 117 }, { "epoch": 0.2248689852310624, "grad_norm": 0.9570940732955933, "learning_rate": 3.933501846281267e-05, "loss": 2.2747, "step": 118 }, { "epoch": 0.22677465459742735, "grad_norm": 0.9292868971824646, "learning_rate": 3.852880399766243e-05, "loss": 2.1949, "step": 119 }, { "epoch": 0.2286803239637923, "grad_norm": 0.9754165410995483, "learning_rate": 3.772572564296005e-05, "loss": 2.2281, "step": 120 }, { "epoch": 0.23058599333015722, "grad_norm": 1.0084413290023804, "learning_rate": 3.6926002952309016e-05, "loss": 2.4057, "step": 121 }, { "epoch": 0.23249166269652216, "grad_norm": 0.9409120678901672, "learning_rate": 3.612985456190778e-05, "loss": 2.2361, "step": 122 }, { "epoch": 0.2343973320628871, "grad_norm": 1.0411348342895508, "learning_rate": 3.533749813077677e-05, "loss": 2.2445, "step": 123 }, { "epoch": 0.23630300142925204, "grad_norm": 0.989592969417572, "learning_rate": 3.4549150281252636e-05, "loss": 2.3068, "step": 124 }, { "epoch": 0.23820867079561697, "grad_norm": 0.9536654353141785, "learning_rate": 3.3765026539765834e-05, "loss": 2.3733, "step": 125 }, { "epoch": 0.2401143401619819, "grad_norm": 1.0774953365325928, "learning_rate": 3.298534127791785e-05, "loss": 2.1134, "step": 126 }, { "epoch": 0.24202000952834682, "grad_norm": 1.1679447889328003, "learning_rate": 3.221030765387417e-05, "loss": 2.3144, "step": 127 }, { "epoch": 0.24392567889471176, "grad_norm": 1.1911946535110474, "learning_rate": 3.144013755408895e-05, "loss": 2.2425, "step": 128 }, { "epoch": 0.2458313482610767, "grad_norm": 1.1719609498977661, "learning_rate": 3.0675041535377405e-05, "loss": 2.4016, "step": 129 }, { "epoch": 0.24773701762744163, "grad_norm": 1.1523762941360474, "learning_rate": 2.991522876735154e-05, "loss": 2.4412, "step": 130 }, { "epoch": 0.24964268699380657, "grad_norm": 1.119282603263855, "learning_rate": 2.916090697523549e-05, "loss": 2.0928, "step": 131 }, { "epoch": 0.25154835636017153, "grad_norm": 1.4509676694869995, "learning_rate": 2.8412282383075363e-05, "loss": 2.282, "step": 132 }, { "epoch": 0.25345402572653647, "grad_norm": 1.133516550064087, "learning_rate": 2.766955965735968e-05, "loss": 2.2687, "step": 133 }, { "epoch": 0.2553596950929014, "grad_norm": 1.1817891597747803, "learning_rate": 2.693294185106562e-05, "loss": 2.2668, "step": 134 }, { "epoch": 0.25726536445926634, "grad_norm": 1.3289722204208374, "learning_rate": 2.6202630348146324e-05, "loss": 2.1709, "step": 135 }, { "epoch": 0.2591710338256313, "grad_norm": 1.2520653009414673, "learning_rate": 2.547882480847461e-05, "loss": 2.285, "step": 136 }, { "epoch": 0.26107670319199616, "grad_norm": 2.837770938873291, "learning_rate": 2.476172311325783e-05, "loss": 2.3204, "step": 137 }, { "epoch": 0.2629823725583611, "grad_norm": 1.3282185792922974, "learning_rate": 2.405152131093926e-05, "loss": 2.2622, "step": 138 }, { "epoch": 0.26488804192472604, "grad_norm": 1.3785673379898071, "learning_rate": 2.3348413563600325e-05, "loss": 2.2191, "step": 139 }, { "epoch": 0.266793711291091, "grad_norm": 1.4621028900146484, "learning_rate": 2.2652592093878666e-05, "loss": 2.1923, "step": 140 }, { "epoch": 0.2686993806574559, "grad_norm": 1.4957550764083862, "learning_rate": 2.196424713241637e-05, "loss": 2.223, "step": 141 }, { "epoch": 0.27060505002382085, "grad_norm": 1.566874623298645, "learning_rate": 2.128356686585282e-05, "loss": 2.1457, "step": 142 }, { "epoch": 0.2725107193901858, "grad_norm": 1.581588625907898, "learning_rate": 2.061073738537635e-05, "loss": 2.3324, "step": 143 }, { "epoch": 0.2744163887565507, "grad_norm": 1.7475031614303589, "learning_rate": 1.9945942635848748e-05, "loss": 2.396, "step": 144 }, { "epoch": 0.27632205812291566, "grad_norm": 2.1765027046203613, "learning_rate": 1.928936436551661e-05, "loss": 2.3309, "step": 145 }, { "epoch": 0.2782277274892806, "grad_norm": 5.112452983856201, "learning_rate": 1.8641182076323148e-05, "loss": 3.0045, "step": 146 }, { "epoch": 0.28013339685564553, "grad_norm": 0.0, "learning_rate": 1.800157297483417e-05, "loss": 0.0, "step": 147 }, { "epoch": 0.28203906622201047, "grad_norm": 0.0, "learning_rate": 1.7370711923791567e-05, "loss": 0.0, "step": 148 }, { "epoch": 0.2839447355883754, "grad_norm": 0.0, "learning_rate": 1.6748771394307585e-05, "loss": 0.0, "step": 149 }, { "epoch": 0.28585040495474034, "grad_norm": 1.6631611585617065, "learning_rate": 1.6135921418712956e-05, "loss": 2.2213, "step": 150 }, { "epoch": 0.28585040495474034, "eval_loss": NaN, "eval_runtime": 123.886, "eval_samples_per_second": 7.136, "eval_steps_per_second": 1.784, "step": 150 }, { "epoch": 0.2877560743211053, "grad_norm": 1.2977925539016724, "learning_rate": 1.553232954407171e-05, "loss": 1.8845, "step": 151 }, { "epoch": 0.2896617436874702, "grad_norm": 1.4739603996276855, "learning_rate": 1.4938160786375572e-05, "loss": 2.1382, "step": 152 }, { "epoch": 0.29156741305383516, "grad_norm": 1.3540451526641846, "learning_rate": 1.435357758543015e-05, "loss": 2.0628, "step": 153 }, { "epoch": 0.2934730824202001, "grad_norm": 1.192720651626587, "learning_rate": 1.3778739760445552e-05, "loss": 2.1649, "step": 154 }, { "epoch": 0.29537875178656503, "grad_norm": 1.1411631107330322, "learning_rate": 1.3213804466343421e-05, "loss": 2.2098, "step": 155 }, { "epoch": 0.29728442115292997, "grad_norm": 1.1121792793273926, "learning_rate": 1.2658926150792322e-05, "loss": 2.3091, "step": 156 }, { "epoch": 0.2991900905192949, "grad_norm": 0.9577322602272034, "learning_rate": 1.2114256511983274e-05, "loss": 2.1103, "step": 157 }, { "epoch": 0.30109575988565984, "grad_norm": 1.0152143239974976, "learning_rate": 1.157994445715706e-05, "loss": 2.1191, "step": 158 }, { "epoch": 0.3030014292520248, "grad_norm": 0.9139170050621033, "learning_rate": 1.1056136061894384e-05, "loss": 2.2654, "step": 159 }, { "epoch": 0.3049070986183897, "grad_norm": 0.8486616611480713, "learning_rate": 1.0542974530180327e-05, "loss": 2.1172, "step": 160 }, { "epoch": 0.30681276798475465, "grad_norm": 0.8078973293304443, "learning_rate": 1.0040600155253765e-05, "loss": 2.1286, "step": 161 }, { "epoch": 0.3087184373511196, "grad_norm": 0.8643123507499695, "learning_rate": 9.549150281252633e-06, "loss": 2.2474, "step": 162 }, { "epoch": 0.3106241067174845, "grad_norm": 0.8092776536941528, "learning_rate": 9.068759265665384e-06, "loss": 2.1999, "step": 163 }, { "epoch": 0.31252977608384946, "grad_norm": 0.8323934674263, "learning_rate": 8.599558442598998e-06, "loss": 2.1502, "step": 164 }, { "epoch": 0.3144354454502144, "grad_norm": 0.8425267934799194, "learning_rate": 8.141676086873572e-06, "loss": 2.2538, "step": 165 }, { "epoch": 0.31634111481657934, "grad_norm": 0.8372315168380737, "learning_rate": 7.695237378953223e-06, "loss": 2.0642, "step": 166 }, { "epoch": 0.3182467841829443, "grad_norm": 0.8230723738670349, "learning_rate": 7.260364370723044e-06, "loss": 2.0993, "step": 167 }, { "epoch": 0.3201524535493092, "grad_norm": 0.8597812056541443, "learning_rate": 6.837175952121306e-06, "loss": 2.1144, "step": 168 }, { "epoch": 0.32205812291567415, "grad_norm": 0.8669524192810059, "learning_rate": 6.425787818636131e-06, "loss": 2.2526, "step": 169 }, { "epoch": 0.3239637922820391, "grad_norm": 0.8641037344932556, "learning_rate": 6.026312439675552e-06, "loss": 2.1673, "step": 170 }, { "epoch": 0.325869461648404, "grad_norm": 0.8340052962303162, "learning_rate": 5.6388590278194096e-06, "loss": 2.2321, "step": 171 }, { "epoch": 0.32777513101476896, "grad_norm": 0.894442081451416, "learning_rate": 5.263533508961827e-06, "loss": 2.0982, "step": 172 }, { "epoch": 0.3296808003811339, "grad_norm": 0.9692354798316956, "learning_rate": 4.900438493352055e-06, "loss": 2.3198, "step": 173 }, { "epoch": 0.33158646974749884, "grad_norm": 0.9209113717079163, "learning_rate": 4.549673247541875e-06, "loss": 2.0399, "step": 174 }, { "epoch": 0.3334921391138637, "grad_norm": 0.9890367984771729, "learning_rate": 4.2113336672471245e-06, "loss": 2.2726, "step": 175 }, { "epoch": 0.33539780848022865, "grad_norm": 1.0139867067337036, "learning_rate": 3.885512251130763e-06, "loss": 2.2643, "step": 176 }, { "epoch": 0.3373034778465936, "grad_norm": 1.0811392068862915, "learning_rate": 3.5722980755146517e-06, "loss": 2.2314, "step": 177 }, { "epoch": 0.33920914721295853, "grad_norm": 0.9895434975624084, "learning_rate": 3.271776770026963e-06, "loss": 2.1285, "step": 178 }, { "epoch": 0.34111481657932347, "grad_norm": 1.0205448865890503, "learning_rate": 2.9840304941919415e-06, "loss": 2.2578, "step": 179 }, { "epoch": 0.3430204859456884, "grad_norm": 1.1815122365951538, "learning_rate": 2.7091379149682685e-06, "loss": 2.2026, "step": 180 }, { "epoch": 0.34492615531205334, "grad_norm": 1.0534892082214355, "learning_rate": 2.4471741852423237e-06, "loss": 2.0888, "step": 181 }, { "epoch": 0.3468318246784183, "grad_norm": 1.1304489374160767, "learning_rate": 2.1982109232821178e-06, "loss": 2.2371, "step": 182 }, { "epoch": 0.3487374940447832, "grad_norm": 1.1426349878311157, "learning_rate": 1.962316193157593e-06, "loss": 2.1717, "step": 183 }, { "epoch": 0.35064316341114815, "grad_norm": 1.275299072265625, "learning_rate": 1.7395544861325718e-06, "loss": 2.312, "step": 184 }, { "epoch": 0.3525488327775131, "grad_norm": 1.1940147876739502, "learning_rate": 1.5299867030334814e-06, "loss": 2.2908, "step": 185 }, { "epoch": 0.354454502143878, "grad_norm": 2.9150757789611816, "learning_rate": 1.333670137599713e-06, "loss": 2.371, "step": 186 }, { "epoch": 0.35636017151024296, "grad_norm": 1.4302035570144653, "learning_rate": 1.1506584608200367e-06, "loss": 2.1789, "step": 187 }, { "epoch": 0.3582658408766079, "grad_norm": 1.333971381187439, "learning_rate": 9.810017062595322e-07, "loss": 2.0292, "step": 188 }, { "epoch": 0.36017151024297284, "grad_norm": 1.4303479194641113, "learning_rate": 8.247462563808817e-07, "loss": 2.0514, "step": 189 }, { "epoch": 0.3620771796093378, "grad_norm": 1.6177750825881958, "learning_rate": 6.819348298638839e-07, "loss": 2.1123, "step": 190 }, { "epoch": 0.3639828489757027, "grad_norm": 1.672381043434143, "learning_rate": 5.526064699265753e-07, "loss": 2.1767, "step": 191 }, { "epoch": 0.36588851834206765, "grad_norm": 1.703062653541565, "learning_rate": 4.367965336512403e-07, "loss": 2.141, "step": 192 }, { "epoch": 0.3677941877084326, "grad_norm": 1.9463555812835693, "learning_rate": 3.3453668231809286e-07, "loss": 2.5574, "step": 193 }, { "epoch": 0.3696998570747975, "grad_norm": 2.1811716556549072, "learning_rate": 2.458548727494292e-07, "loss": 2.4226, "step": 194 }, { "epoch": 0.37160552644116246, "grad_norm": 2.491175889968872, "learning_rate": 1.7077534966650766e-07, "loss": 2.0546, "step": 195 }, { "epoch": 0.3735111958075274, "grad_norm": 0.0, "learning_rate": 1.0931863906127327e-07, "loss": 0.0, "step": 196 }, { "epoch": 0.37541686517389233, "grad_norm": 0.0, "learning_rate": 6.150154258476315e-08, "loss": 0.0, "step": 197 }, { "epoch": 0.37732253454025727, "grad_norm": 0.0, "learning_rate": 2.7337132953697554e-08, "loss": 0.0, "step": 198 }, { "epoch": 0.3792282039066222, "grad_norm": 3.1555798053741455, "learning_rate": 6.834750376549792e-09, "loss": 1.9376, "step": 199 }, { "epoch": 0.38113387327298714, "grad_norm": 0.9697321057319641, "learning_rate": 0.0, "loss": 2.3541, "step": 200 }, { "epoch": 0.38113387327298714, "eval_loss": NaN, "eval_runtime": 124.2367, "eval_samples_per_second": 7.115, "eval_steps_per_second": 1.779, "step": 200 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 3 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.717266463446794e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }