{ "best_metric": 1.6925917863845825, "best_model_checkpoint": "miner_id_24/checkpoint-150", "epoch": 1.9138755980861244, "eval_steps": 50, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009569377990430622, "grad_norm": 0.38082700967788696, "learning_rate": 8.000000000000001e-06, "loss": 1.6292, "step": 1 }, { "epoch": 0.009569377990430622, "eval_loss": 1.831426739692688, "eval_runtime": 5.8102, "eval_samples_per_second": 30.464, "eval_steps_per_second": 7.745, "step": 1 }, { "epoch": 0.019138755980861243, "grad_norm": 0.3949052691459656, "learning_rate": 1.6000000000000003e-05, "loss": 1.5218, "step": 2 }, { "epoch": 0.028708133971291867, "grad_norm": 0.4035767614841461, "learning_rate": 2.4e-05, "loss": 1.5025, "step": 3 }, { "epoch": 0.03827751196172249, "grad_norm": 0.41593310236930847, "learning_rate": 3.2000000000000005e-05, "loss": 1.5595, "step": 4 }, { "epoch": 0.04784688995215311, "grad_norm": 0.40817272663116455, "learning_rate": 4e-05, "loss": 1.4547, "step": 5 }, { "epoch": 0.05741626794258373, "grad_norm": 0.42838597297668457, "learning_rate": 4.8e-05, "loss": 1.6265, "step": 6 }, { "epoch": 0.06698564593301436, "grad_norm": 0.45393508672714233, "learning_rate": 5.6e-05, "loss": 1.6049, "step": 7 }, { "epoch": 0.07655502392344497, "grad_norm": 0.43397629261016846, "learning_rate": 6.400000000000001e-05, "loss": 1.5147, "step": 8 }, { "epoch": 0.0861244019138756, "grad_norm": 0.45276832580566406, "learning_rate": 7.2e-05, "loss": 1.5187, "step": 9 }, { "epoch": 0.09569377990430622, "grad_norm": 0.4829266369342804, "learning_rate": 8e-05, "loss": 1.5023, "step": 10 }, { "epoch": 0.10526315789473684, "grad_norm": 0.5055172443389893, "learning_rate": 7.999453219969877e-05, "loss": 1.5217, "step": 11 }, { "epoch": 0.11483253588516747, "grad_norm": 0.5195921659469604, "learning_rate": 7.997813029363704e-05, "loss": 1.4627, "step": 12 }, { "epoch": 0.12440191387559808, "grad_norm": 0.5627723932266235, "learning_rate": 7.99507987659322e-05, "loss": 1.5943, "step": 13 }, { "epoch": 0.1339712918660287, "grad_norm": 0.561619758605957, "learning_rate": 7.991254508875098e-05, "loss": 1.633, "step": 14 }, { "epoch": 0.14354066985645933, "grad_norm": 0.5991867780685425, "learning_rate": 7.98633797202668e-05, "loss": 1.7317, "step": 15 }, { "epoch": 0.15311004784688995, "grad_norm": 0.6589527130126953, "learning_rate": 7.980331610180046e-05, "loss": 1.7673, "step": 16 }, { "epoch": 0.16267942583732056, "grad_norm": 0.6777735948562622, "learning_rate": 7.973237065414553e-05, "loss": 1.7703, "step": 17 }, { "epoch": 0.1722488038277512, "grad_norm": 0.7253565192222595, "learning_rate": 7.965056277307902e-05, "loss": 1.887, "step": 18 }, { "epoch": 0.18181818181818182, "grad_norm": 0.7007018327713013, "learning_rate": 7.955791482405875e-05, "loss": 1.8466, "step": 19 }, { "epoch": 0.19138755980861244, "grad_norm": 0.7932828068733215, "learning_rate": 7.94544521361089e-05, "loss": 1.9501, "step": 20 }, { "epoch": 0.20095693779904306, "grad_norm": 0.8723483085632324, "learning_rate": 7.93402029948953e-05, "loss": 1.8275, "step": 21 }, { "epoch": 0.21052631578947367, "grad_norm": 0.9875766038894653, "learning_rate": 7.921519863499239e-05, "loss": 1.9756, "step": 22 }, { "epoch": 0.22009569377990432, "grad_norm": 1.1991780996322632, "learning_rate": 7.907947323134398e-05, "loss": 2.0454, "step": 23 }, { "epoch": 0.22966507177033493, "grad_norm": 1.3916829824447632, "learning_rate": 7.893306388992023e-05, "loss": 2.1839, "step": 24 }, { "epoch": 0.23923444976076555, "grad_norm": 2.0087339878082275, "learning_rate": 7.877601063757323e-05, "loss": 1.8768, "step": 25 }, { "epoch": 0.24880382775119617, "grad_norm": 3.617227554321289, "learning_rate": 7.860835641109395e-05, "loss": 2.2618, "step": 26 }, { "epoch": 0.2583732057416268, "grad_norm": 0.3412221074104309, "learning_rate": 7.843014704547393e-05, "loss": 1.3234, "step": 27 }, { "epoch": 0.2679425837320574, "grad_norm": 0.42941951751708984, "learning_rate": 7.824143126137431e-05, "loss": 1.4562, "step": 28 }, { "epoch": 0.27751196172248804, "grad_norm": 0.44896551966667175, "learning_rate": 7.804226065180615e-05, "loss": 1.6191, "step": 29 }, { "epoch": 0.28708133971291866, "grad_norm": 0.4686514437198639, "learning_rate": 7.783268966802539e-05, "loss": 1.5335, "step": 30 }, { "epoch": 0.2966507177033493, "grad_norm": 0.41287559270858765, "learning_rate": 7.761277560464645e-05, "loss": 1.5145, "step": 31 }, { "epoch": 0.3062200956937799, "grad_norm": 0.3899138271808624, "learning_rate": 7.738257858397844e-05, "loss": 1.5435, "step": 32 }, { "epoch": 0.3157894736842105, "grad_norm": 0.38428395986557007, "learning_rate": 7.71421615395883e-05, "loss": 1.4655, "step": 33 }, { "epoch": 0.3253588516746411, "grad_norm": 0.4115948975086212, "learning_rate": 7.68915901990954e-05, "loss": 1.4021, "step": 34 }, { "epoch": 0.3349282296650718, "grad_norm": 0.428444504737854, "learning_rate": 7.663093306620231e-05, "loss": 1.5792, "step": 35 }, { "epoch": 0.3444976076555024, "grad_norm": 0.4278987646102905, "learning_rate": 7.636026140196651e-05, "loss": 1.4603, "step": 36 }, { "epoch": 0.35406698564593303, "grad_norm": 0.45422014594078064, "learning_rate": 7.607964920531837e-05, "loss": 1.5074, "step": 37 }, { "epoch": 0.36363636363636365, "grad_norm": 0.4825778007507324, "learning_rate": 7.578917319283055e-05, "loss": 1.6137, "step": 38 }, { "epoch": 0.37320574162679426, "grad_norm": 0.49149173498153687, "learning_rate": 7.548891277774448e-05, "loss": 1.5871, "step": 39 }, { "epoch": 0.3827751196172249, "grad_norm": 0.5421543121337891, "learning_rate": 7.517895004825956e-05, "loss": 1.6206, "step": 40 }, { "epoch": 0.3923444976076555, "grad_norm": 0.5325545072555542, "learning_rate": 7.48593697450911e-05, "loss": 1.5887, "step": 41 }, { "epoch": 0.4019138755980861, "grad_norm": 0.583401620388031, "learning_rate": 7.453025923830296e-05, "loss": 1.8583, "step": 42 }, { "epoch": 0.41148325358851673, "grad_norm": 0.618493378162384, "learning_rate": 7.419170850342156e-05, "loss": 1.7723, "step": 43 }, { "epoch": 0.42105263157894735, "grad_norm": 0.6902540922164917, "learning_rate": 7.384381009683742e-05, "loss": 1.8642, "step": 44 }, { "epoch": 0.430622009569378, "grad_norm": 0.6965833902359009, "learning_rate": 7.348665913050115e-05, "loss": 1.8881, "step": 45 }, { "epoch": 0.44019138755980863, "grad_norm": 0.7248061299324036, "learning_rate": 7.312035324592081e-05, "loss": 1.8599, "step": 46 }, { "epoch": 0.44976076555023925, "grad_norm": 0.7599021196365356, "learning_rate": 7.274499258746771e-05, "loss": 1.919, "step": 47 }, { "epoch": 0.45933014354066987, "grad_norm": 0.8597122430801392, "learning_rate": 7.236067977499791e-05, "loss": 1.8464, "step": 48 }, { "epoch": 0.4688995215311005, "grad_norm": 0.9372475147247314, "learning_rate": 7.196751987579699e-05, "loss": 1.8919, "step": 49 }, { "epoch": 0.4784688995215311, "grad_norm": 1.1424167156219482, "learning_rate": 7.156562037585576e-05, "loss": 1.875, "step": 50 }, { "epoch": 0.4784688995215311, "eval_loss": 1.7351869344711304, "eval_runtime": 5.7379, "eval_samples_per_second": 30.848, "eval_steps_per_second": 7.843, "step": 50 }, { "epoch": 0.4880382775119617, "grad_norm": 1.656938076019287, "learning_rate": 7.11550911504845e-05, "loss": 1.9099, "step": 51 }, { "epoch": 0.49760765550239233, "grad_norm": 2.4753801822662354, "learning_rate": 7.073604443427437e-05, "loss": 1.6779, "step": 52 }, { "epoch": 0.507177033492823, "grad_norm": 0.29961133003234863, "learning_rate": 7.03085947904134e-05, "loss": 1.4577, "step": 53 }, { "epoch": 0.5167464114832536, "grad_norm": 0.3351958394050598, "learning_rate": 6.987285907936617e-05, "loss": 1.5448, "step": 54 }, { "epoch": 0.5263157894736842, "grad_norm": 0.3630099296569824, "learning_rate": 6.942895642692527e-05, "loss": 1.425, "step": 55 }, { "epoch": 0.5358851674641149, "grad_norm": 0.3872605860233307, "learning_rate": 6.897700819164357e-05, "loss": 1.5064, "step": 56 }, { "epoch": 0.5454545454545454, "grad_norm": 0.3808247148990631, "learning_rate": 6.851713793165589e-05, "loss": 1.5932, "step": 57 }, { "epoch": 0.5550239234449761, "grad_norm": 0.39085471630096436, "learning_rate": 6.804947137089955e-05, "loss": 1.4699, "step": 58 }, { "epoch": 0.5645933014354066, "grad_norm": 0.38267335295677185, "learning_rate": 6.757413636474263e-05, "loss": 1.4573, "step": 59 }, { "epoch": 0.5741626794258373, "grad_norm": 0.3885556757450104, "learning_rate": 6.709126286502965e-05, "loss": 1.477, "step": 60 }, { "epoch": 0.583732057416268, "grad_norm": 0.40890273451805115, "learning_rate": 6.660098288455393e-05, "loss": 1.4652, "step": 61 }, { "epoch": 0.5933014354066986, "grad_norm": 0.4209669828414917, "learning_rate": 6.610343046096674e-05, "loss": 1.5038, "step": 62 }, { "epoch": 0.6028708133971292, "grad_norm": 0.420719176530838, "learning_rate": 6.559874162013267e-05, "loss": 1.5592, "step": 63 }, { "epoch": 0.6124401913875598, "grad_norm": 0.4601193070411682, "learning_rate": 6.508705433894149e-05, "loss": 1.6004, "step": 64 }, { "epoch": 0.6220095693779905, "grad_norm": 0.4896854758262634, "learning_rate": 6.456850850758673e-05, "loss": 1.6569, "step": 65 }, { "epoch": 0.631578947368421, "grad_norm": 0.4936942756175995, "learning_rate": 6.404324589132101e-05, "loss": 1.6673, "step": 66 }, { "epoch": 0.6411483253588517, "grad_norm": 0.5432229042053223, "learning_rate": 6.351141009169893e-05, "loss": 1.6781, "step": 67 }, { "epoch": 0.6507177033492823, "grad_norm": 0.5662530064582825, "learning_rate": 6.297314650731775e-05, "loss": 1.7307, "step": 68 }, { "epoch": 0.6602870813397129, "grad_norm": 0.6015027761459351, "learning_rate": 6.242860229406692e-05, "loss": 1.7204, "step": 69 }, { "epoch": 0.6698564593301436, "grad_norm": 0.632143497467041, "learning_rate": 6.18779263248971e-05, "loss": 1.7661, "step": 70 }, { "epoch": 0.6794258373205742, "grad_norm": 0.6799759268760681, "learning_rate": 6.132126914911976e-05, "loss": 1.8595, "step": 71 }, { "epoch": 0.6889952153110048, "grad_norm": 0.7476505041122437, "learning_rate": 6.075878295124861e-05, "loss": 1.7854, "step": 72 }, { "epoch": 0.6985645933014354, "grad_norm": 0.8099026083946228, "learning_rate": 6.019062150939376e-05, "loss": 1.8917, "step": 73 }, { "epoch": 0.7081339712918661, "grad_norm": 0.8964723348617554, "learning_rate": 5.9616940153220336e-05, "loss": 1.9968, "step": 74 }, { "epoch": 0.7177033492822966, "grad_norm": 0.9372945427894592, "learning_rate": 5.903789572148295e-05, "loss": 2.0208, "step": 75 }, { "epoch": 0.7272727272727273, "grad_norm": 1.2482417821884155, "learning_rate": 5.845364651914752e-05, "loss": 2.1952, "step": 76 }, { "epoch": 0.7368421052631579, "grad_norm": 1.5300464630126953, "learning_rate": 5.786435227411227e-05, "loss": 1.7757, "step": 77 }, { "epoch": 0.7464114832535885, "grad_norm": 2.1026394367218018, "learning_rate": 5.727017409353971e-05, "loss": 1.6011, "step": 78 }, { "epoch": 0.7559808612440191, "grad_norm": 0.2756762206554413, "learning_rate": 5.667127441981162e-05, "loss": 1.4248, "step": 79 }, { "epoch": 0.7655502392344498, "grad_norm": 0.3412489891052246, "learning_rate": 5.606781698611879e-05, "loss": 1.5389, "step": 80 }, { "epoch": 0.7751196172248804, "grad_norm": 0.3573220670223236, "learning_rate": 5.5459966771698096e-05, "loss": 1.506, "step": 81 }, { "epoch": 0.784688995215311, "grad_norm": 0.40763863921165466, "learning_rate": 5.4847889956728834e-05, "loss": 1.5875, "step": 82 }, { "epoch": 0.7942583732057417, "grad_norm": 0.40273529291152954, "learning_rate": 5.423175387690067e-05, "loss": 1.4032, "step": 83 }, { "epoch": 0.8038277511961722, "grad_norm": 0.41646698117256165, "learning_rate": 5.361172697766573e-05, "loss": 1.6233, "step": 84 }, { "epoch": 0.8133971291866029, "grad_norm": 0.40466055274009705, "learning_rate": 5.298797876818735e-05, "loss": 1.4895, "step": 85 }, { "epoch": 0.8229665071770335, "grad_norm": 0.42662107944488525, "learning_rate": 5.23606797749979e-05, "loss": 1.482, "step": 86 }, { "epoch": 0.8325358851674641, "grad_norm": 0.44496554136276245, "learning_rate": 5.17300014953786e-05, "loss": 1.5595, "step": 87 }, { "epoch": 0.8421052631578947, "grad_norm": 0.41382166743278503, "learning_rate": 5.109611635047379e-05, "loss": 1.3534, "step": 88 }, { "epoch": 0.8516746411483254, "grad_norm": 0.44714534282684326, "learning_rate": 5.04591976381528e-05, "loss": 1.4973, "step": 89 }, { "epoch": 0.861244019138756, "grad_norm": 0.4594970643520355, "learning_rate": 4.981941948563197e-05, "loss": 1.7519, "step": 90 }, { "epoch": 0.8708133971291866, "grad_norm": 0.4741780161857605, "learning_rate": 4.9176956801870065e-05, "loss": 1.5083, "step": 91 }, { "epoch": 0.8803827751196173, "grad_norm": 0.49775177240371704, "learning_rate": 4.853198522974988e-05, "loss": 1.6772, "step": 92 }, { "epoch": 0.8899521531100478, "grad_norm": 0.5491089820861816, "learning_rate": 4.788468109805921e-05, "loss": 1.7706, "step": 93 }, { "epoch": 0.8995215311004785, "grad_norm": 0.5795297622680664, "learning_rate": 4.7235221373284407e-05, "loss": 1.7992, "step": 94 }, { "epoch": 0.9090909090909091, "grad_norm": 0.5943451523780823, "learning_rate": 4.658378361122936e-05, "loss": 1.7066, "step": 95 }, { "epoch": 0.9186602870813397, "grad_norm": 0.6468049883842468, "learning_rate": 4.593054590847368e-05, "loss": 1.8196, "step": 96 }, { "epoch": 0.9282296650717703, "grad_norm": 0.7615320086479187, "learning_rate": 4.5275686853682765e-05, "loss": 2.0393, "step": 97 }, { "epoch": 0.937799043062201, "grad_norm": 0.7791320085525513, "learning_rate": 4.4619385478783456e-05, "loss": 1.9799, "step": 98 }, { "epoch": 0.9473684210526315, "grad_norm": 0.870513379573822, "learning_rate": 4.396182121001852e-05, "loss": 2.058, "step": 99 }, { "epoch": 0.9569377990430622, "grad_norm": 0.9276971220970154, "learning_rate": 4.33031738188933e-05, "loss": 1.8242, "step": 100 }, { "epoch": 0.9569377990430622, "eval_loss": 1.6937828063964844, "eval_runtime": 5.7937, "eval_samples_per_second": 30.551, "eval_steps_per_second": 7.767, "step": 100 }, { "epoch": 0.9665071770334929, "grad_norm": 1.0401110649108887, "learning_rate": 4.264362337302798e-05, "loss": 1.9696, "step": 101 }, { "epoch": 0.9760765550239234, "grad_norm": 1.2967274188995361, "learning_rate": 4.1983350186928894e-05, "loss": 2.0238, "step": 102 }, { "epoch": 0.9856459330143541, "grad_norm": 1.6171687841415405, "learning_rate": 4.132253477269233e-05, "loss": 1.7819, "step": 103 }, { "epoch": 0.9952153110047847, "grad_norm": 2.1571907997131348, "learning_rate": 4.0661357790654345e-05, "loss": 1.4887, "step": 104 }, { "epoch": 1.0047846889952152, "grad_norm": 0.6212150454521179, "learning_rate": 4e-05, "loss": 2.2897, "step": 105 }, { "epoch": 1.014354066985646, "grad_norm": 0.2727871239185333, "learning_rate": 3.933864220934566e-05, "loss": 1.3308, "step": 106 }, { "epoch": 1.0239234449760766, "grad_norm": 0.3046855330467224, "learning_rate": 3.8677465227307676e-05, "loss": 1.4615, "step": 107 }, { "epoch": 1.0334928229665072, "grad_norm": 0.31469711661338806, "learning_rate": 3.8016649813071106e-05, "loss": 1.3841, "step": 108 }, { "epoch": 1.0430622009569377, "grad_norm": 0.32190945744514465, "learning_rate": 3.735637662697203e-05, "loss": 1.3789, "step": 109 }, { "epoch": 1.0526315789473684, "grad_norm": 0.3380140960216522, "learning_rate": 3.669682618110671e-05, "loss": 1.437, "step": 110 }, { "epoch": 1.062200956937799, "grad_norm": 0.3436225652694702, "learning_rate": 3.6038178789981494e-05, "loss": 1.2847, "step": 111 }, { "epoch": 1.0717703349282297, "grad_norm": 0.36676591634750366, "learning_rate": 3.538061452121656e-05, "loss": 1.297, "step": 112 }, { "epoch": 1.0813397129186604, "grad_norm": 0.39319589734077454, "learning_rate": 3.472431314631724e-05, "loss": 1.489, "step": 113 }, { "epoch": 1.0909090909090908, "grad_norm": 0.38813480734825134, "learning_rate": 3.406945409152632e-05, "loss": 1.4377, "step": 114 }, { "epoch": 1.1004784688995215, "grad_norm": 0.41558539867401123, "learning_rate": 3.341621638877064e-05, "loss": 1.4287, "step": 115 }, { "epoch": 1.1100478468899522, "grad_norm": 0.40909650921821594, "learning_rate": 3.276477862671562e-05, "loss": 1.441, "step": 116 }, { "epoch": 1.1196172248803828, "grad_norm": 0.45959237217903137, "learning_rate": 3.21153189019408e-05, "loss": 1.6774, "step": 117 }, { "epoch": 1.1291866028708135, "grad_norm": 0.46516451239585876, "learning_rate": 3.146801477025013e-05, "loss": 1.4372, "step": 118 }, { "epoch": 1.138755980861244, "grad_norm": 0.5069877505302429, "learning_rate": 3.082304319812994e-05, "loss": 1.642, "step": 119 }, { "epoch": 1.1483253588516746, "grad_norm": 0.5341264009475708, "learning_rate": 3.0180580514368037e-05, "loss": 1.6585, "step": 120 }, { "epoch": 1.1578947368421053, "grad_norm": 0.56055748462677, "learning_rate": 2.9540802361847212e-05, "loss": 1.6557, "step": 121 }, { "epoch": 1.167464114832536, "grad_norm": 0.5781103372573853, "learning_rate": 2.890388364952623e-05, "loss": 1.6671, "step": 122 }, { "epoch": 1.1770334928229664, "grad_norm": 0.6097198724746704, "learning_rate": 2.8269998504621416e-05, "loss": 1.6451, "step": 123 }, { "epoch": 1.186602870813397, "grad_norm": 0.6518456339836121, "learning_rate": 2.7639320225002108e-05, "loss": 1.6557, "step": 124 }, { "epoch": 1.1961722488038278, "grad_norm": 0.6572650671005249, "learning_rate": 2.7012021231812666e-05, "loss": 1.4739, "step": 125 }, { "epoch": 1.2057416267942584, "grad_norm": 0.783046543598175, "learning_rate": 2.638827302233428e-05, "loss": 1.8914, "step": 126 }, { "epoch": 1.215311004784689, "grad_norm": 0.808749794960022, "learning_rate": 2.576824612309934e-05, "loss": 1.6836, "step": 127 }, { "epoch": 1.2248803827751196, "grad_norm": 0.9802379012107849, "learning_rate": 2.5152110043271166e-05, "loss": 1.6596, "step": 128 }, { "epoch": 1.2344497607655502, "grad_norm": 1.2099603414535522, "learning_rate": 2.454003322830192e-05, "loss": 1.5137, "step": 129 }, { "epoch": 1.244019138755981, "grad_norm": 1.4862251281738281, "learning_rate": 2.393218301388123e-05, "loss": 1.1418, "step": 130 }, { "epoch": 1.2535885167464116, "grad_norm": 1.0337613821029663, "learning_rate": 2.3328725580188395e-05, "loss": 1.2111, "step": 131 }, { "epoch": 1.263157894736842, "grad_norm": 0.33979716897010803, "learning_rate": 2.272982590646029e-05, "loss": 1.4224, "step": 132 }, { "epoch": 1.2727272727272727, "grad_norm": 0.37942126393318176, "learning_rate": 2.2135647725887744e-05, "loss": 1.4692, "step": 133 }, { "epoch": 1.2822966507177034, "grad_norm": 0.3820892572402954, "learning_rate": 2.1546353480852495e-05, "loss": 1.3997, "step": 134 }, { "epoch": 1.291866028708134, "grad_norm": 0.37589502334594727, "learning_rate": 2.096210427851706e-05, "loss": 1.44, "step": 135 }, { "epoch": 1.3014354066985647, "grad_norm": 0.39906907081604004, "learning_rate": 2.038305984677969e-05, "loss": 1.4383, "step": 136 }, { "epoch": 1.3110047846889952, "grad_norm": 0.3934566378593445, "learning_rate": 1.9809378490606264e-05, "loss": 1.3633, "step": 137 }, { "epoch": 1.3205741626794258, "grad_norm": 0.4008135497570038, "learning_rate": 1.9241217048751406e-05, "loss": 1.4147, "step": 138 }, { "epoch": 1.3301435406698565, "grad_norm": 0.4096076488494873, "learning_rate": 1.867873085088026e-05, "loss": 1.3079, "step": 139 }, { "epoch": 1.339712918660287, "grad_norm": 0.4406437873840332, "learning_rate": 1.8122073675102935e-05, "loss": 1.4559, "step": 140 }, { "epoch": 1.3492822966507176, "grad_norm": 0.42370909452438354, "learning_rate": 1.75713977059331e-05, "loss": 1.3705, "step": 141 }, { "epoch": 1.3588516746411483, "grad_norm": 0.4583252966403961, "learning_rate": 1.702685349268226e-05, "loss": 1.4909, "step": 142 }, { "epoch": 1.368421052631579, "grad_norm": 0.44213107228279114, "learning_rate": 1.648858990830108e-05, "loss": 1.3054, "step": 143 }, { "epoch": 1.3779904306220097, "grad_norm": 0.46722662448883057, "learning_rate": 1.5956754108678996e-05, "loss": 1.443, "step": 144 }, { "epoch": 1.38755980861244, "grad_norm": 0.48926499485969543, "learning_rate": 1.5431491492413288e-05, "loss": 1.4217, "step": 145 }, { "epoch": 1.3971291866028708, "grad_norm": 0.5186731219291687, "learning_rate": 1.491294566105852e-05, "loss": 1.567, "step": 146 }, { "epoch": 1.4066985645933014, "grad_norm": 0.5692178010940552, "learning_rate": 1.4401258379867335e-05, "loss": 1.6171, "step": 147 }, { "epoch": 1.4162679425837321, "grad_norm": 0.5609941482543945, "learning_rate": 1.3896569539033253e-05, "loss": 1.5476, "step": 148 }, { "epoch": 1.4258373205741628, "grad_norm": 0.6115127801895142, "learning_rate": 1.3399017115446067e-05, "loss": 1.5292, "step": 149 }, { "epoch": 1.4354066985645932, "grad_norm": 0.6445640325546265, "learning_rate": 1.2908737134970367e-05, "loss": 1.6103, "step": 150 }, { "epoch": 1.4354066985645932, "eval_loss": 1.6925917863845825, "eval_runtime": 5.7417, "eval_samples_per_second": 30.827, "eval_steps_per_second": 7.837, "step": 150 }, { "epoch": 1.444976076555024, "grad_norm": 0.6974514722824097, "learning_rate": 1.242586363525737e-05, "loss": 1.7167, "step": 151 }, { "epoch": 1.4545454545454546, "grad_norm": 0.7227963805198669, "learning_rate": 1.1950528629100457e-05, "loss": 1.6717, "step": 152 }, { "epoch": 1.464114832535885, "grad_norm": 0.8344351053237915, "learning_rate": 1.1482862068344121e-05, "loss": 1.686, "step": 153 }, { "epoch": 1.4736842105263157, "grad_norm": 0.9203869104385376, "learning_rate": 1.1022991808356442e-05, "loss": 1.5159, "step": 154 }, { "epoch": 1.4832535885167464, "grad_norm": 1.1346211433410645, "learning_rate": 1.0571043573074737e-05, "loss": 1.5358, "step": 155 }, { "epoch": 1.492822966507177, "grad_norm": 1.6541485786437988, "learning_rate": 1.0127140920633857e-05, "loss": 1.1705, "step": 156 }, { "epoch": 1.5023923444976077, "grad_norm": 1.685089349746704, "learning_rate": 9.69140520958662e-06, "loss": 1.1268, "step": 157 }, { "epoch": 1.5119617224880382, "grad_norm": 0.30736255645751953, "learning_rate": 9.263955565725648e-06, "loss": 1.3637, "step": 158 }, { "epoch": 1.5215311004784688, "grad_norm": 0.3349371552467346, "learning_rate": 8.844908849515509e-06, "loss": 1.474, "step": 159 }, { "epoch": 1.5311004784688995, "grad_norm": 0.35800015926361084, "learning_rate": 8.434379624144261e-06, "loss": 1.5093, "step": 160 }, { "epoch": 1.5406698564593302, "grad_norm": 0.37739279866218567, "learning_rate": 8.032480124203013e-06, "loss": 1.4268, "step": 161 }, { "epoch": 1.5502392344497609, "grad_norm": 0.40975770354270935, "learning_rate": 7.639320225002106e-06, "loss": 1.5343, "step": 162 }, { "epoch": 1.5598086124401913, "grad_norm": 0.3935057520866394, "learning_rate": 7.255007412532307e-06, "loss": 1.4853, "step": 163 }, { "epoch": 1.569377990430622, "grad_norm": 0.4058459401130676, "learning_rate": 6.8796467540791986e-06, "loss": 1.3923, "step": 164 }, { "epoch": 1.5789473684210527, "grad_norm": 0.4203774631023407, "learning_rate": 6.513340869498859e-06, "loss": 1.3381, "step": 165 }, { "epoch": 1.588516746411483, "grad_norm": 0.42906367778778076, "learning_rate": 6.1561899031625794e-06, "loss": 1.4494, "step": 166 }, { "epoch": 1.598086124401914, "grad_norm": 0.43578189611434937, "learning_rate": 5.808291496578435e-06, "loss": 1.2433, "step": 167 }, { "epoch": 1.6076555023923444, "grad_norm": 0.4657846987247467, "learning_rate": 5.469740761697044e-06, "loss": 1.5992, "step": 168 }, { "epoch": 1.6172248803827751, "grad_norm": 0.49452725052833557, "learning_rate": 5.140630254908905e-06, "loss": 1.392, "step": 169 }, { "epoch": 1.6267942583732058, "grad_norm": 0.5063107013702393, "learning_rate": 4.821049951740442e-06, "loss": 1.6072, "step": 170 }, { "epoch": 1.6363636363636362, "grad_norm": 0.5106216073036194, "learning_rate": 4.511087222255528e-06, "loss": 1.3298, "step": 171 }, { "epoch": 1.6459330143540671, "grad_norm": 0.5595565438270569, "learning_rate": 4.2108268071694616e-06, "loss": 1.5702, "step": 172 }, { "epoch": 1.6555023923444976, "grad_norm": 0.5791851282119751, "learning_rate": 3.9203507946816445e-06, "loss": 1.505, "step": 173 }, { "epoch": 1.6650717703349283, "grad_norm": 0.6414560079574585, "learning_rate": 3.6397385980335e-06, "loss": 1.6376, "step": 174 }, { "epoch": 1.674641148325359, "grad_norm": 0.6914544105529785, "learning_rate": 3.3690669337977e-06, "loss": 1.7964, "step": 175 }, { "epoch": 1.6842105263157894, "grad_norm": 0.7928903698921204, "learning_rate": 3.1084098009046106e-06, "loss": 1.8557, "step": 176 }, { "epoch": 1.69377990430622, "grad_norm": 0.8021854162216187, "learning_rate": 2.8578384604117217e-06, "loss": 1.6362, "step": 177 }, { "epoch": 1.7033492822966507, "grad_norm": 0.7553149461746216, "learning_rate": 2.6174214160215704e-06, "loss": 1.342, "step": 178 }, { "epoch": 1.7129186602870812, "grad_norm": 0.9696720242500305, "learning_rate": 2.3872243953535535e-06, "loss": 1.7634, "step": 179 }, { "epoch": 1.722488038277512, "grad_norm": 1.1475272178649902, "learning_rate": 2.1673103319746146e-06, "loss": 1.7356, "step": 180 }, { "epoch": 1.7320574162679425, "grad_norm": 1.3815126419067383, "learning_rate": 1.957739348193859e-06, "loss": 1.6326, "step": 181 }, { "epoch": 1.7416267942583732, "grad_norm": 2.003146171569824, "learning_rate": 1.7585687386256944e-06, "loss": 1.5472, "step": 182 }, { "epoch": 1.7511961722488039, "grad_norm": 1.2215324640274048, "learning_rate": 1.5698529545260744e-06, "loss": 1.1225, "step": 183 }, { "epoch": 1.7607655502392343, "grad_norm": 0.32720524072647095, "learning_rate": 1.3916435889060575e-06, "loss": 1.4342, "step": 184 }, { "epoch": 1.7703349282296652, "grad_norm": 0.34399574995040894, "learning_rate": 1.2239893624267852e-06, "loss": 1.512, "step": 185 }, { "epoch": 1.7799043062200957, "grad_norm": 0.3406524956226349, "learning_rate": 1.0669361100797704e-06, "loss": 1.2191, "step": 186 }, { "epoch": 1.7894736842105263, "grad_norm": 0.3624950349330902, "learning_rate": 9.205267686560293e-07, "loss": 1.4136, "step": 187 }, { "epoch": 1.799043062200957, "grad_norm": 0.4109777808189392, "learning_rate": 7.848013650076258e-07, "loss": 1.5427, "step": 188 }, { "epoch": 1.8086124401913874, "grad_norm": 0.39658841490745544, "learning_rate": 6.597970051047053e-07, "loss": 1.3736, "step": 189 }, { "epoch": 1.8181818181818183, "grad_norm": 0.41146281361579895, "learning_rate": 5.455478638911071e-07, "loss": 1.445, "step": 190 }, { "epoch": 1.8277511961722488, "grad_norm": 0.410113126039505, "learning_rate": 4.420851759412603e-07, "loss": 1.2919, "step": 191 }, { "epoch": 1.8373205741626795, "grad_norm": 0.43137213587760925, "learning_rate": 3.4943722692099224e-07, "loss": 1.3197, "step": 192 }, { "epoch": 1.8468899521531101, "grad_norm": 0.43355488777160645, "learning_rate": 2.676293458544743e-07, "loss": 1.3578, "step": 193 }, { "epoch": 1.8564593301435406, "grad_norm": 0.4816869795322418, "learning_rate": 1.9668389819954338e-07, "loss": 1.4214, "step": 194 }, { "epoch": 1.8660287081339713, "grad_norm": 0.48422351479530334, "learning_rate": 1.3662027973320614e-07, "loss": 1.3847, "step": 195 }, { "epoch": 1.875598086124402, "grad_norm": 0.522202730178833, "learning_rate": 8.745491124901861e-08, "loss": 1.5238, "step": 196 }, { "epoch": 1.8851674641148324, "grad_norm": 0.5350428223609924, "learning_rate": 4.920123406781052e-08, "loss": 1.5597, "step": 197 }, { "epoch": 1.8947368421052633, "grad_norm": 0.5734149813652039, "learning_rate": 2.1869706362958044e-08, "loss": 1.5932, "step": 198 }, { "epoch": 1.9043062200956937, "grad_norm": 0.6013189554214478, "learning_rate": 5.467800301239834e-09, "loss": 1.6291, "step": 199 }, { "epoch": 1.9138755980861244, "grad_norm": 0.6234821677207947, "learning_rate": 0.0, "loss": 1.4807, "step": 200 }, { "epoch": 1.9138755980861244, "eval_loss": 1.7030987739562988, "eval_runtime": 5.7485, "eval_samples_per_second": 30.79, "eval_steps_per_second": 7.828, "step": 200 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 4, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 1 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.57896697905152e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }