{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.2668889106775174, "eval_steps": 341, "global_step": 1361, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00019609765663300324, "grad_norm": 18.01366424560547, "learning_rate": 2e-05, "loss": 3.0843, "step": 1 }, { "epoch": 0.00019609765663300324, "eval_loss": 1.1017773151397705, "eval_runtime": 79.0856, "eval_samples_per_second": 27.16, "eval_steps_per_second": 13.58, "step": 1 }, { "epoch": 0.0003921953132660065, "grad_norm": 16.80318260192871, "learning_rate": 4e-05, "loss": 3.2221, "step": 2 }, { "epoch": 0.0005882929698990097, "grad_norm": 16.65929412841797, "learning_rate": 6e-05, "loss": 3.8956, "step": 3 }, { "epoch": 0.000784390626532013, "grad_norm": 37.782188415527344, "learning_rate": 8e-05, "loss": 5.185, "step": 4 }, { "epoch": 0.0009804882831650162, "grad_norm": 19.226940155029297, "learning_rate": 0.0001, "loss": 3.1542, "step": 5 }, { "epoch": 0.0011765859397980193, "grad_norm": 26.570402145385742, "learning_rate": 0.00012, "loss": 4.5356, "step": 6 }, { "epoch": 0.0013726835964310226, "grad_norm": 22.43348503112793, "learning_rate": 0.00014, "loss": 3.8177, "step": 7 }, { "epoch": 0.001568781253064026, "grad_norm": 29.817941665649414, "learning_rate": 0.00016, "loss": 4.4013, "step": 8 }, { "epoch": 0.0017648789096970292, "grad_norm": 18.46044158935547, "learning_rate": 0.00018, "loss": 3.374, "step": 9 }, { "epoch": 0.0019609765663300325, "grad_norm": 24.337013244628906, "learning_rate": 0.0002, "loss": 3.2509, "step": 10 }, { "epoch": 0.0021570742229630358, "grad_norm": 18.9931640625, "learning_rate": 0.00019999972962977903, "loss": 2.933, "step": 11 }, { "epoch": 0.0023531718795960386, "grad_norm": 17.7039852142334, "learning_rate": 0.00019999891852057812, "loss": 3.2867, "step": 12 }, { "epoch": 0.002549269536229042, "grad_norm": 15.128417015075684, "learning_rate": 0.0001999975666767833, "loss": 2.2303, "step": 13 }, { "epoch": 0.002745367192862045, "grad_norm": 8.747583389282227, "learning_rate": 0.00019999567410570446, "loss": 1.7348, "step": 14 }, { "epoch": 0.0029414648494950485, "grad_norm": 20.570377349853516, "learning_rate": 0.00019999324081757555, "loss": 3.9284, "step": 15 }, { "epoch": 0.003137562506128052, "grad_norm": 7.86672306060791, "learning_rate": 0.00019999026682555434, "loss": 1.8121, "step": 16 }, { "epoch": 0.003333660162761055, "grad_norm": 10.341567039489746, "learning_rate": 0.0001999867521457224, "loss": 1.5864, "step": 17 }, { "epoch": 0.0035297578193940584, "grad_norm": 7.888909339904785, "learning_rate": 0.00019998269679708504, "loss": 3.1584, "step": 18 }, { "epoch": 0.0037258554760270617, "grad_norm": 15.346771240234375, "learning_rate": 0.00019997810080157113, "loss": 2.0858, "step": 19 }, { "epoch": 0.003921953132660065, "grad_norm": 11.417168617248535, "learning_rate": 0.0001999729641840331, "loss": 3.5646, "step": 20 }, { "epoch": 0.004118050789293068, "grad_norm": 12.86220932006836, "learning_rate": 0.00019996728697224675, "loss": 3.0594, "step": 21 }, { "epoch": 0.0043141484459260715, "grad_norm": 7.418736457824707, "learning_rate": 0.00019996106919691102, "loss": 2.1349, "step": 22 }, { "epoch": 0.004510246102559075, "grad_norm": 4.803528308868408, "learning_rate": 0.00019995431089164795, "loss": 2.3151, "step": 23 }, { "epoch": 0.004706343759192077, "grad_norm": 8.667632102966309, "learning_rate": 0.00019994701209300245, "loss": 1.6791, "step": 24 }, { "epoch": 0.0049024414158250805, "grad_norm": 5.3950724601745605, "learning_rate": 0.00019993917284044202, "loss": 2.7265, "step": 25 }, { "epoch": 0.005098539072458084, "grad_norm": 18.16863441467285, "learning_rate": 0.0001999307931763567, "loss": 1.7289, "step": 26 }, { "epoch": 0.005294636729091087, "grad_norm": 10.66887378692627, "learning_rate": 0.00019992187314605872, "loss": 2.645, "step": 27 }, { "epoch": 0.00549073438572409, "grad_norm": 26.42878532409668, "learning_rate": 0.00019991241279778232, "loss": 3.4603, "step": 28 }, { "epoch": 0.005686832042357094, "grad_norm": 10.017987251281738, "learning_rate": 0.0001999024121826834, "loss": 1.2001, "step": 29 }, { "epoch": 0.005882929698990097, "grad_norm": 11.155564308166504, "learning_rate": 0.00019989187135483933, "loss": 1.5102, "step": 30 }, { "epoch": 0.0060790273556231, "grad_norm": 3.7590131759643555, "learning_rate": 0.00019988079037124864, "loss": 1.0619, "step": 31 }, { "epoch": 0.006275125012256104, "grad_norm": 8.47985553741455, "learning_rate": 0.00019986916929183067, "loss": 2.6256, "step": 32 }, { "epoch": 0.006471222668889107, "grad_norm": 16.655269622802734, "learning_rate": 0.00019985700817942533, "loss": 2.2039, "step": 33 }, { "epoch": 0.00666732032552211, "grad_norm": 9.003829956054688, "learning_rate": 0.00019984430709979264, "loss": 1.7281, "step": 34 }, { "epoch": 0.0068634179821551134, "grad_norm": 36.33080291748047, "learning_rate": 0.0001998310661216125, "loss": 3.0575, "step": 35 }, { "epoch": 0.007059515638788117, "grad_norm": 11.909663200378418, "learning_rate": 0.00019981728531648423, "loss": 1.9623, "step": 36 }, { "epoch": 0.00725561329542112, "grad_norm": 10.967493057250977, "learning_rate": 0.00019980296475892616, "loss": 2.8071, "step": 37 }, { "epoch": 0.007451710952054123, "grad_norm": 11.050918579101562, "learning_rate": 0.00019978810452637543, "loss": 1.8019, "step": 38 }, { "epoch": 0.007647808608687127, "grad_norm": 17.752105712890625, "learning_rate": 0.00019977270469918727, "loss": 3.0332, "step": 39 }, { "epoch": 0.00784390626532013, "grad_norm": 6.67478609085083, "learning_rate": 0.0001997567653606348, "loss": 1.32, "step": 40 }, { "epoch": 0.008040003921953132, "grad_norm": 7.402947902679443, "learning_rate": 0.00019974028659690843, "loss": 1.4442, "step": 41 }, { "epoch": 0.008236101578586136, "grad_norm": 13.006294250488281, "learning_rate": 0.00019972326849711553, "loss": 2.2418, "step": 42 }, { "epoch": 0.008432199235219139, "grad_norm": 5.940983295440674, "learning_rate": 0.00019970571115327985, "loss": 0.9049, "step": 43 }, { "epoch": 0.008628296891852143, "grad_norm": 9.866477012634277, "learning_rate": 0.00019968761466034103, "loss": 2.7203, "step": 44 }, { "epoch": 0.008824394548485145, "grad_norm": 8.239157676696777, "learning_rate": 0.00019966897911615416, "loss": 1.9653, "step": 45 }, { "epoch": 0.00902049220511815, "grad_norm": 16.142181396484375, "learning_rate": 0.0001996498046214891, "loss": 2.7509, "step": 46 }, { "epoch": 0.009216589861751152, "grad_norm": 9.295431137084961, "learning_rate": 0.00019963009128003018, "loss": 2.0133, "step": 47 }, { "epoch": 0.009412687518384154, "grad_norm": 10.362130165100098, "learning_rate": 0.00019960983919837535, "loss": 1.716, "step": 48 }, { "epoch": 0.009608785175017159, "grad_norm": 28.2889461517334, "learning_rate": 0.00019958904848603584, "loss": 2.8961, "step": 49 }, { "epoch": 0.009804882831650161, "grad_norm": 13.199325561523438, "learning_rate": 0.0001995677192554354, "loss": 2.62, "step": 50 }, { "epoch": 0.010000980488283165, "grad_norm": 11.95132827758789, "learning_rate": 0.00019954585162190985, "loss": 2.792, "step": 51 }, { "epoch": 0.010197078144916168, "grad_norm": 16.581575393676758, "learning_rate": 0.0001995234457037063, "loss": 2.7239, "step": 52 }, { "epoch": 0.010393175801549172, "grad_norm": 12.048559188842773, "learning_rate": 0.00019950050162198258, "loss": 1.9892, "step": 53 }, { "epoch": 0.010589273458182174, "grad_norm": 9.297942161560059, "learning_rate": 0.00019947701950080672, "loss": 1.8015, "step": 54 }, { "epoch": 0.010785371114815178, "grad_norm": 6.09962797164917, "learning_rate": 0.00019945299946715596, "loss": 1.1493, "step": 55 }, { "epoch": 0.01098146877144818, "grad_norm": 6.668224811553955, "learning_rate": 0.00019942844165091633, "loss": 0.9968, "step": 56 }, { "epoch": 0.011177566428081185, "grad_norm": 17.000507354736328, "learning_rate": 0.00019940334618488194, "loss": 1.5857, "step": 57 }, { "epoch": 0.011373664084714187, "grad_norm": 11.42551326751709, "learning_rate": 0.00019937771320475406, "loss": 1.5236, "step": 58 }, { "epoch": 0.011569761741347192, "grad_norm": 13.255271911621094, "learning_rate": 0.00019935154284914065, "loss": 1.6174, "step": 59 }, { "epoch": 0.011765859397980194, "grad_norm": 12.664427757263184, "learning_rate": 0.00019932483525955533, "loss": 1.476, "step": 60 }, { "epoch": 0.011961957054613198, "grad_norm": 19.540660858154297, "learning_rate": 0.00019929759058041687, "loss": 1.5251, "step": 61 }, { "epoch": 0.0121580547112462, "grad_norm": 10.44942855834961, "learning_rate": 0.0001992698089590483, "loss": 1.7865, "step": 62 }, { "epoch": 0.012354152367879203, "grad_norm": 13.294017791748047, "learning_rate": 0.00019924149054567606, "loss": 4.1284, "step": 63 }, { "epoch": 0.012550250024512207, "grad_norm": 13.700861930847168, "learning_rate": 0.00019921263549342922, "loss": 2.8987, "step": 64 }, { "epoch": 0.01274634768114521, "grad_norm": 51.7520866394043, "learning_rate": 0.00019918324395833877, "loss": 1.7335, "step": 65 }, { "epoch": 0.012942445337778214, "grad_norm": 25.415748596191406, "learning_rate": 0.00019915331609933657, "loss": 2.6131, "step": 66 }, { "epoch": 0.013138542994411216, "grad_norm": 10.575088500976562, "learning_rate": 0.00019912285207825475, "loss": 1.202, "step": 67 }, { "epoch": 0.01333464065104422, "grad_norm": 7.0860772132873535, "learning_rate": 0.00019909185205982453, "loss": 1.5077, "step": 68 }, { "epoch": 0.013530738307677223, "grad_norm": 8.169393539428711, "learning_rate": 0.00019906031621167553, "loss": 2.4139, "step": 69 }, { "epoch": 0.013726835964310227, "grad_norm": 15.750587463378906, "learning_rate": 0.00019902824470433489, "loss": 2.8999, "step": 70 }, { "epoch": 0.01392293362094323, "grad_norm": 19.615734100341797, "learning_rate": 0.00019899563771122618, "loss": 3.3154, "step": 71 }, { "epoch": 0.014119031277576233, "grad_norm": 24.015810012817383, "learning_rate": 0.0001989624954086686, "loss": 2.6689, "step": 72 }, { "epoch": 0.014315128934209236, "grad_norm": 7.955759525299072, "learning_rate": 0.00019892881797587601, "loss": 1.8847, "step": 73 }, { "epoch": 0.01451122659084224, "grad_norm": 13.804747581481934, "learning_rate": 0.00019889460559495588, "loss": 2.2221, "step": 74 }, { "epoch": 0.014707324247475242, "grad_norm": 19.74148941040039, "learning_rate": 0.0001988598584509084, "loss": 1.9316, "step": 75 }, { "epoch": 0.014903421904108247, "grad_norm": 7.9701385498046875, "learning_rate": 0.00019882457673162543, "loss": 2.1958, "step": 76 }, { "epoch": 0.015099519560741249, "grad_norm": 4.7594170570373535, "learning_rate": 0.00019878876062788954, "loss": 1.3551, "step": 77 }, { "epoch": 0.015295617217374253, "grad_norm": 7.796854496002197, "learning_rate": 0.0001987524103333728, "loss": 2.616, "step": 78 }, { "epoch": 0.015491714874007256, "grad_norm": 6.997422218322754, "learning_rate": 0.00019871552604463602, "loss": 2.908, "step": 79 }, { "epoch": 0.01568781253064026, "grad_norm": 6.436347484588623, "learning_rate": 0.00019867810796112744, "loss": 3.6826, "step": 80 }, { "epoch": 0.01588391018727326, "grad_norm": 10.243407249450684, "learning_rate": 0.00019864015628518175, "loss": 1.3711, "step": 81 }, { "epoch": 0.016080007843906265, "grad_norm": 11.830530166625977, "learning_rate": 0.00019860167122201904, "loss": 2.2325, "step": 82 }, { "epoch": 0.01627610550053927, "grad_norm": 7.456768989562988, "learning_rate": 0.0001985626529797436, "loss": 1.9991, "step": 83 }, { "epoch": 0.016472203157172273, "grad_norm": 50.91776657104492, "learning_rate": 0.00019852310176934288, "loss": 1.294, "step": 84 }, { "epoch": 0.016668300813805274, "grad_norm": 6.4661054611206055, "learning_rate": 0.00019848301780468622, "loss": 2.4052, "step": 85 }, { "epoch": 0.016864398470438278, "grad_norm": 5.4749674797058105, "learning_rate": 0.00019844240130252385, "loss": 2.1508, "step": 86 }, { "epoch": 0.017060496127071282, "grad_norm": 9.457857131958008, "learning_rate": 0.00019840125248248564, "loss": 2.1732, "step": 87 }, { "epoch": 0.017256593783704286, "grad_norm": 6.932736396789551, "learning_rate": 0.00019835957156707988, "loss": 1.0618, "step": 88 }, { "epoch": 0.017452691440337287, "grad_norm": 10.447488784790039, "learning_rate": 0.00019831735878169212, "loss": 1.1214, "step": 89 }, { "epoch": 0.01764878909697029, "grad_norm": 7.159199237823486, "learning_rate": 0.000198274614354584, "loss": 2.495, "step": 90 }, { "epoch": 0.017844886753603295, "grad_norm": 8.753946304321289, "learning_rate": 0.00019823133851689187, "loss": 2.343, "step": 91 }, { "epoch": 0.0180409844102363, "grad_norm": 9.124870300292969, "learning_rate": 0.00019818753150262574, "loss": 1.6556, "step": 92 }, { "epoch": 0.0182370820668693, "grad_norm": 5.470896244049072, "learning_rate": 0.00019814319354866786, "loss": 1.2787, "step": 93 }, { "epoch": 0.018433179723502304, "grad_norm": 8.39749526977539, "learning_rate": 0.00019809832489477142, "loss": 1.6804, "step": 94 }, { "epoch": 0.01862927738013531, "grad_norm": 6.869524955749512, "learning_rate": 0.0001980529257835594, "loss": 1.6563, "step": 95 }, { "epoch": 0.01882537503676831, "grad_norm": 11.144633293151855, "learning_rate": 0.0001980069964605232, "loss": 1.9428, "step": 96 }, { "epoch": 0.019021472693401313, "grad_norm": 11.564960479736328, "learning_rate": 0.00019796053717402118, "loss": 2.2905, "step": 97 }, { "epoch": 0.019217570350034317, "grad_norm": 4.628671169281006, "learning_rate": 0.00019791354817527755, "loss": 1.0654, "step": 98 }, { "epoch": 0.01941366800666732, "grad_norm": 15.970938682556152, "learning_rate": 0.00019786602971838074, "loss": 2.9314, "step": 99 }, { "epoch": 0.019609765663300322, "grad_norm": 5.425403594970703, "learning_rate": 0.00019781798206028239, "loss": 1.7236, "step": 100 }, { "epoch": 0.019805863319933326, "grad_norm": 5.114929676055908, "learning_rate": 0.0001977694054607955, "loss": 2.1674, "step": 101 }, { "epoch": 0.02000196097656633, "grad_norm": 9.564162254333496, "learning_rate": 0.0001977203001825935, "loss": 2.3091, "step": 102 }, { "epoch": 0.020198058633199335, "grad_norm": 7.184217929840088, "learning_rate": 0.00019767066649120838, "loss": 2.5231, "step": 103 }, { "epoch": 0.020394156289832335, "grad_norm": 8.926065444946289, "learning_rate": 0.00019762050465502965, "loss": 1.3343, "step": 104 }, { "epoch": 0.02059025394646534, "grad_norm": 5.511591911315918, "learning_rate": 0.0001975698149453026, "loss": 1.7055, "step": 105 }, { "epoch": 0.020786351603098344, "grad_norm": 8.730749130249023, "learning_rate": 0.00019751859763612704, "loss": 1.7862, "step": 106 }, { "epoch": 0.020982449259731348, "grad_norm": 10.519305229187012, "learning_rate": 0.00019746685300445565, "loss": 1.9938, "step": 107 }, { "epoch": 0.02117854691636435, "grad_norm": 6.307378768920898, "learning_rate": 0.00019741458133009258, "loss": 1.5184, "step": 108 }, { "epoch": 0.021374644572997353, "grad_norm": 10.66562557220459, "learning_rate": 0.00019736178289569186, "loss": 2.0555, "step": 109 }, { "epoch": 0.021570742229630357, "grad_norm": 6.235299587249756, "learning_rate": 0.0001973084579867561, "loss": 1.4547, "step": 110 }, { "epoch": 0.021766839886263357, "grad_norm": 12.061420440673828, "learning_rate": 0.00019725460689163455, "loss": 1.9823, "step": 111 }, { "epoch": 0.02196293754289636, "grad_norm": 8.891043663024902, "learning_rate": 0.0001972002299015219, "loss": 2.4757, "step": 112 }, { "epoch": 0.022159035199529366, "grad_norm": 8.200632095336914, "learning_rate": 0.00019714532731045649, "loss": 1.4906, "step": 113 }, { "epoch": 0.02235513285616237, "grad_norm": 8.106672286987305, "learning_rate": 0.00019708989941531887, "loss": 0.8863, "step": 114 }, { "epoch": 0.02255123051279537, "grad_norm": 6.451066493988037, "learning_rate": 0.0001970339465158301, "loss": 2.1244, "step": 115 }, { "epoch": 0.022747328169428375, "grad_norm": 6.3670220375061035, "learning_rate": 0.0001969774689145501, "loss": 1.3732, "step": 116 }, { "epoch": 0.02294342582606138, "grad_norm": 6.401678562164307, "learning_rate": 0.0001969204669168761, "loss": 1.3963, "step": 117 }, { "epoch": 0.023139523482694383, "grad_norm": 6.917253494262695, "learning_rate": 0.00019686294083104094, "loss": 1.3802, "step": 118 }, { "epoch": 0.023335621139327384, "grad_norm": 4.852232456207275, "learning_rate": 0.00019680489096811149, "loss": 0.7538, "step": 119 }, { "epoch": 0.023531718795960388, "grad_norm": 5.08411169052124, "learning_rate": 0.00019674631764198677, "loss": 1.5128, "step": 120 }, { "epoch": 0.023727816452593392, "grad_norm": 13.64840316772461, "learning_rate": 0.00019668722116939649, "loss": 1.2718, "step": 121 }, { "epoch": 0.023923914109226396, "grad_norm": 7.075209617614746, "learning_rate": 0.00019662760186989913, "loss": 2.4186, "step": 122 }, { "epoch": 0.024120011765859397, "grad_norm": 12.186111450195312, "learning_rate": 0.00019656746006588044, "loss": 2.8495, "step": 123 }, { "epoch": 0.0243161094224924, "grad_norm": 13.221094131469727, "learning_rate": 0.00019650679608255138, "loss": 1.1832, "step": 124 }, { "epoch": 0.024512207079125405, "grad_norm": 4.765728950500488, "learning_rate": 0.0001964456102479467, "loss": 1.1407, "step": 125 }, { "epoch": 0.024708304735758406, "grad_norm": 4.750761032104492, "learning_rate": 0.00019638390289292295, "loss": 1.1096, "step": 126 }, { "epoch": 0.02490440239239141, "grad_norm": 5.133049011230469, "learning_rate": 0.0001963216743511567, "loss": 2.6525, "step": 127 }, { "epoch": 0.025100500049024414, "grad_norm": 14.541696548461914, "learning_rate": 0.0001962589249591429, "loss": 3.3739, "step": 128 }, { "epoch": 0.02529659770565742, "grad_norm": 7.252123832702637, "learning_rate": 0.00019619565505619288, "loss": 2.0899, "step": 129 }, { "epoch": 0.02549269536229042, "grad_norm": 7.14664888381958, "learning_rate": 0.00019613186498443257, "loss": 1.4538, "step": 130 }, { "epoch": 0.025688793018923423, "grad_norm": 7.936334609985352, "learning_rate": 0.0001960675550888007, "loss": 1.351, "step": 131 }, { "epoch": 0.025884890675556427, "grad_norm": 7.465827465057373, "learning_rate": 0.00019600272571704687, "loss": 1.0752, "step": 132 }, { "epoch": 0.02608098833218943, "grad_norm": 12.85190486907959, "learning_rate": 0.00019593737721972977, "loss": 2.5674, "step": 133 }, { "epoch": 0.026277085988822432, "grad_norm": 8.640382766723633, "learning_rate": 0.00019587150995021505, "loss": 2.5631, "step": 134 }, { "epoch": 0.026473183645455436, "grad_norm": 5.070466995239258, "learning_rate": 0.00019580512426467376, "loss": 0.8935, "step": 135 }, { "epoch": 0.02666928130208844, "grad_norm": 5.741654872894287, "learning_rate": 0.00019573822052208013, "loss": 2.1005, "step": 136 }, { "epoch": 0.026865378958721445, "grad_norm": 8.615095138549805, "learning_rate": 0.00019567079908420972, "loss": 2.7478, "step": 137 }, { "epoch": 0.027061476615354445, "grad_norm": 15.559739112854004, "learning_rate": 0.00019560286031563754, "loss": 1.8455, "step": 138 }, { "epoch": 0.02725757427198745, "grad_norm": 5.683060169219971, "learning_rate": 0.000195534404583736, "loss": 1.6896, "step": 139 }, { "epoch": 0.027453671928620454, "grad_norm": 5.791153430938721, "learning_rate": 0.00019546543225867292, "loss": 1.9291, "step": 140 }, { "epoch": 0.027649769585253458, "grad_norm": 6.800760269165039, "learning_rate": 0.0001953959437134095, "loss": 1.789, "step": 141 }, { "epoch": 0.02784586724188646, "grad_norm": 6.912458896636963, "learning_rate": 0.00019532593932369849, "loss": 2.2544, "step": 142 }, { "epoch": 0.028041964898519463, "grad_norm": 6.928767681121826, "learning_rate": 0.00019525541946808188, "loss": 1.2531, "step": 143 }, { "epoch": 0.028238062555152467, "grad_norm": 8.96179485321045, "learning_rate": 0.00019518438452788907, "loss": 2.3403, "step": 144 }, { "epoch": 0.028434160211785468, "grad_norm": 7.004507064819336, "learning_rate": 0.00019511283488723473, "loss": 1.1211, "step": 145 }, { "epoch": 0.028630257868418472, "grad_norm": 8.655360221862793, "learning_rate": 0.00019504077093301665, "loss": 1.6074, "step": 146 }, { "epoch": 0.028826355525051476, "grad_norm": 7.188142776489258, "learning_rate": 0.00019496819305491383, "loss": 1.3564, "step": 147 }, { "epoch": 0.02902245318168448, "grad_norm": 5.618204116821289, "learning_rate": 0.00019489510164538416, "loss": 2.5936, "step": 148 }, { "epoch": 0.02921855083831748, "grad_norm": 6.578945159912109, "learning_rate": 0.00019482149709966246, "loss": 0.9577, "step": 149 }, { "epoch": 0.029414648494950485, "grad_norm": 11.468759536743164, "learning_rate": 0.00019474737981575832, "loss": 2.1746, "step": 150 }, { "epoch": 0.02961074615158349, "grad_norm": 5.663360595703125, "learning_rate": 0.00019467275019445385, "loss": 1.2751, "step": 151 }, { "epoch": 0.029806843808216493, "grad_norm": 6.297888278961182, "learning_rate": 0.00019459760863930155, "loss": 1.574, "step": 152 }, { "epoch": 0.030002941464849494, "grad_norm": 8.760517120361328, "learning_rate": 0.00019452195555662224, "loss": 1.1148, "step": 153 }, { "epoch": 0.030199039121482498, "grad_norm": 8.745152473449707, "learning_rate": 0.00019444579135550273, "loss": 1.4212, "step": 154 }, { "epoch": 0.030395136778115502, "grad_norm": 7.594636917114258, "learning_rate": 0.00019436911644779366, "loss": 0.9161, "step": 155 }, { "epoch": 0.030591234434748506, "grad_norm": 8.563089370727539, "learning_rate": 0.00019429193124810725, "loss": 1.5844, "step": 156 }, { "epoch": 0.030787332091381507, "grad_norm": 16.4445743560791, "learning_rate": 0.00019421423617381508, "loss": 1.4798, "step": 157 }, { "epoch": 0.03098342974801451, "grad_norm": 6.86074161529541, "learning_rate": 0.00019413603164504588, "loss": 2.3323, "step": 158 }, { "epoch": 0.031179527404647515, "grad_norm": 11.293970108032227, "learning_rate": 0.0001940573180846831, "loss": 2.0304, "step": 159 }, { "epoch": 0.03137562506128052, "grad_norm": 9.772544860839844, "learning_rate": 0.00019397809591836286, "loss": 3.1622, "step": 160 }, { "epoch": 0.031571722717913524, "grad_norm": 8.489032745361328, "learning_rate": 0.00019389836557447143, "loss": 1.0113, "step": 161 }, { "epoch": 0.03176782037454652, "grad_norm": 12.992219924926758, "learning_rate": 0.000193818127484143, "loss": 2.5478, "step": 162 }, { "epoch": 0.031963918031179525, "grad_norm": 6.758875846862793, "learning_rate": 0.0001937373820812574, "loss": 1.8935, "step": 163 }, { "epoch": 0.03216001568781253, "grad_norm": 11.88957405090332, "learning_rate": 0.0001936561298024377, "loss": 3.1899, "step": 164 }, { "epoch": 0.03235611334444553, "grad_norm": 8.777862548828125, "learning_rate": 0.00019357437108704777, "loss": 2.7038, "step": 165 }, { "epoch": 0.03255221100107854, "grad_norm": 5.135364532470703, "learning_rate": 0.0001934921063771901, "loss": 0.7644, "step": 166 }, { "epoch": 0.03274830865771154, "grad_norm": 9.942065238952637, "learning_rate": 0.00019340933611770321, "loss": 1.4148, "step": 167 }, { "epoch": 0.032944406314344546, "grad_norm": 7.881731033325195, "learning_rate": 0.0001933260607561594, "loss": 1.7402, "step": 168 }, { "epoch": 0.03314050397097755, "grad_norm": 6.189711570739746, "learning_rate": 0.00019324228074286222, "loss": 1.642, "step": 169 }, { "epoch": 0.03333660162761055, "grad_norm": 4.979406356811523, "learning_rate": 0.00019315799653084404, "loss": 2.425, "step": 170 }, { "epoch": 0.03353269928424355, "grad_norm": 8.461871147155762, "learning_rate": 0.00019307320857586376, "loss": 2.6563, "step": 171 }, { "epoch": 0.033728796940876556, "grad_norm": 8.992694854736328, "learning_rate": 0.00019298791733640406, "loss": 1.7962, "step": 172 }, { "epoch": 0.03392489459750956, "grad_norm": 8.074629783630371, "learning_rate": 0.00019290212327366924, "loss": 1.5342, "step": 173 }, { "epoch": 0.034120992254142564, "grad_norm": 6.5658111572265625, "learning_rate": 0.00019281582685158247, "loss": 0.7919, "step": 174 }, { "epoch": 0.03431708991077557, "grad_norm": 9.946451187133789, "learning_rate": 0.00019272902853678336, "loss": 1.5664, "step": 175 }, { "epoch": 0.03451318756740857, "grad_norm": 6.764862060546875, "learning_rate": 0.00019264172879862552, "loss": 2.1083, "step": 176 }, { "epoch": 0.03470928522404157, "grad_norm": 11.195146560668945, "learning_rate": 0.000192553928109174, "loss": 1.3371, "step": 177 }, { "epoch": 0.034905382880674574, "grad_norm": 6.352316856384277, "learning_rate": 0.00019246562694320255, "loss": 2.543, "step": 178 }, { "epoch": 0.03510148053730758, "grad_norm": 4.030996799468994, "learning_rate": 0.00019237682577819137, "loss": 0.9273, "step": 179 }, { "epoch": 0.03529757819394058, "grad_norm": 6.009339809417725, "learning_rate": 0.00019228752509432417, "loss": 2.1444, "step": 180 }, { "epoch": 0.035493675850573586, "grad_norm": 14.911331176757812, "learning_rate": 0.00019219772537448597, "loss": 1.5989, "step": 181 }, { "epoch": 0.03568977350720659, "grad_norm": 6.454592227935791, "learning_rate": 0.00019210742710426012, "loss": 1.0608, "step": 182 }, { "epoch": 0.035885871163839594, "grad_norm": 11.354242324829102, "learning_rate": 0.00019201663077192586, "loss": 1.7558, "step": 183 }, { "epoch": 0.0360819688204726, "grad_norm": 5.804329872131348, "learning_rate": 0.0001919253368684557, "loss": 1.411, "step": 184 }, { "epoch": 0.036278066477105596, "grad_norm": 6.735583305358887, "learning_rate": 0.00019183354588751271, "loss": 2.4473, "step": 185 }, { "epoch": 0.0364741641337386, "grad_norm": 9.169321060180664, "learning_rate": 0.00019174125832544786, "loss": 1.8947, "step": 186 }, { "epoch": 0.036670261790371604, "grad_norm": 3.465175151824951, "learning_rate": 0.0001916484746812973, "loss": 1.1306, "step": 187 }, { "epoch": 0.03686635944700461, "grad_norm": 10.71740436553955, "learning_rate": 0.0001915551954567797, "loss": 1.0832, "step": 188 }, { "epoch": 0.03706245710363761, "grad_norm": 10.946560859680176, "learning_rate": 0.0001914614211562936, "loss": 1.9391, "step": 189 }, { "epoch": 0.03725855476027062, "grad_norm": 6.847762584686279, "learning_rate": 0.0001913671522869145, "loss": 2.197, "step": 190 }, { "epoch": 0.03745465241690362, "grad_norm": 13.330089569091797, "learning_rate": 0.00019127238935839235, "loss": 2.3539, "step": 191 }, { "epoch": 0.03765075007353662, "grad_norm": 22.94158172607422, "learning_rate": 0.00019117713288314863, "loss": 2.868, "step": 192 }, { "epoch": 0.03784684773016962, "grad_norm": 7.080881595611572, "learning_rate": 0.00019108138337627358, "loss": 1.7925, "step": 193 }, { "epoch": 0.038042945386802626, "grad_norm": 4.726489067077637, "learning_rate": 0.00019098514135552357, "loss": 1.008, "step": 194 }, { "epoch": 0.03823904304343563, "grad_norm": 5.7414870262146, "learning_rate": 0.00019088840734131807, "loss": 0.8934, "step": 195 }, { "epoch": 0.038435140700068635, "grad_norm": 6.781312465667725, "learning_rate": 0.00019079118185673705, "loss": 1.6637, "step": 196 }, { "epoch": 0.03863123835670164, "grad_norm": 12.264861106872559, "learning_rate": 0.00019069346542751803, "loss": 1.6055, "step": 197 }, { "epoch": 0.03882733601333464, "grad_norm": 8.975689888000488, "learning_rate": 0.00019059525858205323, "loss": 2.6467, "step": 198 }, { "epoch": 0.03902343366996765, "grad_norm": 9.027484893798828, "learning_rate": 0.0001904965618513868, "loss": 1.9813, "step": 199 }, { "epoch": 0.039219531326600644, "grad_norm": 9.491011619567871, "learning_rate": 0.0001903973757692119, "loss": 1.4225, "step": 200 }, { "epoch": 0.03941562898323365, "grad_norm": 5.202075481414795, "learning_rate": 0.00019029770087186773, "loss": 1.3524, "step": 201 }, { "epoch": 0.03961172663986665, "grad_norm": 11.139556884765625, "learning_rate": 0.00019019753769833678, "loss": 2.0723, "step": 202 }, { "epoch": 0.03980782429649966, "grad_norm": 11.383284568786621, "learning_rate": 0.0001900968867902419, "loss": 1.5843, "step": 203 }, { "epoch": 0.04000392195313266, "grad_norm": 7.112687110900879, "learning_rate": 0.00018999574869184324, "loss": 1.3899, "step": 204 }, { "epoch": 0.040200019609765665, "grad_norm": 13.50672721862793, "learning_rate": 0.00018989412395003537, "loss": 1.7484, "step": 205 }, { "epoch": 0.04039611726639867, "grad_norm": 13.107057571411133, "learning_rate": 0.00018979201311434434, "loss": 1.6412, "step": 206 }, { "epoch": 0.040592214923031666, "grad_norm": 11.221402168273926, "learning_rate": 0.0001896894167369248, "loss": 2.608, "step": 207 }, { "epoch": 0.04078831257966467, "grad_norm": 4.945010662078857, "learning_rate": 0.0001895863353725568, "loss": 1.3582, "step": 208 }, { "epoch": 0.040984410236297675, "grad_norm": 15.066801071166992, "learning_rate": 0.00018948276957864299, "loss": 1.7296, "step": 209 }, { "epoch": 0.04118050789293068, "grad_norm": 7.497617244720459, "learning_rate": 0.0001893787199152055, "loss": 1.4961, "step": 210 }, { "epoch": 0.04137660554956368, "grad_norm": 4.299473762512207, "learning_rate": 0.00018927418694488296, "loss": 1.7403, "step": 211 }, { "epoch": 0.04157270320619669, "grad_norm": 6.988886833190918, "learning_rate": 0.00018916917123292738, "loss": 2.6546, "step": 212 }, { "epoch": 0.04176880086282969, "grad_norm": 9.461721420288086, "learning_rate": 0.00018906367334720124, "loss": 2.067, "step": 213 }, { "epoch": 0.041964898519462696, "grad_norm": 4.021213054656982, "learning_rate": 0.0001889576938581742, "loss": 1.0274, "step": 214 }, { "epoch": 0.04216099617609569, "grad_norm": 7.254751205444336, "learning_rate": 0.00018885123333892026, "loss": 1.7091, "step": 215 }, { "epoch": 0.0423570938327287, "grad_norm": 10.783350944519043, "learning_rate": 0.00018874429236511448, "loss": 1.3779, "step": 216 }, { "epoch": 0.0425531914893617, "grad_norm": 16.38484764099121, "learning_rate": 0.00018863687151503, "loss": 2.8516, "step": 217 }, { "epoch": 0.042749289145994705, "grad_norm": 5.848017692565918, "learning_rate": 0.00018852897136953473, "loss": 1.6814, "step": 218 }, { "epoch": 0.04294538680262771, "grad_norm": 5.781267166137695, "learning_rate": 0.00018842059251208845, "loss": 2.1672, "step": 219 }, { "epoch": 0.043141484459260714, "grad_norm": 6.244543552398682, "learning_rate": 0.00018831173552873946, "loss": 0.6934, "step": 220 }, { "epoch": 0.04333758211589372, "grad_norm": 6.526815414428711, "learning_rate": 0.0001882024010081215, "loss": 1.6071, "step": 221 }, { "epoch": 0.043533679772526715, "grad_norm": 6.064664363861084, "learning_rate": 0.00018809258954145052, "loss": 1.8964, "step": 222 }, { "epoch": 0.04372977742915972, "grad_norm": 6.710343837738037, "learning_rate": 0.0001879823017225215, "loss": 2.063, "step": 223 }, { "epoch": 0.04392587508579272, "grad_norm": 9.060622215270996, "learning_rate": 0.00018787153814770537, "loss": 1.568, "step": 224 }, { "epoch": 0.04412197274242573, "grad_norm": 4.686062335968018, "learning_rate": 0.00018776029941594552, "loss": 1.1178, "step": 225 }, { "epoch": 0.04431807039905873, "grad_norm": 6.257881164550781, "learning_rate": 0.00018764858612875472, "loss": 2.1195, "step": 226 }, { "epoch": 0.044514168055691736, "grad_norm": 9.814234733581543, "learning_rate": 0.00018753639889021196, "loss": 1.1679, "step": 227 }, { "epoch": 0.04471026571232474, "grad_norm": 13.946937561035156, "learning_rate": 0.00018742373830695898, "loss": 1.6899, "step": 228 }, { "epoch": 0.044906363368957744, "grad_norm": 6.277964115142822, "learning_rate": 0.0001873106049881971, "loss": 1.5076, "step": 229 }, { "epoch": 0.04510246102559074, "grad_norm": 14.092109680175781, "learning_rate": 0.00018719699954568398, "loss": 1.9726, "step": 230 }, { "epoch": 0.045298558682223745, "grad_norm": 6.5708231925964355, "learning_rate": 0.00018708292259373015, "loss": 0.6623, "step": 231 }, { "epoch": 0.04549465633885675, "grad_norm": 6.839974880218506, "learning_rate": 0.00018696837474919582, "loss": 1.8836, "step": 232 }, { "epoch": 0.045690753995489754, "grad_norm": 17.919775009155273, "learning_rate": 0.00018685335663148753, "loss": 2.2343, "step": 233 }, { "epoch": 0.04588685165212276, "grad_norm": 7.140262603759766, "learning_rate": 0.00018673786886255476, "loss": 1.9653, "step": 234 }, { "epoch": 0.04608294930875576, "grad_norm": 11.049707412719727, "learning_rate": 0.00018662191206688658, "loss": 1.8658, "step": 235 }, { "epoch": 0.046279046965388766, "grad_norm": 5.617869853973389, "learning_rate": 0.00018650548687150823, "loss": 1.862, "step": 236 }, { "epoch": 0.04647514462202176, "grad_norm": 6.494004726409912, "learning_rate": 0.00018638859390597792, "loss": 3.787, "step": 237 }, { "epoch": 0.04667124227865477, "grad_norm": 17.628082275390625, "learning_rate": 0.00018627123380238314, "loss": 1.9129, "step": 238 }, { "epoch": 0.04686733993528777, "grad_norm": 5.634758949279785, "learning_rate": 0.0001861534071953374, "loss": 1.4702, "step": 239 }, { "epoch": 0.047063437591920776, "grad_norm": 20.99329948425293, "learning_rate": 0.00018603511472197685, "loss": 2.2196, "step": 240 }, { "epoch": 0.04725953524855378, "grad_norm": 6.650235652923584, "learning_rate": 0.00018591635702195673, "loss": 0.986, "step": 241 }, { "epoch": 0.047455632905186784, "grad_norm": 4.837536334991455, "learning_rate": 0.00018579713473744795, "loss": 1.2033, "step": 242 }, { "epoch": 0.04765173056181979, "grad_norm": 9.390655517578125, "learning_rate": 0.00018567744851313362, "loss": 1.7356, "step": 243 }, { "epoch": 0.04784782821845279, "grad_norm": 8.881529808044434, "learning_rate": 0.0001855572989962056, "loss": 1.0063, "step": 244 }, { "epoch": 0.04804392587508579, "grad_norm": 6.325936317443848, "learning_rate": 0.00018543668683636085, "loss": 1.5957, "step": 245 }, { "epoch": 0.048240023531718794, "grad_norm": 10.866408348083496, "learning_rate": 0.0001853156126857981, "loss": 2.0472, "step": 246 }, { "epoch": 0.0484361211883518, "grad_norm": 6.741912364959717, "learning_rate": 0.00018519407719921427, "loss": 1.6029, "step": 247 }, { "epoch": 0.0486322188449848, "grad_norm": 5.979978561401367, "learning_rate": 0.00018507208103380092, "loss": 0.9353, "step": 248 }, { "epoch": 0.048828316501617806, "grad_norm": 6.375998020172119, "learning_rate": 0.00018494962484924058, "loss": 2.5973, "step": 249 }, { "epoch": 0.04902441415825081, "grad_norm": 8.528829574584961, "learning_rate": 0.00018482670930770342, "loss": 1.1618, "step": 250 }, { "epoch": 0.049220511814883815, "grad_norm": 7.874660491943359, "learning_rate": 0.0001847033350738435, "loss": 1.3043, "step": 251 }, { "epoch": 0.04941660947151681, "grad_norm": 4.421433925628662, "learning_rate": 0.00018457950281479513, "loss": 2.0768, "step": 252 }, { "epoch": 0.049612707128149816, "grad_norm": 8.347919464111328, "learning_rate": 0.00018445521320016944, "loss": 1.0983, "step": 253 }, { "epoch": 0.04980880478478282, "grad_norm": 6.713651657104492, "learning_rate": 0.00018433046690205068, "loss": 0.9891, "step": 254 }, { "epoch": 0.050004902441415824, "grad_norm": 12.359843254089355, "learning_rate": 0.0001842052645949925, "loss": 2.5271, "step": 255 }, { "epoch": 0.05020100009804883, "grad_norm": 4.7271199226379395, "learning_rate": 0.00018407960695601442, "loss": 1.394, "step": 256 }, { "epoch": 0.05039709775468183, "grad_norm": 5.810708522796631, "learning_rate": 0.0001839534946645981, "loss": 1.7354, "step": 257 }, { "epoch": 0.05059319541131484, "grad_norm": 28.575908660888672, "learning_rate": 0.00018382692840268367, "loss": 3.4793, "step": 258 }, { "epoch": 0.05078929306794784, "grad_norm": 5.775376319885254, "learning_rate": 0.00018369990885466617, "loss": 1.4695, "step": 259 }, { "epoch": 0.05098539072458084, "grad_norm": 7.531515121459961, "learning_rate": 0.0001835724367073916, "loss": 2.0772, "step": 260 }, { "epoch": 0.05118148838121384, "grad_norm": 5.099686145782471, "learning_rate": 0.00018344451265015348, "loss": 3.2733, "step": 261 }, { "epoch": 0.05137758603784685, "grad_norm": 5.558218479156494, "learning_rate": 0.00018331613737468887, "loss": 1.7578, "step": 262 }, { "epoch": 0.05157368369447985, "grad_norm": 7.837512016296387, "learning_rate": 0.00018318731157517478, "loss": 2.7413, "step": 263 }, { "epoch": 0.051769781351112855, "grad_norm": 9.013374328613281, "learning_rate": 0.00018305803594822448, "loss": 1.7722, "step": 264 }, { "epoch": 0.05196587900774586, "grad_norm": 7.108828067779541, "learning_rate": 0.00018292831119288348, "loss": 1.31, "step": 265 }, { "epoch": 0.05216197666437886, "grad_norm": 6.387202739715576, "learning_rate": 0.0001827981380106261, "loss": 1.0588, "step": 266 }, { "epoch": 0.05235807432101187, "grad_norm": 13.306784629821777, "learning_rate": 0.00018266751710535131, "loss": 2.6092, "step": 267 }, { "epoch": 0.052554171977644865, "grad_norm": 7.774232864379883, "learning_rate": 0.00018253644918337915, "loss": 1.2318, "step": 268 }, { "epoch": 0.05275026963427787, "grad_norm": 5.396208763122559, "learning_rate": 0.00018240493495344694, "loss": 1.2144, "step": 269 }, { "epoch": 0.05294636729091087, "grad_norm": 13.393839836120605, "learning_rate": 0.0001822729751267053, "loss": 2.4278, "step": 270 }, { "epoch": 0.05314246494754388, "grad_norm": 14.873156547546387, "learning_rate": 0.00018214057041671434, "loss": 2.4015, "step": 271 }, { "epoch": 0.05333856260417688, "grad_norm": 7.224187850952148, "learning_rate": 0.00018200772153943988, "loss": 1.578, "step": 272 }, { "epoch": 0.053534660260809885, "grad_norm": 16.03417205810547, "learning_rate": 0.00018187442921324958, "loss": 2.4634, "step": 273 }, { "epoch": 0.05373075791744289, "grad_norm": 6.284254550933838, "learning_rate": 0.00018174069415890888, "loss": 1.1995, "step": 274 }, { "epoch": 0.05392685557407589, "grad_norm": 5.986271858215332, "learning_rate": 0.00018160651709957736, "loss": 1.8718, "step": 275 }, { "epoch": 0.05412295323070889, "grad_norm": 13.663339614868164, "learning_rate": 0.00018147189876080463, "loss": 1.8664, "step": 276 }, { "epoch": 0.054319050887341895, "grad_norm": 8.843401908874512, "learning_rate": 0.0001813368398705265, "loss": 4.1059, "step": 277 }, { "epoch": 0.0545151485439749, "grad_norm": 6.658843994140625, "learning_rate": 0.00018120134115906096, "loss": 0.8269, "step": 278 }, { "epoch": 0.0547112462006079, "grad_norm": 14.153685569763184, "learning_rate": 0.0001810654033591044, "loss": 1.9537, "step": 279 }, { "epoch": 0.05490734385724091, "grad_norm": 5.585552215576172, "learning_rate": 0.00018092902720572745, "loss": 1.2531, "step": 280 }, { "epoch": 0.05510344151387391, "grad_norm": 6.414156436920166, "learning_rate": 0.00018079221343637113, "loss": 0.9456, "step": 281 }, { "epoch": 0.055299539170506916, "grad_norm": 6.674518585205078, "learning_rate": 0.00018065496279084283, "loss": 1.1899, "step": 282 }, { "epoch": 0.05549563682713991, "grad_norm": 16.878883361816406, "learning_rate": 0.00018051727601131227, "loss": 0.8761, "step": 283 }, { "epoch": 0.05569173448377292, "grad_norm": 31.99020004272461, "learning_rate": 0.0001803791538423076, "loss": 1.7235, "step": 284 }, { "epoch": 0.05588783214040592, "grad_norm": 4.821253776550293, "learning_rate": 0.0001802405970307112, "loss": 1.2902, "step": 285 }, { "epoch": 0.056083929797038926, "grad_norm": 7.2293267250061035, "learning_rate": 0.00018010160632575577, "loss": 1.0864, "step": 286 }, { "epoch": 0.05628002745367193, "grad_norm": 8.262434959411621, "learning_rate": 0.00017996218247902035, "loss": 1.6246, "step": 287 }, { "epoch": 0.056476125110304934, "grad_norm": 7.845222473144531, "learning_rate": 0.00017982232624442595, "loss": 1.7249, "step": 288 }, { "epoch": 0.05667222276693794, "grad_norm": 9.781503677368164, "learning_rate": 0.0001796820383782319, "loss": 1.9911, "step": 289 }, { "epoch": 0.056868320423570935, "grad_norm": 10.7435884475708, "learning_rate": 0.00017954131963903133, "loss": 1.4689, "step": 290 }, { "epoch": 0.05706441808020394, "grad_norm": 8.709835052490234, "learning_rate": 0.00017940017078774747, "loss": 2.3939, "step": 291 }, { "epoch": 0.057260515736836944, "grad_norm": 6.140249252319336, "learning_rate": 0.00017925859258762915, "loss": 2.0753, "step": 292 }, { "epoch": 0.05745661339346995, "grad_norm": 9.725993156433105, "learning_rate": 0.00017911658580424704, "loss": 1.5315, "step": 293 }, { "epoch": 0.05765271105010295, "grad_norm": 20.413393020629883, "learning_rate": 0.00017897415120548917, "loss": 2.4083, "step": 294 }, { "epoch": 0.057848808706735956, "grad_norm": 7.881735324859619, "learning_rate": 0.00017883128956155706, "loss": 3.1099, "step": 295 }, { "epoch": 0.05804490636336896, "grad_norm": 8.856921195983887, "learning_rate": 0.0001786880016449614, "loss": 0.9566, "step": 296 }, { "epoch": 0.058241004020001964, "grad_norm": 25.80061912536621, "learning_rate": 0.0001785442882305179, "loss": 2.6205, "step": 297 }, { "epoch": 0.05843710167663496, "grad_norm": 13.293787956237793, "learning_rate": 0.00017840015009534308, "loss": 1.7317, "step": 298 }, { "epoch": 0.058633199333267966, "grad_norm": 11.313176155090332, "learning_rate": 0.00017825558801885016, "loss": 1.5122, "step": 299 }, { "epoch": 0.05882929698990097, "grad_norm": 9.636054039001465, "learning_rate": 0.00017811060278274474, "loss": 2.1976, "step": 300 }, { "epoch": 0.059025394646533974, "grad_norm": 11.10985279083252, "learning_rate": 0.00017796519517102066, "loss": 2.5926, "step": 301 }, { "epoch": 0.05922149230316698, "grad_norm": 6.27458381652832, "learning_rate": 0.00017781936596995563, "loss": 2.5326, "step": 302 }, { "epoch": 0.05941758995979998, "grad_norm": 10.134029388427734, "learning_rate": 0.00017767311596810715, "loss": 1.4142, "step": 303 }, { "epoch": 0.059613687616432987, "grad_norm": 10.580477714538574, "learning_rate": 0.0001775264459563081, "loss": 2.3389, "step": 304 }, { "epoch": 0.059809785273065984, "grad_norm": 7.545969009399414, "learning_rate": 0.00017737935672766257, "loss": 2.3728, "step": 305 }, { "epoch": 0.06000588292969899, "grad_norm": 6.5107622146606445, "learning_rate": 0.00017723184907754154, "loss": 1.0087, "step": 306 }, { "epoch": 0.06020198058633199, "grad_norm": 10.663747787475586, "learning_rate": 0.00017708392380357845, "loss": 2.2001, "step": 307 }, { "epoch": 0.060398078242964996, "grad_norm": 10.211285591125488, "learning_rate": 0.0001769355817056651, "loss": 2.2537, "step": 308 }, { "epoch": 0.060594175899598, "grad_norm": 8.512606620788574, "learning_rate": 0.00017678682358594728, "loss": 1.0687, "step": 309 }, { "epoch": 0.060790273556231005, "grad_norm": 10.64330768585205, "learning_rate": 0.0001766376502488202, "loss": 2.4672, "step": 310 }, { "epoch": 0.06098637121286401, "grad_norm": 9.972363471984863, "learning_rate": 0.0001764880625009245, "loss": 2.2277, "step": 311 }, { "epoch": 0.06118246886949701, "grad_norm": 10.01198959350586, "learning_rate": 0.0001763380611511416, "loss": 1.906, "step": 312 }, { "epoch": 0.06137856652613001, "grad_norm": 11.719582557678223, "learning_rate": 0.00017618764701058949, "loss": 2.4849, "step": 313 }, { "epoch": 0.061574664182763014, "grad_norm": 9.08901309967041, "learning_rate": 0.0001760368208926182, "loss": 1.4476, "step": 314 }, { "epoch": 0.06177076183939602, "grad_norm": 11.74946403503418, "learning_rate": 0.00017588558361280557, "loss": 1.3607, "step": 315 }, { "epoch": 0.06196685949602902, "grad_norm": 6.549604892730713, "learning_rate": 0.00017573393598895276, "loss": 2.066, "step": 316 }, { "epoch": 0.06216295715266203, "grad_norm": 5.160909652709961, "learning_rate": 0.00017558187884107978, "loss": 0.9799, "step": 317 }, { "epoch": 0.06235905480929503, "grad_norm": 11.00747299194336, "learning_rate": 0.00017542941299142112, "loss": 2.3193, "step": 318 }, { "epoch": 0.06255515246592804, "grad_norm": 4.467062473297119, "learning_rate": 0.00017527653926442135, "loss": 1.9208, "step": 319 }, { "epoch": 0.06275125012256104, "grad_norm": 11.554845809936523, "learning_rate": 0.00017512325848673043, "loss": 1.5197, "step": 320 }, { "epoch": 0.06294734777919404, "grad_norm": 6.626307964324951, "learning_rate": 0.0001749695714871996, "loss": 2.0222, "step": 321 }, { "epoch": 0.06314344543582705, "grad_norm": 5.357675075531006, "learning_rate": 0.00017481547909687658, "loss": 2.3264, "step": 322 }, { "epoch": 0.06333954309246005, "grad_norm": 9.451851844787598, "learning_rate": 0.00017466098214900124, "loss": 1.5699, "step": 323 }, { "epoch": 0.06353564074909304, "grad_norm": 5.937804698944092, "learning_rate": 0.00017450608147900106, "loss": 1.832, "step": 324 }, { "epoch": 0.06373173840572605, "grad_norm": 10.744606971740723, "learning_rate": 0.00017435077792448664, "loss": 1.4665, "step": 325 }, { "epoch": 0.06392783606235905, "grad_norm": 9.900199890136719, "learning_rate": 0.0001741950723252471, "loss": 1.4785, "step": 326 }, { "epoch": 0.06412393371899205, "grad_norm": 13.663531303405762, "learning_rate": 0.00017403896552324553, "loss": 0.9564, "step": 327 }, { "epoch": 0.06432003137562506, "grad_norm": 4.708930015563965, "learning_rate": 0.00017388245836261464, "loss": 0.8293, "step": 328 }, { "epoch": 0.06451612903225806, "grad_norm": 6.656357288360596, "learning_rate": 0.00017372555168965184, "loss": 2.2177, "step": 329 }, { "epoch": 0.06471222668889107, "grad_norm": 6.349363803863525, "learning_rate": 0.00017356824635281502, "loss": 3.3319, "step": 330 }, { "epoch": 0.06490832434552407, "grad_norm": 10.561055183410645, "learning_rate": 0.00017341054320271776, "loss": 1.919, "step": 331 }, { "epoch": 0.06510442200215708, "grad_norm": 10.805617332458496, "learning_rate": 0.00017325244309212475, "loss": 2.9898, "step": 332 }, { "epoch": 0.06530051965879008, "grad_norm": 9.67844009399414, "learning_rate": 0.0001730939468759472, "loss": 3.0727, "step": 333 }, { "epoch": 0.06549661731542308, "grad_norm": 9.487295150756836, "learning_rate": 0.00017293505541123833, "loss": 1.2741, "step": 334 }, { "epoch": 0.06569271497205609, "grad_norm": 5.027645111083984, "learning_rate": 0.00017277576955718847, "loss": 1.6962, "step": 335 }, { "epoch": 0.06588881262868909, "grad_norm": 6.2696685791015625, "learning_rate": 0.0001726160901751207, "loss": 1.4388, "step": 336 }, { "epoch": 0.0660849102853221, "grad_norm": 9.509906768798828, "learning_rate": 0.000172456018128486, "loss": 2.3173, "step": 337 }, { "epoch": 0.0662810079419551, "grad_norm": 5.882188320159912, "learning_rate": 0.00017229555428285864, "loss": 1.1832, "step": 338 }, { "epoch": 0.06647710559858809, "grad_norm": 8.857211112976074, "learning_rate": 0.00017213469950593156, "loss": 2.484, "step": 339 }, { "epoch": 0.0666732032552211, "grad_norm": 10.660289764404297, "learning_rate": 0.00017197345466751158, "loss": 2.5169, "step": 340 }, { "epoch": 0.0668693009118541, "grad_norm": 6.141770839691162, "learning_rate": 0.00017181182063951474, "loss": 1.601, "step": 341 }, { "epoch": 0.0668693009118541, "eval_loss": 0.4308730363845825, "eval_runtime": 77.758, "eval_samples_per_second": 27.624, "eval_steps_per_second": 13.812, "step": 341 }, { "epoch": 0.0670653985684871, "grad_norm": 9.668785095214844, "learning_rate": 0.00017164979829596165, "loss": 3.3758, "step": 342 }, { "epoch": 0.06726149622512011, "grad_norm": 5.6129255294799805, "learning_rate": 0.00017148738851297256, "loss": 1.3369, "step": 343 }, { "epoch": 0.06745759388175311, "grad_norm": 8.042420387268066, "learning_rate": 0.0001713245921687629, "loss": 1.918, "step": 344 }, { "epoch": 0.06765369153838612, "grad_norm": 6.536214351654053, "learning_rate": 0.00017116141014363837, "loss": 0.97, "step": 345 }, { "epoch": 0.06784978919501912, "grad_norm": 10.576717376708984, "learning_rate": 0.0001709978433199901, "loss": 2.0315, "step": 346 }, { "epoch": 0.06804588685165212, "grad_norm": 13.645353317260742, "learning_rate": 0.00017083389258229013, "loss": 2.3226, "step": 347 }, { "epoch": 0.06824198450828513, "grad_norm": 10.135254859924316, "learning_rate": 0.00017066955881708636, "loss": 2.4686, "step": 348 }, { "epoch": 0.06843808216491813, "grad_norm": 5.7091569900512695, "learning_rate": 0.0001705048429129979, "loss": 1.633, "step": 349 }, { "epoch": 0.06863417982155114, "grad_norm": 7.3178935050964355, "learning_rate": 0.0001703397457607103, "loss": 1.5411, "step": 350 }, { "epoch": 0.06883027747818414, "grad_norm": 10.689863204956055, "learning_rate": 0.0001701742682529706, "loss": 2.8238, "step": 351 }, { "epoch": 0.06902637513481714, "grad_norm": 4.083006381988525, "learning_rate": 0.00017000841128458265, "loss": 1.7831, "step": 352 }, { "epoch": 0.06922247279145015, "grad_norm": 4.592381477355957, "learning_rate": 0.0001698421757524021, "loss": 1.1135, "step": 353 }, { "epoch": 0.06941857044808314, "grad_norm": 9.44848346710205, "learning_rate": 0.00016967556255533174, "loss": 1.6301, "step": 354 }, { "epoch": 0.06961466810471614, "grad_norm": 7.35884428024292, "learning_rate": 0.0001695085725943165, "loss": 1.6776, "step": 355 }, { "epoch": 0.06981076576134915, "grad_norm": 7.0129475593566895, "learning_rate": 0.00016934120677233863, "loss": 1.1665, "step": 356 }, { "epoch": 0.07000686341798215, "grad_norm": 11.612259864807129, "learning_rate": 0.00016917346599441276, "loss": 0.9044, "step": 357 }, { "epoch": 0.07020296107461516, "grad_norm": 5.861693382263184, "learning_rate": 0.00016900535116758125, "loss": 1.5738, "step": 358 }, { "epoch": 0.07039905873124816, "grad_norm": 4.579985618591309, "learning_rate": 0.00016883686320090884, "loss": 1.6305, "step": 359 }, { "epoch": 0.07059515638788116, "grad_norm": 6.142041206359863, "learning_rate": 0.00016866800300547813, "loss": 0.6206, "step": 360 }, { "epoch": 0.07079125404451417, "grad_norm": 3.1343085765838623, "learning_rate": 0.0001684987714943845, "loss": 0.7649, "step": 361 }, { "epoch": 0.07098735170114717, "grad_norm": 5.473539352416992, "learning_rate": 0.00016832916958273118, "loss": 1.4839, "step": 362 }, { "epoch": 0.07118344935778018, "grad_norm": 8.52957534790039, "learning_rate": 0.00016815919818762427, "loss": 2.5437, "step": 363 }, { "epoch": 0.07137954701441318, "grad_norm": 6.140387058258057, "learning_rate": 0.00016798885822816786, "loss": 1.2282, "step": 364 }, { "epoch": 0.07157564467104618, "grad_norm": 9.70067024230957, "learning_rate": 0.000167818150625459, "loss": 2.3286, "step": 365 }, { "epoch": 0.07177174232767919, "grad_norm": 6.14666223526001, "learning_rate": 0.0001676470763025827, "loss": 1.3825, "step": 366 }, { "epoch": 0.07196783998431219, "grad_norm": 10.377222061157227, "learning_rate": 0.0001674756361846071, "loss": 1.9278, "step": 367 }, { "epoch": 0.0721639376409452, "grad_norm": 7.09204626083374, "learning_rate": 0.00016730383119857817, "loss": 1.2413, "step": 368 }, { "epoch": 0.07236003529757819, "grad_norm": 12.448009490966797, "learning_rate": 0.00016713166227351497, "loss": 1.7415, "step": 369 }, { "epoch": 0.07255613295421119, "grad_norm": 4.5050811767578125, "learning_rate": 0.00016695913034040454, "loss": 1.7608, "step": 370 }, { "epoch": 0.0727522306108442, "grad_norm": 8.113517761230469, "learning_rate": 0.00016678623633219677, "loss": 1.8336, "step": 371 }, { "epoch": 0.0729483282674772, "grad_norm": 7.662284851074219, "learning_rate": 0.00016661298118379948, "loss": 2.5928, "step": 372 }, { "epoch": 0.0731444259241102, "grad_norm": 7.647708892822266, "learning_rate": 0.00016643936583207337, "loss": 3.006, "step": 373 }, { "epoch": 0.07334052358074321, "grad_norm": 7.049048900604248, "learning_rate": 0.00016626539121582685, "loss": 1.9476, "step": 374 }, { "epoch": 0.07353662123737621, "grad_norm": 4.551039218902588, "learning_rate": 0.000166091058275811, "loss": 2.2753, "step": 375 }, { "epoch": 0.07373271889400922, "grad_norm": 8.680680274963379, "learning_rate": 0.00016591636795471455, "loss": 2.0624, "step": 376 }, { "epoch": 0.07392881655064222, "grad_norm": 10.324525833129883, "learning_rate": 0.0001657413211971587, "loss": 1.6293, "step": 377 }, { "epoch": 0.07412491420727522, "grad_norm": 5.324687957763672, "learning_rate": 0.0001655659189496921, "loss": 1.4987, "step": 378 }, { "epoch": 0.07432101186390823, "grad_norm": 11.023552894592285, "learning_rate": 0.00016539016216078557, "loss": 1.9108, "step": 379 }, { "epoch": 0.07451710952054123, "grad_norm": 6.6586594581604, "learning_rate": 0.00016521405178082722, "loss": 2.4198, "step": 380 }, { "epoch": 0.07471320717717424, "grad_norm": 9.818254470825195, "learning_rate": 0.0001650375887621171, "loss": 2.162, "step": 381 }, { "epoch": 0.07490930483380724, "grad_norm": 6.547168254852295, "learning_rate": 0.000164860774058862, "loss": 1.8307, "step": 382 }, { "epoch": 0.07510540249044025, "grad_norm": 7.256173610687256, "learning_rate": 0.00016468360862717066, "loss": 0.9432, "step": 383 }, { "epoch": 0.07530150014707324, "grad_norm": 10.348817825317383, "learning_rate": 0.00016450609342504813, "loss": 1.7824, "step": 384 }, { "epoch": 0.07549759780370624, "grad_norm": 14.728826522827148, "learning_rate": 0.00016432822941239096, "loss": 3.1737, "step": 385 }, { "epoch": 0.07569369546033924, "grad_norm": 7.894331932067871, "learning_rate": 0.00016415001755098175, "loss": 2.0515, "step": 386 }, { "epoch": 0.07588979311697225, "grad_norm": 7.474819183349609, "learning_rate": 0.00016397145880448416, "loss": 3.1276, "step": 387 }, { "epoch": 0.07608589077360525, "grad_norm": 6.2442626953125, "learning_rate": 0.00016379255413843754, "loss": 0.9294, "step": 388 }, { "epoch": 0.07628198843023826, "grad_norm": 8.229086875915527, "learning_rate": 0.0001636133045202517, "loss": 2.122, "step": 389 }, { "epoch": 0.07647808608687126, "grad_norm": 5.9484100341796875, "learning_rate": 0.0001634337109192019, "loss": 1.3164, "step": 390 }, { "epoch": 0.07667418374350427, "grad_norm": 6.490477561950684, "learning_rate": 0.0001632537743064233, "loss": 1.1598, "step": 391 }, { "epoch": 0.07687028140013727, "grad_norm": 36.34891128540039, "learning_rate": 0.000163073495654906, "loss": 2.4203, "step": 392 }, { "epoch": 0.07706637905677027, "grad_norm": 5.949688911437988, "learning_rate": 0.00016289287593948952, "loss": 2.0893, "step": 393 }, { "epoch": 0.07726247671340328, "grad_norm": 5.93861198425293, "learning_rate": 0.00016271191613685776, "loss": 1.829, "step": 394 }, { "epoch": 0.07745857437003628, "grad_norm": 8.41703987121582, "learning_rate": 0.00016253061722553355, "loss": 2.1544, "step": 395 }, { "epoch": 0.07765467202666929, "grad_norm": 8.92647647857666, "learning_rate": 0.00016234898018587337, "loss": 1.3316, "step": 396 }, { "epoch": 0.07785076968330229, "grad_norm": 4.4784698486328125, "learning_rate": 0.0001621670060000622, "loss": 1.0305, "step": 397 }, { "epoch": 0.0780468673399353, "grad_norm": 10.080864906311035, "learning_rate": 0.00016198469565210805, "loss": 2.1012, "step": 398 }, { "epoch": 0.07824296499656828, "grad_norm": 8.505258560180664, "learning_rate": 0.0001618020501278367, "loss": 2.2348, "step": 399 }, { "epoch": 0.07843906265320129, "grad_norm": 6.3116559982299805, "learning_rate": 0.00016161907041488635, "loss": 1.3726, "step": 400 }, { "epoch": 0.07863516030983429, "grad_norm": 7.111060619354248, "learning_rate": 0.00016143575750270233, "loss": 1.5926, "step": 401 }, { "epoch": 0.0788312579664673, "grad_norm": 5.481679916381836, "learning_rate": 0.0001612521123825317, "loss": 0.8062, "step": 402 }, { "epoch": 0.0790273556231003, "grad_norm": 9.399847984313965, "learning_rate": 0.00016106813604741782, "loss": 1.7107, "step": 403 }, { "epoch": 0.0792234532797333, "grad_norm": 7.784383296966553, "learning_rate": 0.00016088382949219524, "loss": 1.4449, "step": 404 }, { "epoch": 0.07941955093636631, "grad_norm": 10.075730323791504, "learning_rate": 0.000160699193713484, "loss": 2.3277, "step": 405 }, { "epoch": 0.07961564859299931, "grad_norm": 4.376564979553223, "learning_rate": 0.00016051422970968438, "loss": 1.3489, "step": 406 }, { "epoch": 0.07981174624963232, "grad_norm": 28.959577560424805, "learning_rate": 0.00016032893848097165, "loss": 0.9703, "step": 407 }, { "epoch": 0.08000784390626532, "grad_norm": 9.604166030883789, "learning_rate": 0.00016014332102929027, "loss": 2.195, "step": 408 }, { "epoch": 0.08020394156289833, "grad_norm": 5.949897289276123, "learning_rate": 0.00015995737835834906, "loss": 0.8828, "step": 409 }, { "epoch": 0.08040003921953133, "grad_norm": 9.921558380126953, "learning_rate": 0.00015977111147361507, "loss": 2.1569, "step": 410 }, { "epoch": 0.08059613687616433, "grad_norm": 9.299111366271973, "learning_rate": 0.00015958452138230877, "loss": 1.4095, "step": 411 }, { "epoch": 0.08079223453279734, "grad_norm": 8.806063652038574, "learning_rate": 0.00015939760909339823, "loss": 1.7785, "step": 412 }, { "epoch": 0.08098833218943034, "grad_norm": 5.1091108322143555, "learning_rate": 0.00015921037561759377, "loss": 1.8415, "step": 413 }, { "epoch": 0.08118442984606333, "grad_norm": 12.836872100830078, "learning_rate": 0.0001590228219673425, "loss": 1.8644, "step": 414 }, { "epoch": 0.08138052750269634, "grad_norm": 6.567033767700195, "learning_rate": 0.00015883494915682289, "loss": 1.9557, "step": 415 }, { "epoch": 0.08157662515932934, "grad_norm": 8.996990203857422, "learning_rate": 0.00015864675820193922, "loss": 1.888, "step": 416 }, { "epoch": 0.08177272281596235, "grad_norm": 7.494692802429199, "learning_rate": 0.00015845825012031605, "loss": 2.4514, "step": 417 }, { "epoch": 0.08196882047259535, "grad_norm": 5.421525478363037, "learning_rate": 0.0001582694259312928, "loss": 1.0314, "step": 418 }, { "epoch": 0.08216491812922835, "grad_norm": 25.4990291595459, "learning_rate": 0.0001580802866559183, "loss": 0.9981, "step": 419 }, { "epoch": 0.08236101578586136, "grad_norm": 7.055636405944824, "learning_rate": 0.00015789083331694506, "loss": 1.718, "step": 420 }, { "epoch": 0.08255711344249436, "grad_norm": 8.986098289489746, "learning_rate": 0.00015770106693882387, "loss": 1.3962, "step": 421 }, { "epoch": 0.08275321109912737, "grad_norm": 4.7463788986206055, "learning_rate": 0.0001575109885476983, "loss": 1.4039, "step": 422 }, { "epoch": 0.08294930875576037, "grad_norm": 6.555906295776367, "learning_rate": 0.00015732059917139912, "loss": 1.0908, "step": 423 }, { "epoch": 0.08314540641239337, "grad_norm": 7.233346462249756, "learning_rate": 0.00015712989983943863, "loss": 1.9737, "step": 424 }, { "epoch": 0.08334150406902638, "grad_norm": 4.797744274139404, "learning_rate": 0.0001569388915830053, "loss": 1.2191, "step": 425 }, { "epoch": 0.08353760172565938, "grad_norm": 6.303066253662109, "learning_rate": 0.000156747575434958, "loss": 1.8937, "step": 426 }, { "epoch": 0.08373369938229239, "grad_norm": 12.17996597290039, "learning_rate": 0.00015655595242982048, "loss": 4.1888, "step": 427 }, { "epoch": 0.08392979703892539, "grad_norm": 7.345923900604248, "learning_rate": 0.00015636402360377587, "loss": 1.0776, "step": 428 }, { "epoch": 0.08412589469555838, "grad_norm": 7.688333034515381, "learning_rate": 0.0001561717899946609, "loss": 0.9205, "step": 429 }, { "epoch": 0.08432199235219139, "grad_norm": 5.487595558166504, "learning_rate": 0.00015597925264196049, "loss": 1.4256, "step": 430 }, { "epoch": 0.08451809000882439, "grad_norm": 7.871713161468506, "learning_rate": 0.0001557864125868019, "loss": 2.3536, "step": 431 }, { "epoch": 0.0847141876654574, "grad_norm": 5.0042033195495605, "learning_rate": 0.00015559327087194942, "loss": 3.0954, "step": 432 }, { "epoch": 0.0849102853220904, "grad_norm": 6.770071506500244, "learning_rate": 0.0001553998285417983, "loss": 1.5016, "step": 433 }, { "epoch": 0.0851063829787234, "grad_norm": 6.884315490722656, "learning_rate": 0.00015520608664236949, "loss": 0.9995, "step": 434 }, { "epoch": 0.0853024806353564, "grad_norm": 7.716054439544678, "learning_rate": 0.00015501204622130377, "loss": 1.525, "step": 435 }, { "epoch": 0.08549857829198941, "grad_norm": 9.22014045715332, "learning_rate": 0.0001548177083278562, "loss": 1.7368, "step": 436 }, { "epoch": 0.08569467594862241, "grad_norm": 11.081851959228516, "learning_rate": 0.0001546230740128904, "loss": 2.5292, "step": 437 }, { "epoch": 0.08589077360525542, "grad_norm": 3.8471319675445557, "learning_rate": 0.0001544281443288728, "loss": 1.8473, "step": 438 }, { "epoch": 0.08608687126188842, "grad_norm": 7.554915904998779, "learning_rate": 0.00015423292032986695, "loss": 1.4886, "step": 439 }, { "epoch": 0.08628296891852143, "grad_norm": 5.459715843200684, "learning_rate": 0.00015403740307152805, "loss": 2.7721, "step": 440 }, { "epoch": 0.08647906657515443, "grad_norm": 5.094114780426025, "learning_rate": 0.00015384159361109698, "loss": 0.6022, "step": 441 }, { "epoch": 0.08667516423178744, "grad_norm": 4.084317684173584, "learning_rate": 0.00015364549300739467, "loss": 0.8871, "step": 442 }, { "epoch": 0.08687126188842044, "grad_norm": 3.810145378112793, "learning_rate": 0.0001534491023208164, "loss": 2.2058, "step": 443 }, { "epoch": 0.08706735954505343, "grad_norm": 9.405007362365723, "learning_rate": 0.000153252422613326, "loss": 2.414, "step": 444 }, { "epoch": 0.08726345720168643, "grad_norm": 8.687947273254395, "learning_rate": 0.00015305545494845023, "loss": 0.8809, "step": 445 }, { "epoch": 0.08745955485831944, "grad_norm": 6.396414756774902, "learning_rate": 0.00015285820039127293, "loss": 1.488, "step": 446 }, { "epoch": 0.08765565251495244, "grad_norm": 4.447728157043457, "learning_rate": 0.00015266066000842917, "loss": 1.1921, "step": 447 }, { "epoch": 0.08785175017158545, "grad_norm": 9.093881607055664, "learning_rate": 0.00015246283486809977, "loss": 1.6693, "step": 448 }, { "epoch": 0.08804784782821845, "grad_norm": 8.160088539123535, "learning_rate": 0.0001522647260400053, "loss": 2.9597, "step": 449 }, { "epoch": 0.08824394548485145, "grad_norm": 4.759898662567139, "learning_rate": 0.00015206633459540023, "loss": 1.121, "step": 450 }, { "epoch": 0.08844004314148446, "grad_norm": 6.537527561187744, "learning_rate": 0.0001518676616070674, "loss": 1.1842, "step": 451 }, { "epoch": 0.08863614079811746, "grad_norm": 3.6689343452453613, "learning_rate": 0.0001516687081493119, "loss": 1.3464, "step": 452 }, { "epoch": 0.08883223845475047, "grad_norm": 13.292744636535645, "learning_rate": 0.00015146947529795567, "loss": 2.8284, "step": 453 }, { "epoch": 0.08902833611138347, "grad_norm": 3.7040257453918457, "learning_rate": 0.0001512699641303312, "loss": 1.0223, "step": 454 }, { "epoch": 0.08922443376801648, "grad_norm": 8.529545783996582, "learning_rate": 0.00015107017572527616, "loss": 1.9594, "step": 455 }, { "epoch": 0.08942053142464948, "grad_norm": 11.990189552307129, "learning_rate": 0.00015087011116312718, "loss": 1.2631, "step": 456 }, { "epoch": 0.08961662908128248, "grad_norm": 5.5952534675598145, "learning_rate": 0.00015066977152571428, "loss": 0.9935, "step": 457 }, { "epoch": 0.08981272673791549, "grad_norm": 6.168736457824707, "learning_rate": 0.0001504691578963549, "loss": 0.9699, "step": 458 }, { "epoch": 0.09000882439454848, "grad_norm": 6.674266338348389, "learning_rate": 0.0001502682713598481, "loss": 1.3238, "step": 459 }, { "epoch": 0.09020492205118148, "grad_norm": 7.915739059448242, "learning_rate": 0.00015006711300246853, "loss": 1.4752, "step": 460 }, { "epoch": 0.09040101970781449, "grad_norm": 9.858468055725098, "learning_rate": 0.00014986568391196092, "loss": 1.7595, "step": 461 }, { "epoch": 0.09059711736444749, "grad_norm": 7.993055820465088, "learning_rate": 0.0001496639851775337, "loss": 1.6935, "step": 462 }, { "epoch": 0.0907932150210805, "grad_norm": 8.133615493774414, "learning_rate": 0.00014946201788985358, "loss": 1.3022, "step": 463 }, { "epoch": 0.0909893126777135, "grad_norm": 23.21516990661621, "learning_rate": 0.0001492597831410393, "loss": 1.9148, "step": 464 }, { "epoch": 0.0911854103343465, "grad_norm": 5.873632431030273, "learning_rate": 0.00014905728202465595, "loss": 0.8053, "step": 465 }, { "epoch": 0.09138150799097951, "grad_norm": 9.700281143188477, "learning_rate": 0.000148854515635709, "loss": 1.8187, "step": 466 }, { "epoch": 0.09157760564761251, "grad_norm": 9.361727714538574, "learning_rate": 0.00014865148507063833, "loss": 1.5871, "step": 467 }, { "epoch": 0.09177370330424552, "grad_norm": 6.021685600280762, "learning_rate": 0.00014844819142731223, "loss": 0.6632, "step": 468 }, { "epoch": 0.09196980096087852, "grad_norm": 8.407185554504395, "learning_rate": 0.0001482446358050217, "loss": 2.1806, "step": 469 }, { "epoch": 0.09216589861751152, "grad_norm": 6.5181450843811035, "learning_rate": 0.00014804081930447433, "loss": 1.1511, "step": 470 }, { "epoch": 0.09236199627414453, "grad_norm": 6.047727584838867, "learning_rate": 0.00014783674302778832, "loss": 1.4668, "step": 471 }, { "epoch": 0.09255809393077753, "grad_norm": 5.636353969573975, "learning_rate": 0.00014763240807848666, "loss": 1.2796, "step": 472 }, { "epoch": 0.09275419158741054, "grad_norm": 10.02710247039795, "learning_rate": 0.00014742781556149103, "loss": 1.3822, "step": 473 }, { "epoch": 0.09295028924404353, "grad_norm": 5.736359596252441, "learning_rate": 0.00014722296658311595, "loss": 1.6941, "step": 474 }, { "epoch": 0.09314638690067653, "grad_norm": 4.80971622467041, "learning_rate": 0.0001470178622510627, "loss": 2.3505, "step": 475 }, { "epoch": 0.09334248455730954, "grad_norm": 6.357244491577148, "learning_rate": 0.00014681250367441328, "loss": 3.0606, "step": 476 }, { "epoch": 0.09353858221394254, "grad_norm": 6.115423202514648, "learning_rate": 0.00014660689196362462, "loss": 1.4211, "step": 477 }, { "epoch": 0.09373467987057554, "grad_norm": 10.913143157958984, "learning_rate": 0.0001464010282305224, "loss": 1.6531, "step": 478 }, { "epoch": 0.09393077752720855, "grad_norm": 7.621870517730713, "learning_rate": 0.000146194913588295, "loss": 1.964, "step": 479 }, { "epoch": 0.09412687518384155, "grad_norm": 11.227121353149414, "learning_rate": 0.0001459885491514878, "loss": 1.639, "step": 480 }, { "epoch": 0.09432297284047456, "grad_norm": 4.943283557891846, "learning_rate": 0.00014578193603599662, "loss": 0.5043, "step": 481 }, { "epoch": 0.09451907049710756, "grad_norm": 13.006143569946289, "learning_rate": 0.0001455750753590622, "loss": 3.0382, "step": 482 }, { "epoch": 0.09471516815374056, "grad_norm": 6.181257247924805, "learning_rate": 0.00014536796823926386, "loss": 1.3281, "step": 483 }, { "epoch": 0.09491126581037357, "grad_norm": 4.05950403213501, "learning_rate": 0.0001451606157965136, "loss": 1.5701, "step": 484 }, { "epoch": 0.09510736346700657, "grad_norm": 7.532302379608154, "learning_rate": 0.0001449530191520499, "loss": 1.811, "step": 485 }, { "epoch": 0.09530346112363958, "grad_norm": 7.403939247131348, "learning_rate": 0.00014474517942843175, "loss": 1.8269, "step": 486 }, { "epoch": 0.09549955878027258, "grad_norm": 18.15042495727539, "learning_rate": 0.0001445370977495326, "loss": 2.4586, "step": 487 }, { "epoch": 0.09569565643690559, "grad_norm": 13.836175918579102, "learning_rate": 0.00014432877524053427, "loss": 1.7828, "step": 488 }, { "epoch": 0.09589175409353858, "grad_norm": 5.9050822257995605, "learning_rate": 0.0001441202130279208, "loss": 2.7068, "step": 489 }, { "epoch": 0.09608785175017158, "grad_norm": 7.173941612243652, "learning_rate": 0.0001439114122394724, "loss": 2.7424, "step": 490 }, { "epoch": 0.09628394940680458, "grad_norm": 17.888187408447266, "learning_rate": 0.0001437023740042594, "loss": 2.0253, "step": 491 }, { "epoch": 0.09648004706343759, "grad_norm": 6.492856025695801, "learning_rate": 0.00014349309945263606, "loss": 2.1604, "step": 492 }, { "epoch": 0.09667614472007059, "grad_norm": 11.6004638671875, "learning_rate": 0.00014328358971623455, "loss": 2.1492, "step": 493 }, { "epoch": 0.0968722423767036, "grad_norm": 6.262972354888916, "learning_rate": 0.00014307384592795872, "loss": 2.3007, "step": 494 }, { "epoch": 0.0970683400333366, "grad_norm": 6.768704891204834, "learning_rate": 0.00014286386922197805, "loss": 0.6546, "step": 495 }, { "epoch": 0.0972644376899696, "grad_norm": 4.372698783874512, "learning_rate": 0.0001426536607337215, "loss": 2.3647, "step": 496 }, { "epoch": 0.09746053534660261, "grad_norm": 12.042864799499512, "learning_rate": 0.00014244322159987145, "loss": 2.1466, "step": 497 }, { "epoch": 0.09765663300323561, "grad_norm": 5.374109745025635, "learning_rate": 0.0001422325529583573, "loss": 0.8599, "step": 498 }, { "epoch": 0.09785273065986862, "grad_norm": 7.459904193878174, "learning_rate": 0.00014202165594834963, "loss": 1.7119, "step": 499 }, { "epoch": 0.09804882831650162, "grad_norm": 8.25338363647461, "learning_rate": 0.00014181053171025392, "loss": 1.3014, "step": 500 }, { "epoch": 0.09824492597313463, "grad_norm": 8.594246864318848, "learning_rate": 0.00014159918138570424, "loss": 1.9332, "step": 501 }, { "epoch": 0.09844102362976763, "grad_norm": 4.802338600158691, "learning_rate": 0.00014138760611755727, "loss": 1.6591, "step": 502 }, { "epoch": 0.09863712128640063, "grad_norm": 5.049933433532715, "learning_rate": 0.00014117580704988612, "loss": 1.3722, "step": 503 }, { "epoch": 0.09883321894303362, "grad_norm": 9.438610076904297, "learning_rate": 0.00014096378532797393, "loss": 2.1888, "step": 504 }, { "epoch": 0.09902931659966663, "grad_norm": 11.144261360168457, "learning_rate": 0.00014075154209830792, "loss": 1.2526, "step": 505 }, { "epoch": 0.09922541425629963, "grad_norm": 12.456216812133789, "learning_rate": 0.0001405390785085731, "loss": 2.8251, "step": 506 }, { "epoch": 0.09942151191293264, "grad_norm": 13.78912353515625, "learning_rate": 0.00014032639570764593, "loss": 2.3312, "step": 507 }, { "epoch": 0.09961760956956564, "grad_norm": 22.588031768798828, "learning_rate": 0.00014011349484558847, "loss": 1.8865, "step": 508 }, { "epoch": 0.09981370722619864, "grad_norm": 5.357258319854736, "learning_rate": 0.00013990037707364166, "loss": 2.1296, "step": 509 }, { "epoch": 0.10000980488283165, "grad_norm": 13.293752670288086, "learning_rate": 0.00013968704354421952, "loss": 2.0479, "step": 510 }, { "epoch": 0.10020590253946465, "grad_norm": 6.764191150665283, "learning_rate": 0.00013947349541090274, "loss": 1.3908, "step": 511 }, { "epoch": 0.10040200019609766, "grad_norm": 5.701320648193359, "learning_rate": 0.00013925973382843246, "loss": 1.1343, "step": 512 }, { "epoch": 0.10059809785273066, "grad_norm": 6.728299140930176, "learning_rate": 0.0001390457599527039, "loss": 1.3219, "step": 513 }, { "epoch": 0.10079419550936367, "grad_norm": 9.094908714294434, "learning_rate": 0.00013883157494076046, "loss": 1.1282, "step": 514 }, { "epoch": 0.10099029316599667, "grad_norm": 7.321430683135986, "learning_rate": 0.00013861717995078708, "loss": 1.0628, "step": 515 }, { "epoch": 0.10118639082262967, "grad_norm": 5.976141452789307, "learning_rate": 0.00013840257614210414, "loss": 1.5372, "step": 516 }, { "epoch": 0.10138248847926268, "grad_norm": 7.5356926918029785, "learning_rate": 0.00013818776467516125, "loss": 2.0415, "step": 517 }, { "epoch": 0.10157858613589568, "grad_norm": 5.656033992767334, "learning_rate": 0.00013797274671153092, "loss": 1.9171, "step": 518 }, { "epoch": 0.10177468379252867, "grad_norm": 4.822179317474365, "learning_rate": 0.0001377575234139022, "loss": 1.1379, "step": 519 }, { "epoch": 0.10197078144916168, "grad_norm": 2.5650954246520996, "learning_rate": 0.0001375420959460745, "loss": 0.3984, "step": 520 }, { "epoch": 0.10216687910579468, "grad_norm": 6.997866630554199, "learning_rate": 0.00013732646547295126, "loss": 1.2149, "step": 521 }, { "epoch": 0.10236297676242768, "grad_norm": 9.925884246826172, "learning_rate": 0.00013711063316053368, "loss": 1.9542, "step": 522 }, { "epoch": 0.10255907441906069, "grad_norm": 7.913006782531738, "learning_rate": 0.00013689460017591432, "loss": 1.3162, "step": 523 }, { "epoch": 0.1027551720756937, "grad_norm": 8.868382453918457, "learning_rate": 0.00013667836768727092, "loss": 2.8749, "step": 524 }, { "epoch": 0.1029512697323267, "grad_norm": 7.418759346008301, "learning_rate": 0.00013646193686386, "loss": 1.2697, "step": 525 }, { "epoch": 0.1031473673889597, "grad_norm": 9.501752853393555, "learning_rate": 0.00013624530887601055, "loss": 1.1473, "step": 526 }, { "epoch": 0.1033434650455927, "grad_norm": 6.225274562835693, "learning_rate": 0.0001360284848951177, "loss": 1.0762, "step": 527 }, { "epoch": 0.10353956270222571, "grad_norm": 6.502389907836914, "learning_rate": 0.0001358114660936364, "loss": 1.383, "step": 528 }, { "epoch": 0.10373566035885871, "grad_norm": 17.53512954711914, "learning_rate": 0.00013559425364507508, "loss": 1.7968, "step": 529 }, { "epoch": 0.10393175801549172, "grad_norm": 6.0248122215271, "learning_rate": 0.00013537684872398927, "loss": 1.2982, "step": 530 }, { "epoch": 0.10412785567212472, "grad_norm": 4.621517658233643, "learning_rate": 0.00013515925250597537, "loss": 1.5811, "step": 531 }, { "epoch": 0.10432395332875773, "grad_norm": 6.690089702606201, "learning_rate": 0.00013494146616766406, "loss": 1.3376, "step": 532 }, { "epoch": 0.10452005098539073, "grad_norm": 6.418732643127441, "learning_rate": 0.00013472349088671418, "loss": 1.507, "step": 533 }, { "epoch": 0.10471614864202373, "grad_norm": 12.271800994873047, "learning_rate": 0.0001345053278418062, "loss": 1.6329, "step": 534 }, { "epoch": 0.10491224629865673, "grad_norm": 10.394608497619629, "learning_rate": 0.000134286978212636, "loss": 1.8793, "step": 535 }, { "epoch": 0.10510834395528973, "grad_norm": 5.690524101257324, "learning_rate": 0.00013406844317990826, "loss": 2.2874, "step": 536 }, { "epoch": 0.10530444161192273, "grad_norm": 4.313036918640137, "learning_rate": 0.00013384972392533034, "loss": 2.2905, "step": 537 }, { "epoch": 0.10550053926855574, "grad_norm": 6.115211009979248, "learning_rate": 0.0001336308216316056, "loss": 2.3146, "step": 538 }, { "epoch": 0.10569663692518874, "grad_norm": 7.43934965133667, "learning_rate": 0.0001334117374824274, "loss": 2.5837, "step": 539 }, { "epoch": 0.10589273458182175, "grad_norm": 8.397297859191895, "learning_rate": 0.00013319247266247225, "loss": 2.1631, "step": 540 }, { "epoch": 0.10608883223845475, "grad_norm": 5.273848056793213, "learning_rate": 0.00013297302835739376, "loss": 1.261, "step": 541 }, { "epoch": 0.10628492989508775, "grad_norm": 5.789434909820557, "learning_rate": 0.00013275340575381598, "loss": 1.8465, "step": 542 }, { "epoch": 0.10648102755172076, "grad_norm": 8.769668579101562, "learning_rate": 0.0001325336060393272, "loss": 1.3243, "step": 543 }, { "epoch": 0.10667712520835376, "grad_norm": 7.573061943054199, "learning_rate": 0.00013231363040247334, "loss": 1.243, "step": 544 }, { "epoch": 0.10687322286498677, "grad_norm": 9.284170150756836, "learning_rate": 0.00013209348003275165, "loss": 1.6408, "step": 545 }, { "epoch": 0.10706932052161977, "grad_norm": 4.322136878967285, "learning_rate": 0.00013187315612060415, "loss": 3.2532, "step": 546 }, { "epoch": 0.10726541817825277, "grad_norm": 7.865033149719238, "learning_rate": 0.00013165265985741142, "loss": 3.5798, "step": 547 }, { "epoch": 0.10746151583488578, "grad_norm": 3.1747426986694336, "learning_rate": 0.00013143199243548587, "loss": 1.1817, "step": 548 }, { "epoch": 0.10765761349151878, "grad_norm": 6.890042781829834, "learning_rate": 0.00013121115504806553, "loss": 2.0318, "step": 549 }, { "epoch": 0.10785371114815177, "grad_norm": 7.632523059844971, "learning_rate": 0.00013099014888930748, "loss": 1.3925, "step": 550 }, { "epoch": 0.10804980880478478, "grad_norm": 4.939361095428467, "learning_rate": 0.00013076897515428132, "loss": 1.3661, "step": 551 }, { "epoch": 0.10824590646141778, "grad_norm": 6.433002948760986, "learning_rate": 0.00013054763503896294, "loss": 0.9917, "step": 552 }, { "epoch": 0.10844200411805079, "grad_norm": 4.507363319396973, "learning_rate": 0.00013032612974022784, "loss": 0.9992, "step": 553 }, { "epoch": 0.10863810177468379, "grad_norm": 9.477548599243164, "learning_rate": 0.00013010446045584479, "loss": 1.1149, "step": 554 }, { "epoch": 0.1088341994313168, "grad_norm": 5.909473419189453, "learning_rate": 0.00012988262838446922, "loss": 2.0005, "step": 555 }, { "epoch": 0.1090302970879498, "grad_norm": 8.154961585998535, "learning_rate": 0.00012966063472563685, "loss": 2.1173, "step": 556 }, { "epoch": 0.1092263947445828, "grad_norm": 4.8719868659973145, "learning_rate": 0.00012943848067975718, "loss": 2.925, "step": 557 }, { "epoch": 0.1094224924012158, "grad_norm": 5.895204544067383, "learning_rate": 0.00012921616744810698, "loss": 2.0375, "step": 558 }, { "epoch": 0.10961859005784881, "grad_norm": 5.050942897796631, "learning_rate": 0.0001289936962328238, "loss": 1.6125, "step": 559 }, { "epoch": 0.10981468771448182, "grad_norm": 5.890807151794434, "learning_rate": 0.0001287710682368995, "loss": 0.991, "step": 560 }, { "epoch": 0.11001078537111482, "grad_norm": 5.316340446472168, "learning_rate": 0.00012854828466417364, "loss": 1.8846, "step": 561 }, { "epoch": 0.11020688302774782, "grad_norm": 14.81187629699707, "learning_rate": 0.00012832534671932715, "loss": 2.141, "step": 562 }, { "epoch": 0.11040298068438083, "grad_norm": 11.379783630371094, "learning_rate": 0.0001281022556078756, "loss": 2.9432, "step": 563 }, { "epoch": 0.11059907834101383, "grad_norm": 17.28523826599121, "learning_rate": 0.0001278790125361629, "loss": 2.4028, "step": 564 }, { "epoch": 0.11079517599764682, "grad_norm": 6.626383304595947, "learning_rate": 0.0001276556187113546, "loss": 1.1534, "step": 565 }, { "epoch": 0.11099127365427983, "grad_norm": 7.6215128898620605, "learning_rate": 0.00012743207534143144, "loss": 1.6464, "step": 566 }, { "epoch": 0.11118737131091283, "grad_norm": 8.348710060119629, "learning_rate": 0.00012720838363518286, "loss": 1.1823, "step": 567 }, { "epoch": 0.11138346896754583, "grad_norm": 5.307929039001465, "learning_rate": 0.00012698454480220039, "loss": 2.3759, "step": 568 }, { "epoch": 0.11157956662417884, "grad_norm": 4.690614223480225, "learning_rate": 0.00012676056005287106, "loss": 1.3698, "step": 569 }, { "epoch": 0.11177566428081184, "grad_norm": 5.7534871101379395, "learning_rate": 0.00012653643059837107, "loss": 1.9587, "step": 570 }, { "epoch": 0.11197176193744485, "grad_norm": 5.088260650634766, "learning_rate": 0.00012631215765065897, "loss": 1.0195, "step": 571 }, { "epoch": 0.11216785959407785, "grad_norm": 5.987957000732422, "learning_rate": 0.0001260877424224693, "loss": 2.3612, "step": 572 }, { "epoch": 0.11236395725071086, "grad_norm": 5.895928859710693, "learning_rate": 0.000125863186127306, "loss": 2.0393, "step": 573 }, { "epoch": 0.11256005490734386, "grad_norm": 6.654543876647949, "learning_rate": 0.00012563848997943568, "loss": 1.7165, "step": 574 }, { "epoch": 0.11275615256397686, "grad_norm": 13.193924903869629, "learning_rate": 0.00012541365519388138, "loss": 1.3969, "step": 575 }, { "epoch": 0.11295225022060987, "grad_norm": 5.471423149108887, "learning_rate": 0.00012518868298641564, "loss": 1.091, "step": 576 }, { "epoch": 0.11314834787724287, "grad_norm": 9.188228607177734, "learning_rate": 0.00012496357457355422, "loss": 2.0812, "step": 577 }, { "epoch": 0.11334444553387588, "grad_norm": 11.622628211975098, "learning_rate": 0.00012473833117254931, "loss": 2.1117, "step": 578 }, { "epoch": 0.11354054319050888, "grad_norm": 4.972383975982666, "learning_rate": 0.00012451295400138314, "loss": 1.3126, "step": 579 }, { "epoch": 0.11373664084714187, "grad_norm": 7.87300968170166, "learning_rate": 0.0001242874442787611, "loss": 3.2112, "step": 580 }, { "epoch": 0.11393273850377487, "grad_norm": 5.145363807678223, "learning_rate": 0.0001240618032241056, "loss": 1.2819, "step": 581 }, { "epoch": 0.11412883616040788, "grad_norm": 12.539934158325195, "learning_rate": 0.00012383603205754894, "loss": 2.2262, "step": 582 }, { "epoch": 0.11432493381704088, "grad_norm": 5.802661895751953, "learning_rate": 0.00012361013199992726, "loss": 1.9663, "step": 583 }, { "epoch": 0.11452103147367389, "grad_norm": 16.773561477661133, "learning_rate": 0.0001233841042727734, "loss": 2.342, "step": 584 }, { "epoch": 0.11471712913030689, "grad_norm": 6.163604259490967, "learning_rate": 0.0001231579500983108, "loss": 1.3009, "step": 585 }, { "epoch": 0.1149132267869399, "grad_norm": 7.877258777618408, "learning_rate": 0.00012293167069944645, "loss": 1.9388, "step": 586 }, { "epoch": 0.1151093244435729, "grad_norm": 13.269306182861328, "learning_rate": 0.00012270526729976465, "loss": 2.3028, "step": 587 }, { "epoch": 0.1153054221002059, "grad_norm": 9.610801696777344, "learning_rate": 0.00012247874112352009, "loss": 1.0249, "step": 588 }, { "epoch": 0.11550151975683891, "grad_norm": 11.20408821105957, "learning_rate": 0.00012225209339563145, "loss": 2.2807, "step": 589 }, { "epoch": 0.11569761741347191, "grad_norm": 13.390026092529297, "learning_rate": 0.00012202532534167463, "loss": 1.7141, "step": 590 }, { "epoch": 0.11589371507010492, "grad_norm": 4.4530229568481445, "learning_rate": 0.00012179843818787624, "loss": 1.4108, "step": 591 }, { "epoch": 0.11608981272673792, "grad_norm": 12.671486854553223, "learning_rate": 0.00012157143316110684, "loss": 1.6154, "step": 592 }, { "epoch": 0.11628591038337092, "grad_norm": 6.195436477661133, "learning_rate": 0.00012134431148887442, "loss": 3.4713, "step": 593 }, { "epoch": 0.11648200804000393, "grad_norm": 5.860586643218994, "learning_rate": 0.00012111707439931776, "loss": 2.0938, "step": 594 }, { "epoch": 0.11667810569663692, "grad_norm": 7.4623260498046875, "learning_rate": 0.00012088972312119964, "loss": 1.9674, "step": 595 }, { "epoch": 0.11687420335326992, "grad_norm": 7.368410110473633, "learning_rate": 0.00012066225888390042, "loss": 1.3425, "step": 596 }, { "epoch": 0.11707030100990293, "grad_norm": 7.527665138244629, "learning_rate": 0.00012043468291741116, "loss": 1.6366, "step": 597 }, { "epoch": 0.11726639866653593, "grad_norm": 6.0890793800354, "learning_rate": 0.00012020699645232721, "loss": 1.9652, "step": 598 }, { "epoch": 0.11746249632316894, "grad_norm": 6.350461959838867, "learning_rate": 0.00011997920071984133, "loss": 2.2867, "step": 599 }, { "epoch": 0.11765859397980194, "grad_norm": 6.444486618041992, "learning_rate": 0.00011975129695173719, "loss": 0.7242, "step": 600 }, { "epoch": 0.11785469163643494, "grad_norm": 7.270365238189697, "learning_rate": 0.00011952328638038261, "loss": 2.2346, "step": 601 }, { "epoch": 0.11805078929306795, "grad_norm": 5.961056709289551, "learning_rate": 0.00011929517023872298, "loss": 2.4926, "step": 602 }, { "epoch": 0.11824688694970095, "grad_norm": 5.347055435180664, "learning_rate": 0.00011906694976027455, "loss": 1.3998, "step": 603 }, { "epoch": 0.11844298460633396, "grad_norm": 11.233309745788574, "learning_rate": 0.0001188386261791177, "loss": 2.0185, "step": 604 }, { "epoch": 0.11863908226296696, "grad_norm": 4.813647747039795, "learning_rate": 0.0001186102007298904, "loss": 1.3932, "step": 605 }, { "epoch": 0.11883517991959996, "grad_norm": 5.281280040740967, "learning_rate": 0.00011838167464778142, "loss": 1.7445, "step": 606 }, { "epoch": 0.11903127757623297, "grad_norm": 8.755961418151855, "learning_rate": 0.00011815304916852372, "loss": 1.656, "step": 607 }, { "epoch": 0.11922737523286597, "grad_norm": 15.501289367675781, "learning_rate": 0.00011792432552838772, "loss": 2.0894, "step": 608 }, { "epoch": 0.11942347288949898, "grad_norm": 6.1117753982543945, "learning_rate": 0.00011769550496417466, "loss": 1.0473, "step": 609 }, { "epoch": 0.11961957054613197, "grad_norm": 9.669829368591309, "learning_rate": 0.00011746658871320983, "loss": 0.7436, "step": 610 }, { "epoch": 0.11981566820276497, "grad_norm": 10.253652572631836, "learning_rate": 0.00011723757801333601, "loss": 2.7502, "step": 611 }, { "epoch": 0.12001176585939798, "grad_norm": 7.084132671356201, "learning_rate": 0.00011700847410290667, "loss": 2.5537, "step": 612 }, { "epoch": 0.12020786351603098, "grad_norm": 5.936439037322998, "learning_rate": 0.0001167792782207793, "loss": 1.1551, "step": 613 }, { "epoch": 0.12040396117266398, "grad_norm": 7.053248405456543, "learning_rate": 0.0001165499916063087, "loss": 1.2488, "step": 614 }, { "epoch": 0.12060005882929699, "grad_norm": 7.141607284545898, "learning_rate": 0.00011632061549934036, "loss": 1.4092, "step": 615 }, { "epoch": 0.12079615648592999, "grad_norm": 7.723854064941406, "learning_rate": 0.00011609115114020363, "loss": 1.7669, "step": 616 }, { "epoch": 0.120992254142563, "grad_norm": 7.095085144042969, "learning_rate": 0.00011586159976970514, "loss": 1.5018, "step": 617 }, { "epoch": 0.121188351799196, "grad_norm": 6.469241142272949, "learning_rate": 0.00011563196262912193, "loss": 2.3377, "step": 618 }, { "epoch": 0.121384449455829, "grad_norm": 6.16892671585083, "learning_rate": 0.00011540224096019494, "loss": 2.0011, "step": 619 }, { "epoch": 0.12158054711246201, "grad_norm": 4.764594554901123, "learning_rate": 0.00011517243600512209, "loss": 0.8995, "step": 620 }, { "epoch": 0.12177664476909501, "grad_norm": 8.256946563720703, "learning_rate": 0.00011494254900655175, "loss": 1.5867, "step": 621 }, { "epoch": 0.12197274242572802, "grad_norm": 9.220061302185059, "learning_rate": 0.00011471258120757586, "loss": 1.0146, "step": 622 }, { "epoch": 0.12216884008236102, "grad_norm": 6.284097194671631, "learning_rate": 0.00011448253385172335, "loss": 0.8305, "step": 623 }, { "epoch": 0.12236493773899403, "grad_norm": 6.183596134185791, "learning_rate": 0.00011425240818295326, "loss": 0.7749, "step": 624 }, { "epoch": 0.12256103539562702, "grad_norm": 7.550481796264648, "learning_rate": 0.00011402220544564819, "loss": 1.3157, "step": 625 }, { "epoch": 0.12275713305226002, "grad_norm": 4.810642719268799, "learning_rate": 0.0001137919268846074, "loss": 1.6442, "step": 626 }, { "epoch": 0.12295323070889302, "grad_norm": 10.603952407836914, "learning_rate": 0.00011356157374504027, "loss": 1.7275, "step": 627 }, { "epoch": 0.12314932836552603, "grad_norm": 13.009543418884277, "learning_rate": 0.00011333114727255929, "loss": 2.3718, "step": 628 }, { "epoch": 0.12334542602215903, "grad_norm": 8.453059196472168, "learning_rate": 0.00011310064871317366, "loss": 1.6558, "step": 629 }, { "epoch": 0.12354152367879204, "grad_norm": 7.005093574523926, "learning_rate": 0.00011287007931328226, "loss": 1.9709, "step": 630 }, { "epoch": 0.12373762133542504, "grad_norm": 5.710804462432861, "learning_rate": 0.00011263944031966714, "loss": 1.8595, "step": 631 }, { "epoch": 0.12393371899205805, "grad_norm": 3.608022689819336, "learning_rate": 0.00011240873297948653, "loss": 1.7165, "step": 632 }, { "epoch": 0.12412981664869105, "grad_norm": 8.973870277404785, "learning_rate": 0.0001121779585402684, "loss": 1.4924, "step": 633 }, { "epoch": 0.12432591430532405, "grad_norm": 4.735976219177246, "learning_rate": 0.00011194711824990344, "loss": 1.2721, "step": 634 }, { "epoch": 0.12452201196195706, "grad_norm": 4.169306755065918, "learning_rate": 0.00011171621335663844, "loss": 0.718, "step": 635 }, { "epoch": 0.12471810961859006, "grad_norm": 14.129206657409668, "learning_rate": 0.00011148524510906956, "loss": 1.1904, "step": 636 }, { "epoch": 0.12491420727522307, "grad_norm": 12.76352310180664, "learning_rate": 0.00011125421475613554, "loss": 2.4872, "step": 637 }, { "epoch": 0.12511030493185607, "grad_norm": 8.636000633239746, "learning_rate": 0.00011102312354711091, "loss": 0.7629, "step": 638 }, { "epoch": 0.12530640258848907, "grad_norm": 6.22203254699707, "learning_rate": 0.00011079197273159925, "loss": 1.7372, "step": 639 }, { "epoch": 0.12550250024512208, "grad_norm": 4.235958576202393, "learning_rate": 0.0001105607635595266, "loss": 1.8661, "step": 640 }, { "epoch": 0.12569859790175508, "grad_norm": 5.598705768585205, "learning_rate": 0.00011032949728113437, "loss": 1.7918, "step": 641 }, { "epoch": 0.1258946955583881, "grad_norm": 9.218809127807617, "learning_rate": 0.00011009817514697291, "loss": 1.2968, "step": 642 }, { "epoch": 0.1260907932150211, "grad_norm": 7.068932056427002, "learning_rate": 0.00010986679840789451, "loss": 1.8485, "step": 643 }, { "epoch": 0.1262868908716541, "grad_norm": 3.891483783721924, "learning_rate": 0.00010963536831504684, "loss": 0.3918, "step": 644 }, { "epoch": 0.1264829885282871, "grad_norm": 5.959576606750488, "learning_rate": 0.00010940388611986592, "loss": 0.7057, "step": 645 }, { "epoch": 0.1266790861849201, "grad_norm": 11.663713455200195, "learning_rate": 0.00010917235307406966, "loss": 1.5916, "step": 646 }, { "epoch": 0.12687518384155308, "grad_norm": 15.05775260925293, "learning_rate": 0.00010894077042965083, "loss": 2.979, "step": 647 }, { "epoch": 0.12707128149818608, "grad_norm": 7.2112650871276855, "learning_rate": 0.00010870913943887049, "loss": 1.6293, "step": 648 }, { "epoch": 0.1272673791548191, "grad_norm": 13.646367073059082, "learning_rate": 0.00010847746135425102, "loss": 2.5277, "step": 649 }, { "epoch": 0.1274634768114521, "grad_norm": 6.29921293258667, "learning_rate": 0.0001082457374285696, "loss": 2.4015, "step": 650 }, { "epoch": 0.1276595744680851, "grad_norm": 6.337920188903809, "learning_rate": 0.00010801396891485114, "loss": 0.9928, "step": 651 }, { "epoch": 0.1278556721247181, "grad_norm": 7.32103157043457, "learning_rate": 0.00010778215706636177, "loss": 2.4466, "step": 652 }, { "epoch": 0.1280517697813511, "grad_norm": 10.154195785522461, "learning_rate": 0.00010755030313660188, "loss": 2.4858, "step": 653 }, { "epoch": 0.1282478674379841, "grad_norm": 7.129855632781982, "learning_rate": 0.00010731840837929946, "loss": 1.2546, "step": 654 }, { "epoch": 0.1284439650946171, "grad_norm": 5.403641223907471, "learning_rate": 0.00010708647404840319, "loss": 1.1017, "step": 655 }, { "epoch": 0.12864006275125012, "grad_norm": 5.554705619812012, "learning_rate": 0.00010685450139807584, "loss": 2.0789, "step": 656 }, { "epoch": 0.12883616040788312, "grad_norm": 7.935024738311768, "learning_rate": 0.00010662249168268736, "loss": 1.9944, "step": 657 }, { "epoch": 0.12903225806451613, "grad_norm": 4.453607559204102, "learning_rate": 0.00010639044615680809, "loss": 0.8442, "step": 658 }, { "epoch": 0.12922835572114913, "grad_norm": 13.252726554870605, "learning_rate": 0.00010615836607520202, "loss": 1.1374, "step": 659 }, { "epoch": 0.12942445337778213, "grad_norm": 4.426390171051025, "learning_rate": 0.00010592625269282, "loss": 1.0554, "step": 660 }, { "epoch": 0.12962055103441514, "grad_norm": 6.692317008972168, "learning_rate": 0.000105694107264793, "loss": 0.6669, "step": 661 }, { "epoch": 0.12981664869104814, "grad_norm": 6.067706108093262, "learning_rate": 0.00010546193104642519, "loss": 1.488, "step": 662 }, { "epoch": 0.13001274634768115, "grad_norm": 7.749240875244141, "learning_rate": 0.00010522972529318733, "loss": 1.8632, "step": 663 }, { "epoch": 0.13020884400431415, "grad_norm": 4.964468955993652, "learning_rate": 0.00010499749126070979, "loss": 0.5845, "step": 664 }, { "epoch": 0.13040494166094715, "grad_norm": 6.432995319366455, "learning_rate": 0.00010476523020477593, "loss": 1.1617, "step": 665 }, { "epoch": 0.13060103931758016, "grad_norm": 6.4099249839782715, "learning_rate": 0.00010453294338131519, "loss": 1.2821, "step": 666 }, { "epoch": 0.13079713697421316, "grad_norm": 5.935438632965088, "learning_rate": 0.0001043006320463964, "loss": 1.4453, "step": 667 }, { "epoch": 0.13099323463084617, "grad_norm": 10.963349342346191, "learning_rate": 0.00010406829745622085, "loss": 1.6318, "step": 668 }, { "epoch": 0.13118933228747917, "grad_norm": 9.157288551330566, "learning_rate": 0.00010383594086711567, "loss": 1.6981, "step": 669 }, { "epoch": 0.13138542994411218, "grad_norm": 17.15442657470703, "learning_rate": 0.00010360356353552687, "loss": 1.498, "step": 670 }, { "epoch": 0.13158152760074518, "grad_norm": 3.763289451599121, "learning_rate": 0.00010337116671801272, "loss": 2.425, "step": 671 }, { "epoch": 0.13177762525737818, "grad_norm": 8.91812801361084, "learning_rate": 0.00010313875167123672, "loss": 2.3129, "step": 672 }, { "epoch": 0.1319737229140112, "grad_norm": 23.093202590942383, "learning_rate": 0.00010290631965196109, "loss": 2.3172, "step": 673 }, { "epoch": 0.1321698205706442, "grad_norm": 6.487079620361328, "learning_rate": 0.00010267387191703972, "loss": 1.6108, "step": 674 }, { "epoch": 0.1323659182272772, "grad_norm": 3.9093501567840576, "learning_rate": 0.00010244140972341155, "loss": 0.8954, "step": 675 }, { "epoch": 0.1325620158839102, "grad_norm": 11.078289985656738, "learning_rate": 0.00010220893432809365, "loss": 2.0604, "step": 676 }, { "epoch": 0.13275811354054318, "grad_norm": 7.805664539337158, "learning_rate": 0.00010197644698817446, "loss": 1.834, "step": 677 }, { "epoch": 0.13295421119717618, "grad_norm": 6.838125705718994, "learning_rate": 0.00010174394896080713, "loss": 1.4876, "step": 678 }, { "epoch": 0.13315030885380919, "grad_norm": 3.9814906120300293, "learning_rate": 0.00010151144150320247, "loss": 1.3479, "step": 679 }, { "epoch": 0.1333464065104422, "grad_norm": 6.396786212921143, "learning_rate": 0.00010127892587262233, "loss": 1.7642, "step": 680 }, { "epoch": 0.1335425041670752, "grad_norm": 10.133893013000488, "learning_rate": 0.00010104640332637278, "loss": 0.8957, "step": 681 }, { "epoch": 0.1337386018237082, "grad_norm": 6.3815412521362305, "learning_rate": 0.00010081387512179729, "loss": 1.4104, "step": 682 }, { "epoch": 0.1337386018237082, "eval_loss": 0.41039586067199707, "eval_runtime": 77.8442, "eval_samples_per_second": 27.594, "eval_steps_per_second": 13.797, "step": 682 }, { "epoch": 0.1339346994803412, "grad_norm": 11.698119163513184, "learning_rate": 0.00010058134251626983, "loss": 2.1025, "step": 683 }, { "epoch": 0.1341307971369742, "grad_norm": 5.185816287994385, "learning_rate": 0.00010034880676718832, "loss": 1.0986, "step": 684 }, { "epoch": 0.1343268947936072, "grad_norm": 6.936627388000488, "learning_rate": 0.0001001162691319675, "loss": 1.3077, "step": 685 }, { "epoch": 0.13452299245024021, "grad_norm": 4.661152362823486, "learning_rate": 9.988373086803251e-05, "loss": 1.6312, "step": 686 }, { "epoch": 0.13471909010687322, "grad_norm": 4.227341651916504, "learning_rate": 9.965119323281174e-05, "loss": 0.6856, "step": 687 }, { "epoch": 0.13491518776350622, "grad_norm": 24.00637435913086, "learning_rate": 9.941865748373017e-05, "loss": 2.8267, "step": 688 }, { "epoch": 0.13511128542013923, "grad_norm": 4.701898574829102, "learning_rate": 9.918612487820273e-05, "loss": 1.7314, "step": 689 }, { "epoch": 0.13530738307677223, "grad_norm": 7.700753688812256, "learning_rate": 9.895359667362724e-05, "loss": 1.1831, "step": 690 }, { "epoch": 0.13550348073340523, "grad_norm": 5.045268535614014, "learning_rate": 9.87210741273777e-05, "loss": 2.6148, "step": 691 }, { "epoch": 0.13569957839003824, "grad_norm": 8.903031349182129, "learning_rate": 9.848855849679754e-05, "loss": 2.3442, "step": 692 }, { "epoch": 0.13589567604667124, "grad_norm": 6.869392395019531, "learning_rate": 9.82560510391929e-05, "loss": 1.7329, "step": 693 }, { "epoch": 0.13609177370330425, "grad_norm": 5.181933403015137, "learning_rate": 9.802355301182556e-05, "loss": 2.2152, "step": 694 }, { "epoch": 0.13628787135993725, "grad_norm": 9.427200317382812, "learning_rate": 9.779106567190642e-05, "loss": 0.9606, "step": 695 }, { "epoch": 0.13648396901657026, "grad_norm": 5.893560409545898, "learning_rate": 9.755859027658848e-05, "loss": 0.7759, "step": 696 }, { "epoch": 0.13668006667320326, "grad_norm": 3.864348888397217, "learning_rate": 9.73261280829603e-05, "loss": 0.714, "step": 697 }, { "epoch": 0.13687616432983626, "grad_norm": 7.077169418334961, "learning_rate": 9.709368034803892e-05, "loss": 1.9489, "step": 698 }, { "epoch": 0.13707226198646927, "grad_norm": 10.102705955505371, "learning_rate": 9.686124832876328e-05, "loss": 2.2884, "step": 699 }, { "epoch": 0.13726835964310227, "grad_norm": 6.154596328735352, "learning_rate": 9.662883328198731e-05, "loss": 1.1614, "step": 700 }, { "epoch": 0.13746445729973528, "grad_norm": 8.005276679992676, "learning_rate": 9.639643646447316e-05, "loss": 2.5097, "step": 701 }, { "epoch": 0.13766055495636828, "grad_norm": 12.45338249206543, "learning_rate": 9.616405913288438e-05, "loss": 0.8686, "step": 702 }, { "epoch": 0.13785665261300128, "grad_norm": 4.405421257019043, "learning_rate": 9.593170254377916e-05, "loss": 0.8499, "step": 703 }, { "epoch": 0.1380527502696343, "grad_norm": 9.178966522216797, "learning_rate": 9.569936795360363e-05, "loss": 1.7378, "step": 704 }, { "epoch": 0.1382488479262673, "grad_norm": 9.432720184326172, "learning_rate": 9.546705661868484e-05, "loss": 1.5712, "step": 705 }, { "epoch": 0.1384449455829003, "grad_norm": 5.602529048919678, "learning_rate": 9.523476979522411e-05, "loss": 1.8486, "step": 706 }, { "epoch": 0.1386410432395333, "grad_norm": 7.334858417510986, "learning_rate": 9.50025087392902e-05, "loss": 1.3093, "step": 707 }, { "epoch": 0.13883714089616628, "grad_norm": 16.70349884033203, "learning_rate": 9.477027470681269e-05, "loss": 1.8117, "step": 708 }, { "epoch": 0.13903323855279928, "grad_norm": 4.915411472320557, "learning_rate": 9.453806895357482e-05, "loss": 1.296, "step": 709 }, { "epoch": 0.1392293362094323, "grad_norm": 9.790595054626465, "learning_rate": 9.430589273520703e-05, "loss": 1.7763, "step": 710 }, { "epoch": 0.1394254338660653, "grad_norm": 5.754830837249756, "learning_rate": 9.407374730718e-05, "loss": 1.2865, "step": 711 }, { "epoch": 0.1396215315226983, "grad_norm": 6.5165534019470215, "learning_rate": 9.3841633924798e-05, "loss": 1.0498, "step": 712 }, { "epoch": 0.1398176291793313, "grad_norm": 4.821907043457031, "learning_rate": 9.360955384319195e-05, "loss": 1.1823, "step": 713 }, { "epoch": 0.1400137268359643, "grad_norm": 7.226667881011963, "learning_rate": 9.337750831731266e-05, "loss": 1.0176, "step": 714 }, { "epoch": 0.1402098244925973, "grad_norm": 5.560497760772705, "learning_rate": 9.314549860192417e-05, "loss": 1.1645, "step": 715 }, { "epoch": 0.1404059221492303, "grad_norm": 4.834329128265381, "learning_rate": 9.291352595159682e-05, "loss": 1.682, "step": 716 }, { "epoch": 0.14060201980586332, "grad_norm": 6.791970729827881, "learning_rate": 9.268159162070058e-05, "loss": 1.5408, "step": 717 }, { "epoch": 0.14079811746249632, "grad_norm": 5.209107875823975, "learning_rate": 9.244969686339813e-05, "loss": 0.6371, "step": 718 }, { "epoch": 0.14099421511912932, "grad_norm": 7.674073696136475, "learning_rate": 9.221784293363824e-05, "loss": 1.9635, "step": 719 }, { "epoch": 0.14119031277576233, "grad_norm": 13.054336547851562, "learning_rate": 9.198603108514888e-05, "loss": 1.2977, "step": 720 }, { "epoch": 0.14138641043239533, "grad_norm": 7.0277299880981445, "learning_rate": 9.175426257143045e-05, "loss": 1.3063, "step": 721 }, { "epoch": 0.14158250808902834, "grad_norm": 10.085116386413574, "learning_rate": 9.152253864574898e-05, "loss": 1.5895, "step": 722 }, { "epoch": 0.14177860574566134, "grad_norm": 11.588119506835938, "learning_rate": 9.129086056112955e-05, "loss": 1.6528, "step": 723 }, { "epoch": 0.14197470340229434, "grad_norm": 6.269704818725586, "learning_rate": 9.10592295703492e-05, "loss": 1.1952, "step": 724 }, { "epoch": 0.14217080105892735, "grad_norm": 5.680713176727295, "learning_rate": 9.082764692593039e-05, "loss": 2.322, "step": 725 }, { "epoch": 0.14236689871556035, "grad_norm": 6.14839506149292, "learning_rate": 9.059611388013408e-05, "loss": 2.6696, "step": 726 }, { "epoch": 0.14256299637219336, "grad_norm": 13.12667179107666, "learning_rate": 9.03646316849532e-05, "loss": 2.0335, "step": 727 }, { "epoch": 0.14275909402882636, "grad_norm": 7.358575820922852, "learning_rate": 9.01332015921055e-05, "loss": 0.8538, "step": 728 }, { "epoch": 0.14295519168545937, "grad_norm": 11.932861328125, "learning_rate": 8.990182485302713e-05, "loss": 1.4472, "step": 729 }, { "epoch": 0.14315128934209237, "grad_norm": 8.689105987548828, "learning_rate": 8.967050271886564e-05, "loss": 1.8314, "step": 730 }, { "epoch": 0.14334738699872537, "grad_norm": 5.939968585968018, "learning_rate": 8.943923644047342e-05, "loss": 0.8076, "step": 731 }, { "epoch": 0.14354348465535838, "grad_norm": 8.181587219238281, "learning_rate": 8.920802726840076e-05, "loss": 1.4664, "step": 732 }, { "epoch": 0.14373958231199138, "grad_norm": 7.449463844299316, "learning_rate": 8.897687645288914e-05, "loss": 1.0997, "step": 733 }, { "epoch": 0.14393567996862439, "grad_norm": 5.9800848960876465, "learning_rate": 8.874578524386447e-05, "loss": 0.9288, "step": 734 }, { "epoch": 0.1441317776252574, "grad_norm": 8.202332496643066, "learning_rate": 8.851475489093045e-05, "loss": 1.5658, "step": 735 }, { "epoch": 0.1443278752818904, "grad_norm": 4.058600425720215, "learning_rate": 8.828378664336158e-05, "loss": 1.1746, "step": 736 }, { "epoch": 0.1445239729385234, "grad_norm": 5.467014789581299, "learning_rate": 8.805288175009659e-05, "loss": 0.9683, "step": 737 }, { "epoch": 0.14472007059515637, "grad_norm": 8.572188377380371, "learning_rate": 8.782204145973162e-05, "loss": 1.3502, "step": 738 }, { "epoch": 0.14491616825178938, "grad_norm": 6.00791072845459, "learning_rate": 8.759126702051348e-05, "loss": 1.5011, "step": 739 }, { "epoch": 0.14511226590842238, "grad_norm": 7.9531989097595215, "learning_rate": 8.736055968033289e-05, "loss": 2.3835, "step": 740 }, { "epoch": 0.1453083635650554, "grad_norm": 7.390120029449463, "learning_rate": 8.712992068671774e-05, "loss": 1.1469, "step": 741 }, { "epoch": 0.1455044612216884, "grad_norm": 6.503965377807617, "learning_rate": 8.689935128682635e-05, "loss": 1.9587, "step": 742 }, { "epoch": 0.1457005588783214, "grad_norm": 7.921380996704102, "learning_rate": 8.666885272744073e-05, "loss": 1.9997, "step": 743 }, { "epoch": 0.1458966565349544, "grad_norm": 7.979074001312256, "learning_rate": 8.643842625495978e-05, "loss": 1.4087, "step": 744 }, { "epoch": 0.1460927541915874, "grad_norm": 5.653082847595215, "learning_rate": 8.620807311539259e-05, "loss": 1.3974, "step": 745 }, { "epoch": 0.1462888518482204, "grad_norm": 7.764763355255127, "learning_rate": 8.597779455435183e-05, "loss": 1.4269, "step": 746 }, { "epoch": 0.1464849495048534, "grad_norm": 11.947178840637207, "learning_rate": 8.574759181704676e-05, "loss": 2.4265, "step": 747 }, { "epoch": 0.14668104716148642, "grad_norm": 9.1488037109375, "learning_rate": 8.551746614827669e-05, "loss": 2.2206, "step": 748 }, { "epoch": 0.14687714481811942, "grad_norm": 5.008571147918701, "learning_rate": 8.528741879242415e-05, "loss": 2.1277, "step": 749 }, { "epoch": 0.14707324247475242, "grad_norm": 4.040747165679932, "learning_rate": 8.505745099344827e-05, "loss": 0.7625, "step": 750 }, { "epoch": 0.14726934013138543, "grad_norm": 8.826830863952637, "learning_rate": 8.482756399487793e-05, "loss": 1.1674, "step": 751 }, { "epoch": 0.14746543778801843, "grad_norm": 4.931051731109619, "learning_rate": 8.45977590398051e-05, "loss": 1.1831, "step": 752 }, { "epoch": 0.14766153544465144, "grad_norm": 5.232216835021973, "learning_rate": 8.436803737087806e-05, "loss": 1.0998, "step": 753 }, { "epoch": 0.14785763310128444, "grad_norm": 9.30781364440918, "learning_rate": 8.413840023029488e-05, "loss": 2.8714, "step": 754 }, { "epoch": 0.14805373075791745, "grad_norm": 6.524617671966553, "learning_rate": 8.390884885979639e-05, "loss": 1.3297, "step": 755 }, { "epoch": 0.14824982841455045, "grad_norm": 6.3461432456970215, "learning_rate": 8.367938450065967e-05, "loss": 1.6513, "step": 756 }, { "epoch": 0.14844592607118345, "grad_norm": 9.379528999328613, "learning_rate": 8.345000839369132e-05, "loss": 1.3968, "step": 757 }, { "epoch": 0.14864202372781646, "grad_norm": 9.503636360168457, "learning_rate": 8.322072177922073e-05, "loss": 1.0089, "step": 758 }, { "epoch": 0.14883812138444946, "grad_norm": 4.907662868499756, "learning_rate": 8.299152589709336e-05, "loss": 2.4323, "step": 759 }, { "epoch": 0.14903421904108247, "grad_norm": 7.7067084312438965, "learning_rate": 8.2762421986664e-05, "loss": 1.4224, "step": 760 }, { "epoch": 0.14923031669771547, "grad_norm": 5.226773738861084, "learning_rate": 8.253341128679018e-05, "loss": 2.0016, "step": 761 }, { "epoch": 0.14942641435434847, "grad_norm": 8.012574195861816, "learning_rate": 8.230449503582538e-05, "loss": 1.8039, "step": 762 }, { "epoch": 0.14962251201098148, "grad_norm": 6.705967426300049, "learning_rate": 8.207567447161231e-05, "loss": 2.5853, "step": 763 }, { "epoch": 0.14981860966761448, "grad_norm": 9.187888145446777, "learning_rate": 8.184695083147627e-05, "loss": 1.0529, "step": 764 }, { "epoch": 0.1500147073242475, "grad_norm": 4.542299270629883, "learning_rate": 8.16183253522186e-05, "loss": 1.2803, "step": 765 }, { "epoch": 0.1502108049808805, "grad_norm": 5.094648361206055, "learning_rate": 8.138979927010964e-05, "loss": 0.9061, "step": 766 }, { "epoch": 0.1504069026375135, "grad_norm": 13.718326568603516, "learning_rate": 8.116137382088233e-05, "loss": 2.0593, "step": 767 }, { "epoch": 0.15060300029414647, "grad_norm": 6.143137454986572, "learning_rate": 8.093305023972546e-05, "loss": 0.8494, "step": 768 }, { "epoch": 0.15079909795077948, "grad_norm": 6.3516154289245605, "learning_rate": 8.070482976127703e-05, "loss": 2.214, "step": 769 }, { "epoch": 0.15099519560741248, "grad_norm": 9.909046173095703, "learning_rate": 8.047671361961741e-05, "loss": 0.928, "step": 770 }, { "epoch": 0.15119129326404548, "grad_norm": 6.7013373374938965, "learning_rate": 8.024870304826286e-05, "loss": 1.2937, "step": 771 }, { "epoch": 0.1513873909206785, "grad_norm": 7.552654266357422, "learning_rate": 8.002079928015867e-05, "loss": 2.3988, "step": 772 }, { "epoch": 0.1515834885773115, "grad_norm": 4.806221961975098, "learning_rate": 7.97930035476728e-05, "loss": 1.9656, "step": 773 }, { "epoch": 0.1517795862339445, "grad_norm": 13.320466041564941, "learning_rate": 7.956531708258887e-05, "loss": 2.1333, "step": 774 }, { "epoch": 0.1519756838905775, "grad_norm": 7.802790641784668, "learning_rate": 7.933774111609964e-05, "loss": 0.766, "step": 775 }, { "epoch": 0.1521717815472105, "grad_norm": 9.146486282348633, "learning_rate": 7.911027687880037e-05, "loss": 1.0215, "step": 776 }, { "epoch": 0.1523678792038435, "grad_norm": 5.029911518096924, "learning_rate": 7.888292560068226e-05, "loss": 0.6491, "step": 777 }, { "epoch": 0.1525639768604765, "grad_norm": 16.61541175842285, "learning_rate": 7.865568851112559e-05, "loss": 2.3993, "step": 778 }, { "epoch": 0.15276007451710952, "grad_norm": 8.573732376098633, "learning_rate": 7.842856683889321e-05, "loss": 1.5773, "step": 779 }, { "epoch": 0.15295617217374252, "grad_norm": 7.390541076660156, "learning_rate": 7.820156181212379e-05, "loss": 0.6461, "step": 780 }, { "epoch": 0.15315226983037553, "grad_norm": 4.455470085144043, "learning_rate": 7.79746746583254e-05, "loss": 1.9557, "step": 781 }, { "epoch": 0.15334836748700853, "grad_norm": 9.000947952270508, "learning_rate": 7.774790660436858e-05, "loss": 2.5368, "step": 782 }, { "epoch": 0.15354446514364153, "grad_norm": 7.608001232147217, "learning_rate": 7.752125887647992e-05, "loss": 1.607, "step": 783 }, { "epoch": 0.15374056280027454, "grad_norm": 4.594658851623535, "learning_rate": 7.729473270023537e-05, "loss": 0.9085, "step": 784 }, { "epoch": 0.15393666045690754, "grad_norm": 4.750558853149414, "learning_rate": 7.706832930055357e-05, "loss": 1.8871, "step": 785 }, { "epoch": 0.15413275811354055, "grad_norm": 6.5569987297058105, "learning_rate": 7.684204990168925e-05, "loss": 1.4386, "step": 786 }, { "epoch": 0.15432885577017355, "grad_norm": 4.655035495758057, "learning_rate": 7.66158957272266e-05, "loss": 1.9621, "step": 787 }, { "epoch": 0.15452495342680655, "grad_norm": 9.556004524230957, "learning_rate": 7.638986800007277e-05, "loss": 0.9825, "step": 788 }, { "epoch": 0.15472105108343956, "grad_norm": 4.108066082000732, "learning_rate": 7.616396794245107e-05, "loss": 2.0709, "step": 789 }, { "epoch": 0.15491714874007256, "grad_norm": 7.94503116607666, "learning_rate": 7.593819677589445e-05, "loss": 1.1699, "step": 790 }, { "epoch": 0.15511324639670557, "grad_norm": 6.220666408538818, "learning_rate": 7.571255572123889e-05, "loss": 1.1849, "step": 791 }, { "epoch": 0.15530934405333857, "grad_norm": 6.629292964935303, "learning_rate": 7.54870459986169e-05, "loss": 1.3573, "step": 792 }, { "epoch": 0.15550544170997158, "grad_norm": 4.624551296234131, "learning_rate": 7.52616688274507e-05, "loss": 1.1043, "step": 793 }, { "epoch": 0.15570153936660458, "grad_norm": 8.356186866760254, "learning_rate": 7.503642542644581e-05, "loss": 1.3448, "step": 794 }, { "epoch": 0.15589763702323758, "grad_norm": 5.954531192779541, "learning_rate": 7.481131701358434e-05, "loss": 1.2289, "step": 795 }, { "epoch": 0.1560937346798706, "grad_norm": 8.886476516723633, "learning_rate": 7.458634480611864e-05, "loss": 2.0398, "step": 796 }, { "epoch": 0.1562898323365036, "grad_norm": 5.7918524742126465, "learning_rate": 7.436151002056433e-05, "loss": 1.4364, "step": 797 }, { "epoch": 0.15648592999313657, "grad_norm": 9.434075355529785, "learning_rate": 7.413681387269406e-05, "loss": 1.8601, "step": 798 }, { "epoch": 0.15668202764976957, "grad_norm": 5.067231178283691, "learning_rate": 7.391225757753068e-05, "loss": 1.414, "step": 799 }, { "epoch": 0.15687812530640258, "grad_norm": 6.89833927154541, "learning_rate": 7.368784234934104e-05, "loss": 1.6815, "step": 800 }, { "epoch": 0.15707422296303558, "grad_norm": 15.612439155578613, "learning_rate": 7.346356940162895e-05, "loss": 3.0339, "step": 801 }, { "epoch": 0.15727032061966859, "grad_norm": 5.933378219604492, "learning_rate": 7.323943994712895e-05, "loss": 1.6531, "step": 802 }, { "epoch": 0.1574664182763016, "grad_norm": 4.7962727546691895, "learning_rate": 7.301545519779964e-05, "loss": 1.8519, "step": 803 }, { "epoch": 0.1576625159329346, "grad_norm": 7.490151882171631, "learning_rate": 7.279161636481715e-05, "loss": 1.429, "step": 804 }, { "epoch": 0.1578586135895676, "grad_norm": 10.799139976501465, "learning_rate": 7.256792465856858e-05, "loss": 1.391, "step": 805 }, { "epoch": 0.1580547112462006, "grad_norm": 4.731945514678955, "learning_rate": 7.23443812886454e-05, "loss": 1.3097, "step": 806 }, { "epoch": 0.1582508089028336, "grad_norm": 6.669147491455078, "learning_rate": 7.21209874638371e-05, "loss": 2.5443, "step": 807 }, { "epoch": 0.1584469065594666, "grad_norm": 7.046877861022949, "learning_rate": 7.189774439212442e-05, "loss": 1.5809, "step": 808 }, { "epoch": 0.15864300421609961, "grad_norm": 5.848972320556641, "learning_rate": 7.167465328067289e-05, "loss": 1.6708, "step": 809 }, { "epoch": 0.15883910187273262, "grad_norm": 6.136227607727051, "learning_rate": 7.145171533582636e-05, "loss": 1.2122, "step": 810 }, { "epoch": 0.15903519952936562, "grad_norm": 4.7808709144592285, "learning_rate": 7.122893176310052e-05, "loss": 1.8965, "step": 811 }, { "epoch": 0.15923129718599863, "grad_norm": 4.156777381896973, "learning_rate": 7.100630376717621e-05, "loss": 1.2742, "step": 812 }, { "epoch": 0.15942739484263163, "grad_norm": 5.69835090637207, "learning_rate": 7.078383255189304e-05, "loss": 0.7664, "step": 813 }, { "epoch": 0.15962349249926464, "grad_norm": 9.707277297973633, "learning_rate": 7.056151932024282e-05, "loss": 1.3492, "step": 814 }, { "epoch": 0.15981959015589764, "grad_norm": 10.071001052856445, "learning_rate": 7.033936527436318e-05, "loss": 1.2432, "step": 815 }, { "epoch": 0.16001568781253064, "grad_norm": 6.86775541305542, "learning_rate": 7.011737161553081e-05, "loss": 1.7583, "step": 816 }, { "epoch": 0.16021178546916365, "grad_norm": 5.526086330413818, "learning_rate": 6.989553954415524e-05, "loss": 1.5891, "step": 817 }, { "epoch": 0.16040788312579665, "grad_norm": 26.293426513671875, "learning_rate": 6.967387025977215e-05, "loss": 1.7835, "step": 818 }, { "epoch": 0.16060398078242966, "grad_norm": 5.713183879852295, "learning_rate": 6.945236496103707e-05, "loss": 1.7831, "step": 819 }, { "epoch": 0.16080007843906266, "grad_norm": 7.669525146484375, "learning_rate": 6.92310248457187e-05, "loss": 1.8961, "step": 820 }, { "epoch": 0.16099617609569566, "grad_norm": 7.351653099060059, "learning_rate": 6.900985111069256e-05, "loss": 1.8533, "step": 821 }, { "epoch": 0.16119227375232867, "grad_norm": 6.009294509887695, "learning_rate": 6.878884495193448e-05, "loss": 1.7683, "step": 822 }, { "epoch": 0.16138837140896167, "grad_norm": 5.90309476852417, "learning_rate": 6.856800756451414e-05, "loss": 1.7174, "step": 823 }, { "epoch": 0.16158446906559468, "grad_norm": 6.903116703033447, "learning_rate": 6.83473401425886e-05, "loss": 1.6063, "step": 824 }, { "epoch": 0.16178056672222768, "grad_norm": 5.747986793518066, "learning_rate": 6.812684387939585e-05, "loss": 1.7352, "step": 825 }, { "epoch": 0.16197666437886069, "grad_norm": 4.83687686920166, "learning_rate": 6.79065199672484e-05, "loss": 1.0662, "step": 826 }, { "epoch": 0.1621727620354937, "grad_norm": 4.761837482452393, "learning_rate": 6.768636959752669e-05, "loss": 1.4324, "step": 827 }, { "epoch": 0.16236885969212667, "grad_norm": 4.991310119628906, "learning_rate": 6.746639396067283e-05, "loss": 0.6861, "step": 828 }, { "epoch": 0.16256495734875967, "grad_norm": 9.824933052062988, "learning_rate": 6.724659424618401e-05, "loss": 1.2527, "step": 829 }, { "epoch": 0.16276105500539267, "grad_norm": 9.043631553649902, "learning_rate": 6.702697164260626e-05, "loss": 1.2587, "step": 830 }, { "epoch": 0.16295715266202568, "grad_norm": 6.001374244689941, "learning_rate": 6.680752733752777e-05, "loss": 1.0727, "step": 831 }, { "epoch": 0.16315325031865868, "grad_norm": 7.015263557434082, "learning_rate": 6.658826251757264e-05, "loss": 0.963, "step": 832 }, { "epoch": 0.1633493479752917, "grad_norm": 16.830978393554688, "learning_rate": 6.63691783683944e-05, "loss": 1.3392, "step": 833 }, { "epoch": 0.1635454456319247, "grad_norm": 12.247193336486816, "learning_rate": 6.61502760746697e-05, "loss": 2.2754, "step": 834 }, { "epoch": 0.1637415432885577, "grad_norm": 3.836216449737549, "learning_rate": 6.593155682009176e-05, "loss": 2.4028, "step": 835 }, { "epoch": 0.1639376409451907, "grad_norm": 5.886529445648193, "learning_rate": 6.571302178736404e-05, "loss": 1.0402, "step": 836 }, { "epoch": 0.1641337386018237, "grad_norm": 5.377973556518555, "learning_rate": 6.549467215819378e-05, "loss": 1.4784, "step": 837 }, { "epoch": 0.1643298362584567, "grad_norm": 4.867081165313721, "learning_rate": 6.527650911328585e-05, "loss": 1.9024, "step": 838 }, { "epoch": 0.1645259339150897, "grad_norm": 8.56846809387207, "learning_rate": 6.505853383233596e-05, "loss": 1.2514, "step": 839 }, { "epoch": 0.16472203157172272, "grad_norm": 14.424854278564453, "learning_rate": 6.484074749402467e-05, "loss": 1.2373, "step": 840 }, { "epoch": 0.16491812922835572, "grad_norm": 5.4463067054748535, "learning_rate": 6.462315127601071e-05, "loss": 1.5188, "step": 841 }, { "epoch": 0.16511422688498872, "grad_norm": 5.752686500549316, "learning_rate": 6.440574635492493e-05, "loss": 1.0012, "step": 842 }, { "epoch": 0.16531032454162173, "grad_norm": 5.5835442543029785, "learning_rate": 6.418853390636364e-05, "loss": 1.8797, "step": 843 }, { "epoch": 0.16550642219825473, "grad_norm": 8.692926406860352, "learning_rate": 6.39715151048823e-05, "loss": 1.2265, "step": 844 }, { "epoch": 0.16570251985488774, "grad_norm": 5.524386882781982, "learning_rate": 6.375469112398946e-05, "loss": 0.7208, "step": 845 }, { "epoch": 0.16589861751152074, "grad_norm": 5.7656145095825195, "learning_rate": 6.353806313614001e-05, "loss": 1.2716, "step": 846 }, { "epoch": 0.16609471516815374, "grad_norm": 12.103293418884277, "learning_rate": 6.332163231272911e-05, "loss": 3.2642, "step": 847 }, { "epoch": 0.16629081282478675, "grad_norm": 22.349411010742188, "learning_rate": 6.310539982408568e-05, "loss": 2.3537, "step": 848 }, { "epoch": 0.16648691048141975, "grad_norm": 7.782928466796875, "learning_rate": 6.288936683946633e-05, "loss": 1.1905, "step": 849 }, { "epoch": 0.16668300813805276, "grad_norm": 6.026412487030029, "learning_rate": 6.267353452704876e-05, "loss": 1.1583, "step": 850 }, { "epoch": 0.16687910579468576, "grad_norm": 10.18653392791748, "learning_rate": 6.245790405392553e-05, "loss": 2.5888, "step": 851 }, { "epoch": 0.16707520345131877, "grad_norm": 6.149682998657227, "learning_rate": 6.224247658609781e-05, "loss": 1.0514, "step": 852 }, { "epoch": 0.16727130110795177, "grad_norm": 5.528243541717529, "learning_rate": 6.202725328846909e-05, "loss": 0.8576, "step": 853 }, { "epoch": 0.16746739876458477, "grad_norm": 7.682480812072754, "learning_rate": 6.181223532483875e-05, "loss": 1.5959, "step": 854 }, { "epoch": 0.16766349642121778, "grad_norm": 4.653623580932617, "learning_rate": 6.159742385789588e-05, "loss": 0.7513, "step": 855 }, { "epoch": 0.16785959407785078, "grad_norm": 18.469745635986328, "learning_rate": 6.138282004921295e-05, "loss": 3.1673, "step": 856 }, { "epoch": 0.1680556917344838, "grad_norm": 4.403334140777588, "learning_rate": 6.116842505923955e-05, "loss": 0.9741, "step": 857 }, { "epoch": 0.16825178939111676, "grad_norm": 5.073941230773926, "learning_rate": 6.095424004729611e-05, "loss": 0.8638, "step": 858 }, { "epoch": 0.16844788704774977, "grad_norm": 6.09031867980957, "learning_rate": 6.0740266171567606e-05, "loss": 0.7445, "step": 859 }, { "epoch": 0.16864398470438277, "grad_norm": 5.235655784606934, "learning_rate": 6.0526504589097254e-05, "loss": 1.79, "step": 860 }, { "epoch": 0.16884008236101578, "grad_norm": 5.974957466125488, "learning_rate": 6.031295645578049e-05, "loss": 2.638, "step": 861 }, { "epoch": 0.16903618001764878, "grad_norm": 4.034956932067871, "learning_rate": 6.009962292635838e-05, "loss": 1.2214, "step": 862 }, { "epoch": 0.16923227767428178, "grad_norm": 5.100616931915283, "learning_rate": 5.988650515441156e-05, "loss": 1.7694, "step": 863 }, { "epoch": 0.1694283753309148, "grad_norm": 6.878417491912842, "learning_rate": 5.967360429235407e-05, "loss": 1.2678, "step": 864 }, { "epoch": 0.1696244729875478, "grad_norm": 6.435969352722168, "learning_rate": 5.946092149142692e-05, "loss": 1.449, "step": 865 }, { "epoch": 0.1698205706441808, "grad_norm": 5.022701740264893, "learning_rate": 5.924845790169209e-05, "loss": 0.8444, "step": 866 }, { "epoch": 0.1700166683008138, "grad_norm": 4.173498153686523, "learning_rate": 5.903621467202608e-05, "loss": 1.089, "step": 867 }, { "epoch": 0.1702127659574468, "grad_norm": 6.396774768829346, "learning_rate": 5.88241929501139e-05, "loss": 1.1751, "step": 868 }, { "epoch": 0.1704088636140798, "grad_norm": 4.628440856933594, "learning_rate": 5.861239388244273e-05, "loss": 1.4271, "step": 869 }, { "epoch": 0.1706049612707128, "grad_norm": 13.596205711364746, "learning_rate": 5.8400818614295806e-05, "loss": 0.7325, "step": 870 }, { "epoch": 0.17080105892734582, "grad_norm": 8.975894927978516, "learning_rate": 5.818946828974607e-05, "loss": 1.8689, "step": 871 }, { "epoch": 0.17099715658397882, "grad_norm": 5.900057315826416, "learning_rate": 5.797834405165035e-05, "loss": 0.8738, "step": 872 }, { "epoch": 0.17119325424061183, "grad_norm": 6.28643798828125, "learning_rate": 5.776744704164272e-05, "loss": 2.4116, "step": 873 }, { "epoch": 0.17138935189724483, "grad_norm": 6.7851033210754395, "learning_rate": 5.7556778400128583e-05, "loss": 1.5964, "step": 874 }, { "epoch": 0.17158544955387783, "grad_norm": 12.063843727111816, "learning_rate": 5.7346339266278505e-05, "loss": 1.9227, "step": 875 }, { "epoch": 0.17178154721051084, "grad_norm": 6.9710469245910645, "learning_rate": 5.713613077802199e-05, "loss": 1.3143, "step": 876 }, { "epoch": 0.17197764486714384, "grad_norm": 10.355106353759766, "learning_rate": 5.6926154072041295e-05, "loss": 2.005, "step": 877 }, { "epoch": 0.17217374252377685, "grad_norm": 8.473671913146973, "learning_rate": 5.671641028376546e-05, "loss": 1.4054, "step": 878 }, { "epoch": 0.17236984018040985, "grad_norm": 6.026031494140625, "learning_rate": 5.650690054736396e-05, "loss": 0.9981, "step": 879 }, { "epoch": 0.17256593783704285, "grad_norm": 7.832981109619141, "learning_rate": 5.629762599574064e-05, "loss": 1.318, "step": 880 }, { "epoch": 0.17276203549367586, "grad_norm": 8.796941757202148, "learning_rate": 5.6088587760527656e-05, "loss": 1.7696, "step": 881 }, { "epoch": 0.17295813315030886, "grad_norm": 5.9997124671936035, "learning_rate": 5.587978697207927e-05, "loss": 1.5515, "step": 882 }, { "epoch": 0.17315423080694187, "grad_norm": 10.373340606689453, "learning_rate": 5.5671224759465775e-05, "loss": 1.735, "step": 883 }, { "epoch": 0.17335032846357487, "grad_norm": 6.619256973266602, "learning_rate": 5.546290225046744e-05, "loss": 1.3402, "step": 884 }, { "epoch": 0.17354642612020788, "grad_norm": 7.469152927398682, "learning_rate": 5.5254820571568325e-05, "loss": 1.0792, "step": 885 }, { "epoch": 0.17374252377684088, "grad_norm": 4.927844524383545, "learning_rate": 5.504698084795014e-05, "loss": 1.4485, "step": 886 }, { "epoch": 0.17393862143347388, "grad_norm": 4.025864601135254, "learning_rate": 5.483938420348642e-05, "loss": 0.9515, "step": 887 }, { "epoch": 0.17413471909010686, "grad_norm": 14.059304237365723, "learning_rate": 5.4632031760736146e-05, "loss": 1.9276, "step": 888 }, { "epoch": 0.17433081674673986, "grad_norm": 6.336188793182373, "learning_rate": 5.4424924640937824e-05, "loss": 1.7752, "step": 889 }, { "epoch": 0.17452691440337287, "grad_norm": 7.820261478424072, "learning_rate": 5.421806396400339e-05, "loss": 1.4182, "step": 890 }, { "epoch": 0.17472301206000587, "grad_norm": 3.8585610389709473, "learning_rate": 5.4011450848512225e-05, "loss": 1.1074, "step": 891 }, { "epoch": 0.17491910971663888, "grad_norm": 5.7243123054504395, "learning_rate": 5.3805086411704985e-05, "loss": 2.2836, "step": 892 }, { "epoch": 0.17511520737327188, "grad_norm": 3.6346094608306885, "learning_rate": 5.3598971769477655e-05, "loss": 1.2185, "step": 893 }, { "epoch": 0.17531130502990488, "grad_norm": 5.283199310302734, "learning_rate": 5.339310803637538e-05, "loss": 1.0693, "step": 894 }, { "epoch": 0.1755074026865379, "grad_norm": 5.19635009765625, "learning_rate": 5.318749632558674e-05, "loss": 1.3408, "step": 895 }, { "epoch": 0.1757035003431709, "grad_norm": 7.379054546356201, "learning_rate": 5.298213774893735e-05, "loss": 1.6441, "step": 896 }, { "epoch": 0.1758995979998039, "grad_norm": 10.650435447692871, "learning_rate": 5.277703341688407e-05, "loss": 1.1186, "step": 897 }, { "epoch": 0.1760956956564369, "grad_norm": 4.219560623168945, "learning_rate": 5.257218443850895e-05, "loss": 1.1465, "step": 898 }, { "epoch": 0.1762917933130699, "grad_norm": 9.091025352478027, "learning_rate": 5.236759192151336e-05, "loss": 2.0643, "step": 899 }, { "epoch": 0.1764878909697029, "grad_norm": 6.650880336761475, "learning_rate": 5.2163256972211714e-05, "loss": 1.6255, "step": 900 }, { "epoch": 0.1766839886263359, "grad_norm": 5.008049011230469, "learning_rate": 5.195918069552572e-05, "loss": 1.0065, "step": 901 }, { "epoch": 0.17688008628296892, "grad_norm": 6.407417297363281, "learning_rate": 5.175536419497831e-05, "loss": 0.7189, "step": 902 }, { "epoch": 0.17707618393960192, "grad_norm": 4.843297004699707, "learning_rate": 5.155180857268779e-05, "loss": 1.5639, "step": 903 }, { "epoch": 0.17727228159623493, "grad_norm": 3.7099854946136475, "learning_rate": 5.1348514929361714e-05, "loss": 1.0337, "step": 904 }, { "epoch": 0.17746837925286793, "grad_norm": 7.940112590789795, "learning_rate": 5.114548436429098e-05, "loss": 1.9378, "step": 905 }, { "epoch": 0.17766447690950093, "grad_norm": 6.462250232696533, "learning_rate": 5.0942717975344035e-05, "loss": 1.2462, "step": 906 }, { "epoch": 0.17786057456613394, "grad_norm": 3.7596821784973145, "learning_rate": 5.074021685896072e-05, "loss": 0.5025, "step": 907 }, { "epoch": 0.17805667222276694, "grad_norm": 5.532608509063721, "learning_rate": 5.0537982110146464e-05, "loss": 1.2263, "step": 908 }, { "epoch": 0.17825276987939995, "grad_norm": 6.286501884460449, "learning_rate": 5.033601482246628e-05, "loss": 0.9899, "step": 909 }, { "epoch": 0.17844886753603295, "grad_norm": 8.714296340942383, "learning_rate": 5.013431608803909e-05, "loss": 1.0626, "step": 910 }, { "epoch": 0.17864496519266596, "grad_norm": 6.827104568481445, "learning_rate": 4.993288699753146e-05, "loss": 1.1741, "step": 911 }, { "epoch": 0.17884106284929896, "grad_norm": 7.155247688293457, "learning_rate": 4.973172864015193e-05, "loss": 1.9644, "step": 912 }, { "epoch": 0.17903716050593196, "grad_norm": 4.7466139793396, "learning_rate": 4.953084210364508e-05, "loss": 1.2649, "step": 913 }, { "epoch": 0.17923325816256497, "grad_norm": 6.292973041534424, "learning_rate": 4.933022847428571e-05, "loss": 2.3925, "step": 914 }, { "epoch": 0.17942935581919797, "grad_norm": 6.9081950187683105, "learning_rate": 4.912988883687283e-05, "loss": 2.7031, "step": 915 }, { "epoch": 0.17962545347583098, "grad_norm": 10.206450462341309, "learning_rate": 4.892982427472387e-05, "loss": 1.2479, "step": 916 }, { "epoch": 0.17982155113246398, "grad_norm": 9.753491401672363, "learning_rate": 4.873003586966881e-05, "loss": 1.1683, "step": 917 }, { "epoch": 0.18001764878909696, "grad_norm": 5.422173023223877, "learning_rate": 4.853052470204438e-05, "loss": 1.0117, "step": 918 }, { "epoch": 0.18021374644572996, "grad_norm": 4.505789756774902, "learning_rate": 4.8331291850688096e-05, "loss": 0.7031, "step": 919 }, { "epoch": 0.18040984410236297, "grad_norm": 6.756499290466309, "learning_rate": 4.813233839293265e-05, "loss": 1.5295, "step": 920 }, { "epoch": 0.18060594175899597, "grad_norm": 7.535793781280518, "learning_rate": 4.7933665404599804e-05, "loss": 1.2856, "step": 921 }, { "epoch": 0.18080203941562897, "grad_norm": 8.04592227935791, "learning_rate": 4.773527395999474e-05, "loss": 0.9891, "step": 922 }, { "epoch": 0.18099813707226198, "grad_norm": 5.000694274902344, "learning_rate": 4.753716513190024e-05, "loss": 1.3128, "step": 923 }, { "epoch": 0.18119423472889498, "grad_norm": 8.065174102783203, "learning_rate": 4.7339339991570876e-05, "loss": 1.9393, "step": 924 }, { "epoch": 0.18139033238552799, "grad_norm": 6.1632256507873535, "learning_rate": 4.714179960872712e-05, "loss": 1.7183, "step": 925 }, { "epoch": 0.181586430042161, "grad_norm": 6.445769786834717, "learning_rate": 4.69445450515498e-05, "loss": 2.6642, "step": 926 }, { "epoch": 0.181782527698794, "grad_norm": 8.51395320892334, "learning_rate": 4.674757738667405e-05, "loss": 1.2191, "step": 927 }, { "epoch": 0.181978625355427, "grad_norm": 9.159540176391602, "learning_rate": 4.6550897679183625e-05, "loss": 2.1541, "step": 928 }, { "epoch": 0.18217472301206, "grad_norm": 7.939080715179443, "learning_rate": 4.635450699260535e-05, "loss": 1.4875, "step": 929 }, { "epoch": 0.182370820668693, "grad_norm": 11.180910110473633, "learning_rate": 4.615840638890305e-05, "loss": 1.9817, "step": 930 }, { "epoch": 0.182566918325326, "grad_norm": 7.014312744140625, "learning_rate": 4.596259692847198e-05, "loss": 1.2057, "step": 931 }, { "epoch": 0.18276301598195901, "grad_norm": 3.89388108253479, "learning_rate": 4.5767079670133064e-05, "loss": 1.2204, "step": 932 }, { "epoch": 0.18295911363859202, "grad_norm": 7.551772594451904, "learning_rate": 4.557185567112725e-05, "loss": 1.271, "step": 933 }, { "epoch": 0.18315521129522502, "grad_norm": 5.167237758636475, "learning_rate": 4.537692598710962e-05, "loss": 2.0885, "step": 934 }, { "epoch": 0.18335130895185803, "grad_norm": 7.082968235015869, "learning_rate": 4.5182291672143796e-05, "loss": 2.0684, "step": 935 }, { "epoch": 0.18354740660849103, "grad_norm": 5.6844801902771, "learning_rate": 4.498795377869622e-05, "loss": 1.2413, "step": 936 }, { "epoch": 0.18374350426512404, "grad_norm": 8.226604461669922, "learning_rate": 4.479391335763054e-05, "loss": 1.7761, "step": 937 }, { "epoch": 0.18393960192175704, "grad_norm": 6.878251075744629, "learning_rate": 4.4600171458201735e-05, "loss": 1.4162, "step": 938 }, { "epoch": 0.18413569957839004, "grad_norm": 6.791055202484131, "learning_rate": 4.440672912805063e-05, "loss": 1.3623, "step": 939 }, { "epoch": 0.18433179723502305, "grad_norm": 4.67439079284668, "learning_rate": 4.4213587413198085e-05, "loss": 0.9896, "step": 940 }, { "epoch": 0.18452789489165605, "grad_norm": 6.860576629638672, "learning_rate": 4.402074735803955e-05, "loss": 1.7911, "step": 941 }, { "epoch": 0.18472399254828906, "grad_norm": 5.80997896194458, "learning_rate": 4.382821000533913e-05, "loss": 2.8923, "step": 942 }, { "epoch": 0.18492009020492206, "grad_norm": 5.034128189086914, "learning_rate": 4.3635976396224186e-05, "loss": 1.7938, "step": 943 }, { "epoch": 0.18511618786155506, "grad_norm": 4.669545650482178, "learning_rate": 4.3444047570179525e-05, "loss": 0.8484, "step": 944 }, { "epoch": 0.18531228551818807, "grad_norm": 6.147574424743652, "learning_rate": 4.3252424565042017e-05, "loss": 1.1467, "step": 945 }, { "epoch": 0.18550838317482107, "grad_norm": 21.214567184448242, "learning_rate": 4.306110841699471e-05, "loss": 3.0427, "step": 946 }, { "epoch": 0.18570448083145408, "grad_norm": 5.828713893890381, "learning_rate": 4.2870100160561344e-05, "loss": 0.9022, "step": 947 }, { "epoch": 0.18590057848808705, "grad_norm": 8.350537300109863, "learning_rate": 4.267940082860088e-05, "loss": 1.5416, "step": 948 }, { "epoch": 0.18609667614472006, "grad_norm": 3.421091318130493, "learning_rate": 4.2489011452301705e-05, "loss": 0.7294, "step": 949 }, { "epoch": 0.18629277380135306, "grad_norm": 5.414788246154785, "learning_rate": 4.229893306117616e-05, "loss": 0.9339, "step": 950 }, { "epoch": 0.18648887145798607, "grad_norm": 4.0585761070251465, "learning_rate": 4.210916668305496e-05, "loss": 0.9102, "step": 951 }, { "epoch": 0.18668496911461907, "grad_norm": 13.428049087524414, "learning_rate": 4.1919713344081704e-05, "loss": 2.287, "step": 952 }, { "epoch": 0.18688106677125207, "grad_norm": 7.367785453796387, "learning_rate": 4.17305740687072e-05, "loss": 1.5482, "step": 953 }, { "epoch": 0.18707716442788508, "grad_norm": 4.9446635246276855, "learning_rate": 4.154174987968398e-05, "loss": 1.5513, "step": 954 }, { "epoch": 0.18727326208451808, "grad_norm": 6.660252094268799, "learning_rate": 4.135324179806078e-05, "loss": 1.6624, "step": 955 }, { "epoch": 0.1874693597411511, "grad_norm": 8.531050682067871, "learning_rate": 4.11650508431771e-05, "loss": 1.58, "step": 956 }, { "epoch": 0.1876654573977841, "grad_norm": 5.770985126495361, "learning_rate": 4.09771780326575e-05, "loss": 1.2329, "step": 957 }, { "epoch": 0.1878615550544171, "grad_norm": 6.347046375274658, "learning_rate": 4.0789624382406264e-05, "loss": 1.3816, "step": 958 }, { "epoch": 0.1880576527110501, "grad_norm": 10.09997272491455, "learning_rate": 4.060239090660181e-05, "loss": 1.6532, "step": 959 }, { "epoch": 0.1882537503676831, "grad_norm": 5.894500255584717, "learning_rate": 4.041547861769126e-05, "loss": 2.0952, "step": 960 }, { "epoch": 0.1884498480243161, "grad_norm": 3.906419038772583, "learning_rate": 4.0228888526384936e-05, "loss": 1.5555, "step": 961 }, { "epoch": 0.1886459456809491, "grad_norm": 6.6089630126953125, "learning_rate": 4.004262164165098e-05, "loss": 2.0901, "step": 962 }, { "epoch": 0.18884204333758212, "grad_norm": 5.344242095947266, "learning_rate": 3.985667897070974e-05, "loss": 2.2121, "step": 963 }, { "epoch": 0.18903814099421512, "grad_norm": 3.924384117126465, "learning_rate": 3.967106151902842e-05, "loss": 2.3173, "step": 964 }, { "epoch": 0.18923423865084812, "grad_norm": 5.392980098724365, "learning_rate": 3.948577029031566e-05, "loss": 1.3514, "step": 965 }, { "epoch": 0.18943033630748113, "grad_norm": 5.208516597747803, "learning_rate": 3.930080628651607e-05, "loss": 1.1333, "step": 966 }, { "epoch": 0.18962643396411413, "grad_norm": 8.4103364944458, "learning_rate": 3.9116170507804784e-05, "loss": 3.0624, "step": 967 }, { "epoch": 0.18982253162074714, "grad_norm": 4.458950996398926, "learning_rate": 3.89318639525822e-05, "loss": 1.974, "step": 968 }, { "epoch": 0.19001862927738014, "grad_norm": 4.55155086517334, "learning_rate": 3.874788761746836e-05, "loss": 1.2587, "step": 969 }, { "epoch": 0.19021472693401315, "grad_norm": 3.517223834991455, "learning_rate": 3.8564242497297686e-05, "loss": 0.9581, "step": 970 }, { "epoch": 0.19041082459064615, "grad_norm": 6.368021011352539, "learning_rate": 3.8380929585113666e-05, "loss": 1.6365, "step": 971 }, { "epoch": 0.19060692224727915, "grad_norm": 5.153208255767822, "learning_rate": 3.819794987216333e-05, "loss": 0.9107, "step": 972 }, { "epoch": 0.19080301990391216, "grad_norm": 9.487665176391602, "learning_rate": 3.801530434789199e-05, "loss": 2.9295, "step": 973 }, { "epoch": 0.19099911756054516, "grad_norm": 10.716341018676758, "learning_rate": 3.783299399993781e-05, "loss": 1.9429, "step": 974 }, { "epoch": 0.19119521521717817, "grad_norm": 4.834690093994141, "learning_rate": 3.7651019814126654e-05, "loss": 1.7383, "step": 975 }, { "epoch": 0.19139131287381117, "grad_norm": 4.203547954559326, "learning_rate": 3.74693827744665e-05, "loss": 0.7084, "step": 976 }, { "epoch": 0.19158741053044417, "grad_norm": 8.38505744934082, "learning_rate": 3.7288083863142274e-05, "loss": 2.7231, "step": 977 }, { "epoch": 0.19178350818707715, "grad_norm": 6.748301029205322, "learning_rate": 3.710712406051049e-05, "loss": 1.3249, "step": 978 }, { "epoch": 0.19197960584371015, "grad_norm": 5.358412265777588, "learning_rate": 3.692650434509404e-05, "loss": 0.7995, "step": 979 }, { "epoch": 0.19217570350034316, "grad_norm": 3.909745693206787, "learning_rate": 3.674622569357673e-05, "loss": 2.3175, "step": 980 }, { "epoch": 0.19237180115697616, "grad_norm": 11.555536270141602, "learning_rate": 3.6566289080798153e-05, "loss": 1.7112, "step": 981 }, { "epoch": 0.19256789881360917, "grad_norm": 10.718825340270996, "learning_rate": 3.638669547974831e-05, "loss": 1.1124, "step": 982 }, { "epoch": 0.19276399647024217, "grad_norm": 4.483185291290283, "learning_rate": 3.6207445861562494e-05, "loss": 1.85, "step": 983 }, { "epoch": 0.19296009412687518, "grad_norm": 4.881377696990967, "learning_rate": 3.602854119551585e-05, "loss": 0.8783, "step": 984 }, { "epoch": 0.19315619178350818, "grad_norm": 5.172372817993164, "learning_rate": 3.584998244901826e-05, "loss": 1.1851, "step": 985 }, { "epoch": 0.19335228944014118, "grad_norm": 8.540918350219727, "learning_rate": 3.567177058760905e-05, "loss": 0.8138, "step": 986 }, { "epoch": 0.1935483870967742, "grad_norm": 10.905487060546875, "learning_rate": 3.549390657495188e-05, "loss": 2.8036, "step": 987 }, { "epoch": 0.1937444847534072, "grad_norm": 6.448824405670166, "learning_rate": 3.531639137282938e-05, "loss": 2.4083, "step": 988 }, { "epoch": 0.1939405824100402, "grad_norm": 4.409211158752441, "learning_rate": 3.5139225941138e-05, "loss": 1.2686, "step": 989 }, { "epoch": 0.1941366800666732, "grad_norm": 4.925105094909668, "learning_rate": 3.496241123788294e-05, "loss": 3.3955, "step": 990 }, { "epoch": 0.1943327777233062, "grad_norm": 6.541469097137451, "learning_rate": 3.478594821917279e-05, "loss": 1.4906, "step": 991 }, { "epoch": 0.1945288753799392, "grad_norm": 7.749048233032227, "learning_rate": 3.460983783921443e-05, "loss": 1.7907, "step": 992 }, { "epoch": 0.1947249730365722, "grad_norm": 4.833098888397217, "learning_rate": 3.44340810503079e-05, "loss": 1.2777, "step": 993 }, { "epoch": 0.19492107069320522, "grad_norm": 5.138740062713623, "learning_rate": 3.425867880284129e-05, "loss": 1.7267, "step": 994 }, { "epoch": 0.19511716834983822, "grad_norm": 8.098197937011719, "learning_rate": 3.4083632045285466e-05, "loss": 1.7961, "step": 995 }, { "epoch": 0.19531326600647123, "grad_norm": 4.534963130950928, "learning_rate": 3.390894172418903e-05, "loss": 1.773, "step": 996 }, { "epoch": 0.19550936366310423, "grad_norm": 9.311367988586426, "learning_rate": 3.373460878417315e-05, "loss": 1.2173, "step": 997 }, { "epoch": 0.19570546131973723, "grad_norm": 4.8790388107299805, "learning_rate": 3.356063416792662e-05, "loss": 1.3003, "step": 998 }, { "epoch": 0.19590155897637024, "grad_norm": 4.137528419494629, "learning_rate": 3.338701881620052e-05, "loss": 1.8794, "step": 999 }, { "epoch": 0.19609765663300324, "grad_norm": 8.082904815673828, "learning_rate": 3.321376366780326e-05, "loss": 1.3442, "step": 1000 }, { "epoch": 0.19629375428963625, "grad_norm": 7.494518280029297, "learning_rate": 3.3040869659595485e-05, "loss": 2.3671, "step": 1001 }, { "epoch": 0.19648985194626925, "grad_norm": 4.482415676116943, "learning_rate": 3.286833772648504e-05, "loss": 0.7332, "step": 1002 }, { "epoch": 0.19668594960290225, "grad_norm": 6.464245796203613, "learning_rate": 3.269616880142182e-05, "loss": 1.6453, "step": 1003 }, { "epoch": 0.19688204725953526, "grad_norm": 6.342423915863037, "learning_rate": 3.252436381539291e-05, "loss": 1.5742, "step": 1004 }, { "epoch": 0.19707814491616826, "grad_norm": 2.4557156562805176, "learning_rate": 3.235292369741729e-05, "loss": 0.3941, "step": 1005 }, { "epoch": 0.19727424257280127, "grad_norm": 6.024993419647217, "learning_rate": 3.218184937454103e-05, "loss": 1.7341, "step": 1006 }, { "epoch": 0.19747034022943427, "grad_norm": 4.793758392333984, "learning_rate": 3.2011141771832184e-05, "loss": 0.8558, "step": 1007 }, { "epoch": 0.19766643788606725, "grad_norm": 8.792879104614258, "learning_rate": 3.184080181237575e-05, "loss": 0.6646, "step": 1008 }, { "epoch": 0.19786253554270025, "grad_norm": 8.334936141967773, "learning_rate": 3.167083041726886e-05, "loss": 1.1602, "step": 1009 }, { "epoch": 0.19805863319933326, "grad_norm": 5.780764579772949, "learning_rate": 3.1501228505615534e-05, "loss": 2.342, "step": 1010 }, { "epoch": 0.19825473085596626, "grad_norm": 9.708879470825195, "learning_rate": 3.1331996994521915e-05, "loss": 0.981, "step": 1011 }, { "epoch": 0.19845082851259926, "grad_norm": 5.935221195220947, "learning_rate": 3.1163136799091186e-05, "loss": 0.8119, "step": 1012 }, { "epoch": 0.19864692616923227, "grad_norm": 9.303512573242188, "learning_rate": 3.0994648832418784e-05, "loss": 1.5852, "step": 1013 }, { "epoch": 0.19884302382586527, "grad_norm": 7.7904052734375, "learning_rate": 3.082653400558724e-05, "loss": 1.4118, "step": 1014 }, { "epoch": 0.19903912148249828, "grad_norm": 4.866517066955566, "learning_rate": 3.065879322766142e-05, "loss": 0.7066, "step": 1015 }, { "epoch": 0.19923521913913128, "grad_norm": 4.0074896812438965, "learning_rate": 3.0491427405683514e-05, "loss": 1.0785, "step": 1016 }, { "epoch": 0.19943131679576429, "grad_norm": 5.893292427062988, "learning_rate": 3.0324437444668274e-05, "loss": 2.6953, "step": 1017 }, { "epoch": 0.1996274144523973, "grad_norm": 10.129006385803223, "learning_rate": 3.0157824247597922e-05, "loss": 1.4669, "step": 1018 }, { "epoch": 0.1998235121090303, "grad_norm": 6.336552143096924, "learning_rate": 2.9991588715417383e-05, "loss": 1.9093, "step": 1019 }, { "epoch": 0.2000196097656633, "grad_norm": 7.2662482261657715, "learning_rate": 2.9825731747029394e-05, "loss": 1.3408, "step": 1020 }, { "epoch": 0.2002157074222963, "grad_norm": 4.900631427764893, "learning_rate": 2.966025423928972e-05, "loss": 1.2219, "step": 1021 }, { "epoch": 0.2004118050789293, "grad_norm": 10.661945343017578, "learning_rate": 2.9495157087002123e-05, "loss": 2.2735, "step": 1022 }, { "epoch": 0.2006079027355623, "grad_norm": 6.423792362213135, "learning_rate": 2.9330441182913694e-05, "loss": 1.362, "step": 1023 }, { "epoch": 0.2006079027355623, "eval_loss": 0.3885071277618408, "eval_runtime": 77.8128, "eval_samples_per_second": 27.605, "eval_steps_per_second": 13.802, "step": 1023 }, { "epoch": 0.20080400039219531, "grad_norm": 7.308141231536865, "learning_rate": 2.9166107417709888e-05, "loss": 1.0916, "step": 1024 }, { "epoch": 0.20100009804882832, "grad_norm": 3.7046291828155518, "learning_rate": 2.9002156680009906e-05, "loss": 1.7084, "step": 1025 }, { "epoch": 0.20119619570546132, "grad_norm": 3.8992371559143066, "learning_rate": 2.883858985636165e-05, "loss": 1.0185, "step": 1026 }, { "epoch": 0.20139229336209433, "grad_norm": 4.895712852478027, "learning_rate": 2.867540783123711e-05, "loss": 1.1252, "step": 1027 }, { "epoch": 0.20158839101872733, "grad_norm": 10.118478775024414, "learning_rate": 2.851261148702744e-05, "loss": 1.8931, "step": 1028 }, { "epoch": 0.20178448867536034, "grad_norm": 9.701197624206543, "learning_rate": 2.835020170403837e-05, "loss": 2.1721, "step": 1029 }, { "epoch": 0.20198058633199334, "grad_norm": 5.3619184494018555, "learning_rate": 2.8188179360485267e-05, "loss": 1.3209, "step": 1030 }, { "epoch": 0.20217668398862634, "grad_norm": 7.866724967956543, "learning_rate": 2.8026545332488418e-05, "loss": 1.4948, "step": 1031 }, { "epoch": 0.20237278164525935, "grad_norm": 5.446761608123779, "learning_rate": 2.7865300494068435e-05, "loss": 1.041, "step": 1032 }, { "epoch": 0.20256887930189235, "grad_norm": 8.317418098449707, "learning_rate": 2.7704445717141368e-05, "loss": 1.0779, "step": 1033 }, { "epoch": 0.20276497695852536, "grad_norm": 4.434285640716553, "learning_rate": 2.7543981871514023e-05, "loss": 1.9139, "step": 1034 }, { "epoch": 0.20296107461515836, "grad_norm": 14.01183795928955, "learning_rate": 2.7383909824879294e-05, "loss": 2.4565, "step": 1035 }, { "epoch": 0.20315717227179136, "grad_norm": 4.523947715759277, "learning_rate": 2.7224230442811526e-05, "loss": 1.7748, "step": 1036 }, { "epoch": 0.20335326992842437, "grad_norm": 10.790266990661621, "learning_rate": 2.7064944588761688e-05, "loss": 1.9152, "step": 1037 }, { "epoch": 0.20354936758505734, "grad_norm": 7.028364181518555, "learning_rate": 2.6906053124052798e-05, "loss": 1.6433, "step": 1038 }, { "epoch": 0.20374546524169035, "grad_norm": 8.6144437789917, "learning_rate": 2.674755690787526e-05, "loss": 2.3653, "step": 1039 }, { "epoch": 0.20394156289832335, "grad_norm": 4.6696343421936035, "learning_rate": 2.6589456797282253e-05, "loss": 1.0972, "step": 1040 }, { "epoch": 0.20413766055495636, "grad_norm": 7.880300998687744, "learning_rate": 2.6431753647184988e-05, "loss": 1.47, "step": 1041 }, { "epoch": 0.20433375821158936, "grad_norm": 11.923666954040527, "learning_rate": 2.6274448310348178e-05, "loss": 2.1949, "step": 1042 }, { "epoch": 0.20452985586822237, "grad_norm": 11.636192321777344, "learning_rate": 2.6117541637385402e-05, "loss": 1.3088, "step": 1043 }, { "epoch": 0.20472595352485537, "grad_norm": 4.3036885261535645, "learning_rate": 2.5961034476754487e-05, "loss": 1.5687, "step": 1044 }, { "epoch": 0.20492205118148837, "grad_norm": 4.218216419219971, "learning_rate": 2.5804927674752922e-05, "loss": 0.672, "step": 1045 }, { "epoch": 0.20511814883812138, "grad_norm": 4.2901291847229, "learning_rate": 2.564922207551337e-05, "loss": 1.048, "step": 1046 }, { "epoch": 0.20531424649475438, "grad_norm": 6.806373119354248, "learning_rate": 2.549391852099896e-05, "loss": 2.2606, "step": 1047 }, { "epoch": 0.2055103441513874, "grad_norm": 5.39298152923584, "learning_rate": 2.5339017850998802e-05, "loss": 1.5175, "step": 1048 }, { "epoch": 0.2057064418080204, "grad_norm": 7.037325859069824, "learning_rate": 2.5184520903123476e-05, "loss": 1.3757, "step": 1049 }, { "epoch": 0.2059025394646534, "grad_norm": 6.3087239265441895, "learning_rate": 2.503042851280043e-05, "loss": 1.0206, "step": 1050 }, { "epoch": 0.2060986371212864, "grad_norm": 6.11033821105957, "learning_rate": 2.4876741513269597e-05, "loss": 1.99, "step": 1051 }, { "epoch": 0.2062947347779194, "grad_norm": 5.525476932525635, "learning_rate": 2.4723460735578697e-05, "loss": 0.7599, "step": 1052 }, { "epoch": 0.2064908324345524, "grad_norm": 6.608320236206055, "learning_rate": 2.4570587008578895e-05, "loss": 1.2481, "step": 1053 }, { "epoch": 0.2066869300911854, "grad_norm": 3.596764087677002, "learning_rate": 2.4418121158920227e-05, "loss": 0.713, "step": 1054 }, { "epoch": 0.20688302774781842, "grad_norm": 7.637975692749023, "learning_rate": 2.4266064011047264e-05, "loss": 1.3772, "step": 1055 }, { "epoch": 0.20707912540445142, "grad_norm": 4.366866111755371, "learning_rate": 2.411441638719445e-05, "loss": 0.5525, "step": 1056 }, { "epoch": 0.20727522306108442, "grad_norm": 5.603202819824219, "learning_rate": 2.396317910738184e-05, "loss": 0.9035, "step": 1057 }, { "epoch": 0.20747132071771743, "grad_norm": 7.776876449584961, "learning_rate": 2.3812352989410537e-05, "loss": 1.6217, "step": 1058 }, { "epoch": 0.20766741837435043, "grad_norm": 9.07315444946289, "learning_rate": 2.366193884885841e-05, "loss": 2.2412, "step": 1059 }, { "epoch": 0.20786351603098344, "grad_norm": 7.473457336425781, "learning_rate": 2.3511937499075508e-05, "loss": 1.6256, "step": 1060 }, { "epoch": 0.20805961368761644, "grad_norm": 10.729793548583984, "learning_rate": 2.3362349751179825e-05, "loss": 1.247, "step": 1061 }, { "epoch": 0.20825571134424944, "grad_norm": 6.086622714996338, "learning_rate": 2.3213176414052738e-05, "loss": 1.613, "step": 1062 }, { "epoch": 0.20845180900088245, "grad_norm": 6.6213836669921875, "learning_rate": 2.30644182943349e-05, "loss": 0.9007, "step": 1063 }, { "epoch": 0.20864790665751545, "grad_norm": 4.5187602043151855, "learning_rate": 2.2916076196421587e-05, "loss": 1.3616, "step": 1064 }, { "epoch": 0.20884400431414846, "grad_norm": 8.176258087158203, "learning_rate": 2.276815092245851e-05, "loss": 1.1822, "step": 1065 }, { "epoch": 0.20904010197078146, "grad_norm": 6.9451494216918945, "learning_rate": 2.2620643272337426e-05, "loss": 1.2958, "step": 1066 }, { "epoch": 0.20923619962741447, "grad_norm": 4.266102313995361, "learning_rate": 2.2473554043691913e-05, "loss": 1.5274, "step": 1067 }, { "epoch": 0.20943229728404747, "grad_norm": 7.066012382507324, "learning_rate": 2.2326884031892882e-05, "loss": 1.0549, "step": 1068 }, { "epoch": 0.20962839494068045, "grad_norm": 11.074556350708008, "learning_rate": 2.21806340300444e-05, "loss": 1.3012, "step": 1069 }, { "epoch": 0.20982449259731345, "grad_norm": 5.717066287994385, "learning_rate": 2.203480482897935e-05, "loss": 1.4355, "step": 1070 }, { "epoch": 0.21002059025394645, "grad_norm": 5.230037689208984, "learning_rate": 2.188939721725526e-05, "loss": 1.7493, "step": 1071 }, { "epoch": 0.21021668791057946, "grad_norm": 4.878988265991211, "learning_rate": 2.174441198114985e-05, "loss": 0.7765, "step": 1072 }, { "epoch": 0.21041278556721246, "grad_norm": 4.619425296783447, "learning_rate": 2.159984990465691e-05, "loss": 1.8523, "step": 1073 }, { "epoch": 0.21060888322384547, "grad_norm": 4.031160354614258, "learning_rate": 2.1455711769482113e-05, "loss": 0.9848, "step": 1074 }, { "epoch": 0.21080498088047847, "grad_norm": 7.096545219421387, "learning_rate": 2.131199835503861e-05, "loss": 1.4097, "step": 1075 }, { "epoch": 0.21100107853711147, "grad_norm": 5.833085060119629, "learning_rate": 2.1168710438442953e-05, "loss": 0.9551, "step": 1076 }, { "epoch": 0.21119717619374448, "grad_norm": 5.613104343414307, "learning_rate": 2.1025848794510815e-05, "loss": 1.9112, "step": 1077 }, { "epoch": 0.21139327385037748, "grad_norm": 7.113245010375977, "learning_rate": 2.0883414195752972e-05, "loss": 1.2791, "step": 1078 }, { "epoch": 0.2115893715070105, "grad_norm": 5.34241247177124, "learning_rate": 2.0741407412370838e-05, "loss": 1.4305, "step": 1079 }, { "epoch": 0.2117854691636435, "grad_norm": 4.97748327255249, "learning_rate": 2.0599829212252553e-05, "loss": 1.0597, "step": 1080 }, { "epoch": 0.2119815668202765, "grad_norm": 6.11815881729126, "learning_rate": 2.045868036096864e-05, "loss": 1.3793, "step": 1081 }, { "epoch": 0.2121776644769095, "grad_norm": 7.822902202606201, "learning_rate": 2.031796162176811e-05, "loss": 1.7602, "step": 1082 }, { "epoch": 0.2123737621335425, "grad_norm": 5.7074875831604, "learning_rate": 2.0177673755574044e-05, "loss": 1.9033, "step": 1083 }, { "epoch": 0.2125698597901755, "grad_norm": 5.141830921173096, "learning_rate": 2.0037817520979686e-05, "loss": 0.6569, "step": 1084 }, { "epoch": 0.2127659574468085, "grad_norm": 8.139151573181152, "learning_rate": 1.9898393674244243e-05, "loss": 1.6038, "step": 1085 }, { "epoch": 0.21296205510344152, "grad_norm": 6.489431858062744, "learning_rate": 1.975940296928882e-05, "loss": 1.7581, "step": 1086 }, { "epoch": 0.21315815276007452, "grad_norm": 9.372897148132324, "learning_rate": 1.9620846157692418e-05, "loss": 2.5077, "step": 1087 }, { "epoch": 0.21335425041670752, "grad_norm": 11.757204055786133, "learning_rate": 1.9482723988687734e-05, "loss": 2.0571, "step": 1088 }, { "epoch": 0.21355034807334053, "grad_norm": 4.995403289794922, "learning_rate": 1.934503720915719e-05, "loss": 1.3159, "step": 1089 }, { "epoch": 0.21374644572997353, "grad_norm": 4.217586994171143, "learning_rate": 1.9207786563628894e-05, "loss": 0.7797, "step": 1090 }, { "epoch": 0.21394254338660654, "grad_norm": 5.487307548522949, "learning_rate": 1.907097279427258e-05, "loss": 1.2424, "step": 1091 }, { "epoch": 0.21413864104323954, "grad_norm": 5.5197014808654785, "learning_rate": 1.8934596640895607e-05, "loss": 1.6582, "step": 1092 }, { "epoch": 0.21433473869987255, "grad_norm": 9.223129272460938, "learning_rate": 1.8798658840939055e-05, "loss": 1.5407, "step": 1093 }, { "epoch": 0.21453083635650555, "grad_norm": 6.948054790496826, "learning_rate": 1.8663160129473535e-05, "loss": 0.7758, "step": 1094 }, { "epoch": 0.21472693401313855, "grad_norm": 5.582278251647949, "learning_rate": 1.8528101239195395e-05, "loss": 2.0049, "step": 1095 }, { "epoch": 0.21492303166977156, "grad_norm": 4.116588115692139, "learning_rate": 1.8393482900422644e-05, "loss": 1.8004, "step": 1096 }, { "epoch": 0.21511912932640456, "grad_norm": 6.368351936340332, "learning_rate": 1.8259305841091123e-05, "loss": 1.4293, "step": 1097 }, { "epoch": 0.21531522698303757, "grad_norm": 5.690855503082275, "learning_rate": 1.8125570786750455e-05, "loss": 1.1567, "step": 1098 }, { "epoch": 0.21551132463967054, "grad_norm": 7.080834865570068, "learning_rate": 1.7992278460560152e-05, "loss": 1.1252, "step": 1099 }, { "epoch": 0.21570742229630355, "grad_norm": 3.924699544906616, "learning_rate": 1.785942958328569e-05, "loss": 1.9084, "step": 1100 }, { "epoch": 0.21590351995293655, "grad_norm": 4.245781421661377, "learning_rate": 1.772702487329474e-05, "loss": 1.2304, "step": 1101 }, { "epoch": 0.21609961760956956, "grad_norm": 11.68691349029541, "learning_rate": 1.7595065046553085e-05, "loss": 1.0978, "step": 1102 }, { "epoch": 0.21629571526620256, "grad_norm": 5.583014011383057, "learning_rate": 1.7463550816620876e-05, "loss": 2.6622, "step": 1103 }, { "epoch": 0.21649181292283556, "grad_norm": 6.454019546508789, "learning_rate": 1.7332482894648726e-05, "loss": 1.7374, "step": 1104 }, { "epoch": 0.21668791057946857, "grad_norm": 9.824748992919922, "learning_rate": 1.7201861989373926e-05, "loss": 2.4338, "step": 1105 }, { "epoch": 0.21688400823610157, "grad_norm": 4.780369758605957, "learning_rate": 1.7071688807116525e-05, "loss": 0.9636, "step": 1106 }, { "epoch": 0.21708010589273458, "grad_norm": 6.861015796661377, "learning_rate": 1.694196405177556e-05, "loss": 1.1233, "step": 1107 }, { "epoch": 0.21727620354936758, "grad_norm": 4.883321762084961, "learning_rate": 1.681268842482523e-05, "loss": 0.8476, "step": 1108 }, { "epoch": 0.21747230120600058, "grad_norm": 4.856523513793945, "learning_rate": 1.6683862625311164e-05, "loss": 1.4524, "step": 1109 }, { "epoch": 0.2176683988626336, "grad_norm": 5.7625017166137695, "learning_rate": 1.6555487349846544e-05, "loss": 1.5877, "step": 1110 }, { "epoch": 0.2178644965192666, "grad_norm": 6.9636359214782715, "learning_rate": 1.6427563292608416e-05, "loss": 1.3804, "step": 1111 }, { "epoch": 0.2180605941758996, "grad_norm": 5.53889799118042, "learning_rate": 1.630009114533384e-05, "loss": 1.4286, "step": 1112 }, { "epoch": 0.2182566918325326, "grad_norm": 8.081591606140137, "learning_rate": 1.6173071597316335e-05, "loss": 2.5756, "step": 1113 }, { "epoch": 0.2184527894891656, "grad_norm": 9.321168899536133, "learning_rate": 1.604650533540194e-05, "loss": 2.107, "step": 1114 }, { "epoch": 0.2186488871457986, "grad_norm": 7.095883846282959, "learning_rate": 1.5920393043985593e-05, "loss": 0.9074, "step": 1115 }, { "epoch": 0.2188449848024316, "grad_norm": 10.197619438171387, "learning_rate": 1.579473540500751e-05, "loss": 1.7506, "step": 1116 }, { "epoch": 0.21904108245906462, "grad_norm": 8.685538291931152, "learning_rate": 1.5669533097949328e-05, "loss": 1.6941, "step": 1117 }, { "epoch": 0.21923718011569762, "grad_norm": 6.081611156463623, "learning_rate": 1.5544786799830568e-05, "loss": 2.3827, "step": 1118 }, { "epoch": 0.21943327777233063, "grad_norm": 4.3908162117004395, "learning_rate": 1.5420497185204873e-05, "loss": 1.1959, "step": 1119 }, { "epoch": 0.21962937542896363, "grad_norm": 13.473033905029297, "learning_rate": 1.5296664926156525e-05, "loss": 2.7042, "step": 1120 }, { "epoch": 0.21982547308559663, "grad_norm": 7.351769924163818, "learning_rate": 1.5173290692296582e-05, "loss": 2.4473, "step": 1121 }, { "epoch": 0.22002157074222964, "grad_norm": 4.283702373504639, "learning_rate": 1.5050375150759433e-05, "loss": 1.2878, "step": 1122 }, { "epoch": 0.22021766839886264, "grad_norm": 6.013007164001465, "learning_rate": 1.4927918966199094e-05, "loss": 1.7944, "step": 1123 }, { "epoch": 0.22041376605549565, "grad_norm": 6.42008638381958, "learning_rate": 1.4805922800785733e-05, "loss": 1.727, "step": 1124 }, { "epoch": 0.22060986371212865, "grad_norm": 8.836663246154785, "learning_rate": 1.4684387314201919e-05, "loss": 2.6804, "step": 1125 }, { "epoch": 0.22080596136876166, "grad_norm": 6.903241157531738, "learning_rate": 1.4563313163639192e-05, "loss": 1.1425, "step": 1126 }, { "epoch": 0.22100205902539466, "grad_norm": 25.018720626831055, "learning_rate": 1.4442701003794434e-05, "loss": 2.5687, "step": 1127 }, { "epoch": 0.22119815668202766, "grad_norm": 4.393866062164307, "learning_rate": 1.4322551486866364e-05, "loss": 1.2885, "step": 1128 }, { "epoch": 0.22139425433866064, "grad_norm": 10.157917976379395, "learning_rate": 1.4202865262552045e-05, "loss": 2.6514, "step": 1129 }, { "epoch": 0.22159035199529364, "grad_norm": 8.52134895324707, "learning_rate": 1.4083642978043277e-05, "loss": 1.5537, "step": 1130 }, { "epoch": 0.22178644965192665, "grad_norm": 7.985105991363525, "learning_rate": 1.3964885278023176e-05, "loss": 1.631, "step": 1131 }, { "epoch": 0.22198254730855965, "grad_norm": 6.039412021636963, "learning_rate": 1.3846592804662638e-05, "loss": 0.9408, "step": 1132 }, { "epoch": 0.22217864496519266, "grad_norm": 6.0496697425842285, "learning_rate": 1.3728766197616905e-05, "loss": 2.2771, "step": 1133 }, { "epoch": 0.22237474262182566, "grad_norm": 6.791964530944824, "learning_rate": 1.3611406094022095e-05, "loss": 1.5164, "step": 1134 }, { "epoch": 0.22257084027845866, "grad_norm": 5.74752950668335, "learning_rate": 1.349451312849177e-05, "loss": 1.7676, "step": 1135 }, { "epoch": 0.22276693793509167, "grad_norm": 5.087756633758545, "learning_rate": 1.337808793311346e-05, "loss": 0.8442, "step": 1136 }, { "epoch": 0.22296303559172467, "grad_norm": 6.2235426902771, "learning_rate": 1.3262131137445266e-05, "loss": 1.7659, "step": 1137 }, { "epoch": 0.22315913324835768, "grad_norm": 5.494975566864014, "learning_rate": 1.314664336851248e-05, "loss": 2.0008, "step": 1138 }, { "epoch": 0.22335523090499068, "grad_norm": 7.123453617095947, "learning_rate": 1.3031625250804191e-05, "loss": 1.3864, "step": 1139 }, { "epoch": 0.22355132856162369, "grad_norm": 4.093453407287598, "learning_rate": 1.2917077406269872e-05, "loss": 0.8456, "step": 1140 }, { "epoch": 0.2237474262182567, "grad_norm": 9.605305671691895, "learning_rate": 1.280300045431605e-05, "loss": 2.1936, "step": 1141 }, { "epoch": 0.2239435238748897, "grad_norm": 5.773606777191162, "learning_rate": 1.2689395011802896e-05, "loss": 1.3258, "step": 1142 }, { "epoch": 0.2241396215315227, "grad_norm": 5.294035911560059, "learning_rate": 1.2576261693041036e-05, "loss": 1.6611, "step": 1143 }, { "epoch": 0.2243357191881557, "grad_norm": 8.730361938476562, "learning_rate": 1.2463601109788058e-05, "loss": 2.3499, "step": 1144 }, { "epoch": 0.2245318168447887, "grad_norm": 7.3704304695129395, "learning_rate": 1.235141387124531e-05, "loss": 1.6554, "step": 1145 }, { "epoch": 0.2247279145014217, "grad_norm": 6.475876331329346, "learning_rate": 1.2239700584054514e-05, "loss": 1.5567, "step": 1146 }, { "epoch": 0.22492401215805471, "grad_norm": 5.318692207336426, "learning_rate": 1.2128461852294647e-05, "loss": 1.6012, "step": 1147 }, { "epoch": 0.22512010981468772, "grad_norm": 8.062721252441406, "learning_rate": 1.2017698277478506e-05, "loss": 1.0389, "step": 1148 }, { "epoch": 0.22531620747132072, "grad_norm": 8.7105131149292, "learning_rate": 1.1907410458549517e-05, "loss": 3.1607, "step": 1149 }, { "epoch": 0.22551230512795373, "grad_norm": 8.629892349243164, "learning_rate": 1.1797598991878512e-05, "loss": 1.8994, "step": 1150 }, { "epoch": 0.22570840278458673, "grad_norm": 8.824653625488281, "learning_rate": 1.1688264471260546e-05, "loss": 0.8091, "step": 1151 }, { "epoch": 0.22590450044121974, "grad_norm": 4.759090900421143, "learning_rate": 1.1579407487911564e-05, "loss": 1.3273, "step": 1152 }, { "epoch": 0.22610059809785274, "grad_norm": 4.78541374206543, "learning_rate": 1.1471028630465285e-05, "loss": 0.8232, "step": 1153 }, { "epoch": 0.22629669575448574, "grad_norm": 6.233590602874756, "learning_rate": 1.136312848497002e-05, "loss": 2.525, "step": 1154 }, { "epoch": 0.22649279341111875, "grad_norm": 7.085466384887695, "learning_rate": 1.1255707634885526e-05, "loss": 1.6632, "step": 1155 }, { "epoch": 0.22668889106775175, "grad_norm": 4.938376426696777, "learning_rate": 1.1148766661079768e-05, "loss": 0.6869, "step": 1156 }, { "epoch": 0.22688498872438476, "grad_norm": 9.032883644104004, "learning_rate": 1.1042306141825808e-05, "loss": 1.7692, "step": 1157 }, { "epoch": 0.22708108638101776, "grad_norm": 7.324281215667725, "learning_rate": 1.0936326652798779e-05, "loss": 1.5287, "step": 1158 }, { "epoch": 0.22727718403765074, "grad_norm": 9.270583152770996, "learning_rate": 1.083082876707262e-05, "loss": 2.5909, "step": 1159 }, { "epoch": 0.22747328169428374, "grad_norm": 8.368681907653809, "learning_rate": 1.0725813055117051e-05, "loss": 1.3548, "step": 1160 }, { "epoch": 0.22766937935091675, "grad_norm": 5.176713943481445, "learning_rate": 1.0621280084794483e-05, "loss": 0.6788, "step": 1161 }, { "epoch": 0.22786547700754975, "grad_norm": 16.383853912353516, "learning_rate": 1.0517230421357016e-05, "loss": 1.4931, "step": 1162 }, { "epoch": 0.22806157466418275, "grad_norm": 6.631777286529541, "learning_rate": 1.041366462744322e-05, "loss": 1.8152, "step": 1163 }, { "epoch": 0.22825767232081576, "grad_norm": 6.580327033996582, "learning_rate": 1.0310583263075236e-05, "loss": 1.2307, "step": 1164 }, { "epoch": 0.22845376997744876, "grad_norm": 3.8760287761688232, "learning_rate": 1.0207986885655662e-05, "loss": 1.398, "step": 1165 }, { "epoch": 0.22864986763408177, "grad_norm": 4.494516849517822, "learning_rate": 1.0105876049964658e-05, "loss": 1.9966, "step": 1166 }, { "epoch": 0.22884596529071477, "grad_norm": 4.6880879402160645, "learning_rate": 1.0004251308156776e-05, "loss": 1.7868, "step": 1167 }, { "epoch": 0.22904206294734777, "grad_norm": 4.97517204284668, "learning_rate": 9.903113209758096e-06, "loss": 0.9129, "step": 1168 }, { "epoch": 0.22923816060398078, "grad_norm": 6.1455183029174805, "learning_rate": 9.802462301663218e-06, "loss": 1.0401, "step": 1169 }, { "epoch": 0.22943425826061378, "grad_norm": 5.264155387878418, "learning_rate": 9.702299128132286e-06, "loss": 0.6803, "step": 1170 }, { "epoch": 0.2296303559172468, "grad_norm": 7.0842108726501465, "learning_rate": 9.602624230788127e-06, "loss": 1.2855, "step": 1171 }, { "epoch": 0.2298264535738798, "grad_norm": 23.354158401489258, "learning_rate": 9.503438148613208e-06, "loss": 2.12, "step": 1172 }, { "epoch": 0.2300225512305128, "grad_norm": 4.803404331207275, "learning_rate": 9.404741417946783e-06, "loss": 1.6464, "step": 1173 }, { "epoch": 0.2302186488871458, "grad_norm": 5.195870876312256, "learning_rate": 9.306534572481996e-06, "loss": 2.3011, "step": 1174 }, { "epoch": 0.2304147465437788, "grad_norm": 4.52567720413208, "learning_rate": 9.208818143262965e-06, "loss": 0.6482, "step": 1175 }, { "epoch": 0.2306108442004118, "grad_norm": 6.062211990356445, "learning_rate": 9.111592658681933e-06, "loss": 1.5717, "step": 1176 }, { "epoch": 0.2308069418570448, "grad_norm": 6.672834396362305, "learning_rate": 9.014858644476444e-06, "loss": 1.468, "step": 1177 }, { "epoch": 0.23100303951367782, "grad_norm": 8.963277816772461, "learning_rate": 8.918616623726428e-06, "loss": 1.5148, "step": 1178 }, { "epoch": 0.23119913717031082, "grad_norm": 4.698469638824463, "learning_rate": 8.822867116851396e-06, "loss": 2.3368, "step": 1179 }, { "epoch": 0.23139523482694382, "grad_norm": 7.142313003540039, "learning_rate": 8.727610641607642e-06, "loss": 2.7635, "step": 1180 }, { "epoch": 0.23159133248357683, "grad_norm": 5.125174045562744, "learning_rate": 8.632847713085502e-06, "loss": 1.3292, "step": 1181 }, { "epoch": 0.23178743014020983, "grad_norm": 5.394103527069092, "learning_rate": 8.538578843706423e-06, "loss": 1.7812, "step": 1182 }, { "epoch": 0.23198352779684284, "grad_norm": 6.396340370178223, "learning_rate": 8.444804543220308e-06, "loss": 1.4261, "step": 1183 }, { "epoch": 0.23217962545347584, "grad_norm": 10.947870254516602, "learning_rate": 8.351525318702702e-06, "loss": 0.9685, "step": 1184 }, { "epoch": 0.23237572311010884, "grad_norm": 8.984673500061035, "learning_rate": 8.25874167455214e-06, "loss": 1.0452, "step": 1185 }, { "epoch": 0.23257182076674185, "grad_norm": 9.12095832824707, "learning_rate": 8.166454112487288e-06, "loss": 1.5966, "step": 1186 }, { "epoch": 0.23276791842337485, "grad_norm": 17.694997787475586, "learning_rate": 8.074663131544313e-06, "loss": 2.1847, "step": 1187 }, { "epoch": 0.23296401608000786, "grad_norm": 7.42760705947876, "learning_rate": 7.983369228074155e-06, "loss": 2.1763, "step": 1188 }, { "epoch": 0.23316011373664083, "grad_norm": 6.491481304168701, "learning_rate": 7.892572895739913e-06, "loss": 1.6988, "step": 1189 }, { "epoch": 0.23335621139327384, "grad_norm": 8.320481300354004, "learning_rate": 7.80227462551405e-06, "loss": 1.4067, "step": 1190 }, { "epoch": 0.23355230904990684, "grad_norm": 6.234914302825928, "learning_rate": 7.712474905675837e-06, "loss": 1.675, "step": 1191 }, { "epoch": 0.23374840670653985, "grad_norm": 6.310087203979492, "learning_rate": 7.62317422180866e-06, "loss": 2.4978, "step": 1192 }, { "epoch": 0.23394450436317285, "grad_norm": 7.758471488952637, "learning_rate": 7.53437305679745e-06, "loss": 1.8757, "step": 1193 }, { "epoch": 0.23414060201980585, "grad_norm": 4.8770318031311035, "learning_rate": 7.446071890826023e-06, "loss": 1.5988, "step": 1194 }, { "epoch": 0.23433669967643886, "grad_norm": 3.4752848148345947, "learning_rate": 7.358271201374478e-06, "loss": 1.526, "step": 1195 }, { "epoch": 0.23453279733307186, "grad_norm": 4.924532890319824, "learning_rate": 7.270971463216658e-06, "loss": 0.7961, "step": 1196 }, { "epoch": 0.23472889498970487, "grad_norm": 5.1697306632995605, "learning_rate": 7.184173148417561e-06, "loss": 1.114, "step": 1197 }, { "epoch": 0.23492499264633787, "grad_norm": 6.321658134460449, "learning_rate": 7.0978767263307764e-06, "loss": 1.0523, "step": 1198 }, { "epoch": 0.23512109030297088, "grad_norm": 4.862210273742676, "learning_rate": 7.012082663595931e-06, "loss": 2.3474, "step": 1199 }, { "epoch": 0.23531718795960388, "grad_norm": 6.735289096832275, "learning_rate": 6.9267914241362585e-06, "loss": 2.2363, "step": 1200 }, { "epoch": 0.23551328561623688, "grad_norm": 6.494600772857666, "learning_rate": 6.842003469155955e-06, "loss": 1.4925, "step": 1201 }, { "epoch": 0.2357093832728699, "grad_norm": 6.873384952545166, "learning_rate": 6.757719257137807e-06, "loss": 1.7298, "step": 1202 }, { "epoch": 0.2359054809295029, "grad_norm": 11.259360313415527, "learning_rate": 6.673939243840599e-06, "loss": 1.0081, "step": 1203 }, { "epoch": 0.2361015785861359, "grad_norm": 4.887358665466309, "learning_rate": 6.590663882296788e-06, "loss": 2.1434, "step": 1204 }, { "epoch": 0.2362976762427689, "grad_norm": 4.775937557220459, "learning_rate": 6.507893622809913e-06, "loss": 0.6404, "step": 1205 }, { "epoch": 0.2364937738994019, "grad_norm": 4.322073459625244, "learning_rate": 6.425628912952242e-06, "loss": 0.9223, "step": 1206 }, { "epoch": 0.2366898715560349, "grad_norm": 5.388957500457764, "learning_rate": 6.343870197562307e-06, "loss": 1.7039, "step": 1207 }, { "epoch": 0.2368859692126679, "grad_norm": 7.30275297164917, "learning_rate": 6.262617918742586e-06, "loss": 1.7324, "step": 1208 }, { "epoch": 0.23708206686930092, "grad_norm": 6.703151702880859, "learning_rate": 6.181872515857001e-06, "loss": 2.2659, "step": 1209 }, { "epoch": 0.23727816452593392, "grad_norm": 7.984116554260254, "learning_rate": 6.10163442552858e-06, "loss": 1.5091, "step": 1210 }, { "epoch": 0.23747426218256693, "grad_norm": 6.996639728546143, "learning_rate": 6.021904081637142e-06, "loss": 1.9216, "step": 1211 }, { "epoch": 0.23767035983919993, "grad_norm": 9.186728477478027, "learning_rate": 5.942681915316894e-06, "loss": 2.6348, "step": 1212 }, { "epoch": 0.23786645749583293, "grad_norm": 12.796127319335938, "learning_rate": 5.863968354954141e-06, "loss": 1.6889, "step": 1213 }, { "epoch": 0.23806255515246594, "grad_norm": 10.490285873413086, "learning_rate": 5.7857638261849314e-06, "loss": 2.4966, "step": 1214 }, { "epoch": 0.23825865280909894, "grad_norm": 4.777184009552002, "learning_rate": 5.70806875189277e-06, "loss": 1.7654, "step": 1215 }, { "epoch": 0.23845475046573195, "grad_norm": 5.094268321990967, "learning_rate": 5.630883552206367e-06, "loss": 0.7742, "step": 1216 }, { "epoch": 0.23865084812236495, "grad_norm": 3.6413767337799072, "learning_rate": 5.554208644497283e-06, "loss": 0.7798, "step": 1217 }, { "epoch": 0.23884694577899795, "grad_norm": 11.46200942993164, "learning_rate": 5.478044443377761e-06, "loss": 1.899, "step": 1218 }, { "epoch": 0.23904304343563093, "grad_norm": 6.592775821685791, "learning_rate": 5.402391360698456e-06, "loss": 0.7878, "step": 1219 }, { "epoch": 0.23923914109226393, "grad_norm": 4.292765140533447, "learning_rate": 5.327249805546175e-06, "loss": 1.1989, "step": 1220 }, { "epoch": 0.23943523874889694, "grad_norm": 5.002085208892822, "learning_rate": 5.2526201842416965e-06, "loss": 1.0088, "step": 1221 }, { "epoch": 0.23963133640552994, "grad_norm": 8.515301704406738, "learning_rate": 5.178502900337534e-06, "loss": 2.1646, "step": 1222 }, { "epoch": 0.23982743406216295, "grad_norm": 5.089056491851807, "learning_rate": 5.104898354615861e-06, "loss": 1.0671, "step": 1223 }, { "epoch": 0.24002353171879595, "grad_norm": 4.508873462677002, "learning_rate": 5.031806945086204e-06, "loss": 2.017, "step": 1224 }, { "epoch": 0.24021962937542896, "grad_norm": 5.015408992767334, "learning_rate": 4.959229066983373e-06, "loss": 2.8555, "step": 1225 }, { "epoch": 0.24041572703206196, "grad_norm": 5.895279407501221, "learning_rate": 4.887165112765291e-06, "loss": 0.8382, "step": 1226 }, { "epoch": 0.24061182468869496, "grad_norm": 4.138760566711426, "learning_rate": 4.8156154721109345e-06, "loss": 0.9274, "step": 1227 }, { "epoch": 0.24080792234532797, "grad_norm": 6.234400749206543, "learning_rate": 4.74458053191813e-06, "loss": 1.9066, "step": 1228 }, { "epoch": 0.24100402000196097, "grad_norm": 6.2309441566467285, "learning_rate": 4.674060676301528e-06, "loss": 1.2137, "step": 1229 }, { "epoch": 0.24120011765859398, "grad_norm": 6.614995956420898, "learning_rate": 4.604056286590497e-06, "loss": 2.1118, "step": 1230 }, { "epoch": 0.24139621531522698, "grad_norm": 7.630300045013428, "learning_rate": 4.534567741327112e-06, "loss": 1.028, "step": 1231 }, { "epoch": 0.24159231297185998, "grad_norm": 19.701765060424805, "learning_rate": 4.465595416264012e-06, "loss": 2.3872, "step": 1232 }, { "epoch": 0.241788410628493, "grad_norm": 9.740580558776855, "learning_rate": 4.397139684362462e-06, "loss": 1.8809, "step": 1233 }, { "epoch": 0.241984508285126, "grad_norm": 5.803571701049805, "learning_rate": 4.329200915790288e-06, "loss": 2.1341, "step": 1234 }, { "epoch": 0.242180605941759, "grad_norm": 5.827723503112793, "learning_rate": 4.261779477919892e-06, "loss": 0.8315, "step": 1235 }, { "epoch": 0.242376703598392, "grad_norm": 11.323637008666992, "learning_rate": 4.194875735326253e-06, "loss": 3.126, "step": 1236 }, { "epoch": 0.242572801255025, "grad_norm": 6.157869815826416, "learning_rate": 4.12849004978495e-06, "loss": 1.5295, "step": 1237 }, { "epoch": 0.242768898911658, "grad_norm": 3.730557680130005, "learning_rate": 4.062622780270253e-06, "loss": 0.8116, "step": 1238 }, { "epoch": 0.24296499656829101, "grad_norm": 7.499046325683594, "learning_rate": 3.9972742829531225e-06, "loss": 1.5176, "step": 1239 }, { "epoch": 0.24316109422492402, "grad_norm": 8.390069007873535, "learning_rate": 3.932444911199318e-06, "loss": 0.7114, "step": 1240 }, { "epoch": 0.24335719188155702, "grad_norm": 3.6422150135040283, "learning_rate": 3.8681350155674315e-06, "loss": 0.8022, "step": 1241 }, { "epoch": 0.24355328953819003, "grad_norm": 11.20263957977295, "learning_rate": 3.8043449438071256e-06, "loss": 1.6755, "step": 1242 }, { "epoch": 0.24374938719482303, "grad_norm": 8.747391700744629, "learning_rate": 3.7410750408571005e-06, "loss": 2.4434, "step": 1243 }, { "epoch": 0.24394548485145603, "grad_norm": 4.528903961181641, "learning_rate": 3.678325648843306e-06, "loss": 1.1245, "step": 1244 }, { "epoch": 0.24414158250808904, "grad_norm": 3.3688809871673584, "learning_rate": 3.6160971070770654e-06, "loss": 1.2526, "step": 1245 }, { "epoch": 0.24433768016472204, "grad_norm": 7.626687049865723, "learning_rate": 3.5543897520533e-06, "loss": 2.2074, "step": 1246 }, { "epoch": 0.24453377782135505, "grad_norm": 6.176375389099121, "learning_rate": 3.4932039174486174e-06, "loss": 1.9458, "step": 1247 }, { "epoch": 0.24472987547798805, "grad_norm": 3.6054394245147705, "learning_rate": 3.432539934119572e-06, "loss": 1.6104, "step": 1248 }, { "epoch": 0.24492597313462103, "grad_norm": 4.255305767059326, "learning_rate": 3.372398130100851e-06, "loss": 1.4093, "step": 1249 }, { "epoch": 0.24512207079125403, "grad_norm": 7.934302806854248, "learning_rate": 3.3127788306035134e-06, "loss": 1.5826, "step": 1250 }, { "epoch": 0.24531816844788704, "grad_norm": 3.8543338775634766, "learning_rate": 3.253682358013244e-06, "loss": 0.5287, "step": 1251 }, { "epoch": 0.24551426610452004, "grad_norm": 5.766990661621094, "learning_rate": 3.1951090318885434e-06, "loss": 2.0189, "step": 1252 }, { "epoch": 0.24571036376115304, "grad_norm": 5.633528232574463, "learning_rate": 3.1370591689590777e-06, "loss": 1.1827, "step": 1253 }, { "epoch": 0.24590646141778605, "grad_norm": 4.339439868927002, "learning_rate": 3.0795330831239313e-06, "loss": 1.5764, "step": 1254 }, { "epoch": 0.24610255907441905, "grad_norm": 7.025827407836914, "learning_rate": 3.022531085449931e-06, "loss": 1.94, "step": 1255 }, { "epoch": 0.24629865673105206, "grad_norm": 7.727778434753418, "learning_rate": 2.966053484169917e-06, "loss": 1.6985, "step": 1256 }, { "epoch": 0.24649475438768506, "grad_norm": 4.980061054229736, "learning_rate": 2.9101005846811304e-06, "loss": 0.6095, "step": 1257 }, { "epoch": 0.24669085204431807, "grad_norm": 9.262031555175781, "learning_rate": 2.854672689543514e-06, "loss": 1.9517, "step": 1258 }, { "epoch": 0.24688694970095107, "grad_norm": 10.269953727722168, "learning_rate": 2.7997700984781272e-06, "loss": 1.8526, "step": 1259 }, { "epoch": 0.24708304735758407, "grad_norm": 5.118125915527344, "learning_rate": 2.745393108365457e-06, "loss": 1.2967, "step": 1260 }, { "epoch": 0.24727914501421708, "grad_norm": 7.1361403465271, "learning_rate": 2.6915420132439085e-06, "loss": 1.9573, "step": 1261 }, { "epoch": 0.24747524267085008, "grad_norm": 5.416757583618164, "learning_rate": 2.638217104308127e-06, "loss": 1.2684, "step": 1262 }, { "epoch": 0.24767134032748309, "grad_norm": 5.381096363067627, "learning_rate": 2.585418669907458e-06, "loss": 0.922, "step": 1263 }, { "epoch": 0.2478674379841161, "grad_norm": 4.663347244262695, "learning_rate": 2.5331469955443778e-06, "loss": 0.8759, "step": 1264 }, { "epoch": 0.2480635356407491, "grad_norm": 12.182421684265137, "learning_rate": 2.4814023638729757e-06, "loss": 0.8895, "step": 1265 }, { "epoch": 0.2482596332973821, "grad_norm": 6.225351333618164, "learning_rate": 2.430185054697409e-06, "loss": 1.8041, "step": 1266 }, { "epoch": 0.2484557309540151, "grad_norm": 9.036142349243164, "learning_rate": 2.3794953449703837e-06, "loss": 2.0539, "step": 1267 }, { "epoch": 0.2486518286106481, "grad_norm": 6.467651844024658, "learning_rate": 2.3293335087916314e-06, "loss": 1.5472, "step": 1268 }, { "epoch": 0.2488479262672811, "grad_norm": 7.765258312225342, "learning_rate": 2.279699817406533e-06, "loss": 1.8689, "step": 1269 }, { "epoch": 0.24904402392391412, "grad_norm": 9.24152946472168, "learning_rate": 2.230594539204489e-06, "loss": 1.9637, "step": 1270 }, { "epoch": 0.24924012158054712, "grad_norm": 4.302461624145508, "learning_rate": 2.1820179397176287e-06, "loss": 0.8332, "step": 1271 }, { "epoch": 0.24943621923718012, "grad_norm": 5.98826789855957, "learning_rate": 2.133970281619246e-06, "loss": 3.038, "step": 1272 }, { "epoch": 0.24963231689381313, "grad_norm": 5.860217094421387, "learning_rate": 2.0864518247224797e-06, "loss": 1.7476, "step": 1273 }, { "epoch": 0.24982841455044613, "grad_norm": 7.621738910675049, "learning_rate": 2.039462825978822e-06, "loss": 1.7148, "step": 1274 }, { "epoch": 0.2500245122070791, "grad_norm": 6.9399614334106445, "learning_rate": 1.9930035394768233e-06, "loss": 1.3601, "step": 1275 }, { "epoch": 0.25022060986371214, "grad_norm": 17.67597007751465, "learning_rate": 1.947074216440592e-06, "loss": 3.0125, "step": 1276 }, { "epoch": 0.2504167075203451, "grad_norm": 4.356191158294678, "learning_rate": 1.9016751052285953e-06, "loss": 1.0786, "step": 1277 }, { "epoch": 0.25061280517697815, "grad_norm": 5.882324695587158, "learning_rate": 1.8568064513321715e-06, "loss": 1.2133, "step": 1278 }, { "epoch": 0.2508089028336111, "grad_norm": 4.599434852600098, "learning_rate": 1.8124684973742534e-06, "loss": 1.3988, "step": 1279 }, { "epoch": 0.25100500049024416, "grad_norm": 10.744729995727539, "learning_rate": 1.768661483108136e-06, "loss": 2.4072, "step": 1280 }, { "epoch": 0.25120109814687713, "grad_norm": 4.716323375701904, "learning_rate": 1.7253856454160333e-06, "loss": 2.1812, "step": 1281 }, { "epoch": 0.25139719580351017, "grad_norm": 3.4088003635406494, "learning_rate": 1.682641218307901e-06, "loss": 0.8226, "step": 1282 }, { "epoch": 0.25159329346014314, "grad_norm": 5.354026794433594, "learning_rate": 1.640428432920138e-06, "loss": 1.3017, "step": 1283 }, { "epoch": 0.2517893911167762, "grad_norm": 4.884855270385742, "learning_rate": 1.5987475175143651e-06, "loss": 0.6766, "step": 1284 }, { "epoch": 0.25198548877340915, "grad_norm": 8.052160263061523, "learning_rate": 1.557598697476148e-06, "loss": 2.2202, "step": 1285 }, { "epoch": 0.2521815864300422, "grad_norm": 5.809813976287842, "learning_rate": 1.5169821953137875e-06, "loss": 1.1507, "step": 1286 }, { "epoch": 0.25237768408667516, "grad_norm": 5.726134300231934, "learning_rate": 1.47689823065712e-06, "loss": 1.3755, "step": 1287 }, { "epoch": 0.2525737817433082, "grad_norm": 6.9269022941589355, "learning_rate": 1.4373470202563855e-06, "loss": 1.9541, "step": 1288 }, { "epoch": 0.25276987939994117, "grad_norm": 6.82175874710083, "learning_rate": 1.398328777980973e-06, "loss": 0.9483, "step": 1289 }, { "epoch": 0.2529659770565742, "grad_norm": 5.6992058753967285, "learning_rate": 1.3598437148182652e-06, "loss": 1.8013, "step": 1290 }, { "epoch": 0.2531620747132072, "grad_norm": 6.446459770202637, "learning_rate": 1.3218920388725853e-06, "loss": 1.059, "step": 1291 }, { "epoch": 0.2533581723698402, "grad_norm": 5.6021223068237305, "learning_rate": 1.2844739553640073e-06, "loss": 2.3983, "step": 1292 }, { "epoch": 0.2535542700264732, "grad_norm": 5.945677757263184, "learning_rate": 1.2475896666272136e-06, "loss": 1.3391, "step": 1293 }, { "epoch": 0.25375036768310616, "grad_norm": 8.367269515991211, "learning_rate": 1.2112393721104843e-06, "loss": 0.9355, "step": 1294 }, { "epoch": 0.2539464653397392, "grad_norm": 4.452106952667236, "learning_rate": 1.1754232683745537e-06, "loss": 1.3174, "step": 1295 }, { "epoch": 0.25414256299637217, "grad_norm": 5.392490863800049, "learning_rate": 1.1401415490916e-06, "loss": 0.7378, "step": 1296 }, { "epoch": 0.2543386606530052, "grad_norm": 7.21303129196167, "learning_rate": 1.1053944050441245e-06, "loss": 2.9174, "step": 1297 }, { "epoch": 0.2545347583096382, "grad_norm": 8.784597396850586, "learning_rate": 1.0711820241240067e-06, "loss": 3.5646, "step": 1298 }, { "epoch": 0.2547308559662712, "grad_norm": 6.628826141357422, "learning_rate": 1.0375045913314063e-06, "loss": 1.2581, "step": 1299 }, { "epoch": 0.2549269536229042, "grad_norm": 11.671515464782715, "learning_rate": 1.0043622887738413e-06, "loss": 2.1951, "step": 1300 }, { "epoch": 0.2551230512795372, "grad_norm": 4.008133888244629, "learning_rate": 9.717552956651331e-07, "loss": 2.3149, "step": 1301 }, { "epoch": 0.2553191489361702, "grad_norm": 8.170031547546387, "learning_rate": 9.396837883244746e-07, "loss": 1.576, "step": 1302 }, { "epoch": 0.2555152465928032, "grad_norm": 7.534716606140137, "learning_rate": 9.081479401754966e-07, "loss": 2.1834, "step": 1303 }, { "epoch": 0.2557113442494362, "grad_norm": 4.619508266448975, "learning_rate": 8.771479217452694e-07, "loss": 1.6227, "step": 1304 }, { "epoch": 0.25590744190606923, "grad_norm": 4.464870452880859, "learning_rate": 8.466839006634364e-07, "loss": 0.6817, "step": 1305 }, { "epoch": 0.2561035395627022, "grad_norm": 7.0456743240356445, "learning_rate": 8.167560416612596e-07, "loss": 2.6702, "step": 1306 }, { "epoch": 0.25629963721933524, "grad_norm": 3.647512674331665, "learning_rate": 7.873645065708091e-07, "loss": 2.1032, "step": 1307 }, { "epoch": 0.2564957348759682, "grad_norm": 10.428240776062012, "learning_rate": 7.585094543239857e-07, "loss": 1.6973, "step": 1308 }, { "epoch": 0.25669183253260125, "grad_norm": 4.8686299324035645, "learning_rate": 7.301910409517221e-07, "loss": 1.8198, "step": 1309 }, { "epoch": 0.2568879301892342, "grad_norm": 13.215803146362305, "learning_rate": 7.024094195831277e-07, "loss": 1.791, "step": 1310 }, { "epoch": 0.25708402784586726, "grad_norm": 7.418087959289551, "learning_rate": 6.751647404446781e-07, "loss": 1.0254, "step": 1311 }, { "epoch": 0.25728012550250023, "grad_norm": 3.0393893718719482, "learning_rate": 6.484571508593718e-07, "loss": 0.429, "step": 1312 }, { "epoch": 0.25747622315913327, "grad_norm": 8.804031372070312, "learning_rate": 6.222867952459299e-07, "loss": 1.6825, "step": 1313 }, { "epoch": 0.25767232081576624, "grad_norm": 13.433954238891602, "learning_rate": 5.966538151180645e-07, "loss": 1.9529, "step": 1314 }, { "epoch": 0.2578684184723993, "grad_norm": 5.364668369293213, "learning_rate": 5.715583490836673e-07, "loss": 1.2584, "step": 1315 }, { "epoch": 0.25806451612903225, "grad_norm": 4.066149711608887, "learning_rate": 5.470005328440664e-07, "loss": 1.6191, "step": 1316 }, { "epoch": 0.2582606137856653, "grad_norm": 8.135047912597656, "learning_rate": 5.22980499193304e-07, "loss": 2.0681, "step": 1317 }, { "epoch": 0.25845671144229826, "grad_norm": 3.4767017364501953, "learning_rate": 4.994983780174156e-07, "loss": 1.0348, "step": 1318 }, { "epoch": 0.2586528090989313, "grad_norm": 6.77761173248291, "learning_rate": 4.7655429629372973e-07, "loss": 0.7133, "step": 1319 }, { "epoch": 0.25884890675556427, "grad_norm": 4.188867092132568, "learning_rate": 4.5414837809018007e-07, "loss": 1.6745, "step": 1320 }, { "epoch": 0.2590450044121973, "grad_norm": 3.1510398387908936, "learning_rate": 4.322807445646171e-07, "loss": 0.7392, "step": 1321 }, { "epoch": 0.2592411020688303, "grad_norm": 14.09803581237793, "learning_rate": 4.1095151396418617e-07, "loss": 2.5941, "step": 1322 }, { "epoch": 0.2594371997254633, "grad_norm": 7.8199028968811035, "learning_rate": 3.9016080162466164e-07, "loss": 1.2837, "step": 1323 }, { "epoch": 0.2596332973820963, "grad_norm": 4.440105438232422, "learning_rate": 3.699087199698359e-07, "loss": 1.7956, "step": 1324 }, { "epoch": 0.25982939503872926, "grad_norm": 7.9481048583984375, "learning_rate": 3.50195378510898e-07, "loss": 0.7979, "step": 1325 }, { "epoch": 0.2600254926953623, "grad_norm": 6.204482555389404, "learning_rate": 3.310208838458562e-07, "loss": 2.1294, "step": 1326 }, { "epoch": 0.26022159035199527, "grad_norm": 6.299440860748291, "learning_rate": 3.1238533965897156e-07, "loss": 1.4474, "step": 1327 }, { "epoch": 0.2604176880086283, "grad_norm": 5.203184604644775, "learning_rate": 2.9428884672015876e-07, "loss": 1.7164, "step": 1328 }, { "epoch": 0.2606137856652613, "grad_norm": 5.018675804138184, "learning_rate": 2.7673150288447526e-07, "loss": 1.0184, "step": 1329 }, { "epoch": 0.2608098833218943, "grad_norm": 5.618327617645264, "learning_rate": 2.597134030915771e-07, "loss": 1.4616, "step": 1330 }, { "epoch": 0.2610059809785273, "grad_norm": 10.888134002685547, "learning_rate": 2.432346393652196e-07, "loss": 2.6353, "step": 1331 }, { "epoch": 0.2612020786351603, "grad_norm": 4.126128196716309, "learning_rate": 2.2729530081273542e-07, "loss": 1.0003, "step": 1332 }, { "epoch": 0.2613981762917933, "grad_norm": 7.105251789093018, "learning_rate": 2.118954736245682e-07, "loss": 1.4611, "step": 1333 }, { "epoch": 0.2615942739484263, "grad_norm": 4.438936233520508, "learning_rate": 1.9703524107382855e-07, "loss": 1.0069, "step": 1334 }, { "epoch": 0.2617903716050593, "grad_norm": 8.27482795715332, "learning_rate": 1.8271468351579446e-07, "loss": 2.8264, "step": 1335 }, { "epoch": 0.26198646926169233, "grad_norm": 7.972108840942383, "learning_rate": 1.6893387838750053e-07, "loss": 2.4485, "step": 1336 }, { "epoch": 0.2621825669183253, "grad_norm": 5.920525074005127, "learning_rate": 1.5569290020736039e-07, "loss": 1.4498, "step": 1337 }, { "epoch": 0.26237866457495834, "grad_norm": 5.699307918548584, "learning_rate": 1.429918205746672e-07, "loss": 2.0144, "step": 1338 }, { "epoch": 0.2625747622315913, "grad_norm": 2.4948651790618896, "learning_rate": 1.308307081693272e-07, "loss": 0.7282, "step": 1339 }, { "epoch": 0.26277085988822435, "grad_norm": 9.36095142364502, "learning_rate": 1.192096287513711e-07, "loss": 2.026, "step": 1340 }, { "epoch": 0.2629669575448573, "grad_norm": 11.627699851989746, "learning_rate": 1.0812864516067667e-07, "loss": 1.6524, "step": 1341 }, { "epoch": 0.26316305520149036, "grad_norm": 4.287498474121094, "learning_rate": 9.758781731661337e-08, "loss": 1.1361, "step": 1342 }, { "epoch": 0.26335915285812334, "grad_norm": 3.8631551265716553, "learning_rate": 8.758720221768713e-08, "loss": 1.0284, "step": 1343 }, { "epoch": 0.26355525051475637, "grad_norm": 7.8086323738098145, "learning_rate": 7.812685394127384e-08, "loss": 1.3352, "step": 1344 }, { "epoch": 0.26375134817138934, "grad_norm": 7.071966171264648, "learning_rate": 6.920682364330855e-08, "loss": 1.783, "step": 1345 }, { "epoch": 0.2639474458280224, "grad_norm": 6.961006164550781, "learning_rate": 6.082715955800789e-08, "loss": 0.666, "step": 1346 }, { "epoch": 0.26414354348465535, "grad_norm": 10.395101547241211, "learning_rate": 5.298790699758138e-08, "loss": 1.6436, "step": 1347 }, { "epoch": 0.2643396411412884, "grad_norm": 6.384310722351074, "learning_rate": 4.5689108352053867e-08, "loss": 1.5711, "step": 1348 }, { "epoch": 0.26453573879792136, "grad_norm": 5.545608997344971, "learning_rate": 3.893080308898789e-08, "loss": 1.2522, "step": 1349 }, { "epoch": 0.2647318364545544, "grad_norm": 17.074024200439453, "learning_rate": 3.271302775325058e-08, "loss": 3.1042, "step": 1350 }, { "epoch": 0.26492793411118737, "grad_norm": 7.239038944244385, "learning_rate": 2.7035815966891532e-08, "loss": 1.658, "step": 1351 }, { "epoch": 0.2651240317678204, "grad_norm": 6.944170951843262, "learning_rate": 2.1899198428876333e-08, "loss": 1.5184, "step": 1352 }, { "epoch": 0.2653201294244534, "grad_norm": 5.548795700073242, "learning_rate": 1.730320291498666e-08, "loss": 1.6825, "step": 1353 }, { "epoch": 0.26551622708108635, "grad_norm": 8.901213645935059, "learning_rate": 1.3247854277609328e-08, "loss": 1.1894, "step": 1354 }, { "epoch": 0.2657123247377194, "grad_norm": 9.556551933288574, "learning_rate": 9.73317444566968e-09, "loss": 2.2414, "step": 1355 }, { "epoch": 0.26590842239435236, "grad_norm": 6.794140338897705, "learning_rate": 6.759182424453947e-09, "loss": 1.3999, "step": 1356 }, { "epoch": 0.2661045200509854, "grad_norm": 5.626242637634277, "learning_rate": 4.325894295553745e-09, "loss": 1.5497, "step": 1357 }, { "epoch": 0.26630061770761837, "grad_norm": 6.506687164306641, "learning_rate": 2.433323216721739e-09, "loss": 2.8993, "step": 1358 }, { "epoch": 0.2664967153642514, "grad_norm": 5.048938274383545, "learning_rate": 1.081479421871645e-09, "loss": 1.7226, "step": 1359 }, { "epoch": 0.2666928130208844, "grad_norm": 5.443484306335449, "learning_rate": 2.7037022096720876e-10, "loss": 0.9488, "step": 1360 }, { "epoch": 0.2668889106775174, "grad_norm": 17.070817947387695, "learning_rate": 0.0, "loss": 2.8164, "step": 1361 } ], "logging_steps": 1, "max_steps": 1361, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 341, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.914693202093015e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }