{ "best_metric": 0.8778727054595947, "best_model_checkpoint": "miner_id_24/checkpoint-500", "epoch": 0.32663726931242854, "eval_steps": 100, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006532745386248571, "grad_norm": 11.547571182250977, "learning_rate": 5e-06, "loss": 4.6466, "step": 1 }, { "epoch": 0.0006532745386248571, "eval_loss": 1.2989917993545532, "eval_runtime": 190.9993, "eval_samples_per_second": 13.497, "eval_steps_per_second": 3.377, "step": 1 }, { "epoch": 0.0013065490772497142, "grad_norm": 11.975031852722168, "learning_rate": 1e-05, "loss": 5.0526, "step": 2 }, { "epoch": 0.0019598236158745713, "grad_norm": 10.511598587036133, "learning_rate": 1.5e-05, "loss": 4.8616, "step": 3 }, { "epoch": 0.0026130981544994283, "grad_norm": 8.188464164733887, "learning_rate": 2e-05, "loss": 4.739, "step": 4 }, { "epoch": 0.0032663726931242854, "grad_norm": 4.608264446258545, "learning_rate": 2.5e-05, "loss": 4.2898, "step": 5 }, { "epoch": 0.0039196472317491425, "grad_norm": 4.9322381019592285, "learning_rate": 3e-05, "loss": 4.4307, "step": 6 }, { "epoch": 0.004572921770374, "grad_norm": 5.750156402587891, "learning_rate": 3.5e-05, "loss": 4.3233, "step": 7 }, { "epoch": 0.005226196308998857, "grad_norm": 5.375987529754639, "learning_rate": 4e-05, "loss": 4.235, "step": 8 }, { "epoch": 0.005879470847623714, "grad_norm": 4.336426734924316, "learning_rate": 4.5e-05, "loss": 4.2463, "step": 9 }, { "epoch": 0.006532745386248571, "grad_norm": 3.931546211242676, "learning_rate": 5e-05, "loss": 4.0456, "step": 10 }, { "epoch": 0.007186019924873428, "grad_norm": 4.456380844116211, "learning_rate": 5.500000000000001e-05, "loss": 4.2932, "step": 11 }, { "epoch": 0.007839294463498285, "grad_norm": 4.070809364318848, "learning_rate": 6e-05, "loss": 3.968, "step": 12 }, { "epoch": 0.008492569002123142, "grad_norm": 3.567103147506714, "learning_rate": 6.500000000000001e-05, "loss": 4.1319, "step": 13 }, { "epoch": 0.009145843540748, "grad_norm": 3.41229248046875, "learning_rate": 7e-05, "loss": 3.9014, "step": 14 }, { "epoch": 0.009799118079372856, "grad_norm": 3.476752281188965, "learning_rate": 7.500000000000001e-05, "loss": 3.9473, "step": 15 }, { "epoch": 0.010452392617997713, "grad_norm": 3.421405076980591, "learning_rate": 8e-05, "loss": 3.6684, "step": 16 }, { "epoch": 0.01110566715662257, "grad_norm": 3.7692196369171143, "learning_rate": 8.5e-05, "loss": 3.6834, "step": 17 }, { "epoch": 0.011758941695247428, "grad_norm": 3.8790102005004883, "learning_rate": 9e-05, "loss": 4.1705, "step": 18 }, { "epoch": 0.012412216233872285, "grad_norm": 4.199179172515869, "learning_rate": 9.5e-05, "loss": 4.0168, "step": 19 }, { "epoch": 0.013065490772497142, "grad_norm": 4.3811516761779785, "learning_rate": 0.0001, "loss": 4.2636, "step": 20 }, { "epoch": 0.013718765311121999, "grad_norm": 4.086483478546143, "learning_rate": 9.999892908320647e-05, "loss": 3.386, "step": 21 }, { "epoch": 0.014372039849746856, "grad_norm": 4.643093585968018, "learning_rate": 9.999571637870036e-05, "loss": 4.1357, "step": 22 }, { "epoch": 0.015025314388371713, "grad_norm": 4.607544422149658, "learning_rate": 9.999036202410325e-05, "loss": 3.4956, "step": 23 }, { "epoch": 0.01567858892699657, "grad_norm": 4.911647319793701, "learning_rate": 9.998286624877786e-05, "loss": 3.5906, "step": 24 }, { "epoch": 0.01633186346562143, "grad_norm": 5.430598735809326, "learning_rate": 9.997322937381829e-05, "loss": 4.3089, "step": 25 }, { "epoch": 0.016985138004246284, "grad_norm": 5.196270942687988, "learning_rate": 9.996145181203615e-05, "loss": 3.7026, "step": 26 }, { "epoch": 0.017638412542871143, "grad_norm": 6.21434211730957, "learning_rate": 9.994753406794301e-05, "loss": 3.8458, "step": 27 }, { "epoch": 0.018291687081496, "grad_norm": 6.019957542419434, "learning_rate": 9.99314767377287e-05, "loss": 3.638, "step": 28 }, { "epoch": 0.018944961620120857, "grad_norm": 7.344250679016113, "learning_rate": 9.991328050923581e-05, "loss": 3.5139, "step": 29 }, { "epoch": 0.019598236158745713, "grad_norm": 8.33799934387207, "learning_rate": 9.989294616193017e-05, "loss": 3.8276, "step": 30 }, { "epoch": 0.02025151069737057, "grad_norm": 7.272817611694336, "learning_rate": 9.98704745668676e-05, "loss": 3.7295, "step": 31 }, { "epoch": 0.020904785235995427, "grad_norm": 8.105911254882812, "learning_rate": 9.98458666866564e-05, "loss": 3.6633, "step": 32 }, { "epoch": 0.021558059774620286, "grad_norm": 10.303070068359375, "learning_rate": 9.981912357541627e-05, "loss": 4.163, "step": 33 }, { "epoch": 0.02221133431324514, "grad_norm": 11.288853645324707, "learning_rate": 9.97902463787331e-05, "loss": 3.8152, "step": 34 }, { "epoch": 0.02286460885187, "grad_norm": 11.985376358032227, "learning_rate": 9.975923633360985e-05, "loss": 3.189, "step": 35 }, { "epoch": 0.023517883390494855, "grad_norm": 11.919270515441895, "learning_rate": 9.972609476841367e-05, "loss": 3.3431, "step": 36 }, { "epoch": 0.024171157929119714, "grad_norm": 11.31579303741455, "learning_rate": 9.969082310281891e-05, "loss": 3.6613, "step": 37 }, { "epoch": 0.02482443246774457, "grad_norm": 13.086581230163574, "learning_rate": 9.965342284774632e-05, "loss": 3.2378, "step": 38 }, { "epoch": 0.025477707006369428, "grad_norm": 13.448277473449707, "learning_rate": 9.961389560529836e-05, "loss": 3.7662, "step": 39 }, { "epoch": 0.026130981544994283, "grad_norm": 15.159260749816895, "learning_rate": 9.957224306869053e-05, "loss": 3.0795, "step": 40 }, { "epoch": 0.026784256083619142, "grad_norm": 16.633817672729492, "learning_rate": 9.952846702217886e-05, "loss": 3.2916, "step": 41 }, { "epoch": 0.027437530622243998, "grad_norm": 17.962066650390625, "learning_rate": 9.948256934098352e-05, "loss": 3.0719, "step": 42 }, { "epoch": 0.028090805160868856, "grad_norm": 19.26710319519043, "learning_rate": 9.943455199120837e-05, "loss": 3.4875, "step": 43 }, { "epoch": 0.02874407969949371, "grad_norm": 23.785097122192383, "learning_rate": 9.938441702975689e-05, "loss": 3.5956, "step": 44 }, { "epoch": 0.02939735423811857, "grad_norm": 25.271921157836914, "learning_rate": 9.933216660424395e-05, "loss": 3.0942, "step": 45 }, { "epoch": 0.030050628776743426, "grad_norm": 26.448965072631836, "learning_rate": 9.927780295290389e-05, "loss": 3.391, "step": 46 }, { "epoch": 0.030703903315368285, "grad_norm": 21.440593719482422, "learning_rate": 9.922132840449459e-05, "loss": 2.6666, "step": 47 }, { "epoch": 0.03135717785399314, "grad_norm": 25.863950729370117, "learning_rate": 9.916274537819775e-05, "loss": 2.983, "step": 48 }, { "epoch": 0.032010452392617995, "grad_norm": 31.86448097229004, "learning_rate": 9.91020563835152e-05, "loss": 3.6128, "step": 49 }, { "epoch": 0.03266372693124286, "grad_norm": 47.512001037597656, "learning_rate": 9.903926402016153e-05, "loss": 5.6647, "step": 50 }, { "epoch": 0.03331700146986771, "grad_norm": 7.3868536949157715, "learning_rate": 9.897437097795257e-05, "loss": 4.1604, "step": 51 }, { "epoch": 0.03397027600849257, "grad_norm": 4.953103065490723, "learning_rate": 9.890738003669029e-05, "loss": 4.5837, "step": 52 }, { "epoch": 0.034623550547117424, "grad_norm": 3.1169042587280273, "learning_rate": 9.883829406604363e-05, "loss": 4.6018, "step": 53 }, { "epoch": 0.035276825085742286, "grad_norm": 2.86468505859375, "learning_rate": 9.876711602542563e-05, "loss": 4.3774, "step": 54 }, { "epoch": 0.03593009962436714, "grad_norm": 2.981433868408203, "learning_rate": 9.869384896386668e-05, "loss": 4.2887, "step": 55 }, { "epoch": 0.036583374162992, "grad_norm": 2.8601815700531006, "learning_rate": 9.861849601988383e-05, "loss": 3.9852, "step": 56 }, { "epoch": 0.03723664870161685, "grad_norm": 2.5838541984558105, "learning_rate": 9.854106042134641e-05, "loss": 3.9778, "step": 57 }, { "epoch": 0.037889923240241714, "grad_norm": 2.4947381019592285, "learning_rate": 9.846154548533773e-05, "loss": 4.0099, "step": 58 }, { "epoch": 0.03854319777886657, "grad_norm": 2.7639665603637695, "learning_rate": 9.837995461801299e-05, "loss": 4.0092, "step": 59 }, { "epoch": 0.039196472317491425, "grad_norm": 2.836315631866455, "learning_rate": 9.829629131445342e-05, "loss": 3.7736, "step": 60 }, { "epoch": 0.03984974685611628, "grad_norm": 2.828981637954712, "learning_rate": 9.821055915851647e-05, "loss": 3.8882, "step": 61 }, { "epoch": 0.04050302139474114, "grad_norm": 3.3011579513549805, "learning_rate": 9.812276182268236e-05, "loss": 4.1102, "step": 62 }, { "epoch": 0.041156295933366, "grad_norm": 2.943286180496216, "learning_rate": 9.803290306789676e-05, "loss": 4.0913, "step": 63 }, { "epoch": 0.04180957047199085, "grad_norm": 2.9910857677459717, "learning_rate": 9.794098674340965e-05, "loss": 3.8973, "step": 64 }, { "epoch": 0.04246284501061571, "grad_norm": 3.0220143795013428, "learning_rate": 9.784701678661045e-05, "loss": 3.8295, "step": 65 }, { "epoch": 0.04311611954924057, "grad_norm": 3.3190722465515137, "learning_rate": 9.775099722285935e-05, "loss": 4.1248, "step": 66 }, { "epoch": 0.043769394087865426, "grad_norm": 3.1170575618743896, "learning_rate": 9.765293216531486e-05, "loss": 3.876, "step": 67 }, { "epoch": 0.04442266862649028, "grad_norm": 3.3601784706115723, "learning_rate": 9.755282581475769e-05, "loss": 4.2187, "step": 68 }, { "epoch": 0.04507594316511514, "grad_norm": 3.230534553527832, "learning_rate": 9.74506824594107e-05, "loss": 4.0142, "step": 69 }, { "epoch": 0.04572921770374, "grad_norm": 3.33896541595459, "learning_rate": 9.73465064747553e-05, "loss": 3.869, "step": 70 }, { "epoch": 0.046382492242364855, "grad_norm": 3.4622280597686768, "learning_rate": 9.724030232334391e-05, "loss": 3.9758, "step": 71 }, { "epoch": 0.04703576678098971, "grad_norm": 3.7361583709716797, "learning_rate": 9.713207455460894e-05, "loss": 3.7177, "step": 72 }, { "epoch": 0.047689041319614565, "grad_norm": 4.231381416320801, "learning_rate": 9.702182780466775e-05, "loss": 4.1549, "step": 73 }, { "epoch": 0.04834231585823943, "grad_norm": 4.004554271697998, "learning_rate": 9.690956679612421e-05, "loss": 3.6364, "step": 74 }, { "epoch": 0.04899559039686428, "grad_norm": 3.9489033222198486, "learning_rate": 9.67952963378663e-05, "loss": 3.6562, "step": 75 }, { "epoch": 0.04964886493548914, "grad_norm": 4.792303562164307, "learning_rate": 9.667902132486009e-05, "loss": 4.2874, "step": 76 }, { "epoch": 0.050302139474113994, "grad_norm": 5.080928802490234, "learning_rate": 9.656074673794018e-05, "loss": 3.4885, "step": 77 }, { "epoch": 0.050955414012738856, "grad_norm": 5.956547737121582, "learning_rate": 9.644047764359622e-05, "loss": 3.8888, "step": 78 }, { "epoch": 0.05160868855136371, "grad_norm": 5.492791652679443, "learning_rate": 9.631821919375591e-05, "loss": 3.5125, "step": 79 }, { "epoch": 0.05226196308998857, "grad_norm": 5.798184394836426, "learning_rate": 9.619397662556435e-05, "loss": 3.6023, "step": 80 }, { "epoch": 0.05291523762861342, "grad_norm": 6.541443824768066, "learning_rate": 9.606775526115963e-05, "loss": 3.9795, "step": 81 }, { "epoch": 0.053568512167238284, "grad_norm": 6.9007248878479, "learning_rate": 9.593956050744492e-05, "loss": 3.9131, "step": 82 }, { "epoch": 0.05422178670586314, "grad_norm": 7.156078815460205, "learning_rate": 9.580939785585681e-05, "loss": 3.5494, "step": 83 }, { "epoch": 0.054875061244487995, "grad_norm": 7.655080318450928, "learning_rate": 9.567727288213005e-05, "loss": 3.526, "step": 84 }, { "epoch": 0.05552833578311285, "grad_norm": 8.81391429901123, "learning_rate": 9.554319124605879e-05, "loss": 3.9594, "step": 85 }, { "epoch": 0.05618161032173771, "grad_norm": 9.645224571228027, "learning_rate": 9.540715869125407e-05, "loss": 3.8675, "step": 86 }, { "epoch": 0.05683488486036257, "grad_norm": 9.505653381347656, "learning_rate": 9.526918104489777e-05, "loss": 3.3039, "step": 87 }, { "epoch": 0.05748815939898742, "grad_norm": 11.144811630249023, "learning_rate": 9.512926421749304e-05, "loss": 3.289, "step": 88 }, { "epoch": 0.05814143393761228, "grad_norm": 11.904159545898438, "learning_rate": 9.498741420261108e-05, "loss": 3.5135, "step": 89 }, { "epoch": 0.05879470847623714, "grad_norm": 11.668728828430176, "learning_rate": 9.484363707663442e-05, "loss": 3.1836, "step": 90 }, { "epoch": 0.059447983014861996, "grad_norm": 13.778961181640625, "learning_rate": 9.469793899849661e-05, "loss": 3.6804, "step": 91 }, { "epoch": 0.06010125755348685, "grad_norm": 13.867650985717773, "learning_rate": 9.45503262094184e-05, "loss": 3.4685, "step": 92 }, { "epoch": 0.06075453209211171, "grad_norm": 17.512781143188477, "learning_rate": 9.440080503264037e-05, "loss": 3.3502, "step": 93 }, { "epoch": 0.06140780663073657, "grad_norm": 24.185264587402344, "learning_rate": 9.42493818731521e-05, "loss": 2.7727, "step": 94 }, { "epoch": 0.062061081169361425, "grad_norm": 19.322153091430664, "learning_rate": 9.409606321741775e-05, "loss": 3.4637, "step": 95 }, { "epoch": 0.06271435570798628, "grad_norm": 20.148555755615234, "learning_rate": 9.394085563309827e-05, "loss": 2.7134, "step": 96 }, { "epoch": 0.06336763024661114, "grad_norm": 19.66134262084961, "learning_rate": 9.378376576876999e-05, "loss": 2.1593, "step": 97 }, { "epoch": 0.06402090478523599, "grad_norm": 21.868074417114258, "learning_rate": 9.362480035363986e-05, "loss": 3.1943, "step": 98 }, { "epoch": 0.06467417932386085, "grad_norm": 24.685863494873047, "learning_rate": 9.34639661972572e-05, "loss": 2.8522, "step": 99 }, { "epoch": 0.06532745386248572, "grad_norm": 32.843997955322266, "learning_rate": 9.330127018922194e-05, "loss": 3.7582, "step": 100 }, { "epoch": 0.06532745386248572, "eval_loss": 1.087826132774353, "eval_runtime": 193.6149, "eval_samples_per_second": 13.315, "eval_steps_per_second": 3.331, "step": 100 }, { "epoch": 0.06598072840111056, "grad_norm": 6.478489875793457, "learning_rate": 9.31367192988896e-05, "loss": 4.3501, "step": 101 }, { "epoch": 0.06663400293973543, "grad_norm": 4.0702080726623535, "learning_rate": 9.297032057507264e-05, "loss": 4.2697, "step": 102 }, { "epoch": 0.06728727747836027, "grad_norm": 2.496466875076294, "learning_rate": 9.280208114573859e-05, "loss": 3.8445, "step": 103 }, { "epoch": 0.06794055201698514, "grad_norm": 2.428135633468628, "learning_rate": 9.263200821770461e-05, "loss": 4.0675, "step": 104 }, { "epoch": 0.06859382655561, "grad_norm": 2.5381500720977783, "learning_rate": 9.246010907632895e-05, "loss": 4.223, "step": 105 }, { "epoch": 0.06924710109423485, "grad_norm": 2.4143152236938477, "learning_rate": 9.228639108519868e-05, "loss": 3.9627, "step": 106 }, { "epoch": 0.06990037563285971, "grad_norm": 2.6589813232421875, "learning_rate": 9.211086168581433e-05, "loss": 4.0724, "step": 107 }, { "epoch": 0.07055365017148457, "grad_norm": 2.906999349594116, "learning_rate": 9.193352839727121e-05, "loss": 4.0853, "step": 108 }, { "epoch": 0.07120692471010942, "grad_norm": 2.604168653488159, "learning_rate": 9.175439881593716e-05, "loss": 3.9059, "step": 109 }, { "epoch": 0.07186019924873428, "grad_norm": 2.6407856941223145, "learning_rate": 9.157348061512727e-05, "loss": 4.1648, "step": 110 }, { "epoch": 0.07251347378735913, "grad_norm": 2.5845110416412354, "learning_rate": 9.139078154477512e-05, "loss": 3.9705, "step": 111 }, { "epoch": 0.073166748325984, "grad_norm": 2.725156784057617, "learning_rate": 9.120630943110077e-05, "loss": 3.8773, "step": 112 }, { "epoch": 0.07382002286460886, "grad_norm": 2.776826858520508, "learning_rate": 9.102007217627568e-05, "loss": 4.1886, "step": 113 }, { "epoch": 0.0744732974032337, "grad_norm": 2.7510571479797363, "learning_rate": 9.083207775808396e-05, "loss": 3.8013, "step": 114 }, { "epoch": 0.07512657194185857, "grad_norm": 2.9689526557922363, "learning_rate": 9.064233422958077e-05, "loss": 3.8871, "step": 115 }, { "epoch": 0.07577984648048343, "grad_norm": 2.9751687049865723, "learning_rate": 9.045084971874738e-05, "loss": 4.0687, "step": 116 }, { "epoch": 0.07643312101910828, "grad_norm": 3.208277702331543, "learning_rate": 9.025763242814291e-05, "loss": 4.0952, "step": 117 }, { "epoch": 0.07708639555773314, "grad_norm": 2.978297233581543, "learning_rate": 9.006269063455304e-05, "loss": 3.7542, "step": 118 }, { "epoch": 0.07773967009635799, "grad_norm": 3.079564332962036, "learning_rate": 8.986603268863536e-05, "loss": 3.7339, "step": 119 }, { "epoch": 0.07839294463498285, "grad_norm": 3.3465781211853027, "learning_rate": 8.966766701456177e-05, "loss": 4.1308, "step": 120 }, { "epoch": 0.07904621917360771, "grad_norm": 3.38362717628479, "learning_rate": 8.94676021096575e-05, "loss": 3.6047, "step": 121 }, { "epoch": 0.07969949371223256, "grad_norm": 3.4256107807159424, "learning_rate": 8.926584654403724e-05, "loss": 4.0278, "step": 122 }, { "epoch": 0.08035276825085742, "grad_norm": 3.728395700454712, "learning_rate": 8.906240896023794e-05, "loss": 3.5618, "step": 123 }, { "epoch": 0.08100604278948229, "grad_norm": 4.323002338409424, "learning_rate": 8.885729807284856e-05, "loss": 3.7879, "step": 124 }, { "epoch": 0.08165931732810713, "grad_norm": 4.338598728179932, "learning_rate": 8.865052266813685e-05, "loss": 3.4469, "step": 125 }, { "epoch": 0.082312591866732, "grad_norm": 4.841395854949951, "learning_rate": 8.844209160367299e-05, "loss": 3.6624, "step": 126 }, { "epoch": 0.08296586640535684, "grad_norm": 5.293141841888428, "learning_rate": 8.823201380795001e-05, "loss": 3.6083, "step": 127 }, { "epoch": 0.0836191409439817, "grad_norm": 6.072646141052246, "learning_rate": 8.802029828000156e-05, "loss": 3.7721, "step": 128 }, { "epoch": 0.08427241548260657, "grad_norm": 5.880528450012207, "learning_rate": 8.780695408901613e-05, "loss": 3.6489, "step": 129 }, { "epoch": 0.08492569002123142, "grad_norm": 6.2913055419921875, "learning_rate": 8.759199037394887e-05, "loss": 4.0706, "step": 130 }, { "epoch": 0.08557896455985628, "grad_norm": 6.430176258087158, "learning_rate": 8.737541634312985e-05, "loss": 3.6765, "step": 131 }, { "epoch": 0.08623223909848114, "grad_norm": 6.8034987449646, "learning_rate": 8.715724127386972e-05, "loss": 3.7984, "step": 132 }, { "epoch": 0.08688551363710599, "grad_norm": 8.133237838745117, "learning_rate": 8.693747451206232e-05, "loss": 3.7513, "step": 133 }, { "epoch": 0.08753878817573085, "grad_norm": 9.387150764465332, "learning_rate": 8.671612547178428e-05, "loss": 3.943, "step": 134 }, { "epoch": 0.0881920627143557, "grad_norm": 9.971433639526367, "learning_rate": 8.649320363489179e-05, "loss": 3.5509, "step": 135 }, { "epoch": 0.08884533725298056, "grad_norm": 8.350509643554688, "learning_rate": 8.626871855061438e-05, "loss": 2.8905, "step": 136 }, { "epoch": 0.08949861179160543, "grad_norm": 10.003087043762207, "learning_rate": 8.604267983514594e-05, "loss": 3.369, "step": 137 }, { "epoch": 0.09015188633023027, "grad_norm": 11.017991065979004, "learning_rate": 8.581509717123273e-05, "loss": 3.826, "step": 138 }, { "epoch": 0.09080516086885514, "grad_norm": 11.706796646118164, "learning_rate": 8.558598030775857e-05, "loss": 2.9855, "step": 139 }, { "epoch": 0.09145843540748, "grad_norm": 12.903501510620117, "learning_rate": 8.535533905932738e-05, "loss": 2.6285, "step": 140 }, { "epoch": 0.09211170994610485, "grad_norm": 12.504664421081543, "learning_rate": 8.51231833058426e-05, "loss": 3.0194, "step": 141 }, { "epoch": 0.09276498448472971, "grad_norm": 14.484477043151855, "learning_rate": 8.488952299208401e-05, "loss": 3.1958, "step": 142 }, { "epoch": 0.09341825902335456, "grad_norm": 17.707530975341797, "learning_rate": 8.46543681272818e-05, "loss": 3.957, "step": 143 }, { "epoch": 0.09407153356197942, "grad_norm": 20.62574005126953, "learning_rate": 8.44177287846877e-05, "loss": 3.2235, "step": 144 }, { "epoch": 0.09472480810060428, "grad_norm": 18.29495620727539, "learning_rate": 8.417961510114356e-05, "loss": 2.3518, "step": 145 }, { "epoch": 0.09537808263922913, "grad_norm": 28.382152557373047, "learning_rate": 8.39400372766471e-05, "loss": 4.1565, "step": 146 }, { "epoch": 0.096031357177854, "grad_norm": 18.329111099243164, "learning_rate": 8.36990055739149e-05, "loss": 2.1261, "step": 147 }, { "epoch": 0.09668463171647886, "grad_norm": 18.869550704956055, "learning_rate": 8.345653031794292e-05, "loss": 2.6734, "step": 148 }, { "epoch": 0.0973379062551037, "grad_norm": 23.500282287597656, "learning_rate": 8.321262189556409e-05, "loss": 3.4861, "step": 149 }, { "epoch": 0.09799118079372857, "grad_norm": 38.71291732788086, "learning_rate": 8.296729075500344e-05, "loss": 5.1638, "step": 150 }, { "epoch": 0.09864445533235343, "grad_norm": 3.8299853801727295, "learning_rate": 8.272054740543052e-05, "loss": 4.0414, "step": 151 }, { "epoch": 0.09929772987097828, "grad_norm": 3.310225248336792, "learning_rate": 8.247240241650918e-05, "loss": 4.6365, "step": 152 }, { "epoch": 0.09995100440960314, "grad_norm": 2.5045037269592285, "learning_rate": 8.222286641794488e-05, "loss": 4.4531, "step": 153 }, { "epoch": 0.10060427894822799, "grad_norm": 2.3877153396606445, "learning_rate": 8.197195009902924e-05, "loss": 4.7702, "step": 154 }, { "epoch": 0.10125755348685285, "grad_norm": 2.4667954444885254, "learning_rate": 8.171966420818228e-05, "loss": 4.1018, "step": 155 }, { "epoch": 0.10191082802547771, "grad_norm": 2.424144983291626, "learning_rate": 8.146601955249188e-05, "loss": 4.1768, "step": 156 }, { "epoch": 0.10256410256410256, "grad_norm": 2.476624011993408, "learning_rate": 8.121102699725089e-05, "loss": 3.963, "step": 157 }, { "epoch": 0.10321737710272742, "grad_norm": 2.590161085128784, "learning_rate": 8.095469746549172e-05, "loss": 3.9657, "step": 158 }, { "epoch": 0.10387065164135229, "grad_norm": 2.8294732570648193, "learning_rate": 8.069704193751832e-05, "loss": 4.19, "step": 159 }, { "epoch": 0.10452392617997713, "grad_norm": 2.526905059814453, "learning_rate": 8.043807145043604e-05, "loss": 3.7846, "step": 160 }, { "epoch": 0.105177200718602, "grad_norm": 2.666771411895752, "learning_rate": 8.017779709767858e-05, "loss": 4.2755, "step": 161 }, { "epoch": 0.10583047525722684, "grad_norm": 2.6683263778686523, "learning_rate": 7.991623002853296e-05, "loss": 3.8973, "step": 162 }, { "epoch": 0.1064837497958517, "grad_norm": 2.753737688064575, "learning_rate": 7.965338144766186e-05, "loss": 4.0508, "step": 163 }, { "epoch": 0.10713702433447657, "grad_norm": 2.6762168407440186, "learning_rate": 7.938926261462366e-05, "loss": 3.8199, "step": 164 }, { "epoch": 0.10779029887310142, "grad_norm": 2.7907073497772217, "learning_rate": 7.912388484339012e-05, "loss": 4.0572, "step": 165 }, { "epoch": 0.10844357341172628, "grad_norm": 2.9300825595855713, "learning_rate": 7.88572595018617e-05, "loss": 3.85, "step": 166 }, { "epoch": 0.10909684795035114, "grad_norm": 2.946000337600708, "learning_rate": 7.858939801138061e-05, "loss": 4.0084, "step": 167 }, { "epoch": 0.10975012248897599, "grad_norm": 3.150700569152832, "learning_rate": 7.832031184624164e-05, "loss": 4.0213, "step": 168 }, { "epoch": 0.11040339702760085, "grad_norm": 3.130871057510376, "learning_rate": 7.80500125332005e-05, "loss": 3.7754, "step": 169 }, { "epoch": 0.1110566715662257, "grad_norm": 3.2514853477478027, "learning_rate": 7.777851165098012e-05, "loss": 3.8823, "step": 170 }, { "epoch": 0.11170994610485056, "grad_norm": 3.6097805500030518, "learning_rate": 7.750582082977467e-05, "loss": 3.8464, "step": 171 }, { "epoch": 0.11236322064347543, "grad_norm": 3.4300525188446045, "learning_rate": 7.723195175075136e-05, "loss": 3.587, "step": 172 }, { "epoch": 0.11301649518210027, "grad_norm": 3.962552070617676, "learning_rate": 7.695691614555003e-05, "loss": 3.9351, "step": 173 }, { "epoch": 0.11366976972072514, "grad_norm": 3.7221381664276123, "learning_rate": 7.668072579578058e-05, "loss": 3.7333, "step": 174 }, { "epoch": 0.11432304425935, "grad_norm": 3.8638551235198975, "learning_rate": 7.64033925325184e-05, "loss": 3.2305, "step": 175 }, { "epoch": 0.11497631879797485, "grad_norm": 4.228523254394531, "learning_rate": 7.612492823579745e-05, "loss": 3.7991, "step": 176 }, { "epoch": 0.11562959333659971, "grad_norm": 4.375904083251953, "learning_rate": 7.584534483410137e-05, "loss": 3.3348, "step": 177 }, { "epoch": 0.11628286787522456, "grad_norm": 5.0705156326293945, "learning_rate": 7.55646543038526e-05, "loss": 4.2036, "step": 178 }, { "epoch": 0.11693614241384942, "grad_norm": 5.519981384277344, "learning_rate": 7.528286866889924e-05, "loss": 3.987, "step": 179 }, { "epoch": 0.11758941695247428, "grad_norm": 5.943816184997559, "learning_rate": 7.500000000000001e-05, "loss": 3.4623, "step": 180 }, { "epoch": 0.11824269149109913, "grad_norm": 6.2165751457214355, "learning_rate": 7.471606041430723e-05, "loss": 3.4462, "step": 181 }, { "epoch": 0.11889596602972399, "grad_norm": 6.183679580688477, "learning_rate": 7.443106207484776e-05, "loss": 3.6863, "step": 182 }, { "epoch": 0.11954924056834886, "grad_norm": 7.212845802307129, "learning_rate": 7.414501719000187e-05, "loss": 3.4877, "step": 183 }, { "epoch": 0.1202025151069737, "grad_norm": 7.299278736114502, "learning_rate": 7.385793801298042e-05, "loss": 3.1686, "step": 184 }, { "epoch": 0.12085578964559857, "grad_norm": 7.535823345184326, "learning_rate": 7.35698368412999e-05, "loss": 3.7468, "step": 185 }, { "epoch": 0.12150906418422341, "grad_norm": 9.03598690032959, "learning_rate": 7.328072601625557e-05, "loss": 3.3324, "step": 186 }, { "epoch": 0.12216233872284828, "grad_norm": 8.97368335723877, "learning_rate": 7.2990617922393e-05, "loss": 3.0458, "step": 187 }, { "epoch": 0.12281561326147314, "grad_norm": 11.850152015686035, "learning_rate": 7.269952498697734e-05, "loss": 3.6011, "step": 188 }, { "epoch": 0.12346888780009799, "grad_norm": 12.661114692687988, "learning_rate": 7.240745967946113e-05, "loss": 3.4378, "step": 189 }, { "epoch": 0.12412216233872285, "grad_norm": 15.797039985656738, "learning_rate": 7.211443451095007e-05, "loss": 3.9093, "step": 190 }, { "epoch": 0.12477543687734771, "grad_norm": 13.158418655395508, "learning_rate": 7.18204620336671e-05, "loss": 3.2219, "step": 191 }, { "epoch": 0.12542871141597256, "grad_norm": 14.715095520019531, "learning_rate": 7.152555484041476e-05, "loss": 2.8733, "step": 192 }, { "epoch": 0.1260819859545974, "grad_norm": 15.411552429199219, "learning_rate": 7.122972556403567e-05, "loss": 3.4753, "step": 193 }, { "epoch": 0.12673526049322228, "grad_norm": 15.468315124511719, "learning_rate": 7.09329868768714e-05, "loss": 3.6554, "step": 194 }, { "epoch": 0.12738853503184713, "grad_norm": 17.27188491821289, "learning_rate": 7.063535149021973e-05, "loss": 3.4804, "step": 195 }, { "epoch": 0.12804180957047198, "grad_norm": 21.4166316986084, "learning_rate": 7.033683215379002e-05, "loss": 2.967, "step": 196 }, { "epoch": 0.12869508410909686, "grad_norm": 18.70254135131836, "learning_rate": 7.003744165515705e-05, "loss": 2.5688, "step": 197 }, { "epoch": 0.1293483586477217, "grad_norm": 22.17182159423828, "learning_rate": 6.973719281921335e-05, "loss": 3.8092, "step": 198 }, { "epoch": 0.13000163318634655, "grad_norm": 21.631132125854492, "learning_rate": 6.943609850761979e-05, "loss": 3.4749, "step": 199 }, { "epoch": 0.13065490772497143, "grad_norm": 27.764705657958984, "learning_rate": 6.91341716182545e-05, "loss": 3.2158, "step": 200 }, { "epoch": 0.13065490772497143, "eval_loss": 1.0087602138519287, "eval_runtime": 193.6858, "eval_samples_per_second": 13.31, "eval_steps_per_second": 3.33, "step": 200 }, { "epoch": 0.13130818226359628, "grad_norm": 2.8136210441589355, "learning_rate": 6.883142508466054e-05, "loss": 3.6461, "step": 201 }, { "epoch": 0.13196145680222113, "grad_norm": 3.2069029808044434, "learning_rate": 6.852787187549182e-05, "loss": 4.9158, "step": 202 }, { "epoch": 0.132614731340846, "grad_norm": 2.4724817276000977, "learning_rate": 6.82235249939575e-05, "loss": 3.7633, "step": 203 }, { "epoch": 0.13326800587947085, "grad_norm": 2.367910146713257, "learning_rate": 6.7918397477265e-05, "loss": 4.6979, "step": 204 }, { "epoch": 0.1339212804180957, "grad_norm": 2.252790689468384, "learning_rate": 6.761250239606169e-05, "loss": 4.5078, "step": 205 }, { "epoch": 0.13457455495672055, "grad_norm": 2.2511966228485107, "learning_rate": 6.730585285387465e-05, "loss": 4.0116, "step": 206 }, { "epoch": 0.13522782949534543, "grad_norm": 2.3549845218658447, "learning_rate": 6.699846198654971e-05, "loss": 4.0373, "step": 207 }, { "epoch": 0.13588110403397027, "grad_norm": 2.3142459392547607, "learning_rate": 6.669034296168855e-05, "loss": 3.7566, "step": 208 }, { "epoch": 0.13653437857259512, "grad_norm": 2.4672322273254395, "learning_rate": 6.638150897808468e-05, "loss": 3.8954, "step": 209 }, { "epoch": 0.13718765311122, "grad_norm": 2.590864896774292, "learning_rate": 6.607197326515808e-05, "loss": 4.3348, "step": 210 }, { "epoch": 0.13784092764984485, "grad_norm": 2.5048024654388428, "learning_rate": 6.57617490823885e-05, "loss": 4.0438, "step": 211 }, { "epoch": 0.1384942021884697, "grad_norm": 2.490389108657837, "learning_rate": 6.545084971874738e-05, "loss": 3.7406, "step": 212 }, { "epoch": 0.13914747672709457, "grad_norm": 2.4702444076538086, "learning_rate": 6.513928849212873e-05, "loss": 3.8706, "step": 213 }, { "epoch": 0.13980075126571942, "grad_norm": 2.634340763092041, "learning_rate": 6.482707874877854e-05, "loss": 4.0907, "step": 214 }, { "epoch": 0.14045402580434427, "grad_norm": 2.5504937171936035, "learning_rate": 6.451423386272312e-05, "loss": 3.6381, "step": 215 }, { "epoch": 0.14110730034296914, "grad_norm": 2.8212695121765137, "learning_rate": 6.420076723519614e-05, "loss": 4.3291, "step": 216 }, { "epoch": 0.141760574881594, "grad_norm": 3.0288095474243164, "learning_rate": 6.388669229406462e-05, "loss": 4.1332, "step": 217 }, { "epoch": 0.14241384942021884, "grad_norm": 2.837364435195923, "learning_rate": 6.357202249325371e-05, "loss": 3.5461, "step": 218 }, { "epoch": 0.14306712395884372, "grad_norm": 3.0349831581115723, "learning_rate": 6.32567713121704e-05, "loss": 4.0614, "step": 219 }, { "epoch": 0.14372039849746857, "grad_norm": 3.3535542488098145, "learning_rate": 6.294095225512603e-05, "loss": 3.998, "step": 220 }, { "epoch": 0.14437367303609341, "grad_norm": 3.5414798259735107, "learning_rate": 6.26245788507579e-05, "loss": 3.9925, "step": 221 }, { "epoch": 0.14502694757471826, "grad_norm": 3.5899298191070557, "learning_rate": 6.230766465144967e-05, "loss": 4.4236, "step": 222 }, { "epoch": 0.14568022211334314, "grad_norm": 4.002371788024902, "learning_rate": 6.199022323275083e-05, "loss": 4.1189, "step": 223 }, { "epoch": 0.146333496651968, "grad_norm": 4.0465006828308105, "learning_rate": 6.167226819279528e-05, "loss": 3.6562, "step": 224 }, { "epoch": 0.14698677119059284, "grad_norm": 4.780096530914307, "learning_rate": 6.135381315171867e-05, "loss": 4.0265, "step": 225 }, { "epoch": 0.1476400457292177, "grad_norm": 4.929449081420898, "learning_rate": 6.103487175107507e-05, "loss": 3.3766, "step": 226 }, { "epoch": 0.14829332026784256, "grad_norm": 4.737488746643066, "learning_rate": 6.071545765325254e-05, "loss": 3.2223, "step": 227 }, { "epoch": 0.1489465948064674, "grad_norm": 5.096072673797607, "learning_rate": 6.0395584540887963e-05, "loss": 4.0753, "step": 228 }, { "epoch": 0.14959986934509228, "grad_norm": 5.692816734313965, "learning_rate": 6.007526611628086e-05, "loss": 3.2582, "step": 229 }, { "epoch": 0.15025314388371713, "grad_norm": 6.038702011108398, "learning_rate": 5.9754516100806423e-05, "loss": 4.0465, "step": 230 }, { "epoch": 0.15090641842234198, "grad_norm": 5.955114841461182, "learning_rate": 5.9433348234327765e-05, "loss": 3.5113, "step": 231 }, { "epoch": 0.15155969296096686, "grad_norm": 7.852488994598389, "learning_rate": 5.911177627460739e-05, "loss": 3.946, "step": 232 }, { "epoch": 0.1522129674995917, "grad_norm": 7.21500825881958, "learning_rate": 5.8789813996717736e-05, "loss": 3.2654, "step": 233 }, { "epoch": 0.15286624203821655, "grad_norm": 7.874154090881348, "learning_rate": 5.8467475192451226e-05, "loss": 3.5822, "step": 234 }, { "epoch": 0.15351951657684143, "grad_norm": 8.220457077026367, "learning_rate": 5.814477366972945e-05, "loss": 3.5959, "step": 235 }, { "epoch": 0.15417279111546628, "grad_norm": 9.068872451782227, "learning_rate": 5.782172325201155e-05, "loss": 3.5749, "step": 236 }, { "epoch": 0.15482606565409113, "grad_norm": 10.15458869934082, "learning_rate": 5.749833777770225e-05, "loss": 3.7681, "step": 237 }, { "epoch": 0.15547934019271598, "grad_norm": 10.914835929870605, "learning_rate": 5.717463109955896e-05, "loss": 3.6604, "step": 238 }, { "epoch": 0.15613261473134085, "grad_norm": 11.057694435119629, "learning_rate": 5.685061708409841e-05, "loss": 3.1422, "step": 239 }, { "epoch": 0.1567858892699657, "grad_norm": 12.083439826965332, "learning_rate": 5.6526309611002594e-05, "loss": 3.7388, "step": 240 }, { "epoch": 0.15743916380859055, "grad_norm": 11.058955192565918, "learning_rate": 5.6201722572524275e-05, "loss": 3.0529, "step": 241 }, { "epoch": 0.15809243834721542, "grad_norm": 12.308931350708008, "learning_rate": 5.587686987289189e-05, "loss": 2.9113, "step": 242 }, { "epoch": 0.15874571288584027, "grad_norm": 12.517452239990234, "learning_rate": 5.5551765427713884e-05, "loss": 2.6273, "step": 243 }, { "epoch": 0.15939898742446512, "grad_norm": 13.3101167678833, "learning_rate": 5.522642316338268e-05, "loss": 2.5797, "step": 244 }, { "epoch": 0.16005226196309, "grad_norm": 14.140732765197754, "learning_rate": 5.490085701647805e-05, "loss": 2.689, "step": 245 }, { "epoch": 0.16070553650171485, "grad_norm": 17.422927856445312, "learning_rate": 5.457508093317013e-05, "loss": 2.6193, "step": 246 }, { "epoch": 0.1613588110403397, "grad_norm": 19.202436447143555, "learning_rate": 5.4249108868622086e-05, "loss": 3.1502, "step": 247 }, { "epoch": 0.16201208557896457, "grad_norm": 20.904285430908203, "learning_rate": 5.392295478639225e-05, "loss": 3.32, "step": 248 }, { "epoch": 0.16266536011758942, "grad_norm": 22.786781311035156, "learning_rate": 5.359663265783598e-05, "loss": 3.4771, "step": 249 }, { "epoch": 0.16331863465621427, "grad_norm": 31.27584457397461, "learning_rate": 5.327015646150716e-05, "loss": 4.7491, "step": 250 }, { "epoch": 0.16397190919483914, "grad_norm": 2.7455053329467773, "learning_rate": 5.294354018255945e-05, "loss": 4.2161, "step": 251 }, { "epoch": 0.164625183733464, "grad_norm": 2.7494430541992188, "learning_rate": 5.26167978121472e-05, "loss": 4.3482, "step": 252 }, { "epoch": 0.16527845827208884, "grad_norm": 2.4116597175598145, "learning_rate": 5.228994334682604e-05, "loss": 4.2538, "step": 253 }, { "epoch": 0.1659317328107137, "grad_norm": 2.3236160278320312, "learning_rate": 5.196299078795344e-05, "loss": 3.9966, "step": 254 }, { "epoch": 0.16658500734933857, "grad_norm": 2.252965211868286, "learning_rate": 5.1635954141088813e-05, "loss": 3.9031, "step": 255 }, { "epoch": 0.1672382818879634, "grad_norm": 2.2256863117218018, "learning_rate": 5.1308847415393666e-05, "loss": 3.8695, "step": 256 }, { "epoch": 0.16789155642658826, "grad_norm": 2.294617176055908, "learning_rate": 5.0981684623031415e-05, "loss": 4.0556, "step": 257 }, { "epoch": 0.16854483096521314, "grad_norm": 2.3014414310455322, "learning_rate": 5.0654479778567223e-05, "loss": 3.731, "step": 258 }, { "epoch": 0.169198105503838, "grad_norm": 2.2849111557006836, "learning_rate": 5.0327246898367597e-05, "loss": 3.9152, "step": 259 }, { "epoch": 0.16985138004246284, "grad_norm": 2.435673475265503, "learning_rate": 5e-05, "loss": 4.1635, "step": 260 }, { "epoch": 0.1705046545810877, "grad_norm": 2.5315775871276855, "learning_rate": 4.9672753101632415e-05, "loss": 4.0719, "step": 261 }, { "epoch": 0.17115792911971256, "grad_norm": 2.5360610485076904, "learning_rate": 4.934552022143279e-05, "loss": 4.1543, "step": 262 }, { "epoch": 0.1718112036583374, "grad_norm": 2.621554374694824, "learning_rate": 4.901831537696859e-05, "loss": 4.1552, "step": 263 }, { "epoch": 0.17246447819696228, "grad_norm": 2.6578054428100586, "learning_rate": 4.869115258460635e-05, "loss": 3.8845, "step": 264 }, { "epoch": 0.17311775273558713, "grad_norm": 2.6418116092681885, "learning_rate": 4.83640458589112e-05, "loss": 3.9069, "step": 265 }, { "epoch": 0.17377102727421198, "grad_norm": 2.6990976333618164, "learning_rate": 4.8037009212046586e-05, "loss": 4.0657, "step": 266 }, { "epoch": 0.17442430181283686, "grad_norm": 2.7651925086975098, "learning_rate": 4.7710056653173976e-05, "loss": 4.0418, "step": 267 }, { "epoch": 0.1750775763514617, "grad_norm": 2.7356722354888916, "learning_rate": 4.738320218785281e-05, "loss": 4.0282, "step": 268 }, { "epoch": 0.17573085089008655, "grad_norm": 2.977963924407959, "learning_rate": 4.7056459817440544e-05, "loss": 3.8089, "step": 269 }, { "epoch": 0.1763841254287114, "grad_norm": 3.2176568508148193, "learning_rate": 4.6729843538492847e-05, "loss": 4.392, "step": 270 }, { "epoch": 0.17703739996733628, "grad_norm": 3.1798744201660156, "learning_rate": 4.640336734216403e-05, "loss": 4.0641, "step": 271 }, { "epoch": 0.17769067450596113, "grad_norm": 3.06921648979187, "learning_rate": 4.607704521360776e-05, "loss": 3.7511, "step": 272 }, { "epoch": 0.17834394904458598, "grad_norm": 3.3475852012634277, "learning_rate": 4.575089113137792e-05, "loss": 3.8854, "step": 273 }, { "epoch": 0.17899722358321085, "grad_norm": 3.8393049240112305, "learning_rate": 4.542491906682989e-05, "loss": 4.1801, "step": 274 }, { "epoch": 0.1796504981218357, "grad_norm": 3.7368972301483154, "learning_rate": 4.509914298352197e-05, "loss": 3.4228, "step": 275 }, { "epoch": 0.18030377266046055, "grad_norm": 3.9486243724823, "learning_rate": 4.477357683661734e-05, "loss": 3.4336, "step": 276 }, { "epoch": 0.18095704719908542, "grad_norm": 4.318992614746094, "learning_rate": 4.444823457228612e-05, "loss": 3.1958, "step": 277 }, { "epoch": 0.18161032173771027, "grad_norm": 4.600458145141602, "learning_rate": 4.412313012710813e-05, "loss": 3.3415, "step": 278 }, { "epoch": 0.18226359627633512, "grad_norm": 5.333650588989258, "learning_rate": 4.379827742747575e-05, "loss": 3.744, "step": 279 }, { "epoch": 0.18291687081496, "grad_norm": 5.62832498550415, "learning_rate": 4.347369038899744e-05, "loss": 3.9508, "step": 280 }, { "epoch": 0.18357014535358485, "grad_norm": 5.859029293060303, "learning_rate": 4.3149382915901606e-05, "loss": 3.8297, "step": 281 }, { "epoch": 0.1842234198922097, "grad_norm": 6.399336338043213, "learning_rate": 4.282536890044104e-05, "loss": 3.8776, "step": 282 }, { "epoch": 0.18487669443083457, "grad_norm": 6.635499000549316, "learning_rate": 4.250166222229774e-05, "loss": 4.042, "step": 283 }, { "epoch": 0.18552996896945942, "grad_norm": 6.871761322021484, "learning_rate": 4.2178276747988446e-05, "loss": 3.5912, "step": 284 }, { "epoch": 0.18618324350808427, "grad_norm": 7.2283935546875, "learning_rate": 4.185522633027057e-05, "loss": 3.3263, "step": 285 }, { "epoch": 0.18683651804670912, "grad_norm": 7.931128978729248, "learning_rate": 4.153252480754877e-05, "loss": 3.6491, "step": 286 }, { "epoch": 0.187489792585334, "grad_norm": 8.871136665344238, "learning_rate": 4.1210186003282275e-05, "loss": 3.3801, "step": 287 }, { "epoch": 0.18814306712395884, "grad_norm": 9.464795112609863, "learning_rate": 4.088822372539263e-05, "loss": 3.3301, "step": 288 }, { "epoch": 0.1887963416625837, "grad_norm": 11.650923728942871, "learning_rate": 4.0566651765672246e-05, "loss": 3.6579, "step": 289 }, { "epoch": 0.18944961620120856, "grad_norm": 12.516644477844238, "learning_rate": 4.0245483899193595e-05, "loss": 3.1808, "step": 290 }, { "epoch": 0.1901028907398334, "grad_norm": 12.424504280090332, "learning_rate": 3.992473388371915e-05, "loss": 3.5156, "step": 291 }, { "epoch": 0.19075616527845826, "grad_norm": 17.3675479888916, "learning_rate": 3.960441545911204e-05, "loss": 4.341, "step": 292 }, { "epoch": 0.19140943981708314, "grad_norm": 13.690961837768555, "learning_rate": 3.928454234674747e-05, "loss": 2.6183, "step": 293 }, { "epoch": 0.192062714355708, "grad_norm": 16.243967056274414, "learning_rate": 3.896512824892495e-05, "loss": 3.6009, "step": 294 }, { "epoch": 0.19271598889433283, "grad_norm": 14.707989692687988, "learning_rate": 3.864618684828134e-05, "loss": 2.5393, "step": 295 }, { "epoch": 0.1933692634329577, "grad_norm": 14.387151718139648, "learning_rate": 3.832773180720475e-05, "loss": 2.3307, "step": 296 }, { "epoch": 0.19402253797158256, "grad_norm": 20.52593994140625, "learning_rate": 3.800977676724919e-05, "loss": 2.721, "step": 297 }, { "epoch": 0.1946758125102074, "grad_norm": 17.45858383178711, "learning_rate": 3.769233534855035e-05, "loss": 3.0486, "step": 298 }, { "epoch": 0.19532908704883228, "grad_norm": 19.98239517211914, "learning_rate": 3.73754211492421e-05, "loss": 2.4033, "step": 299 }, { "epoch": 0.19598236158745713, "grad_norm": 26.76961898803711, "learning_rate": 3.705904774487396e-05, "loss": 3.0636, "step": 300 }, { "epoch": 0.19598236158745713, "eval_loss": 0.916141927242279, "eval_runtime": 193.7367, "eval_samples_per_second": 13.307, "eval_steps_per_second": 3.329, "step": 300 }, { "epoch": 0.19663563612608198, "grad_norm": 2.2509307861328125, "learning_rate": 3.6743228687829595e-05, "loss": 4.1521, "step": 301 }, { "epoch": 0.19728891066470686, "grad_norm": 2.2536087036132812, "learning_rate": 3.642797750674629e-05, "loss": 4.0593, "step": 302 }, { "epoch": 0.1979421852033317, "grad_norm": 2.2652878761291504, "learning_rate": 3.6113307705935396e-05, "loss": 4.1726, "step": 303 }, { "epoch": 0.19859545974195655, "grad_norm": 2.3236594200134277, "learning_rate": 3.579923276480387e-05, "loss": 4.1662, "step": 304 }, { "epoch": 0.1992487342805814, "grad_norm": 2.248436689376831, "learning_rate": 3.5485766137276894e-05, "loss": 3.9428, "step": 305 }, { "epoch": 0.19990200881920628, "grad_norm": 2.3227086067199707, "learning_rate": 3.5172921251221455e-05, "loss": 4.159, "step": 306 }, { "epoch": 0.20055528335783113, "grad_norm": 2.3091423511505127, "learning_rate": 3.486071150787128e-05, "loss": 3.8457, "step": 307 }, { "epoch": 0.20120855789645598, "grad_norm": 2.2768592834472656, "learning_rate": 3.4549150281252636e-05, "loss": 3.6557, "step": 308 }, { "epoch": 0.20186183243508085, "grad_norm": 2.4113123416900635, "learning_rate": 3.423825091761153e-05, "loss": 3.9928, "step": 309 }, { "epoch": 0.2025151069737057, "grad_norm": 2.337956666946411, "learning_rate": 3.392802673484193e-05, "loss": 3.8495, "step": 310 }, { "epoch": 0.20316838151233055, "grad_norm": 2.3353707790374756, "learning_rate": 3.361849102191533e-05, "loss": 3.7771, "step": 311 }, { "epoch": 0.20382165605095542, "grad_norm": 2.4087674617767334, "learning_rate": 3.330965703831146e-05, "loss": 4.1044, "step": 312 }, { "epoch": 0.20447493058958027, "grad_norm": 2.4701943397521973, "learning_rate": 3.300153801345028e-05, "loss": 3.5573, "step": 313 }, { "epoch": 0.20512820512820512, "grad_norm": 2.612384557723999, "learning_rate": 3.2694147146125345e-05, "loss": 3.8407, "step": 314 }, { "epoch": 0.20578147966683, "grad_norm": 2.717681884765625, "learning_rate": 3.2387497603938326e-05, "loss": 4.164, "step": 315 }, { "epoch": 0.20643475420545485, "grad_norm": 2.7894811630249023, "learning_rate": 3.2081602522734986e-05, "loss": 3.6986, "step": 316 }, { "epoch": 0.2070880287440797, "grad_norm": 2.887406826019287, "learning_rate": 3.177647500604252e-05, "loss": 4.1417, "step": 317 }, { "epoch": 0.20774130328270457, "grad_norm": 2.995980739593506, "learning_rate": 3.147212812450819e-05, "loss": 4.0842, "step": 318 }, { "epoch": 0.20839457782132942, "grad_norm": 3.2103445529937744, "learning_rate": 3.116857491533947e-05, "loss": 4.2465, "step": 319 }, { "epoch": 0.20904785235995427, "grad_norm": 3.404813766479492, "learning_rate": 3.086582838174551e-05, "loss": 3.4803, "step": 320 }, { "epoch": 0.20970112689857912, "grad_norm": 3.4752864837646484, "learning_rate": 3.056390149238022e-05, "loss": 4.2533, "step": 321 }, { "epoch": 0.210354401437204, "grad_norm": 3.54484486579895, "learning_rate": 3.0262807180786647e-05, "loss": 3.9656, "step": 322 }, { "epoch": 0.21100767597582884, "grad_norm": 3.658374547958374, "learning_rate": 2.996255834484296e-05, "loss": 3.5864, "step": 323 }, { "epoch": 0.2116609505144537, "grad_norm": 3.7134335041046143, "learning_rate": 2.9663167846209998e-05, "loss": 3.587, "step": 324 }, { "epoch": 0.21231422505307856, "grad_norm": 3.8654329776763916, "learning_rate": 2.936464850978027e-05, "loss": 3.8962, "step": 325 }, { "epoch": 0.2129674995917034, "grad_norm": 4.069035530090332, "learning_rate": 2.9067013123128613e-05, "loss": 3.6819, "step": 326 }, { "epoch": 0.21362077413032826, "grad_norm": 4.072150230407715, "learning_rate": 2.8770274435964355e-05, "loss": 3.3375, "step": 327 }, { "epoch": 0.21427404866895314, "grad_norm": 4.656280040740967, "learning_rate": 2.8474445159585235e-05, "loss": 3.5886, "step": 328 }, { "epoch": 0.21492732320757799, "grad_norm": 4.830178737640381, "learning_rate": 2.8179537966332887e-05, "loss": 3.2974, "step": 329 }, { "epoch": 0.21558059774620283, "grad_norm": 5.456755638122559, "learning_rate": 2.7885565489049946e-05, "loss": 3.8774, "step": 330 }, { "epoch": 0.2162338722848277, "grad_norm": 5.759149551391602, "learning_rate": 2.759254032053888e-05, "loss": 3.6094, "step": 331 }, { "epoch": 0.21688714682345256, "grad_norm": 5.614565372467041, "learning_rate": 2.7300475013022663e-05, "loss": 3.608, "step": 332 }, { "epoch": 0.2175404213620774, "grad_norm": 6.076308250427246, "learning_rate": 2.700938207760701e-05, "loss": 3.5142, "step": 333 }, { "epoch": 0.21819369590070228, "grad_norm": 6.575270175933838, "learning_rate": 2.671927398374443e-05, "loss": 3.3849, "step": 334 }, { "epoch": 0.21884697043932713, "grad_norm": 7.774899482727051, "learning_rate": 2.6430163158700115e-05, "loss": 3.4158, "step": 335 }, { "epoch": 0.21950024497795198, "grad_norm": 7.88145637512207, "learning_rate": 2.6142061987019577e-05, "loss": 2.9199, "step": 336 }, { "epoch": 0.22015351951657683, "grad_norm": 10.132671356201172, "learning_rate": 2.5854982809998153e-05, "loss": 4.0171, "step": 337 }, { "epoch": 0.2208067940552017, "grad_norm": 11.227150917053223, "learning_rate": 2.556893792515227e-05, "loss": 2.9164, "step": 338 }, { "epoch": 0.22146006859382655, "grad_norm": 10.218441009521484, "learning_rate": 2.5283939585692783e-05, "loss": 2.9718, "step": 339 }, { "epoch": 0.2221133431324514, "grad_norm": 11.796672821044922, "learning_rate": 2.500000000000001e-05, "loss": 3.0505, "step": 340 }, { "epoch": 0.22276661767107628, "grad_norm": 13.644339561462402, "learning_rate": 2.471713133110078e-05, "loss": 3.9324, "step": 341 }, { "epoch": 0.22341989220970113, "grad_norm": 13.900707244873047, "learning_rate": 2.4435345696147403e-05, "loss": 3.3911, "step": 342 }, { "epoch": 0.22407316674832597, "grad_norm": 18.582195281982422, "learning_rate": 2.4154655165898627e-05, "loss": 2.6938, "step": 343 }, { "epoch": 0.22472644128695085, "grad_norm": 14.891613006591797, "learning_rate": 2.3875071764202563e-05, "loss": 2.8819, "step": 344 }, { "epoch": 0.2253797158255757, "grad_norm": 17.06895637512207, "learning_rate": 2.3596607467481603e-05, "loss": 3.0545, "step": 345 }, { "epoch": 0.22603299036420055, "grad_norm": 18.432598114013672, "learning_rate": 2.3319274204219428e-05, "loss": 3.5887, "step": 346 }, { "epoch": 0.22668626490282542, "grad_norm": 22.221372604370117, "learning_rate": 2.3043083854449988e-05, "loss": 2.5782, "step": 347 }, { "epoch": 0.22733953944145027, "grad_norm": 23.88602638244629, "learning_rate": 2.2768048249248648e-05, "loss": 3.1699, "step": 348 }, { "epoch": 0.22799281398007512, "grad_norm": 17.298885345458984, "learning_rate": 2.2494179170225333e-05, "loss": 2.5779, "step": 349 }, { "epoch": 0.2286460885187, "grad_norm": 34.774845123291016, "learning_rate": 2.2221488349019903e-05, "loss": 4.6921, "step": 350 }, { "epoch": 0.22929936305732485, "grad_norm": 2.0609993934631348, "learning_rate": 2.194998746679952e-05, "loss": 4.05, "step": 351 }, { "epoch": 0.2299526375959497, "grad_norm": 2.18334698677063, "learning_rate": 2.167968815375837e-05, "loss": 4.2886, "step": 352 }, { "epoch": 0.23060591213457454, "grad_norm": 2.1787545680999756, "learning_rate": 2.1410601988619394e-05, "loss": 4.3155, "step": 353 }, { "epoch": 0.23125918667319942, "grad_norm": 2.1137516498565674, "learning_rate": 2.1142740498138324e-05, "loss": 3.7972, "step": 354 }, { "epoch": 0.23191246121182427, "grad_norm": 2.1815781593322754, "learning_rate": 2.08761151566099e-05, "loss": 3.9406, "step": 355 }, { "epoch": 0.23256573575044912, "grad_norm": 2.2473208904266357, "learning_rate": 2.061073738537635e-05, "loss": 3.94, "step": 356 }, { "epoch": 0.233219010289074, "grad_norm": 2.286938428878784, "learning_rate": 2.034661855233815e-05, "loss": 3.8518, "step": 357 }, { "epoch": 0.23387228482769884, "grad_norm": 2.282809257507324, "learning_rate": 2.008376997146705e-05, "loss": 4.0518, "step": 358 }, { "epoch": 0.2345255593663237, "grad_norm": 2.15885066986084, "learning_rate": 1.982220290232143e-05, "loss": 3.7982, "step": 359 }, { "epoch": 0.23517883390494856, "grad_norm": 2.232218027114868, "learning_rate": 1.9561928549563968e-05, "loss": 3.9533, "step": 360 }, { "epoch": 0.2358321084435734, "grad_norm": 2.28951358795166, "learning_rate": 1.9302958062481673e-05, "loss": 3.907, "step": 361 }, { "epoch": 0.23648538298219826, "grad_norm": 2.3329083919525146, "learning_rate": 1.9045302534508297e-05, "loss": 4.0736, "step": 362 }, { "epoch": 0.23713865752082314, "grad_norm": 2.4136269092559814, "learning_rate": 1.8788973002749112e-05, "loss": 3.8261, "step": 363 }, { "epoch": 0.23779193205944799, "grad_norm": 2.6540565490722656, "learning_rate": 1.8533980447508137e-05, "loss": 4.0891, "step": 364 }, { "epoch": 0.23844520659807283, "grad_norm": 2.4953386783599854, "learning_rate": 1.8280335791817733e-05, "loss": 3.9206, "step": 365 }, { "epoch": 0.2390984811366977, "grad_norm": 2.7900853157043457, "learning_rate": 1.8028049900970767e-05, "loss": 3.7279, "step": 366 }, { "epoch": 0.23975175567532256, "grad_norm": 2.7315573692321777, "learning_rate": 1.777713358205514e-05, "loss": 4.0381, "step": 367 }, { "epoch": 0.2404050302139474, "grad_norm": 2.8006503582000732, "learning_rate": 1.7527597583490822e-05, "loss": 3.7893, "step": 368 }, { "epoch": 0.24105830475257226, "grad_norm": 3.0075833797454834, "learning_rate": 1.7279452594569483e-05, "loss": 3.9685, "step": 369 }, { "epoch": 0.24171157929119713, "grad_norm": 3.305966854095459, "learning_rate": 1.703270924499656e-05, "loss": 4.3688, "step": 370 }, { "epoch": 0.24236485382982198, "grad_norm": 3.364837646484375, "learning_rate": 1.678737810443593e-05, "loss": 4.0271, "step": 371 }, { "epoch": 0.24301812836844683, "grad_norm": 3.410656213760376, "learning_rate": 1.6543469682057106e-05, "loss": 3.913, "step": 372 }, { "epoch": 0.2436714029070717, "grad_norm": 3.7187421321868896, "learning_rate": 1.6300994426085103e-05, "loss": 3.6959, "step": 373 }, { "epoch": 0.24432467744569655, "grad_norm": 3.892642021179199, "learning_rate": 1.605996272335291e-05, "loss": 3.8044, "step": 374 }, { "epoch": 0.2449779519843214, "grad_norm": 3.7945556640625, "learning_rate": 1.5820384898856434e-05, "loss": 3.2933, "step": 375 }, { "epoch": 0.24563122652294628, "grad_norm": 4.004775047302246, "learning_rate": 1.5582271215312294e-05, "loss": 3.3998, "step": 376 }, { "epoch": 0.24628450106157113, "grad_norm": 4.55605936050415, "learning_rate": 1.5345631872718214e-05, "loss": 4.0181, "step": 377 }, { "epoch": 0.24693777560019597, "grad_norm": 4.659095287322998, "learning_rate": 1.5110477007916001e-05, "loss": 3.6973, "step": 378 }, { "epoch": 0.24759105013882085, "grad_norm": 4.90739107131958, "learning_rate": 1.4876816694157419e-05, "loss": 3.4013, "step": 379 }, { "epoch": 0.2482443246774457, "grad_norm": 5.735424995422363, "learning_rate": 1.4644660940672627e-05, "loss": 3.6558, "step": 380 }, { "epoch": 0.24889759921607055, "grad_norm": 5.81011962890625, "learning_rate": 1.4414019692241437e-05, "loss": 3.5117, "step": 381 }, { "epoch": 0.24955087375469542, "grad_norm": 5.775092601776123, "learning_rate": 1.4184902828767287e-05, "loss": 3.5385, "step": 382 }, { "epoch": 0.25020414829332027, "grad_norm": 6.017894268035889, "learning_rate": 1.3957320164854059e-05, "loss": 3.2936, "step": 383 }, { "epoch": 0.2508574228319451, "grad_norm": 6.8194708824157715, "learning_rate": 1.373128144938563e-05, "loss": 3.1461, "step": 384 }, { "epoch": 0.25151069737056997, "grad_norm": 7.288778781890869, "learning_rate": 1.3506796365108232e-05, "loss": 3.711, "step": 385 }, { "epoch": 0.2521639719091948, "grad_norm": 7.597186088562012, "learning_rate": 1.3283874528215733e-05, "loss": 3.68, "step": 386 }, { "epoch": 0.2528172464478197, "grad_norm": 9.038799285888672, "learning_rate": 1.3062525487937699e-05, "loss": 3.1288, "step": 387 }, { "epoch": 0.25347052098644457, "grad_norm": 9.388714790344238, "learning_rate": 1.2842758726130283e-05, "loss": 3.068, "step": 388 }, { "epoch": 0.2541237955250694, "grad_norm": 11.19589614868164, "learning_rate": 1.2624583656870154e-05, "loss": 3.2682, "step": 389 }, { "epoch": 0.25477707006369427, "grad_norm": 10.183329582214355, "learning_rate": 1.2408009626051137e-05, "loss": 2.604, "step": 390 }, { "epoch": 0.2554303446023191, "grad_norm": 12.288995742797852, "learning_rate": 1.2193045910983863e-05, "loss": 3.0615, "step": 391 }, { "epoch": 0.25608361914094396, "grad_norm": 10.170268058776855, "learning_rate": 1.1979701719998453e-05, "loss": 3.0368, "step": 392 }, { "epoch": 0.2567368936795688, "grad_norm": 12.902504920959473, "learning_rate": 1.1767986192049984e-05, "loss": 2.6595, "step": 393 }, { "epoch": 0.2573901682181937, "grad_norm": 14.612078666687012, "learning_rate": 1.1557908396327028e-05, "loss": 3.1923, "step": 394 }, { "epoch": 0.25804344275681856, "grad_norm": 14.616671562194824, "learning_rate": 1.134947733186315e-05, "loss": 3.3654, "step": 395 }, { "epoch": 0.2586967172954434, "grad_norm": 15.300880432128906, "learning_rate": 1.1142701927151456e-05, "loss": 2.3551, "step": 396 }, { "epoch": 0.25934999183406826, "grad_norm": 16.621246337890625, "learning_rate": 1.0937591039762085e-05, "loss": 2.6044, "step": 397 }, { "epoch": 0.2600032663726931, "grad_norm": 20.40192985534668, "learning_rate": 1.0734153455962765e-05, "loss": 2.9109, "step": 398 }, { "epoch": 0.26065654091131796, "grad_norm": 21.176897048950195, "learning_rate": 1.0532397890342505e-05, "loss": 2.6236, "step": 399 }, { "epoch": 0.26130981544994286, "grad_norm": 36.7708740234375, "learning_rate": 1.0332332985438248e-05, "loss": 5.0028, "step": 400 }, { "epoch": 0.26130981544994286, "eval_loss": 0.8865355849266052, "eval_runtime": 193.5082, "eval_samples_per_second": 13.322, "eval_steps_per_second": 3.333, "step": 400 }, { "epoch": 0.2619630899885677, "grad_norm": 2.0284175872802734, "learning_rate": 1.013396731136465e-05, "loss": 3.2076, "step": 401 }, { "epoch": 0.26261636452719256, "grad_norm": 2.200660467147827, "learning_rate": 9.937309365446973e-06, "loss": 4.3066, "step": 402 }, { "epoch": 0.2632696390658174, "grad_norm": 2.2217345237731934, "learning_rate": 9.742367571857091e-06, "loss": 4.3663, "step": 403 }, { "epoch": 0.26392291360444226, "grad_norm": 2.1234660148620605, "learning_rate": 9.549150281252633e-06, "loss": 4.0273, "step": 404 }, { "epoch": 0.2645761881430671, "grad_norm": 2.140134572982788, "learning_rate": 9.357665770419244e-06, "loss": 4.0097, "step": 405 }, { "epoch": 0.265229462681692, "grad_norm": 2.241824150085449, "learning_rate": 9.167922241916055e-06, "loss": 3.8709, "step": 406 }, { "epoch": 0.26588273722031686, "grad_norm": 2.207245111465454, "learning_rate": 8.97992782372432e-06, "loss": 3.7465, "step": 407 }, { "epoch": 0.2665360117589417, "grad_norm": 2.2445242404937744, "learning_rate": 8.793690568899216e-06, "loss": 3.9702, "step": 408 }, { "epoch": 0.26718928629756655, "grad_norm": 2.209073781967163, "learning_rate": 8.609218455224893e-06, "loss": 3.6522, "step": 409 }, { "epoch": 0.2678425608361914, "grad_norm": 2.3187787532806396, "learning_rate": 8.426519384872733e-06, "loss": 3.9904, "step": 410 }, { "epoch": 0.26849583537481625, "grad_norm": 2.3060576915740967, "learning_rate": 8.245601184062852e-06, "loss": 3.645, "step": 411 }, { "epoch": 0.2691491099134411, "grad_norm": 2.4985568523406982, "learning_rate": 8.066471602728803e-06, "loss": 3.94, "step": 412 }, { "epoch": 0.269802384452066, "grad_norm": 2.4821956157684326, "learning_rate": 7.889138314185678e-06, "loss": 3.8423, "step": 413 }, { "epoch": 0.27045565899069085, "grad_norm": 2.536126136779785, "learning_rate": 7.71360891480134e-06, "loss": 3.8344, "step": 414 }, { "epoch": 0.2711089335293157, "grad_norm": 2.5374646186828613, "learning_rate": 7.539890923671062e-06, "loss": 3.9225, "step": 415 }, { "epoch": 0.27176220806794055, "grad_norm": 2.6912612915039062, "learning_rate": 7.367991782295391e-06, "loss": 3.8863, "step": 416 }, { "epoch": 0.2724154826065654, "grad_norm": 2.7828660011291504, "learning_rate": 7.197918854261432e-06, "loss": 3.8709, "step": 417 }, { "epoch": 0.27306875714519024, "grad_norm": 2.831202507019043, "learning_rate": 7.029679424927365e-06, "loss": 3.7255, "step": 418 }, { "epoch": 0.27372203168381515, "grad_norm": 3.060506582260132, "learning_rate": 6.863280701110408e-06, "loss": 4.0642, "step": 419 }, { "epoch": 0.27437530622244, "grad_norm": 3.0359199047088623, "learning_rate": 6.698729810778065e-06, "loss": 3.7836, "step": 420 }, { "epoch": 0.27502858076106484, "grad_norm": 3.1587436199188232, "learning_rate": 6.536033802742813e-06, "loss": 4.0368, "step": 421 }, { "epoch": 0.2756818552996897, "grad_norm": 3.3620858192443848, "learning_rate": 6.375199646360142e-06, "loss": 3.7547, "step": 422 }, { "epoch": 0.27633512983831454, "grad_norm": 3.528315782546997, "learning_rate": 6.216234231230012e-06, "loss": 3.7245, "step": 423 }, { "epoch": 0.2769884043769394, "grad_norm": 3.4212453365325928, "learning_rate": 6.059144366901736e-06, "loss": 3.3729, "step": 424 }, { "epoch": 0.27764167891556424, "grad_norm": 3.6946189403533936, "learning_rate": 5.903936782582253e-06, "loss": 3.5213, "step": 425 }, { "epoch": 0.27829495345418914, "grad_norm": 3.8112924098968506, "learning_rate": 5.750618126847912e-06, "loss": 3.5261, "step": 426 }, { "epoch": 0.278948227992814, "grad_norm": 4.669442176818848, "learning_rate": 5.599194967359639e-06, "loss": 3.9578, "step": 427 }, { "epoch": 0.27960150253143884, "grad_norm": 4.836587905883789, "learning_rate": 5.449673790581611e-06, "loss": 4.1958, "step": 428 }, { "epoch": 0.2802547770700637, "grad_norm": 4.8962507247924805, "learning_rate": 5.302061001503394e-06, "loss": 3.6376, "step": 429 }, { "epoch": 0.28090805160868854, "grad_norm": 5.25094747543335, "learning_rate": 5.156362923365588e-06, "loss": 3.6825, "step": 430 }, { "epoch": 0.2815613261473134, "grad_norm": 5.470578193664551, "learning_rate": 5.012585797388936e-06, "loss": 3.3548, "step": 431 }, { "epoch": 0.2822146006859383, "grad_norm": 6.3519415855407715, "learning_rate": 4.87073578250698e-06, "loss": 3.4006, "step": 432 }, { "epoch": 0.28286787522456314, "grad_norm": 6.419662952423096, "learning_rate": 4.730818955102234e-06, "loss": 3.493, "step": 433 }, { "epoch": 0.283521149763188, "grad_norm": 6.477165699005127, "learning_rate": 4.592841308745932e-06, "loss": 3.3161, "step": 434 }, { "epoch": 0.28417442430181283, "grad_norm": 6.963860034942627, "learning_rate": 4.456808753941205e-06, "loss": 3.4243, "step": 435 }, { "epoch": 0.2848276988404377, "grad_norm": 7.6668548583984375, "learning_rate": 4.322727117869951e-06, "loss": 2.5756, "step": 436 }, { "epoch": 0.28548097337906253, "grad_norm": 8.136970520019531, "learning_rate": 4.190602144143207e-06, "loss": 3.3489, "step": 437 }, { "epoch": 0.28613424791768743, "grad_norm": 10.233713150024414, "learning_rate": 4.06043949255509e-06, "loss": 3.6143, "step": 438 }, { "epoch": 0.2867875224563123, "grad_norm": 10.85151481628418, "learning_rate": 3.932244738840379e-06, "loss": 3.1783, "step": 439 }, { "epoch": 0.28744079699493713, "grad_norm": 9.790265083312988, "learning_rate": 3.8060233744356633e-06, "loss": 2.58, "step": 440 }, { "epoch": 0.288094071533562, "grad_norm": 10.893715858459473, "learning_rate": 3.681780806244095e-06, "loss": 3.0611, "step": 441 }, { "epoch": 0.28874734607218683, "grad_norm": 12.682134628295898, "learning_rate": 3.5595223564037884e-06, "loss": 2.8948, "step": 442 }, { "epoch": 0.2894006206108117, "grad_norm": 13.385756492614746, "learning_rate": 3.4392532620598216e-06, "loss": 2.6341, "step": 443 }, { "epoch": 0.2900538951494365, "grad_norm": 16.10133934020996, "learning_rate": 3.3209786751399187e-06, "loss": 3.5305, "step": 444 }, { "epoch": 0.29070716968806143, "grad_norm": 12.668194770812988, "learning_rate": 3.2047036621337236e-06, "loss": 2.3386, "step": 445 }, { "epoch": 0.2913604442266863, "grad_norm": 14.395100593566895, "learning_rate": 3.0904332038757977e-06, "loss": 2.0309, "step": 446 }, { "epoch": 0.2920137187653111, "grad_norm": 17.13104820251465, "learning_rate": 2.978172195332263e-06, "loss": 2.4272, "step": 447 }, { "epoch": 0.292666993303936, "grad_norm": 17.783287048339844, "learning_rate": 2.8679254453910785e-06, "loss": 2.7686, "step": 448 }, { "epoch": 0.2933202678425608, "grad_norm": 22.626319885253906, "learning_rate": 2.759697676656098e-06, "loss": 2.9195, "step": 449 }, { "epoch": 0.29397354238118567, "grad_norm": 35.02473831176758, "learning_rate": 2.653493525244721e-06, "loss": 4.8236, "step": 450 }, { "epoch": 0.2946268169198106, "grad_norm": 2.0250165462493896, "learning_rate": 2.549317540589308e-06, "loss": 4.0218, "step": 451 }, { "epoch": 0.2952800914584354, "grad_norm": 1.948515772819519, "learning_rate": 2.4471741852423237e-06, "loss": 3.9039, "step": 452 }, { "epoch": 0.29593336599706027, "grad_norm": 2.03169322013855, "learning_rate": 2.3470678346851518e-06, "loss": 4.3663, "step": 453 }, { "epoch": 0.2965866405356851, "grad_norm": 2.1053688526153564, "learning_rate": 2.2490027771406687e-06, "loss": 4.0127, "step": 454 }, { "epoch": 0.29723991507430997, "grad_norm": 2.2202868461608887, "learning_rate": 2.152983213389559e-06, "loss": 4.0635, "step": 455 }, { "epoch": 0.2978931896129348, "grad_norm": 2.141794443130493, "learning_rate": 2.0590132565903476e-06, "loss": 3.9512, "step": 456 }, { "epoch": 0.2985464641515597, "grad_norm": 2.1424460411071777, "learning_rate": 1.9670969321032407e-06, "loss": 3.928, "step": 457 }, { "epoch": 0.29919973869018457, "grad_norm": 2.2259063720703125, "learning_rate": 1.8772381773176417e-06, "loss": 3.9537, "step": 458 }, { "epoch": 0.2998530132288094, "grad_norm": 2.1677327156066895, "learning_rate": 1.7894408414835362e-06, "loss": 3.7739, "step": 459 }, { "epoch": 0.30050628776743427, "grad_norm": 2.260230541229248, "learning_rate": 1.70370868554659e-06, "loss": 4.0705, "step": 460 }, { "epoch": 0.3011595623060591, "grad_norm": 2.394634246826172, "learning_rate": 1.620045381987012e-06, "loss": 4.1479, "step": 461 }, { "epoch": 0.30181283684468396, "grad_norm": 2.3297996520996094, "learning_rate": 1.5384545146622852e-06, "loss": 3.6953, "step": 462 }, { "epoch": 0.3024661113833088, "grad_norm": 2.3964977264404297, "learning_rate": 1.4589395786535953e-06, "loss": 3.9611, "step": 463 }, { "epoch": 0.3031193859219337, "grad_norm": 2.533637046813965, "learning_rate": 1.3815039801161721e-06, "loss": 3.9468, "step": 464 }, { "epoch": 0.30377266046055856, "grad_norm": 2.529632329940796, "learning_rate": 1.3061510361333185e-06, "loss": 3.6358, "step": 465 }, { "epoch": 0.3044259349991834, "grad_norm": 2.569408655166626, "learning_rate": 1.232883974574367e-06, "loss": 3.5356, "step": 466 }, { "epoch": 0.30507920953780826, "grad_norm": 2.727712869644165, "learning_rate": 1.1617059339563807e-06, "loss": 3.7934, "step": 467 }, { "epoch": 0.3057324840764331, "grad_norm": 2.8594043254852295, "learning_rate": 1.0926199633097157e-06, "loss": 3.933, "step": 468 }, { "epoch": 0.30638575861505796, "grad_norm": 2.939509868621826, "learning_rate": 1.0256290220474307e-06, "loss": 3.5852, "step": 469 }, { "epoch": 0.30703903315368286, "grad_norm": 3.0464086532592773, "learning_rate": 9.607359798384785e-07, "loss": 3.7145, "step": 470 }, { "epoch": 0.3076923076923077, "grad_norm": 3.3523106575012207, "learning_rate": 8.979436164848088e-07, "loss": 3.8206, "step": 471 }, { "epoch": 0.30834558223093256, "grad_norm": 3.190808057785034, "learning_rate": 8.372546218022747e-07, "loss": 3.8822, "step": 472 }, { "epoch": 0.3089988567695574, "grad_norm": 3.408557891845703, "learning_rate": 7.786715955054203e-07, "loss": 3.5316, "step": 473 }, { "epoch": 0.30965213130818225, "grad_norm": 3.75345778465271, "learning_rate": 7.221970470961125e-07, "loss": 3.8423, "step": 474 }, { "epoch": 0.3103054058468071, "grad_norm": 4.055991172790527, "learning_rate": 6.678333957560512e-07, "loss": 3.6669, "step": 475 }, { "epoch": 0.31095868038543195, "grad_norm": 4.102751731872559, "learning_rate": 6.15582970243117e-07, "loss": 3.4249, "step": 476 }, { "epoch": 0.31161195492405686, "grad_norm": 4.2560014724731445, "learning_rate": 5.654480087916303e-07, "loss": 3.4731, "step": 477 }, { "epoch": 0.3122652294626817, "grad_norm": 4.73115873336792, "learning_rate": 5.174306590164879e-07, "loss": 3.3286, "step": 478 }, { "epoch": 0.31291850400130655, "grad_norm": 5.362375259399414, "learning_rate": 4.715329778211375e-07, "loss": 3.2443, "step": 479 }, { "epoch": 0.3135717785399314, "grad_norm": 5.289713382720947, "learning_rate": 4.277569313094809e-07, "loss": 3.2602, "step": 480 }, { "epoch": 0.31422505307855625, "grad_norm": 5.72009801864624, "learning_rate": 3.8610439470164737e-07, "loss": 3.1795, "step": 481 }, { "epoch": 0.3148783276171811, "grad_norm": 6.249419212341309, "learning_rate": 3.465771522536854e-07, "loss": 3.2568, "step": 482 }, { "epoch": 0.315531602155806, "grad_norm": 6.340628147125244, "learning_rate": 3.09176897181096e-07, "loss": 3.7829, "step": 483 }, { "epoch": 0.31618487669443085, "grad_norm": 6.803210735321045, "learning_rate": 2.7390523158633554e-07, "loss": 3.2678, "step": 484 }, { "epoch": 0.3168381512330557, "grad_norm": 7.402984619140625, "learning_rate": 2.407636663901591e-07, "loss": 3.2089, "step": 485 }, { "epoch": 0.31749142577168055, "grad_norm": 8.722670555114746, "learning_rate": 2.0975362126691712e-07, "loss": 3.4273, "step": 486 }, { "epoch": 0.3181447003103054, "grad_norm": 8.667076110839844, "learning_rate": 1.8087642458373134e-07, "loss": 2.9022, "step": 487 }, { "epoch": 0.31879797484893024, "grad_norm": 9.627767562866211, "learning_rate": 1.5413331334360182e-07, "loss": 3.0261, "step": 488 }, { "epoch": 0.31945124938755515, "grad_norm": 10.799980163574219, "learning_rate": 1.2952543313240472e-07, "loss": 2.731, "step": 489 }, { "epoch": 0.32010452392618, "grad_norm": 12.3025484085083, "learning_rate": 1.0705383806982606e-07, "loss": 3.1692, "step": 490 }, { "epoch": 0.32075779846480484, "grad_norm": 11.816473007202148, "learning_rate": 8.671949076420882e-08, "loss": 2.1674, "step": 491 }, { "epoch": 0.3214110730034297, "grad_norm": 13.557419776916504, "learning_rate": 6.852326227130834e-08, "loss": 2.846, "step": 492 }, { "epoch": 0.32206434754205454, "grad_norm": 11.210492134094238, "learning_rate": 5.246593205699424e-08, "loss": 2.5836, "step": 493 }, { "epoch": 0.3227176220806794, "grad_norm": 13.040658950805664, "learning_rate": 3.8548187963854956e-08, "loss": 3.2673, "step": 494 }, { "epoch": 0.32337089661930424, "grad_norm": 15.785333633422852, "learning_rate": 2.6770626181715773e-08, "loss": 2.518, "step": 495 }, { "epoch": 0.32402417115792914, "grad_norm": 15.67574405670166, "learning_rate": 1.7133751222137007e-08, "loss": 2.387, "step": 496 }, { "epoch": 0.324677445696554, "grad_norm": 15.306534767150879, "learning_rate": 9.637975896759077e-09, "loss": 2.3683, "step": 497 }, { "epoch": 0.32533072023517884, "grad_norm": 20.011672973632812, "learning_rate": 4.2836212996499865e-09, "loss": 3.7281, "step": 498 }, { "epoch": 0.3259839947738037, "grad_norm": 21.42987823486328, "learning_rate": 1.0709167935385455e-09, "loss": 2.4191, "step": 499 }, { "epoch": 0.32663726931242854, "grad_norm": 32.15013885498047, "learning_rate": 0.0, "loss": 3.3215, "step": 500 }, { "epoch": 0.32663726931242854, "eval_loss": 0.8778727054595947, "eval_runtime": 193.3536, "eval_samples_per_second": 13.333, "eval_steps_per_second": 3.336, "step": 500 } ], "logging_steps": 1, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.243979410448056e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }