abaddon182's picture
Training in progress, step 500, checkpoint
5f55337 verified
{
"best_metric": 0.8778727054595947,
"best_model_checkpoint": "miner_id_24/checkpoint-500",
"epoch": 0.32663726931242854,
"eval_steps": 100,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0006532745386248571,
"grad_norm": 11.547571182250977,
"learning_rate": 5e-06,
"loss": 4.6466,
"step": 1
},
{
"epoch": 0.0006532745386248571,
"eval_loss": 1.2989917993545532,
"eval_runtime": 190.9993,
"eval_samples_per_second": 13.497,
"eval_steps_per_second": 3.377,
"step": 1
},
{
"epoch": 0.0013065490772497142,
"grad_norm": 11.975031852722168,
"learning_rate": 1e-05,
"loss": 5.0526,
"step": 2
},
{
"epoch": 0.0019598236158745713,
"grad_norm": 10.511598587036133,
"learning_rate": 1.5e-05,
"loss": 4.8616,
"step": 3
},
{
"epoch": 0.0026130981544994283,
"grad_norm": 8.188464164733887,
"learning_rate": 2e-05,
"loss": 4.739,
"step": 4
},
{
"epoch": 0.0032663726931242854,
"grad_norm": 4.608264446258545,
"learning_rate": 2.5e-05,
"loss": 4.2898,
"step": 5
},
{
"epoch": 0.0039196472317491425,
"grad_norm": 4.9322381019592285,
"learning_rate": 3e-05,
"loss": 4.4307,
"step": 6
},
{
"epoch": 0.004572921770374,
"grad_norm": 5.750156402587891,
"learning_rate": 3.5e-05,
"loss": 4.3233,
"step": 7
},
{
"epoch": 0.005226196308998857,
"grad_norm": 5.375987529754639,
"learning_rate": 4e-05,
"loss": 4.235,
"step": 8
},
{
"epoch": 0.005879470847623714,
"grad_norm": 4.336426734924316,
"learning_rate": 4.5e-05,
"loss": 4.2463,
"step": 9
},
{
"epoch": 0.006532745386248571,
"grad_norm": 3.931546211242676,
"learning_rate": 5e-05,
"loss": 4.0456,
"step": 10
},
{
"epoch": 0.007186019924873428,
"grad_norm": 4.456380844116211,
"learning_rate": 5.500000000000001e-05,
"loss": 4.2932,
"step": 11
},
{
"epoch": 0.007839294463498285,
"grad_norm": 4.070809364318848,
"learning_rate": 6e-05,
"loss": 3.968,
"step": 12
},
{
"epoch": 0.008492569002123142,
"grad_norm": 3.567103147506714,
"learning_rate": 6.500000000000001e-05,
"loss": 4.1319,
"step": 13
},
{
"epoch": 0.009145843540748,
"grad_norm": 3.41229248046875,
"learning_rate": 7e-05,
"loss": 3.9014,
"step": 14
},
{
"epoch": 0.009799118079372856,
"grad_norm": 3.476752281188965,
"learning_rate": 7.500000000000001e-05,
"loss": 3.9473,
"step": 15
},
{
"epoch": 0.010452392617997713,
"grad_norm": 3.421405076980591,
"learning_rate": 8e-05,
"loss": 3.6684,
"step": 16
},
{
"epoch": 0.01110566715662257,
"grad_norm": 3.7692196369171143,
"learning_rate": 8.5e-05,
"loss": 3.6834,
"step": 17
},
{
"epoch": 0.011758941695247428,
"grad_norm": 3.8790102005004883,
"learning_rate": 9e-05,
"loss": 4.1705,
"step": 18
},
{
"epoch": 0.012412216233872285,
"grad_norm": 4.199179172515869,
"learning_rate": 9.5e-05,
"loss": 4.0168,
"step": 19
},
{
"epoch": 0.013065490772497142,
"grad_norm": 4.3811516761779785,
"learning_rate": 0.0001,
"loss": 4.2636,
"step": 20
},
{
"epoch": 0.013718765311121999,
"grad_norm": 4.086483478546143,
"learning_rate": 9.999892908320647e-05,
"loss": 3.386,
"step": 21
},
{
"epoch": 0.014372039849746856,
"grad_norm": 4.643093585968018,
"learning_rate": 9.999571637870036e-05,
"loss": 4.1357,
"step": 22
},
{
"epoch": 0.015025314388371713,
"grad_norm": 4.607544422149658,
"learning_rate": 9.999036202410325e-05,
"loss": 3.4956,
"step": 23
},
{
"epoch": 0.01567858892699657,
"grad_norm": 4.911647319793701,
"learning_rate": 9.998286624877786e-05,
"loss": 3.5906,
"step": 24
},
{
"epoch": 0.01633186346562143,
"grad_norm": 5.430598735809326,
"learning_rate": 9.997322937381829e-05,
"loss": 4.3089,
"step": 25
},
{
"epoch": 0.016985138004246284,
"grad_norm": 5.196270942687988,
"learning_rate": 9.996145181203615e-05,
"loss": 3.7026,
"step": 26
},
{
"epoch": 0.017638412542871143,
"grad_norm": 6.21434211730957,
"learning_rate": 9.994753406794301e-05,
"loss": 3.8458,
"step": 27
},
{
"epoch": 0.018291687081496,
"grad_norm": 6.019957542419434,
"learning_rate": 9.99314767377287e-05,
"loss": 3.638,
"step": 28
},
{
"epoch": 0.018944961620120857,
"grad_norm": 7.344250679016113,
"learning_rate": 9.991328050923581e-05,
"loss": 3.5139,
"step": 29
},
{
"epoch": 0.019598236158745713,
"grad_norm": 8.33799934387207,
"learning_rate": 9.989294616193017e-05,
"loss": 3.8276,
"step": 30
},
{
"epoch": 0.02025151069737057,
"grad_norm": 7.272817611694336,
"learning_rate": 9.98704745668676e-05,
"loss": 3.7295,
"step": 31
},
{
"epoch": 0.020904785235995427,
"grad_norm": 8.105911254882812,
"learning_rate": 9.98458666866564e-05,
"loss": 3.6633,
"step": 32
},
{
"epoch": 0.021558059774620286,
"grad_norm": 10.303070068359375,
"learning_rate": 9.981912357541627e-05,
"loss": 4.163,
"step": 33
},
{
"epoch": 0.02221133431324514,
"grad_norm": 11.288853645324707,
"learning_rate": 9.97902463787331e-05,
"loss": 3.8152,
"step": 34
},
{
"epoch": 0.02286460885187,
"grad_norm": 11.985376358032227,
"learning_rate": 9.975923633360985e-05,
"loss": 3.189,
"step": 35
},
{
"epoch": 0.023517883390494855,
"grad_norm": 11.919270515441895,
"learning_rate": 9.972609476841367e-05,
"loss": 3.3431,
"step": 36
},
{
"epoch": 0.024171157929119714,
"grad_norm": 11.31579303741455,
"learning_rate": 9.969082310281891e-05,
"loss": 3.6613,
"step": 37
},
{
"epoch": 0.02482443246774457,
"grad_norm": 13.086581230163574,
"learning_rate": 9.965342284774632e-05,
"loss": 3.2378,
"step": 38
},
{
"epoch": 0.025477707006369428,
"grad_norm": 13.448277473449707,
"learning_rate": 9.961389560529836e-05,
"loss": 3.7662,
"step": 39
},
{
"epoch": 0.026130981544994283,
"grad_norm": 15.159260749816895,
"learning_rate": 9.957224306869053e-05,
"loss": 3.0795,
"step": 40
},
{
"epoch": 0.026784256083619142,
"grad_norm": 16.633817672729492,
"learning_rate": 9.952846702217886e-05,
"loss": 3.2916,
"step": 41
},
{
"epoch": 0.027437530622243998,
"grad_norm": 17.962066650390625,
"learning_rate": 9.948256934098352e-05,
"loss": 3.0719,
"step": 42
},
{
"epoch": 0.028090805160868856,
"grad_norm": 19.26710319519043,
"learning_rate": 9.943455199120837e-05,
"loss": 3.4875,
"step": 43
},
{
"epoch": 0.02874407969949371,
"grad_norm": 23.785097122192383,
"learning_rate": 9.938441702975689e-05,
"loss": 3.5956,
"step": 44
},
{
"epoch": 0.02939735423811857,
"grad_norm": 25.271921157836914,
"learning_rate": 9.933216660424395e-05,
"loss": 3.0942,
"step": 45
},
{
"epoch": 0.030050628776743426,
"grad_norm": 26.448965072631836,
"learning_rate": 9.927780295290389e-05,
"loss": 3.391,
"step": 46
},
{
"epoch": 0.030703903315368285,
"grad_norm": 21.440593719482422,
"learning_rate": 9.922132840449459e-05,
"loss": 2.6666,
"step": 47
},
{
"epoch": 0.03135717785399314,
"grad_norm": 25.863950729370117,
"learning_rate": 9.916274537819775e-05,
"loss": 2.983,
"step": 48
},
{
"epoch": 0.032010452392617995,
"grad_norm": 31.86448097229004,
"learning_rate": 9.91020563835152e-05,
"loss": 3.6128,
"step": 49
},
{
"epoch": 0.03266372693124286,
"grad_norm": 47.512001037597656,
"learning_rate": 9.903926402016153e-05,
"loss": 5.6647,
"step": 50
},
{
"epoch": 0.03331700146986771,
"grad_norm": 7.3868536949157715,
"learning_rate": 9.897437097795257e-05,
"loss": 4.1604,
"step": 51
},
{
"epoch": 0.03397027600849257,
"grad_norm": 4.953103065490723,
"learning_rate": 9.890738003669029e-05,
"loss": 4.5837,
"step": 52
},
{
"epoch": 0.034623550547117424,
"grad_norm": 3.1169042587280273,
"learning_rate": 9.883829406604363e-05,
"loss": 4.6018,
"step": 53
},
{
"epoch": 0.035276825085742286,
"grad_norm": 2.86468505859375,
"learning_rate": 9.876711602542563e-05,
"loss": 4.3774,
"step": 54
},
{
"epoch": 0.03593009962436714,
"grad_norm": 2.981433868408203,
"learning_rate": 9.869384896386668e-05,
"loss": 4.2887,
"step": 55
},
{
"epoch": 0.036583374162992,
"grad_norm": 2.8601815700531006,
"learning_rate": 9.861849601988383e-05,
"loss": 3.9852,
"step": 56
},
{
"epoch": 0.03723664870161685,
"grad_norm": 2.5838541984558105,
"learning_rate": 9.854106042134641e-05,
"loss": 3.9778,
"step": 57
},
{
"epoch": 0.037889923240241714,
"grad_norm": 2.4947381019592285,
"learning_rate": 9.846154548533773e-05,
"loss": 4.0099,
"step": 58
},
{
"epoch": 0.03854319777886657,
"grad_norm": 2.7639665603637695,
"learning_rate": 9.837995461801299e-05,
"loss": 4.0092,
"step": 59
},
{
"epoch": 0.039196472317491425,
"grad_norm": 2.836315631866455,
"learning_rate": 9.829629131445342e-05,
"loss": 3.7736,
"step": 60
},
{
"epoch": 0.03984974685611628,
"grad_norm": 2.828981637954712,
"learning_rate": 9.821055915851647e-05,
"loss": 3.8882,
"step": 61
},
{
"epoch": 0.04050302139474114,
"grad_norm": 3.3011579513549805,
"learning_rate": 9.812276182268236e-05,
"loss": 4.1102,
"step": 62
},
{
"epoch": 0.041156295933366,
"grad_norm": 2.943286180496216,
"learning_rate": 9.803290306789676e-05,
"loss": 4.0913,
"step": 63
},
{
"epoch": 0.04180957047199085,
"grad_norm": 2.9910857677459717,
"learning_rate": 9.794098674340965e-05,
"loss": 3.8973,
"step": 64
},
{
"epoch": 0.04246284501061571,
"grad_norm": 3.0220143795013428,
"learning_rate": 9.784701678661045e-05,
"loss": 3.8295,
"step": 65
},
{
"epoch": 0.04311611954924057,
"grad_norm": 3.3190722465515137,
"learning_rate": 9.775099722285935e-05,
"loss": 4.1248,
"step": 66
},
{
"epoch": 0.043769394087865426,
"grad_norm": 3.1170575618743896,
"learning_rate": 9.765293216531486e-05,
"loss": 3.876,
"step": 67
},
{
"epoch": 0.04442266862649028,
"grad_norm": 3.3601784706115723,
"learning_rate": 9.755282581475769e-05,
"loss": 4.2187,
"step": 68
},
{
"epoch": 0.04507594316511514,
"grad_norm": 3.230534553527832,
"learning_rate": 9.74506824594107e-05,
"loss": 4.0142,
"step": 69
},
{
"epoch": 0.04572921770374,
"grad_norm": 3.33896541595459,
"learning_rate": 9.73465064747553e-05,
"loss": 3.869,
"step": 70
},
{
"epoch": 0.046382492242364855,
"grad_norm": 3.4622280597686768,
"learning_rate": 9.724030232334391e-05,
"loss": 3.9758,
"step": 71
},
{
"epoch": 0.04703576678098971,
"grad_norm": 3.7361583709716797,
"learning_rate": 9.713207455460894e-05,
"loss": 3.7177,
"step": 72
},
{
"epoch": 0.047689041319614565,
"grad_norm": 4.231381416320801,
"learning_rate": 9.702182780466775e-05,
"loss": 4.1549,
"step": 73
},
{
"epoch": 0.04834231585823943,
"grad_norm": 4.004554271697998,
"learning_rate": 9.690956679612421e-05,
"loss": 3.6364,
"step": 74
},
{
"epoch": 0.04899559039686428,
"grad_norm": 3.9489033222198486,
"learning_rate": 9.67952963378663e-05,
"loss": 3.6562,
"step": 75
},
{
"epoch": 0.04964886493548914,
"grad_norm": 4.792303562164307,
"learning_rate": 9.667902132486009e-05,
"loss": 4.2874,
"step": 76
},
{
"epoch": 0.050302139474113994,
"grad_norm": 5.080928802490234,
"learning_rate": 9.656074673794018e-05,
"loss": 3.4885,
"step": 77
},
{
"epoch": 0.050955414012738856,
"grad_norm": 5.956547737121582,
"learning_rate": 9.644047764359622e-05,
"loss": 3.8888,
"step": 78
},
{
"epoch": 0.05160868855136371,
"grad_norm": 5.492791652679443,
"learning_rate": 9.631821919375591e-05,
"loss": 3.5125,
"step": 79
},
{
"epoch": 0.05226196308998857,
"grad_norm": 5.798184394836426,
"learning_rate": 9.619397662556435e-05,
"loss": 3.6023,
"step": 80
},
{
"epoch": 0.05291523762861342,
"grad_norm": 6.541443824768066,
"learning_rate": 9.606775526115963e-05,
"loss": 3.9795,
"step": 81
},
{
"epoch": 0.053568512167238284,
"grad_norm": 6.9007248878479,
"learning_rate": 9.593956050744492e-05,
"loss": 3.9131,
"step": 82
},
{
"epoch": 0.05422178670586314,
"grad_norm": 7.156078815460205,
"learning_rate": 9.580939785585681e-05,
"loss": 3.5494,
"step": 83
},
{
"epoch": 0.054875061244487995,
"grad_norm": 7.655080318450928,
"learning_rate": 9.567727288213005e-05,
"loss": 3.526,
"step": 84
},
{
"epoch": 0.05552833578311285,
"grad_norm": 8.81391429901123,
"learning_rate": 9.554319124605879e-05,
"loss": 3.9594,
"step": 85
},
{
"epoch": 0.05618161032173771,
"grad_norm": 9.645224571228027,
"learning_rate": 9.540715869125407e-05,
"loss": 3.8675,
"step": 86
},
{
"epoch": 0.05683488486036257,
"grad_norm": 9.505653381347656,
"learning_rate": 9.526918104489777e-05,
"loss": 3.3039,
"step": 87
},
{
"epoch": 0.05748815939898742,
"grad_norm": 11.144811630249023,
"learning_rate": 9.512926421749304e-05,
"loss": 3.289,
"step": 88
},
{
"epoch": 0.05814143393761228,
"grad_norm": 11.904159545898438,
"learning_rate": 9.498741420261108e-05,
"loss": 3.5135,
"step": 89
},
{
"epoch": 0.05879470847623714,
"grad_norm": 11.668728828430176,
"learning_rate": 9.484363707663442e-05,
"loss": 3.1836,
"step": 90
},
{
"epoch": 0.059447983014861996,
"grad_norm": 13.778961181640625,
"learning_rate": 9.469793899849661e-05,
"loss": 3.6804,
"step": 91
},
{
"epoch": 0.06010125755348685,
"grad_norm": 13.867650985717773,
"learning_rate": 9.45503262094184e-05,
"loss": 3.4685,
"step": 92
},
{
"epoch": 0.06075453209211171,
"grad_norm": 17.512781143188477,
"learning_rate": 9.440080503264037e-05,
"loss": 3.3502,
"step": 93
},
{
"epoch": 0.06140780663073657,
"grad_norm": 24.185264587402344,
"learning_rate": 9.42493818731521e-05,
"loss": 2.7727,
"step": 94
},
{
"epoch": 0.062061081169361425,
"grad_norm": 19.322153091430664,
"learning_rate": 9.409606321741775e-05,
"loss": 3.4637,
"step": 95
},
{
"epoch": 0.06271435570798628,
"grad_norm": 20.148555755615234,
"learning_rate": 9.394085563309827e-05,
"loss": 2.7134,
"step": 96
},
{
"epoch": 0.06336763024661114,
"grad_norm": 19.66134262084961,
"learning_rate": 9.378376576876999e-05,
"loss": 2.1593,
"step": 97
},
{
"epoch": 0.06402090478523599,
"grad_norm": 21.868074417114258,
"learning_rate": 9.362480035363986e-05,
"loss": 3.1943,
"step": 98
},
{
"epoch": 0.06467417932386085,
"grad_norm": 24.685863494873047,
"learning_rate": 9.34639661972572e-05,
"loss": 2.8522,
"step": 99
},
{
"epoch": 0.06532745386248572,
"grad_norm": 32.843997955322266,
"learning_rate": 9.330127018922194e-05,
"loss": 3.7582,
"step": 100
},
{
"epoch": 0.06532745386248572,
"eval_loss": 1.087826132774353,
"eval_runtime": 193.6149,
"eval_samples_per_second": 13.315,
"eval_steps_per_second": 3.331,
"step": 100
},
{
"epoch": 0.06598072840111056,
"grad_norm": 6.478489875793457,
"learning_rate": 9.31367192988896e-05,
"loss": 4.3501,
"step": 101
},
{
"epoch": 0.06663400293973543,
"grad_norm": 4.0702080726623535,
"learning_rate": 9.297032057507264e-05,
"loss": 4.2697,
"step": 102
},
{
"epoch": 0.06728727747836027,
"grad_norm": 2.496466875076294,
"learning_rate": 9.280208114573859e-05,
"loss": 3.8445,
"step": 103
},
{
"epoch": 0.06794055201698514,
"grad_norm": 2.428135633468628,
"learning_rate": 9.263200821770461e-05,
"loss": 4.0675,
"step": 104
},
{
"epoch": 0.06859382655561,
"grad_norm": 2.5381500720977783,
"learning_rate": 9.246010907632895e-05,
"loss": 4.223,
"step": 105
},
{
"epoch": 0.06924710109423485,
"grad_norm": 2.4143152236938477,
"learning_rate": 9.228639108519868e-05,
"loss": 3.9627,
"step": 106
},
{
"epoch": 0.06990037563285971,
"grad_norm": 2.6589813232421875,
"learning_rate": 9.211086168581433e-05,
"loss": 4.0724,
"step": 107
},
{
"epoch": 0.07055365017148457,
"grad_norm": 2.906999349594116,
"learning_rate": 9.193352839727121e-05,
"loss": 4.0853,
"step": 108
},
{
"epoch": 0.07120692471010942,
"grad_norm": 2.604168653488159,
"learning_rate": 9.175439881593716e-05,
"loss": 3.9059,
"step": 109
},
{
"epoch": 0.07186019924873428,
"grad_norm": 2.6407856941223145,
"learning_rate": 9.157348061512727e-05,
"loss": 4.1648,
"step": 110
},
{
"epoch": 0.07251347378735913,
"grad_norm": 2.5845110416412354,
"learning_rate": 9.139078154477512e-05,
"loss": 3.9705,
"step": 111
},
{
"epoch": 0.073166748325984,
"grad_norm": 2.725156784057617,
"learning_rate": 9.120630943110077e-05,
"loss": 3.8773,
"step": 112
},
{
"epoch": 0.07382002286460886,
"grad_norm": 2.776826858520508,
"learning_rate": 9.102007217627568e-05,
"loss": 4.1886,
"step": 113
},
{
"epoch": 0.0744732974032337,
"grad_norm": 2.7510571479797363,
"learning_rate": 9.083207775808396e-05,
"loss": 3.8013,
"step": 114
},
{
"epoch": 0.07512657194185857,
"grad_norm": 2.9689526557922363,
"learning_rate": 9.064233422958077e-05,
"loss": 3.8871,
"step": 115
},
{
"epoch": 0.07577984648048343,
"grad_norm": 2.9751687049865723,
"learning_rate": 9.045084971874738e-05,
"loss": 4.0687,
"step": 116
},
{
"epoch": 0.07643312101910828,
"grad_norm": 3.208277702331543,
"learning_rate": 9.025763242814291e-05,
"loss": 4.0952,
"step": 117
},
{
"epoch": 0.07708639555773314,
"grad_norm": 2.978297233581543,
"learning_rate": 9.006269063455304e-05,
"loss": 3.7542,
"step": 118
},
{
"epoch": 0.07773967009635799,
"grad_norm": 3.079564332962036,
"learning_rate": 8.986603268863536e-05,
"loss": 3.7339,
"step": 119
},
{
"epoch": 0.07839294463498285,
"grad_norm": 3.3465781211853027,
"learning_rate": 8.966766701456177e-05,
"loss": 4.1308,
"step": 120
},
{
"epoch": 0.07904621917360771,
"grad_norm": 3.38362717628479,
"learning_rate": 8.94676021096575e-05,
"loss": 3.6047,
"step": 121
},
{
"epoch": 0.07969949371223256,
"grad_norm": 3.4256107807159424,
"learning_rate": 8.926584654403724e-05,
"loss": 4.0278,
"step": 122
},
{
"epoch": 0.08035276825085742,
"grad_norm": 3.728395700454712,
"learning_rate": 8.906240896023794e-05,
"loss": 3.5618,
"step": 123
},
{
"epoch": 0.08100604278948229,
"grad_norm": 4.323002338409424,
"learning_rate": 8.885729807284856e-05,
"loss": 3.7879,
"step": 124
},
{
"epoch": 0.08165931732810713,
"grad_norm": 4.338598728179932,
"learning_rate": 8.865052266813685e-05,
"loss": 3.4469,
"step": 125
},
{
"epoch": 0.082312591866732,
"grad_norm": 4.841395854949951,
"learning_rate": 8.844209160367299e-05,
"loss": 3.6624,
"step": 126
},
{
"epoch": 0.08296586640535684,
"grad_norm": 5.293141841888428,
"learning_rate": 8.823201380795001e-05,
"loss": 3.6083,
"step": 127
},
{
"epoch": 0.0836191409439817,
"grad_norm": 6.072646141052246,
"learning_rate": 8.802029828000156e-05,
"loss": 3.7721,
"step": 128
},
{
"epoch": 0.08427241548260657,
"grad_norm": 5.880528450012207,
"learning_rate": 8.780695408901613e-05,
"loss": 3.6489,
"step": 129
},
{
"epoch": 0.08492569002123142,
"grad_norm": 6.2913055419921875,
"learning_rate": 8.759199037394887e-05,
"loss": 4.0706,
"step": 130
},
{
"epoch": 0.08557896455985628,
"grad_norm": 6.430176258087158,
"learning_rate": 8.737541634312985e-05,
"loss": 3.6765,
"step": 131
},
{
"epoch": 0.08623223909848114,
"grad_norm": 6.8034987449646,
"learning_rate": 8.715724127386972e-05,
"loss": 3.7984,
"step": 132
},
{
"epoch": 0.08688551363710599,
"grad_norm": 8.133237838745117,
"learning_rate": 8.693747451206232e-05,
"loss": 3.7513,
"step": 133
},
{
"epoch": 0.08753878817573085,
"grad_norm": 9.387150764465332,
"learning_rate": 8.671612547178428e-05,
"loss": 3.943,
"step": 134
},
{
"epoch": 0.0881920627143557,
"grad_norm": 9.971433639526367,
"learning_rate": 8.649320363489179e-05,
"loss": 3.5509,
"step": 135
},
{
"epoch": 0.08884533725298056,
"grad_norm": 8.350509643554688,
"learning_rate": 8.626871855061438e-05,
"loss": 2.8905,
"step": 136
},
{
"epoch": 0.08949861179160543,
"grad_norm": 10.003087043762207,
"learning_rate": 8.604267983514594e-05,
"loss": 3.369,
"step": 137
},
{
"epoch": 0.09015188633023027,
"grad_norm": 11.017991065979004,
"learning_rate": 8.581509717123273e-05,
"loss": 3.826,
"step": 138
},
{
"epoch": 0.09080516086885514,
"grad_norm": 11.706796646118164,
"learning_rate": 8.558598030775857e-05,
"loss": 2.9855,
"step": 139
},
{
"epoch": 0.09145843540748,
"grad_norm": 12.903501510620117,
"learning_rate": 8.535533905932738e-05,
"loss": 2.6285,
"step": 140
},
{
"epoch": 0.09211170994610485,
"grad_norm": 12.504664421081543,
"learning_rate": 8.51231833058426e-05,
"loss": 3.0194,
"step": 141
},
{
"epoch": 0.09276498448472971,
"grad_norm": 14.484477043151855,
"learning_rate": 8.488952299208401e-05,
"loss": 3.1958,
"step": 142
},
{
"epoch": 0.09341825902335456,
"grad_norm": 17.707530975341797,
"learning_rate": 8.46543681272818e-05,
"loss": 3.957,
"step": 143
},
{
"epoch": 0.09407153356197942,
"grad_norm": 20.62574005126953,
"learning_rate": 8.44177287846877e-05,
"loss": 3.2235,
"step": 144
},
{
"epoch": 0.09472480810060428,
"grad_norm": 18.29495620727539,
"learning_rate": 8.417961510114356e-05,
"loss": 2.3518,
"step": 145
},
{
"epoch": 0.09537808263922913,
"grad_norm": 28.382152557373047,
"learning_rate": 8.39400372766471e-05,
"loss": 4.1565,
"step": 146
},
{
"epoch": 0.096031357177854,
"grad_norm": 18.329111099243164,
"learning_rate": 8.36990055739149e-05,
"loss": 2.1261,
"step": 147
},
{
"epoch": 0.09668463171647886,
"grad_norm": 18.869550704956055,
"learning_rate": 8.345653031794292e-05,
"loss": 2.6734,
"step": 148
},
{
"epoch": 0.0973379062551037,
"grad_norm": 23.500282287597656,
"learning_rate": 8.321262189556409e-05,
"loss": 3.4861,
"step": 149
},
{
"epoch": 0.09799118079372857,
"grad_norm": 38.71291732788086,
"learning_rate": 8.296729075500344e-05,
"loss": 5.1638,
"step": 150
},
{
"epoch": 0.09864445533235343,
"grad_norm": 3.8299853801727295,
"learning_rate": 8.272054740543052e-05,
"loss": 4.0414,
"step": 151
},
{
"epoch": 0.09929772987097828,
"grad_norm": 3.310225248336792,
"learning_rate": 8.247240241650918e-05,
"loss": 4.6365,
"step": 152
},
{
"epoch": 0.09995100440960314,
"grad_norm": 2.5045037269592285,
"learning_rate": 8.222286641794488e-05,
"loss": 4.4531,
"step": 153
},
{
"epoch": 0.10060427894822799,
"grad_norm": 2.3877153396606445,
"learning_rate": 8.197195009902924e-05,
"loss": 4.7702,
"step": 154
},
{
"epoch": 0.10125755348685285,
"grad_norm": 2.4667954444885254,
"learning_rate": 8.171966420818228e-05,
"loss": 4.1018,
"step": 155
},
{
"epoch": 0.10191082802547771,
"grad_norm": 2.424144983291626,
"learning_rate": 8.146601955249188e-05,
"loss": 4.1768,
"step": 156
},
{
"epoch": 0.10256410256410256,
"grad_norm": 2.476624011993408,
"learning_rate": 8.121102699725089e-05,
"loss": 3.963,
"step": 157
},
{
"epoch": 0.10321737710272742,
"grad_norm": 2.590161085128784,
"learning_rate": 8.095469746549172e-05,
"loss": 3.9657,
"step": 158
},
{
"epoch": 0.10387065164135229,
"grad_norm": 2.8294732570648193,
"learning_rate": 8.069704193751832e-05,
"loss": 4.19,
"step": 159
},
{
"epoch": 0.10452392617997713,
"grad_norm": 2.526905059814453,
"learning_rate": 8.043807145043604e-05,
"loss": 3.7846,
"step": 160
},
{
"epoch": 0.105177200718602,
"grad_norm": 2.666771411895752,
"learning_rate": 8.017779709767858e-05,
"loss": 4.2755,
"step": 161
},
{
"epoch": 0.10583047525722684,
"grad_norm": 2.6683263778686523,
"learning_rate": 7.991623002853296e-05,
"loss": 3.8973,
"step": 162
},
{
"epoch": 0.1064837497958517,
"grad_norm": 2.753737688064575,
"learning_rate": 7.965338144766186e-05,
"loss": 4.0508,
"step": 163
},
{
"epoch": 0.10713702433447657,
"grad_norm": 2.6762168407440186,
"learning_rate": 7.938926261462366e-05,
"loss": 3.8199,
"step": 164
},
{
"epoch": 0.10779029887310142,
"grad_norm": 2.7907073497772217,
"learning_rate": 7.912388484339012e-05,
"loss": 4.0572,
"step": 165
},
{
"epoch": 0.10844357341172628,
"grad_norm": 2.9300825595855713,
"learning_rate": 7.88572595018617e-05,
"loss": 3.85,
"step": 166
},
{
"epoch": 0.10909684795035114,
"grad_norm": 2.946000337600708,
"learning_rate": 7.858939801138061e-05,
"loss": 4.0084,
"step": 167
},
{
"epoch": 0.10975012248897599,
"grad_norm": 3.150700569152832,
"learning_rate": 7.832031184624164e-05,
"loss": 4.0213,
"step": 168
},
{
"epoch": 0.11040339702760085,
"grad_norm": 3.130871057510376,
"learning_rate": 7.80500125332005e-05,
"loss": 3.7754,
"step": 169
},
{
"epoch": 0.1110566715662257,
"grad_norm": 3.2514853477478027,
"learning_rate": 7.777851165098012e-05,
"loss": 3.8823,
"step": 170
},
{
"epoch": 0.11170994610485056,
"grad_norm": 3.6097805500030518,
"learning_rate": 7.750582082977467e-05,
"loss": 3.8464,
"step": 171
},
{
"epoch": 0.11236322064347543,
"grad_norm": 3.4300525188446045,
"learning_rate": 7.723195175075136e-05,
"loss": 3.587,
"step": 172
},
{
"epoch": 0.11301649518210027,
"grad_norm": 3.962552070617676,
"learning_rate": 7.695691614555003e-05,
"loss": 3.9351,
"step": 173
},
{
"epoch": 0.11366976972072514,
"grad_norm": 3.7221381664276123,
"learning_rate": 7.668072579578058e-05,
"loss": 3.7333,
"step": 174
},
{
"epoch": 0.11432304425935,
"grad_norm": 3.8638551235198975,
"learning_rate": 7.64033925325184e-05,
"loss": 3.2305,
"step": 175
},
{
"epoch": 0.11497631879797485,
"grad_norm": 4.228523254394531,
"learning_rate": 7.612492823579745e-05,
"loss": 3.7991,
"step": 176
},
{
"epoch": 0.11562959333659971,
"grad_norm": 4.375904083251953,
"learning_rate": 7.584534483410137e-05,
"loss": 3.3348,
"step": 177
},
{
"epoch": 0.11628286787522456,
"grad_norm": 5.0705156326293945,
"learning_rate": 7.55646543038526e-05,
"loss": 4.2036,
"step": 178
},
{
"epoch": 0.11693614241384942,
"grad_norm": 5.519981384277344,
"learning_rate": 7.528286866889924e-05,
"loss": 3.987,
"step": 179
},
{
"epoch": 0.11758941695247428,
"grad_norm": 5.943816184997559,
"learning_rate": 7.500000000000001e-05,
"loss": 3.4623,
"step": 180
},
{
"epoch": 0.11824269149109913,
"grad_norm": 6.2165751457214355,
"learning_rate": 7.471606041430723e-05,
"loss": 3.4462,
"step": 181
},
{
"epoch": 0.11889596602972399,
"grad_norm": 6.183679580688477,
"learning_rate": 7.443106207484776e-05,
"loss": 3.6863,
"step": 182
},
{
"epoch": 0.11954924056834886,
"grad_norm": 7.212845802307129,
"learning_rate": 7.414501719000187e-05,
"loss": 3.4877,
"step": 183
},
{
"epoch": 0.1202025151069737,
"grad_norm": 7.299278736114502,
"learning_rate": 7.385793801298042e-05,
"loss": 3.1686,
"step": 184
},
{
"epoch": 0.12085578964559857,
"grad_norm": 7.535823345184326,
"learning_rate": 7.35698368412999e-05,
"loss": 3.7468,
"step": 185
},
{
"epoch": 0.12150906418422341,
"grad_norm": 9.03598690032959,
"learning_rate": 7.328072601625557e-05,
"loss": 3.3324,
"step": 186
},
{
"epoch": 0.12216233872284828,
"grad_norm": 8.97368335723877,
"learning_rate": 7.2990617922393e-05,
"loss": 3.0458,
"step": 187
},
{
"epoch": 0.12281561326147314,
"grad_norm": 11.850152015686035,
"learning_rate": 7.269952498697734e-05,
"loss": 3.6011,
"step": 188
},
{
"epoch": 0.12346888780009799,
"grad_norm": 12.661114692687988,
"learning_rate": 7.240745967946113e-05,
"loss": 3.4378,
"step": 189
},
{
"epoch": 0.12412216233872285,
"grad_norm": 15.797039985656738,
"learning_rate": 7.211443451095007e-05,
"loss": 3.9093,
"step": 190
},
{
"epoch": 0.12477543687734771,
"grad_norm": 13.158418655395508,
"learning_rate": 7.18204620336671e-05,
"loss": 3.2219,
"step": 191
},
{
"epoch": 0.12542871141597256,
"grad_norm": 14.715095520019531,
"learning_rate": 7.152555484041476e-05,
"loss": 2.8733,
"step": 192
},
{
"epoch": 0.1260819859545974,
"grad_norm": 15.411552429199219,
"learning_rate": 7.122972556403567e-05,
"loss": 3.4753,
"step": 193
},
{
"epoch": 0.12673526049322228,
"grad_norm": 15.468315124511719,
"learning_rate": 7.09329868768714e-05,
"loss": 3.6554,
"step": 194
},
{
"epoch": 0.12738853503184713,
"grad_norm": 17.27188491821289,
"learning_rate": 7.063535149021973e-05,
"loss": 3.4804,
"step": 195
},
{
"epoch": 0.12804180957047198,
"grad_norm": 21.4166316986084,
"learning_rate": 7.033683215379002e-05,
"loss": 2.967,
"step": 196
},
{
"epoch": 0.12869508410909686,
"grad_norm": 18.70254135131836,
"learning_rate": 7.003744165515705e-05,
"loss": 2.5688,
"step": 197
},
{
"epoch": 0.1293483586477217,
"grad_norm": 22.17182159423828,
"learning_rate": 6.973719281921335e-05,
"loss": 3.8092,
"step": 198
},
{
"epoch": 0.13000163318634655,
"grad_norm": 21.631132125854492,
"learning_rate": 6.943609850761979e-05,
"loss": 3.4749,
"step": 199
},
{
"epoch": 0.13065490772497143,
"grad_norm": 27.764705657958984,
"learning_rate": 6.91341716182545e-05,
"loss": 3.2158,
"step": 200
},
{
"epoch": 0.13065490772497143,
"eval_loss": 1.0087602138519287,
"eval_runtime": 193.6858,
"eval_samples_per_second": 13.31,
"eval_steps_per_second": 3.33,
"step": 200
},
{
"epoch": 0.13130818226359628,
"grad_norm": 2.8136210441589355,
"learning_rate": 6.883142508466054e-05,
"loss": 3.6461,
"step": 201
},
{
"epoch": 0.13196145680222113,
"grad_norm": 3.2069029808044434,
"learning_rate": 6.852787187549182e-05,
"loss": 4.9158,
"step": 202
},
{
"epoch": 0.132614731340846,
"grad_norm": 2.4724817276000977,
"learning_rate": 6.82235249939575e-05,
"loss": 3.7633,
"step": 203
},
{
"epoch": 0.13326800587947085,
"grad_norm": 2.367910146713257,
"learning_rate": 6.7918397477265e-05,
"loss": 4.6979,
"step": 204
},
{
"epoch": 0.1339212804180957,
"grad_norm": 2.252790689468384,
"learning_rate": 6.761250239606169e-05,
"loss": 4.5078,
"step": 205
},
{
"epoch": 0.13457455495672055,
"grad_norm": 2.2511966228485107,
"learning_rate": 6.730585285387465e-05,
"loss": 4.0116,
"step": 206
},
{
"epoch": 0.13522782949534543,
"grad_norm": 2.3549845218658447,
"learning_rate": 6.699846198654971e-05,
"loss": 4.0373,
"step": 207
},
{
"epoch": 0.13588110403397027,
"grad_norm": 2.3142459392547607,
"learning_rate": 6.669034296168855e-05,
"loss": 3.7566,
"step": 208
},
{
"epoch": 0.13653437857259512,
"grad_norm": 2.4672322273254395,
"learning_rate": 6.638150897808468e-05,
"loss": 3.8954,
"step": 209
},
{
"epoch": 0.13718765311122,
"grad_norm": 2.590864896774292,
"learning_rate": 6.607197326515808e-05,
"loss": 4.3348,
"step": 210
},
{
"epoch": 0.13784092764984485,
"grad_norm": 2.5048024654388428,
"learning_rate": 6.57617490823885e-05,
"loss": 4.0438,
"step": 211
},
{
"epoch": 0.1384942021884697,
"grad_norm": 2.490389108657837,
"learning_rate": 6.545084971874738e-05,
"loss": 3.7406,
"step": 212
},
{
"epoch": 0.13914747672709457,
"grad_norm": 2.4702444076538086,
"learning_rate": 6.513928849212873e-05,
"loss": 3.8706,
"step": 213
},
{
"epoch": 0.13980075126571942,
"grad_norm": 2.634340763092041,
"learning_rate": 6.482707874877854e-05,
"loss": 4.0907,
"step": 214
},
{
"epoch": 0.14045402580434427,
"grad_norm": 2.5504937171936035,
"learning_rate": 6.451423386272312e-05,
"loss": 3.6381,
"step": 215
},
{
"epoch": 0.14110730034296914,
"grad_norm": 2.8212695121765137,
"learning_rate": 6.420076723519614e-05,
"loss": 4.3291,
"step": 216
},
{
"epoch": 0.141760574881594,
"grad_norm": 3.0288095474243164,
"learning_rate": 6.388669229406462e-05,
"loss": 4.1332,
"step": 217
},
{
"epoch": 0.14241384942021884,
"grad_norm": 2.837364435195923,
"learning_rate": 6.357202249325371e-05,
"loss": 3.5461,
"step": 218
},
{
"epoch": 0.14306712395884372,
"grad_norm": 3.0349831581115723,
"learning_rate": 6.32567713121704e-05,
"loss": 4.0614,
"step": 219
},
{
"epoch": 0.14372039849746857,
"grad_norm": 3.3535542488098145,
"learning_rate": 6.294095225512603e-05,
"loss": 3.998,
"step": 220
},
{
"epoch": 0.14437367303609341,
"grad_norm": 3.5414798259735107,
"learning_rate": 6.26245788507579e-05,
"loss": 3.9925,
"step": 221
},
{
"epoch": 0.14502694757471826,
"grad_norm": 3.5899298191070557,
"learning_rate": 6.230766465144967e-05,
"loss": 4.4236,
"step": 222
},
{
"epoch": 0.14568022211334314,
"grad_norm": 4.002371788024902,
"learning_rate": 6.199022323275083e-05,
"loss": 4.1189,
"step": 223
},
{
"epoch": 0.146333496651968,
"grad_norm": 4.0465006828308105,
"learning_rate": 6.167226819279528e-05,
"loss": 3.6562,
"step": 224
},
{
"epoch": 0.14698677119059284,
"grad_norm": 4.780096530914307,
"learning_rate": 6.135381315171867e-05,
"loss": 4.0265,
"step": 225
},
{
"epoch": 0.1476400457292177,
"grad_norm": 4.929449081420898,
"learning_rate": 6.103487175107507e-05,
"loss": 3.3766,
"step": 226
},
{
"epoch": 0.14829332026784256,
"grad_norm": 4.737488746643066,
"learning_rate": 6.071545765325254e-05,
"loss": 3.2223,
"step": 227
},
{
"epoch": 0.1489465948064674,
"grad_norm": 5.096072673797607,
"learning_rate": 6.0395584540887963e-05,
"loss": 4.0753,
"step": 228
},
{
"epoch": 0.14959986934509228,
"grad_norm": 5.692816734313965,
"learning_rate": 6.007526611628086e-05,
"loss": 3.2582,
"step": 229
},
{
"epoch": 0.15025314388371713,
"grad_norm": 6.038702011108398,
"learning_rate": 5.9754516100806423e-05,
"loss": 4.0465,
"step": 230
},
{
"epoch": 0.15090641842234198,
"grad_norm": 5.955114841461182,
"learning_rate": 5.9433348234327765e-05,
"loss": 3.5113,
"step": 231
},
{
"epoch": 0.15155969296096686,
"grad_norm": 7.852488994598389,
"learning_rate": 5.911177627460739e-05,
"loss": 3.946,
"step": 232
},
{
"epoch": 0.1522129674995917,
"grad_norm": 7.21500825881958,
"learning_rate": 5.8789813996717736e-05,
"loss": 3.2654,
"step": 233
},
{
"epoch": 0.15286624203821655,
"grad_norm": 7.874154090881348,
"learning_rate": 5.8467475192451226e-05,
"loss": 3.5822,
"step": 234
},
{
"epoch": 0.15351951657684143,
"grad_norm": 8.220457077026367,
"learning_rate": 5.814477366972945e-05,
"loss": 3.5959,
"step": 235
},
{
"epoch": 0.15417279111546628,
"grad_norm": 9.068872451782227,
"learning_rate": 5.782172325201155e-05,
"loss": 3.5749,
"step": 236
},
{
"epoch": 0.15482606565409113,
"grad_norm": 10.15458869934082,
"learning_rate": 5.749833777770225e-05,
"loss": 3.7681,
"step": 237
},
{
"epoch": 0.15547934019271598,
"grad_norm": 10.914835929870605,
"learning_rate": 5.717463109955896e-05,
"loss": 3.6604,
"step": 238
},
{
"epoch": 0.15613261473134085,
"grad_norm": 11.057694435119629,
"learning_rate": 5.685061708409841e-05,
"loss": 3.1422,
"step": 239
},
{
"epoch": 0.1567858892699657,
"grad_norm": 12.083439826965332,
"learning_rate": 5.6526309611002594e-05,
"loss": 3.7388,
"step": 240
},
{
"epoch": 0.15743916380859055,
"grad_norm": 11.058955192565918,
"learning_rate": 5.6201722572524275e-05,
"loss": 3.0529,
"step": 241
},
{
"epoch": 0.15809243834721542,
"grad_norm": 12.308931350708008,
"learning_rate": 5.587686987289189e-05,
"loss": 2.9113,
"step": 242
},
{
"epoch": 0.15874571288584027,
"grad_norm": 12.517452239990234,
"learning_rate": 5.5551765427713884e-05,
"loss": 2.6273,
"step": 243
},
{
"epoch": 0.15939898742446512,
"grad_norm": 13.3101167678833,
"learning_rate": 5.522642316338268e-05,
"loss": 2.5797,
"step": 244
},
{
"epoch": 0.16005226196309,
"grad_norm": 14.140732765197754,
"learning_rate": 5.490085701647805e-05,
"loss": 2.689,
"step": 245
},
{
"epoch": 0.16070553650171485,
"grad_norm": 17.422927856445312,
"learning_rate": 5.457508093317013e-05,
"loss": 2.6193,
"step": 246
},
{
"epoch": 0.1613588110403397,
"grad_norm": 19.202436447143555,
"learning_rate": 5.4249108868622086e-05,
"loss": 3.1502,
"step": 247
},
{
"epoch": 0.16201208557896457,
"grad_norm": 20.904285430908203,
"learning_rate": 5.392295478639225e-05,
"loss": 3.32,
"step": 248
},
{
"epoch": 0.16266536011758942,
"grad_norm": 22.786781311035156,
"learning_rate": 5.359663265783598e-05,
"loss": 3.4771,
"step": 249
},
{
"epoch": 0.16331863465621427,
"grad_norm": 31.27584457397461,
"learning_rate": 5.327015646150716e-05,
"loss": 4.7491,
"step": 250
},
{
"epoch": 0.16397190919483914,
"grad_norm": 2.7455053329467773,
"learning_rate": 5.294354018255945e-05,
"loss": 4.2161,
"step": 251
},
{
"epoch": 0.164625183733464,
"grad_norm": 2.7494430541992188,
"learning_rate": 5.26167978121472e-05,
"loss": 4.3482,
"step": 252
},
{
"epoch": 0.16527845827208884,
"grad_norm": 2.4116597175598145,
"learning_rate": 5.228994334682604e-05,
"loss": 4.2538,
"step": 253
},
{
"epoch": 0.1659317328107137,
"grad_norm": 2.3236160278320312,
"learning_rate": 5.196299078795344e-05,
"loss": 3.9966,
"step": 254
},
{
"epoch": 0.16658500734933857,
"grad_norm": 2.252965211868286,
"learning_rate": 5.1635954141088813e-05,
"loss": 3.9031,
"step": 255
},
{
"epoch": 0.1672382818879634,
"grad_norm": 2.2256863117218018,
"learning_rate": 5.1308847415393666e-05,
"loss": 3.8695,
"step": 256
},
{
"epoch": 0.16789155642658826,
"grad_norm": 2.294617176055908,
"learning_rate": 5.0981684623031415e-05,
"loss": 4.0556,
"step": 257
},
{
"epoch": 0.16854483096521314,
"grad_norm": 2.3014414310455322,
"learning_rate": 5.0654479778567223e-05,
"loss": 3.731,
"step": 258
},
{
"epoch": 0.169198105503838,
"grad_norm": 2.2849111557006836,
"learning_rate": 5.0327246898367597e-05,
"loss": 3.9152,
"step": 259
},
{
"epoch": 0.16985138004246284,
"grad_norm": 2.435673475265503,
"learning_rate": 5e-05,
"loss": 4.1635,
"step": 260
},
{
"epoch": 0.1705046545810877,
"grad_norm": 2.5315775871276855,
"learning_rate": 4.9672753101632415e-05,
"loss": 4.0719,
"step": 261
},
{
"epoch": 0.17115792911971256,
"grad_norm": 2.5360610485076904,
"learning_rate": 4.934552022143279e-05,
"loss": 4.1543,
"step": 262
},
{
"epoch": 0.1718112036583374,
"grad_norm": 2.621554374694824,
"learning_rate": 4.901831537696859e-05,
"loss": 4.1552,
"step": 263
},
{
"epoch": 0.17246447819696228,
"grad_norm": 2.6578054428100586,
"learning_rate": 4.869115258460635e-05,
"loss": 3.8845,
"step": 264
},
{
"epoch": 0.17311775273558713,
"grad_norm": 2.6418116092681885,
"learning_rate": 4.83640458589112e-05,
"loss": 3.9069,
"step": 265
},
{
"epoch": 0.17377102727421198,
"grad_norm": 2.6990976333618164,
"learning_rate": 4.8037009212046586e-05,
"loss": 4.0657,
"step": 266
},
{
"epoch": 0.17442430181283686,
"grad_norm": 2.7651925086975098,
"learning_rate": 4.7710056653173976e-05,
"loss": 4.0418,
"step": 267
},
{
"epoch": 0.1750775763514617,
"grad_norm": 2.7356722354888916,
"learning_rate": 4.738320218785281e-05,
"loss": 4.0282,
"step": 268
},
{
"epoch": 0.17573085089008655,
"grad_norm": 2.977963924407959,
"learning_rate": 4.7056459817440544e-05,
"loss": 3.8089,
"step": 269
},
{
"epoch": 0.1763841254287114,
"grad_norm": 3.2176568508148193,
"learning_rate": 4.6729843538492847e-05,
"loss": 4.392,
"step": 270
},
{
"epoch": 0.17703739996733628,
"grad_norm": 3.1798744201660156,
"learning_rate": 4.640336734216403e-05,
"loss": 4.0641,
"step": 271
},
{
"epoch": 0.17769067450596113,
"grad_norm": 3.06921648979187,
"learning_rate": 4.607704521360776e-05,
"loss": 3.7511,
"step": 272
},
{
"epoch": 0.17834394904458598,
"grad_norm": 3.3475852012634277,
"learning_rate": 4.575089113137792e-05,
"loss": 3.8854,
"step": 273
},
{
"epoch": 0.17899722358321085,
"grad_norm": 3.8393049240112305,
"learning_rate": 4.542491906682989e-05,
"loss": 4.1801,
"step": 274
},
{
"epoch": 0.1796504981218357,
"grad_norm": 3.7368972301483154,
"learning_rate": 4.509914298352197e-05,
"loss": 3.4228,
"step": 275
},
{
"epoch": 0.18030377266046055,
"grad_norm": 3.9486243724823,
"learning_rate": 4.477357683661734e-05,
"loss": 3.4336,
"step": 276
},
{
"epoch": 0.18095704719908542,
"grad_norm": 4.318992614746094,
"learning_rate": 4.444823457228612e-05,
"loss": 3.1958,
"step": 277
},
{
"epoch": 0.18161032173771027,
"grad_norm": 4.600458145141602,
"learning_rate": 4.412313012710813e-05,
"loss": 3.3415,
"step": 278
},
{
"epoch": 0.18226359627633512,
"grad_norm": 5.333650588989258,
"learning_rate": 4.379827742747575e-05,
"loss": 3.744,
"step": 279
},
{
"epoch": 0.18291687081496,
"grad_norm": 5.62832498550415,
"learning_rate": 4.347369038899744e-05,
"loss": 3.9508,
"step": 280
},
{
"epoch": 0.18357014535358485,
"grad_norm": 5.859029293060303,
"learning_rate": 4.3149382915901606e-05,
"loss": 3.8297,
"step": 281
},
{
"epoch": 0.1842234198922097,
"grad_norm": 6.399336338043213,
"learning_rate": 4.282536890044104e-05,
"loss": 3.8776,
"step": 282
},
{
"epoch": 0.18487669443083457,
"grad_norm": 6.635499000549316,
"learning_rate": 4.250166222229774e-05,
"loss": 4.042,
"step": 283
},
{
"epoch": 0.18552996896945942,
"grad_norm": 6.871761322021484,
"learning_rate": 4.2178276747988446e-05,
"loss": 3.5912,
"step": 284
},
{
"epoch": 0.18618324350808427,
"grad_norm": 7.2283935546875,
"learning_rate": 4.185522633027057e-05,
"loss": 3.3263,
"step": 285
},
{
"epoch": 0.18683651804670912,
"grad_norm": 7.931128978729248,
"learning_rate": 4.153252480754877e-05,
"loss": 3.6491,
"step": 286
},
{
"epoch": 0.187489792585334,
"grad_norm": 8.871136665344238,
"learning_rate": 4.1210186003282275e-05,
"loss": 3.3801,
"step": 287
},
{
"epoch": 0.18814306712395884,
"grad_norm": 9.464795112609863,
"learning_rate": 4.088822372539263e-05,
"loss": 3.3301,
"step": 288
},
{
"epoch": 0.1887963416625837,
"grad_norm": 11.650923728942871,
"learning_rate": 4.0566651765672246e-05,
"loss": 3.6579,
"step": 289
},
{
"epoch": 0.18944961620120856,
"grad_norm": 12.516644477844238,
"learning_rate": 4.0245483899193595e-05,
"loss": 3.1808,
"step": 290
},
{
"epoch": 0.1901028907398334,
"grad_norm": 12.424504280090332,
"learning_rate": 3.992473388371915e-05,
"loss": 3.5156,
"step": 291
},
{
"epoch": 0.19075616527845826,
"grad_norm": 17.3675479888916,
"learning_rate": 3.960441545911204e-05,
"loss": 4.341,
"step": 292
},
{
"epoch": 0.19140943981708314,
"grad_norm": 13.690961837768555,
"learning_rate": 3.928454234674747e-05,
"loss": 2.6183,
"step": 293
},
{
"epoch": 0.192062714355708,
"grad_norm": 16.243967056274414,
"learning_rate": 3.896512824892495e-05,
"loss": 3.6009,
"step": 294
},
{
"epoch": 0.19271598889433283,
"grad_norm": 14.707989692687988,
"learning_rate": 3.864618684828134e-05,
"loss": 2.5393,
"step": 295
},
{
"epoch": 0.1933692634329577,
"grad_norm": 14.387151718139648,
"learning_rate": 3.832773180720475e-05,
"loss": 2.3307,
"step": 296
},
{
"epoch": 0.19402253797158256,
"grad_norm": 20.52593994140625,
"learning_rate": 3.800977676724919e-05,
"loss": 2.721,
"step": 297
},
{
"epoch": 0.1946758125102074,
"grad_norm": 17.45858383178711,
"learning_rate": 3.769233534855035e-05,
"loss": 3.0486,
"step": 298
},
{
"epoch": 0.19532908704883228,
"grad_norm": 19.98239517211914,
"learning_rate": 3.73754211492421e-05,
"loss": 2.4033,
"step": 299
},
{
"epoch": 0.19598236158745713,
"grad_norm": 26.76961898803711,
"learning_rate": 3.705904774487396e-05,
"loss": 3.0636,
"step": 300
},
{
"epoch": 0.19598236158745713,
"eval_loss": 0.916141927242279,
"eval_runtime": 193.7367,
"eval_samples_per_second": 13.307,
"eval_steps_per_second": 3.329,
"step": 300
},
{
"epoch": 0.19663563612608198,
"grad_norm": 2.2509307861328125,
"learning_rate": 3.6743228687829595e-05,
"loss": 4.1521,
"step": 301
},
{
"epoch": 0.19728891066470686,
"grad_norm": 2.2536087036132812,
"learning_rate": 3.642797750674629e-05,
"loss": 4.0593,
"step": 302
},
{
"epoch": 0.1979421852033317,
"grad_norm": 2.2652878761291504,
"learning_rate": 3.6113307705935396e-05,
"loss": 4.1726,
"step": 303
},
{
"epoch": 0.19859545974195655,
"grad_norm": 2.3236594200134277,
"learning_rate": 3.579923276480387e-05,
"loss": 4.1662,
"step": 304
},
{
"epoch": 0.1992487342805814,
"grad_norm": 2.248436689376831,
"learning_rate": 3.5485766137276894e-05,
"loss": 3.9428,
"step": 305
},
{
"epoch": 0.19990200881920628,
"grad_norm": 2.3227086067199707,
"learning_rate": 3.5172921251221455e-05,
"loss": 4.159,
"step": 306
},
{
"epoch": 0.20055528335783113,
"grad_norm": 2.3091423511505127,
"learning_rate": 3.486071150787128e-05,
"loss": 3.8457,
"step": 307
},
{
"epoch": 0.20120855789645598,
"grad_norm": 2.2768592834472656,
"learning_rate": 3.4549150281252636e-05,
"loss": 3.6557,
"step": 308
},
{
"epoch": 0.20186183243508085,
"grad_norm": 2.4113123416900635,
"learning_rate": 3.423825091761153e-05,
"loss": 3.9928,
"step": 309
},
{
"epoch": 0.2025151069737057,
"grad_norm": 2.337956666946411,
"learning_rate": 3.392802673484193e-05,
"loss": 3.8495,
"step": 310
},
{
"epoch": 0.20316838151233055,
"grad_norm": 2.3353707790374756,
"learning_rate": 3.361849102191533e-05,
"loss": 3.7771,
"step": 311
},
{
"epoch": 0.20382165605095542,
"grad_norm": 2.4087674617767334,
"learning_rate": 3.330965703831146e-05,
"loss": 4.1044,
"step": 312
},
{
"epoch": 0.20447493058958027,
"grad_norm": 2.4701943397521973,
"learning_rate": 3.300153801345028e-05,
"loss": 3.5573,
"step": 313
},
{
"epoch": 0.20512820512820512,
"grad_norm": 2.612384557723999,
"learning_rate": 3.2694147146125345e-05,
"loss": 3.8407,
"step": 314
},
{
"epoch": 0.20578147966683,
"grad_norm": 2.717681884765625,
"learning_rate": 3.2387497603938326e-05,
"loss": 4.164,
"step": 315
},
{
"epoch": 0.20643475420545485,
"grad_norm": 2.7894811630249023,
"learning_rate": 3.2081602522734986e-05,
"loss": 3.6986,
"step": 316
},
{
"epoch": 0.2070880287440797,
"grad_norm": 2.887406826019287,
"learning_rate": 3.177647500604252e-05,
"loss": 4.1417,
"step": 317
},
{
"epoch": 0.20774130328270457,
"grad_norm": 2.995980739593506,
"learning_rate": 3.147212812450819e-05,
"loss": 4.0842,
"step": 318
},
{
"epoch": 0.20839457782132942,
"grad_norm": 3.2103445529937744,
"learning_rate": 3.116857491533947e-05,
"loss": 4.2465,
"step": 319
},
{
"epoch": 0.20904785235995427,
"grad_norm": 3.404813766479492,
"learning_rate": 3.086582838174551e-05,
"loss": 3.4803,
"step": 320
},
{
"epoch": 0.20970112689857912,
"grad_norm": 3.4752864837646484,
"learning_rate": 3.056390149238022e-05,
"loss": 4.2533,
"step": 321
},
{
"epoch": 0.210354401437204,
"grad_norm": 3.54484486579895,
"learning_rate": 3.0262807180786647e-05,
"loss": 3.9656,
"step": 322
},
{
"epoch": 0.21100767597582884,
"grad_norm": 3.658374547958374,
"learning_rate": 2.996255834484296e-05,
"loss": 3.5864,
"step": 323
},
{
"epoch": 0.2116609505144537,
"grad_norm": 3.7134335041046143,
"learning_rate": 2.9663167846209998e-05,
"loss": 3.587,
"step": 324
},
{
"epoch": 0.21231422505307856,
"grad_norm": 3.8654329776763916,
"learning_rate": 2.936464850978027e-05,
"loss": 3.8962,
"step": 325
},
{
"epoch": 0.2129674995917034,
"grad_norm": 4.069035530090332,
"learning_rate": 2.9067013123128613e-05,
"loss": 3.6819,
"step": 326
},
{
"epoch": 0.21362077413032826,
"grad_norm": 4.072150230407715,
"learning_rate": 2.8770274435964355e-05,
"loss": 3.3375,
"step": 327
},
{
"epoch": 0.21427404866895314,
"grad_norm": 4.656280040740967,
"learning_rate": 2.8474445159585235e-05,
"loss": 3.5886,
"step": 328
},
{
"epoch": 0.21492732320757799,
"grad_norm": 4.830178737640381,
"learning_rate": 2.8179537966332887e-05,
"loss": 3.2974,
"step": 329
},
{
"epoch": 0.21558059774620283,
"grad_norm": 5.456755638122559,
"learning_rate": 2.7885565489049946e-05,
"loss": 3.8774,
"step": 330
},
{
"epoch": 0.2162338722848277,
"grad_norm": 5.759149551391602,
"learning_rate": 2.759254032053888e-05,
"loss": 3.6094,
"step": 331
},
{
"epoch": 0.21688714682345256,
"grad_norm": 5.614565372467041,
"learning_rate": 2.7300475013022663e-05,
"loss": 3.608,
"step": 332
},
{
"epoch": 0.2175404213620774,
"grad_norm": 6.076308250427246,
"learning_rate": 2.700938207760701e-05,
"loss": 3.5142,
"step": 333
},
{
"epoch": 0.21819369590070228,
"grad_norm": 6.575270175933838,
"learning_rate": 2.671927398374443e-05,
"loss": 3.3849,
"step": 334
},
{
"epoch": 0.21884697043932713,
"grad_norm": 7.774899482727051,
"learning_rate": 2.6430163158700115e-05,
"loss": 3.4158,
"step": 335
},
{
"epoch": 0.21950024497795198,
"grad_norm": 7.88145637512207,
"learning_rate": 2.6142061987019577e-05,
"loss": 2.9199,
"step": 336
},
{
"epoch": 0.22015351951657683,
"grad_norm": 10.132671356201172,
"learning_rate": 2.5854982809998153e-05,
"loss": 4.0171,
"step": 337
},
{
"epoch": 0.2208067940552017,
"grad_norm": 11.227150917053223,
"learning_rate": 2.556893792515227e-05,
"loss": 2.9164,
"step": 338
},
{
"epoch": 0.22146006859382655,
"grad_norm": 10.218441009521484,
"learning_rate": 2.5283939585692783e-05,
"loss": 2.9718,
"step": 339
},
{
"epoch": 0.2221133431324514,
"grad_norm": 11.796672821044922,
"learning_rate": 2.500000000000001e-05,
"loss": 3.0505,
"step": 340
},
{
"epoch": 0.22276661767107628,
"grad_norm": 13.644339561462402,
"learning_rate": 2.471713133110078e-05,
"loss": 3.9324,
"step": 341
},
{
"epoch": 0.22341989220970113,
"grad_norm": 13.900707244873047,
"learning_rate": 2.4435345696147403e-05,
"loss": 3.3911,
"step": 342
},
{
"epoch": 0.22407316674832597,
"grad_norm": 18.582195281982422,
"learning_rate": 2.4154655165898627e-05,
"loss": 2.6938,
"step": 343
},
{
"epoch": 0.22472644128695085,
"grad_norm": 14.891613006591797,
"learning_rate": 2.3875071764202563e-05,
"loss": 2.8819,
"step": 344
},
{
"epoch": 0.2253797158255757,
"grad_norm": 17.06895637512207,
"learning_rate": 2.3596607467481603e-05,
"loss": 3.0545,
"step": 345
},
{
"epoch": 0.22603299036420055,
"grad_norm": 18.432598114013672,
"learning_rate": 2.3319274204219428e-05,
"loss": 3.5887,
"step": 346
},
{
"epoch": 0.22668626490282542,
"grad_norm": 22.221372604370117,
"learning_rate": 2.3043083854449988e-05,
"loss": 2.5782,
"step": 347
},
{
"epoch": 0.22733953944145027,
"grad_norm": 23.88602638244629,
"learning_rate": 2.2768048249248648e-05,
"loss": 3.1699,
"step": 348
},
{
"epoch": 0.22799281398007512,
"grad_norm": 17.298885345458984,
"learning_rate": 2.2494179170225333e-05,
"loss": 2.5779,
"step": 349
},
{
"epoch": 0.2286460885187,
"grad_norm": 34.774845123291016,
"learning_rate": 2.2221488349019903e-05,
"loss": 4.6921,
"step": 350
},
{
"epoch": 0.22929936305732485,
"grad_norm": 2.0609993934631348,
"learning_rate": 2.194998746679952e-05,
"loss": 4.05,
"step": 351
},
{
"epoch": 0.2299526375959497,
"grad_norm": 2.18334698677063,
"learning_rate": 2.167968815375837e-05,
"loss": 4.2886,
"step": 352
},
{
"epoch": 0.23060591213457454,
"grad_norm": 2.1787545680999756,
"learning_rate": 2.1410601988619394e-05,
"loss": 4.3155,
"step": 353
},
{
"epoch": 0.23125918667319942,
"grad_norm": 2.1137516498565674,
"learning_rate": 2.1142740498138324e-05,
"loss": 3.7972,
"step": 354
},
{
"epoch": 0.23191246121182427,
"grad_norm": 2.1815781593322754,
"learning_rate": 2.08761151566099e-05,
"loss": 3.9406,
"step": 355
},
{
"epoch": 0.23256573575044912,
"grad_norm": 2.2473208904266357,
"learning_rate": 2.061073738537635e-05,
"loss": 3.94,
"step": 356
},
{
"epoch": 0.233219010289074,
"grad_norm": 2.286938428878784,
"learning_rate": 2.034661855233815e-05,
"loss": 3.8518,
"step": 357
},
{
"epoch": 0.23387228482769884,
"grad_norm": 2.282809257507324,
"learning_rate": 2.008376997146705e-05,
"loss": 4.0518,
"step": 358
},
{
"epoch": 0.2345255593663237,
"grad_norm": 2.15885066986084,
"learning_rate": 1.982220290232143e-05,
"loss": 3.7982,
"step": 359
},
{
"epoch": 0.23517883390494856,
"grad_norm": 2.232218027114868,
"learning_rate": 1.9561928549563968e-05,
"loss": 3.9533,
"step": 360
},
{
"epoch": 0.2358321084435734,
"grad_norm": 2.28951358795166,
"learning_rate": 1.9302958062481673e-05,
"loss": 3.907,
"step": 361
},
{
"epoch": 0.23648538298219826,
"grad_norm": 2.3329083919525146,
"learning_rate": 1.9045302534508297e-05,
"loss": 4.0736,
"step": 362
},
{
"epoch": 0.23713865752082314,
"grad_norm": 2.4136269092559814,
"learning_rate": 1.8788973002749112e-05,
"loss": 3.8261,
"step": 363
},
{
"epoch": 0.23779193205944799,
"grad_norm": 2.6540565490722656,
"learning_rate": 1.8533980447508137e-05,
"loss": 4.0891,
"step": 364
},
{
"epoch": 0.23844520659807283,
"grad_norm": 2.4953386783599854,
"learning_rate": 1.8280335791817733e-05,
"loss": 3.9206,
"step": 365
},
{
"epoch": 0.2390984811366977,
"grad_norm": 2.7900853157043457,
"learning_rate": 1.8028049900970767e-05,
"loss": 3.7279,
"step": 366
},
{
"epoch": 0.23975175567532256,
"grad_norm": 2.7315573692321777,
"learning_rate": 1.777713358205514e-05,
"loss": 4.0381,
"step": 367
},
{
"epoch": 0.2404050302139474,
"grad_norm": 2.8006503582000732,
"learning_rate": 1.7527597583490822e-05,
"loss": 3.7893,
"step": 368
},
{
"epoch": 0.24105830475257226,
"grad_norm": 3.0075833797454834,
"learning_rate": 1.7279452594569483e-05,
"loss": 3.9685,
"step": 369
},
{
"epoch": 0.24171157929119713,
"grad_norm": 3.305966854095459,
"learning_rate": 1.703270924499656e-05,
"loss": 4.3688,
"step": 370
},
{
"epoch": 0.24236485382982198,
"grad_norm": 3.364837646484375,
"learning_rate": 1.678737810443593e-05,
"loss": 4.0271,
"step": 371
},
{
"epoch": 0.24301812836844683,
"grad_norm": 3.410656213760376,
"learning_rate": 1.6543469682057106e-05,
"loss": 3.913,
"step": 372
},
{
"epoch": 0.2436714029070717,
"grad_norm": 3.7187421321868896,
"learning_rate": 1.6300994426085103e-05,
"loss": 3.6959,
"step": 373
},
{
"epoch": 0.24432467744569655,
"grad_norm": 3.892642021179199,
"learning_rate": 1.605996272335291e-05,
"loss": 3.8044,
"step": 374
},
{
"epoch": 0.2449779519843214,
"grad_norm": 3.7945556640625,
"learning_rate": 1.5820384898856434e-05,
"loss": 3.2933,
"step": 375
},
{
"epoch": 0.24563122652294628,
"grad_norm": 4.004775047302246,
"learning_rate": 1.5582271215312294e-05,
"loss": 3.3998,
"step": 376
},
{
"epoch": 0.24628450106157113,
"grad_norm": 4.55605936050415,
"learning_rate": 1.5345631872718214e-05,
"loss": 4.0181,
"step": 377
},
{
"epoch": 0.24693777560019597,
"grad_norm": 4.659095287322998,
"learning_rate": 1.5110477007916001e-05,
"loss": 3.6973,
"step": 378
},
{
"epoch": 0.24759105013882085,
"grad_norm": 4.90739107131958,
"learning_rate": 1.4876816694157419e-05,
"loss": 3.4013,
"step": 379
},
{
"epoch": 0.2482443246774457,
"grad_norm": 5.735424995422363,
"learning_rate": 1.4644660940672627e-05,
"loss": 3.6558,
"step": 380
},
{
"epoch": 0.24889759921607055,
"grad_norm": 5.81011962890625,
"learning_rate": 1.4414019692241437e-05,
"loss": 3.5117,
"step": 381
},
{
"epoch": 0.24955087375469542,
"grad_norm": 5.775092601776123,
"learning_rate": 1.4184902828767287e-05,
"loss": 3.5385,
"step": 382
},
{
"epoch": 0.25020414829332027,
"grad_norm": 6.017894268035889,
"learning_rate": 1.3957320164854059e-05,
"loss": 3.2936,
"step": 383
},
{
"epoch": 0.2508574228319451,
"grad_norm": 6.8194708824157715,
"learning_rate": 1.373128144938563e-05,
"loss": 3.1461,
"step": 384
},
{
"epoch": 0.25151069737056997,
"grad_norm": 7.288778781890869,
"learning_rate": 1.3506796365108232e-05,
"loss": 3.711,
"step": 385
},
{
"epoch": 0.2521639719091948,
"grad_norm": 7.597186088562012,
"learning_rate": 1.3283874528215733e-05,
"loss": 3.68,
"step": 386
},
{
"epoch": 0.2528172464478197,
"grad_norm": 9.038799285888672,
"learning_rate": 1.3062525487937699e-05,
"loss": 3.1288,
"step": 387
},
{
"epoch": 0.25347052098644457,
"grad_norm": 9.388714790344238,
"learning_rate": 1.2842758726130283e-05,
"loss": 3.068,
"step": 388
},
{
"epoch": 0.2541237955250694,
"grad_norm": 11.19589614868164,
"learning_rate": 1.2624583656870154e-05,
"loss": 3.2682,
"step": 389
},
{
"epoch": 0.25477707006369427,
"grad_norm": 10.183329582214355,
"learning_rate": 1.2408009626051137e-05,
"loss": 2.604,
"step": 390
},
{
"epoch": 0.2554303446023191,
"grad_norm": 12.288995742797852,
"learning_rate": 1.2193045910983863e-05,
"loss": 3.0615,
"step": 391
},
{
"epoch": 0.25608361914094396,
"grad_norm": 10.170268058776855,
"learning_rate": 1.1979701719998453e-05,
"loss": 3.0368,
"step": 392
},
{
"epoch": 0.2567368936795688,
"grad_norm": 12.902504920959473,
"learning_rate": 1.1767986192049984e-05,
"loss": 2.6595,
"step": 393
},
{
"epoch": 0.2573901682181937,
"grad_norm": 14.612078666687012,
"learning_rate": 1.1557908396327028e-05,
"loss": 3.1923,
"step": 394
},
{
"epoch": 0.25804344275681856,
"grad_norm": 14.616671562194824,
"learning_rate": 1.134947733186315e-05,
"loss": 3.3654,
"step": 395
},
{
"epoch": 0.2586967172954434,
"grad_norm": 15.300880432128906,
"learning_rate": 1.1142701927151456e-05,
"loss": 2.3551,
"step": 396
},
{
"epoch": 0.25934999183406826,
"grad_norm": 16.621246337890625,
"learning_rate": 1.0937591039762085e-05,
"loss": 2.6044,
"step": 397
},
{
"epoch": 0.2600032663726931,
"grad_norm": 20.40192985534668,
"learning_rate": 1.0734153455962765e-05,
"loss": 2.9109,
"step": 398
},
{
"epoch": 0.26065654091131796,
"grad_norm": 21.176897048950195,
"learning_rate": 1.0532397890342505e-05,
"loss": 2.6236,
"step": 399
},
{
"epoch": 0.26130981544994286,
"grad_norm": 36.7708740234375,
"learning_rate": 1.0332332985438248e-05,
"loss": 5.0028,
"step": 400
},
{
"epoch": 0.26130981544994286,
"eval_loss": 0.8865355849266052,
"eval_runtime": 193.5082,
"eval_samples_per_second": 13.322,
"eval_steps_per_second": 3.333,
"step": 400
},
{
"epoch": 0.2619630899885677,
"grad_norm": 2.0284175872802734,
"learning_rate": 1.013396731136465e-05,
"loss": 3.2076,
"step": 401
},
{
"epoch": 0.26261636452719256,
"grad_norm": 2.200660467147827,
"learning_rate": 9.937309365446973e-06,
"loss": 4.3066,
"step": 402
},
{
"epoch": 0.2632696390658174,
"grad_norm": 2.2217345237731934,
"learning_rate": 9.742367571857091e-06,
"loss": 4.3663,
"step": 403
},
{
"epoch": 0.26392291360444226,
"grad_norm": 2.1234660148620605,
"learning_rate": 9.549150281252633e-06,
"loss": 4.0273,
"step": 404
},
{
"epoch": 0.2645761881430671,
"grad_norm": 2.140134572982788,
"learning_rate": 9.357665770419244e-06,
"loss": 4.0097,
"step": 405
},
{
"epoch": 0.265229462681692,
"grad_norm": 2.241824150085449,
"learning_rate": 9.167922241916055e-06,
"loss": 3.8709,
"step": 406
},
{
"epoch": 0.26588273722031686,
"grad_norm": 2.207245111465454,
"learning_rate": 8.97992782372432e-06,
"loss": 3.7465,
"step": 407
},
{
"epoch": 0.2665360117589417,
"grad_norm": 2.2445242404937744,
"learning_rate": 8.793690568899216e-06,
"loss": 3.9702,
"step": 408
},
{
"epoch": 0.26718928629756655,
"grad_norm": 2.209073781967163,
"learning_rate": 8.609218455224893e-06,
"loss": 3.6522,
"step": 409
},
{
"epoch": 0.2678425608361914,
"grad_norm": 2.3187787532806396,
"learning_rate": 8.426519384872733e-06,
"loss": 3.9904,
"step": 410
},
{
"epoch": 0.26849583537481625,
"grad_norm": 2.3060576915740967,
"learning_rate": 8.245601184062852e-06,
"loss": 3.645,
"step": 411
},
{
"epoch": 0.2691491099134411,
"grad_norm": 2.4985568523406982,
"learning_rate": 8.066471602728803e-06,
"loss": 3.94,
"step": 412
},
{
"epoch": 0.269802384452066,
"grad_norm": 2.4821956157684326,
"learning_rate": 7.889138314185678e-06,
"loss": 3.8423,
"step": 413
},
{
"epoch": 0.27045565899069085,
"grad_norm": 2.536126136779785,
"learning_rate": 7.71360891480134e-06,
"loss": 3.8344,
"step": 414
},
{
"epoch": 0.2711089335293157,
"grad_norm": 2.5374646186828613,
"learning_rate": 7.539890923671062e-06,
"loss": 3.9225,
"step": 415
},
{
"epoch": 0.27176220806794055,
"grad_norm": 2.6912612915039062,
"learning_rate": 7.367991782295391e-06,
"loss": 3.8863,
"step": 416
},
{
"epoch": 0.2724154826065654,
"grad_norm": 2.7828660011291504,
"learning_rate": 7.197918854261432e-06,
"loss": 3.8709,
"step": 417
},
{
"epoch": 0.27306875714519024,
"grad_norm": 2.831202507019043,
"learning_rate": 7.029679424927365e-06,
"loss": 3.7255,
"step": 418
},
{
"epoch": 0.27372203168381515,
"grad_norm": 3.060506582260132,
"learning_rate": 6.863280701110408e-06,
"loss": 4.0642,
"step": 419
},
{
"epoch": 0.27437530622244,
"grad_norm": 3.0359199047088623,
"learning_rate": 6.698729810778065e-06,
"loss": 3.7836,
"step": 420
},
{
"epoch": 0.27502858076106484,
"grad_norm": 3.1587436199188232,
"learning_rate": 6.536033802742813e-06,
"loss": 4.0368,
"step": 421
},
{
"epoch": 0.2756818552996897,
"grad_norm": 3.3620858192443848,
"learning_rate": 6.375199646360142e-06,
"loss": 3.7547,
"step": 422
},
{
"epoch": 0.27633512983831454,
"grad_norm": 3.528315782546997,
"learning_rate": 6.216234231230012e-06,
"loss": 3.7245,
"step": 423
},
{
"epoch": 0.2769884043769394,
"grad_norm": 3.4212453365325928,
"learning_rate": 6.059144366901736e-06,
"loss": 3.3729,
"step": 424
},
{
"epoch": 0.27764167891556424,
"grad_norm": 3.6946189403533936,
"learning_rate": 5.903936782582253e-06,
"loss": 3.5213,
"step": 425
},
{
"epoch": 0.27829495345418914,
"grad_norm": 3.8112924098968506,
"learning_rate": 5.750618126847912e-06,
"loss": 3.5261,
"step": 426
},
{
"epoch": 0.278948227992814,
"grad_norm": 4.669442176818848,
"learning_rate": 5.599194967359639e-06,
"loss": 3.9578,
"step": 427
},
{
"epoch": 0.27960150253143884,
"grad_norm": 4.836587905883789,
"learning_rate": 5.449673790581611e-06,
"loss": 4.1958,
"step": 428
},
{
"epoch": 0.2802547770700637,
"grad_norm": 4.8962507247924805,
"learning_rate": 5.302061001503394e-06,
"loss": 3.6376,
"step": 429
},
{
"epoch": 0.28090805160868854,
"grad_norm": 5.25094747543335,
"learning_rate": 5.156362923365588e-06,
"loss": 3.6825,
"step": 430
},
{
"epoch": 0.2815613261473134,
"grad_norm": 5.470578193664551,
"learning_rate": 5.012585797388936e-06,
"loss": 3.3548,
"step": 431
},
{
"epoch": 0.2822146006859383,
"grad_norm": 6.3519415855407715,
"learning_rate": 4.87073578250698e-06,
"loss": 3.4006,
"step": 432
},
{
"epoch": 0.28286787522456314,
"grad_norm": 6.419662952423096,
"learning_rate": 4.730818955102234e-06,
"loss": 3.493,
"step": 433
},
{
"epoch": 0.283521149763188,
"grad_norm": 6.477165699005127,
"learning_rate": 4.592841308745932e-06,
"loss": 3.3161,
"step": 434
},
{
"epoch": 0.28417442430181283,
"grad_norm": 6.963860034942627,
"learning_rate": 4.456808753941205e-06,
"loss": 3.4243,
"step": 435
},
{
"epoch": 0.2848276988404377,
"grad_norm": 7.6668548583984375,
"learning_rate": 4.322727117869951e-06,
"loss": 2.5756,
"step": 436
},
{
"epoch": 0.28548097337906253,
"grad_norm": 8.136970520019531,
"learning_rate": 4.190602144143207e-06,
"loss": 3.3489,
"step": 437
},
{
"epoch": 0.28613424791768743,
"grad_norm": 10.233713150024414,
"learning_rate": 4.06043949255509e-06,
"loss": 3.6143,
"step": 438
},
{
"epoch": 0.2867875224563123,
"grad_norm": 10.85151481628418,
"learning_rate": 3.932244738840379e-06,
"loss": 3.1783,
"step": 439
},
{
"epoch": 0.28744079699493713,
"grad_norm": 9.790265083312988,
"learning_rate": 3.8060233744356633e-06,
"loss": 2.58,
"step": 440
},
{
"epoch": 0.288094071533562,
"grad_norm": 10.893715858459473,
"learning_rate": 3.681780806244095e-06,
"loss": 3.0611,
"step": 441
},
{
"epoch": 0.28874734607218683,
"grad_norm": 12.682134628295898,
"learning_rate": 3.5595223564037884e-06,
"loss": 2.8948,
"step": 442
},
{
"epoch": 0.2894006206108117,
"grad_norm": 13.385756492614746,
"learning_rate": 3.4392532620598216e-06,
"loss": 2.6341,
"step": 443
},
{
"epoch": 0.2900538951494365,
"grad_norm": 16.10133934020996,
"learning_rate": 3.3209786751399187e-06,
"loss": 3.5305,
"step": 444
},
{
"epoch": 0.29070716968806143,
"grad_norm": 12.668194770812988,
"learning_rate": 3.2047036621337236e-06,
"loss": 2.3386,
"step": 445
},
{
"epoch": 0.2913604442266863,
"grad_norm": 14.395100593566895,
"learning_rate": 3.0904332038757977e-06,
"loss": 2.0309,
"step": 446
},
{
"epoch": 0.2920137187653111,
"grad_norm": 17.13104820251465,
"learning_rate": 2.978172195332263e-06,
"loss": 2.4272,
"step": 447
},
{
"epoch": 0.292666993303936,
"grad_norm": 17.783287048339844,
"learning_rate": 2.8679254453910785e-06,
"loss": 2.7686,
"step": 448
},
{
"epoch": 0.2933202678425608,
"grad_norm": 22.626319885253906,
"learning_rate": 2.759697676656098e-06,
"loss": 2.9195,
"step": 449
},
{
"epoch": 0.29397354238118567,
"grad_norm": 35.02473831176758,
"learning_rate": 2.653493525244721e-06,
"loss": 4.8236,
"step": 450
},
{
"epoch": 0.2946268169198106,
"grad_norm": 2.0250165462493896,
"learning_rate": 2.549317540589308e-06,
"loss": 4.0218,
"step": 451
},
{
"epoch": 0.2952800914584354,
"grad_norm": 1.948515772819519,
"learning_rate": 2.4471741852423237e-06,
"loss": 3.9039,
"step": 452
},
{
"epoch": 0.29593336599706027,
"grad_norm": 2.03169322013855,
"learning_rate": 2.3470678346851518e-06,
"loss": 4.3663,
"step": 453
},
{
"epoch": 0.2965866405356851,
"grad_norm": 2.1053688526153564,
"learning_rate": 2.2490027771406687e-06,
"loss": 4.0127,
"step": 454
},
{
"epoch": 0.29723991507430997,
"grad_norm": 2.2202868461608887,
"learning_rate": 2.152983213389559e-06,
"loss": 4.0635,
"step": 455
},
{
"epoch": 0.2978931896129348,
"grad_norm": 2.141794443130493,
"learning_rate": 2.0590132565903476e-06,
"loss": 3.9512,
"step": 456
},
{
"epoch": 0.2985464641515597,
"grad_norm": 2.1424460411071777,
"learning_rate": 1.9670969321032407e-06,
"loss": 3.928,
"step": 457
},
{
"epoch": 0.29919973869018457,
"grad_norm": 2.2259063720703125,
"learning_rate": 1.8772381773176417e-06,
"loss": 3.9537,
"step": 458
},
{
"epoch": 0.2998530132288094,
"grad_norm": 2.1677327156066895,
"learning_rate": 1.7894408414835362e-06,
"loss": 3.7739,
"step": 459
},
{
"epoch": 0.30050628776743427,
"grad_norm": 2.260230541229248,
"learning_rate": 1.70370868554659e-06,
"loss": 4.0705,
"step": 460
},
{
"epoch": 0.3011595623060591,
"grad_norm": 2.394634246826172,
"learning_rate": 1.620045381987012e-06,
"loss": 4.1479,
"step": 461
},
{
"epoch": 0.30181283684468396,
"grad_norm": 2.3297996520996094,
"learning_rate": 1.5384545146622852e-06,
"loss": 3.6953,
"step": 462
},
{
"epoch": 0.3024661113833088,
"grad_norm": 2.3964977264404297,
"learning_rate": 1.4589395786535953e-06,
"loss": 3.9611,
"step": 463
},
{
"epoch": 0.3031193859219337,
"grad_norm": 2.533637046813965,
"learning_rate": 1.3815039801161721e-06,
"loss": 3.9468,
"step": 464
},
{
"epoch": 0.30377266046055856,
"grad_norm": 2.529632329940796,
"learning_rate": 1.3061510361333185e-06,
"loss": 3.6358,
"step": 465
},
{
"epoch": 0.3044259349991834,
"grad_norm": 2.569408655166626,
"learning_rate": 1.232883974574367e-06,
"loss": 3.5356,
"step": 466
},
{
"epoch": 0.30507920953780826,
"grad_norm": 2.727712869644165,
"learning_rate": 1.1617059339563807e-06,
"loss": 3.7934,
"step": 467
},
{
"epoch": 0.3057324840764331,
"grad_norm": 2.8594043254852295,
"learning_rate": 1.0926199633097157e-06,
"loss": 3.933,
"step": 468
},
{
"epoch": 0.30638575861505796,
"grad_norm": 2.939509868621826,
"learning_rate": 1.0256290220474307e-06,
"loss": 3.5852,
"step": 469
},
{
"epoch": 0.30703903315368286,
"grad_norm": 3.0464086532592773,
"learning_rate": 9.607359798384785e-07,
"loss": 3.7145,
"step": 470
},
{
"epoch": 0.3076923076923077,
"grad_norm": 3.3523106575012207,
"learning_rate": 8.979436164848088e-07,
"loss": 3.8206,
"step": 471
},
{
"epoch": 0.30834558223093256,
"grad_norm": 3.190808057785034,
"learning_rate": 8.372546218022747e-07,
"loss": 3.8822,
"step": 472
},
{
"epoch": 0.3089988567695574,
"grad_norm": 3.408557891845703,
"learning_rate": 7.786715955054203e-07,
"loss": 3.5316,
"step": 473
},
{
"epoch": 0.30965213130818225,
"grad_norm": 3.75345778465271,
"learning_rate": 7.221970470961125e-07,
"loss": 3.8423,
"step": 474
},
{
"epoch": 0.3103054058468071,
"grad_norm": 4.055991172790527,
"learning_rate": 6.678333957560512e-07,
"loss": 3.6669,
"step": 475
},
{
"epoch": 0.31095868038543195,
"grad_norm": 4.102751731872559,
"learning_rate": 6.15582970243117e-07,
"loss": 3.4249,
"step": 476
},
{
"epoch": 0.31161195492405686,
"grad_norm": 4.2560014724731445,
"learning_rate": 5.654480087916303e-07,
"loss": 3.4731,
"step": 477
},
{
"epoch": 0.3122652294626817,
"grad_norm": 4.73115873336792,
"learning_rate": 5.174306590164879e-07,
"loss": 3.3286,
"step": 478
},
{
"epoch": 0.31291850400130655,
"grad_norm": 5.362375259399414,
"learning_rate": 4.715329778211375e-07,
"loss": 3.2443,
"step": 479
},
{
"epoch": 0.3135717785399314,
"grad_norm": 5.289713382720947,
"learning_rate": 4.277569313094809e-07,
"loss": 3.2602,
"step": 480
},
{
"epoch": 0.31422505307855625,
"grad_norm": 5.72009801864624,
"learning_rate": 3.8610439470164737e-07,
"loss": 3.1795,
"step": 481
},
{
"epoch": 0.3148783276171811,
"grad_norm": 6.249419212341309,
"learning_rate": 3.465771522536854e-07,
"loss": 3.2568,
"step": 482
},
{
"epoch": 0.315531602155806,
"grad_norm": 6.340628147125244,
"learning_rate": 3.09176897181096e-07,
"loss": 3.7829,
"step": 483
},
{
"epoch": 0.31618487669443085,
"grad_norm": 6.803210735321045,
"learning_rate": 2.7390523158633554e-07,
"loss": 3.2678,
"step": 484
},
{
"epoch": 0.3168381512330557,
"grad_norm": 7.402984619140625,
"learning_rate": 2.407636663901591e-07,
"loss": 3.2089,
"step": 485
},
{
"epoch": 0.31749142577168055,
"grad_norm": 8.722670555114746,
"learning_rate": 2.0975362126691712e-07,
"loss": 3.4273,
"step": 486
},
{
"epoch": 0.3181447003103054,
"grad_norm": 8.667076110839844,
"learning_rate": 1.8087642458373134e-07,
"loss": 2.9022,
"step": 487
},
{
"epoch": 0.31879797484893024,
"grad_norm": 9.627767562866211,
"learning_rate": 1.5413331334360182e-07,
"loss": 3.0261,
"step": 488
},
{
"epoch": 0.31945124938755515,
"grad_norm": 10.799980163574219,
"learning_rate": 1.2952543313240472e-07,
"loss": 2.731,
"step": 489
},
{
"epoch": 0.32010452392618,
"grad_norm": 12.3025484085083,
"learning_rate": 1.0705383806982606e-07,
"loss": 3.1692,
"step": 490
},
{
"epoch": 0.32075779846480484,
"grad_norm": 11.816473007202148,
"learning_rate": 8.671949076420882e-08,
"loss": 2.1674,
"step": 491
},
{
"epoch": 0.3214110730034297,
"grad_norm": 13.557419776916504,
"learning_rate": 6.852326227130834e-08,
"loss": 2.846,
"step": 492
},
{
"epoch": 0.32206434754205454,
"grad_norm": 11.210492134094238,
"learning_rate": 5.246593205699424e-08,
"loss": 2.5836,
"step": 493
},
{
"epoch": 0.3227176220806794,
"grad_norm": 13.040658950805664,
"learning_rate": 3.8548187963854956e-08,
"loss": 3.2673,
"step": 494
},
{
"epoch": 0.32337089661930424,
"grad_norm": 15.785333633422852,
"learning_rate": 2.6770626181715773e-08,
"loss": 2.518,
"step": 495
},
{
"epoch": 0.32402417115792914,
"grad_norm": 15.67574405670166,
"learning_rate": 1.7133751222137007e-08,
"loss": 2.387,
"step": 496
},
{
"epoch": 0.324677445696554,
"grad_norm": 15.306534767150879,
"learning_rate": 9.637975896759077e-09,
"loss": 2.3683,
"step": 497
},
{
"epoch": 0.32533072023517884,
"grad_norm": 20.011672973632812,
"learning_rate": 4.2836212996499865e-09,
"loss": 3.7281,
"step": 498
},
{
"epoch": 0.3259839947738037,
"grad_norm": 21.42987823486328,
"learning_rate": 1.0709167935385455e-09,
"loss": 2.4191,
"step": 499
},
{
"epoch": 0.32663726931242854,
"grad_norm": 32.15013885498047,
"learning_rate": 0.0,
"loss": 3.3215,
"step": 500
},
{
"epoch": 0.32663726931242854,
"eval_loss": 0.8778727054595947,
"eval_runtime": 193.3536,
"eval_samples_per_second": 13.333,
"eval_steps_per_second": 3.336,
"step": 500
}
],
"logging_steps": 1,
"max_steps": 500,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.243979410448056e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}