ErrorAI's picture
Training in progress, step 348, checkpoint
5cc7c78 verified
raw
history blame
61.6 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.06824198450828513,
"eval_steps": 348,
"global_step": 348,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00019609765663300324,
"grad_norm": 20.33372688293457,
"learning_rate": 2e-05,
"loss": 3.0843,
"step": 1
},
{
"epoch": 0.00019609765663300324,
"eval_loss": 1.1017773151397705,
"eval_runtime": 79.9135,
"eval_samples_per_second": 26.879,
"eval_steps_per_second": 13.44,
"step": 1
},
{
"epoch": 0.0003921953132660065,
"grad_norm": 19.32895278930664,
"learning_rate": 4e-05,
"loss": 3.2221,
"step": 2
},
{
"epoch": 0.0005882929698990097,
"grad_norm": 18.6882266998291,
"learning_rate": 6e-05,
"loss": 3.8951,
"step": 3
},
{
"epoch": 0.000784390626532013,
"grad_norm": 43.008060455322266,
"learning_rate": 8e-05,
"loss": 5.167,
"step": 4
},
{
"epoch": 0.0009804882831650162,
"grad_norm": 21.642993927001953,
"learning_rate": 0.0001,
"loss": 3.1304,
"step": 5
},
{
"epoch": 0.0011765859397980193,
"grad_norm": 29.79266929626465,
"learning_rate": 0.00012,
"loss": 4.5153,
"step": 6
},
{
"epoch": 0.0013726835964310226,
"grad_norm": 25.503681182861328,
"learning_rate": 0.00014,
"loss": 3.8083,
"step": 7
},
{
"epoch": 0.001568781253064026,
"grad_norm": 32.35524368286133,
"learning_rate": 0.00016,
"loss": 4.253,
"step": 8
},
{
"epoch": 0.0017648789096970292,
"grad_norm": 21.053390502929688,
"learning_rate": 0.00018,
"loss": 3.3757,
"step": 9
},
{
"epoch": 0.0019609765663300325,
"grad_norm": 25.7067928314209,
"learning_rate": 0.0002,
"loss": 3.2484,
"step": 10
},
{
"epoch": 0.0021570742229630358,
"grad_norm": 22.57227897644043,
"learning_rate": 0.00019999974049780868,
"loss": 2.8378,
"step": 11
},
{
"epoch": 0.0023531718795960386,
"grad_norm": 19.06597900390625,
"learning_rate": 0.00019999896199258152,
"loss": 3.231,
"step": 12
},
{
"epoch": 0.002549269536229042,
"grad_norm": 17.590620040893555,
"learning_rate": 0.000199997664488359,
"loss": 2.2391,
"step": 13
},
{
"epoch": 0.002745367192862045,
"grad_norm": 8.627043724060059,
"learning_rate": 0.00019999584799187522,
"loss": 1.7095,
"step": 14
},
{
"epoch": 0.0029414648494950485,
"grad_norm": 21.60858917236328,
"learning_rate": 0.0001999935125125579,
"loss": 3.9299,
"step": 15
},
{
"epoch": 0.003137562506128052,
"grad_norm": 8.075380325317383,
"learning_rate": 0.00019999065806252829,
"loss": 1.7939,
"step": 16
},
{
"epoch": 0.003333660162761055,
"grad_norm": 11.393594741821289,
"learning_rate": 0.00019998728465660105,
"loss": 1.601,
"step": 17
},
{
"epoch": 0.0035297578193940584,
"grad_norm": 8.256339073181152,
"learning_rate": 0.00019998339231228434,
"loss": 3.1556,
"step": 18
},
{
"epoch": 0.0037258554760270617,
"grad_norm": 20.03615951538086,
"learning_rate": 0.0001999789810497796,
"loss": 2.0883,
"step": 19
},
{
"epoch": 0.003921953132660065,
"grad_norm": 10.166353225708008,
"learning_rate": 0.0001999740508919815,
"loss": 3.5616,
"step": 20
},
{
"epoch": 0.004118050789293068,
"grad_norm": 15.80553913116455,
"learning_rate": 0.0001999686018644777,
"loss": 3.0344,
"step": 21
},
{
"epoch": 0.0043141484459260715,
"grad_norm": 7.451974391937256,
"learning_rate": 0.00019996263399554897,
"loss": 2.1049,
"step": 22
},
{
"epoch": 0.004510246102559075,
"grad_norm": 5.434274673461914,
"learning_rate": 0.00019995614731616875,
"loss": 2.3178,
"step": 23
},
{
"epoch": 0.004706343759192077,
"grad_norm": 10.594315528869629,
"learning_rate": 0.00019994914186000328,
"loss": 1.7096,
"step": 24
},
{
"epoch": 0.0049024414158250805,
"grad_norm": 5.348718166351318,
"learning_rate": 0.0001999416176634111,
"loss": 2.695,
"step": 25
},
{
"epoch": 0.005098539072458084,
"grad_norm": 17.776073455810547,
"learning_rate": 0.00019993357476544312,
"loss": 1.7411,
"step": 26
},
{
"epoch": 0.005294636729091087,
"grad_norm": 10.051606178283691,
"learning_rate": 0.0001999250132078424,
"loss": 2.6161,
"step": 27
},
{
"epoch": 0.00549073438572409,
"grad_norm": 26.03020668029785,
"learning_rate": 0.00019991593303504376,
"loss": 3.3977,
"step": 28
},
{
"epoch": 0.005686832042357094,
"grad_norm": 10.213540077209473,
"learning_rate": 0.00019990633429417363,
"loss": 1.2442,
"step": 29
},
{
"epoch": 0.005882929698990097,
"grad_norm": 11.69288444519043,
"learning_rate": 0.00019989621703505,
"loss": 1.4702,
"step": 30
},
{
"epoch": 0.0060790273556231,
"grad_norm": 4.343452453613281,
"learning_rate": 0.00019988558131018186,
"loss": 1.0779,
"step": 31
},
{
"epoch": 0.006275125012256104,
"grad_norm": 9.106976509094238,
"learning_rate": 0.00019987442717476906,
"loss": 2.5887,
"step": 32
},
{
"epoch": 0.006471222668889107,
"grad_norm": 17.658370971679688,
"learning_rate": 0.00019986275468670205,
"loss": 2.2258,
"step": 33
},
{
"epoch": 0.00666732032552211,
"grad_norm": 6.7451090812683105,
"learning_rate": 0.00019985056390656162,
"loss": 1.7206,
"step": 34
},
{
"epoch": 0.0068634179821551134,
"grad_norm": 28.07065200805664,
"learning_rate": 0.00019983785489761837,
"loss": 2.7356,
"step": 35
},
{
"epoch": 0.007059515638788117,
"grad_norm": 11.387879371643066,
"learning_rate": 0.00019982462772583266,
"loss": 1.973,
"step": 36
},
{
"epoch": 0.00725561329542112,
"grad_norm": 9.64372444152832,
"learning_rate": 0.00019981088245985408,
"loss": 2.7339,
"step": 37
},
{
"epoch": 0.007451710952054123,
"grad_norm": 9.302544593811035,
"learning_rate": 0.00019979661917102115,
"loss": 1.7498,
"step": 38
},
{
"epoch": 0.007647808608687127,
"grad_norm": 15.064400672912598,
"learning_rate": 0.000199781837933361,
"loss": 3.0109,
"step": 39
},
{
"epoch": 0.00784390626532013,
"grad_norm": 7.281099319458008,
"learning_rate": 0.00019976653882358884,
"loss": 1.3118,
"step": 40
},
{
"epoch": 0.008040003921953132,
"grad_norm": 6.4474873542785645,
"learning_rate": 0.0001997507219211078,
"loss": 1.408,
"step": 41
},
{
"epoch": 0.008236101578586136,
"grad_norm": 13.101079940795898,
"learning_rate": 0.00019973438730800822,
"loss": 2.3367,
"step": 42
},
{
"epoch": 0.008432199235219139,
"grad_norm": 5.951049327850342,
"learning_rate": 0.00019971753506906753,
"loss": 0.9101,
"step": 43
},
{
"epoch": 0.008628296891852143,
"grad_norm": 11.212276458740234,
"learning_rate": 0.00019970016529174947,
"loss": 2.7058,
"step": 44
},
{
"epoch": 0.008824394548485145,
"grad_norm": 8.68136978149414,
"learning_rate": 0.0001996822780662041,
"loss": 2.0276,
"step": 45
},
{
"epoch": 0.00902049220511815,
"grad_norm": 17.70038414001465,
"learning_rate": 0.00019966387348526683,
"loss": 2.7989,
"step": 46
},
{
"epoch": 0.009216589861751152,
"grad_norm": 10.247598648071289,
"learning_rate": 0.00019964495164445824,
"loss": 1.9618,
"step": 47
},
{
"epoch": 0.009412687518384154,
"grad_norm": 10.378255844116211,
"learning_rate": 0.0001996255126419835,
"loss": 1.8003,
"step": 48
},
{
"epoch": 0.009608785175017159,
"grad_norm": 31.620820999145508,
"learning_rate": 0.0001996055565787319,
"loss": 2.8785,
"step": 49
},
{
"epoch": 0.009804882831650161,
"grad_norm": 9.976147651672363,
"learning_rate": 0.0001995850835582763,
"loss": 2.5605,
"step": 50
},
{
"epoch": 0.010000980488283165,
"grad_norm": 11.751899719238281,
"learning_rate": 0.00019956409368687258,
"loss": 2.7556,
"step": 51
},
{
"epoch": 0.010197078144916168,
"grad_norm": 15.828932762145996,
"learning_rate": 0.000199542587073459,
"loss": 2.7773,
"step": 52
},
{
"epoch": 0.010393175801549172,
"grad_norm": 10.772979736328125,
"learning_rate": 0.00019952056382965597,
"loss": 1.9553,
"step": 53
},
{
"epoch": 0.010589273458182174,
"grad_norm": 10.821427345275879,
"learning_rate": 0.00019949802406976495,
"loss": 1.8528,
"step": 54
},
{
"epoch": 0.010785371114815178,
"grad_norm": 7.228662490844727,
"learning_rate": 0.00019947496791076837,
"loss": 1.1844,
"step": 55
},
{
"epoch": 0.01098146877144818,
"grad_norm": 7.164773941040039,
"learning_rate": 0.00019945139547232872,
"loss": 1.0291,
"step": 56
},
{
"epoch": 0.011177566428081185,
"grad_norm": 13.927733421325684,
"learning_rate": 0.0001994273068767879,
"loss": 1.5417,
"step": 57
},
{
"epoch": 0.011373664084714187,
"grad_norm": 10.366493225097656,
"learning_rate": 0.00019940270224916688,
"loss": 1.5122,
"step": 58
},
{
"epoch": 0.011569761741347192,
"grad_norm": 11.2214994430542,
"learning_rate": 0.00019937758171716468,
"loss": 1.6003,
"step": 59
},
{
"epoch": 0.011765859397980194,
"grad_norm": 14.360090255737305,
"learning_rate": 0.000199351945411158,
"loss": 1.5651,
"step": 60
},
{
"epoch": 0.011961957054613198,
"grad_norm": 17.97150993347168,
"learning_rate": 0.00019932579346420038,
"loss": 1.6064,
"step": 61
},
{
"epoch": 0.0121580547112462,
"grad_norm": 10.190518379211426,
"learning_rate": 0.00019929912601202151,
"loss": 1.9151,
"step": 62
},
{
"epoch": 0.012354152367879203,
"grad_norm": 13.573248863220215,
"learning_rate": 0.00019927194319302677,
"loss": 4.0602,
"step": 63
},
{
"epoch": 0.012550250024512207,
"grad_norm": 16.919841766357422,
"learning_rate": 0.00019924424514829606,
"loss": 2.8292,
"step": 64
},
{
"epoch": 0.01274634768114521,
"grad_norm": 58.470252990722656,
"learning_rate": 0.00019921603202158354,
"loss": 1.9637,
"step": 65
},
{
"epoch": 0.012942445337778214,
"grad_norm": 18.334800720214844,
"learning_rate": 0.00019918730395931649,
"loss": 2.5609,
"step": 66
},
{
"epoch": 0.013138542994411216,
"grad_norm": 12.280759811401367,
"learning_rate": 0.00019915806111059486,
"loss": 1.2495,
"step": 67
},
{
"epoch": 0.01333464065104422,
"grad_norm": 8.015874862670898,
"learning_rate": 0.0001991283036271903,
"loss": 1.505,
"step": 68
},
{
"epoch": 0.013530738307677223,
"grad_norm": 7.713284969329834,
"learning_rate": 0.0001990980316635455,
"loss": 2.3898,
"step": 69
},
{
"epoch": 0.013726835964310227,
"grad_norm": 18.01800537109375,
"learning_rate": 0.00019906724537677316,
"loss": 3.0263,
"step": 70
},
{
"epoch": 0.01392293362094323,
"grad_norm": 21.270421981811523,
"learning_rate": 0.00019903594492665558,
"loss": 3.2547,
"step": 71
},
{
"epoch": 0.014119031277576233,
"grad_norm": 21.60205841064453,
"learning_rate": 0.0001990041304756434,
"loss": 2.577,
"step": 72
},
{
"epoch": 0.014315128934209236,
"grad_norm": 10.01419734954834,
"learning_rate": 0.00019897180218885507,
"loss": 1.9092,
"step": 73
},
{
"epoch": 0.01451122659084224,
"grad_norm": 14.10943603515625,
"learning_rate": 0.00019893896023407578,
"loss": 2.2377,
"step": 74
},
{
"epoch": 0.014707324247475242,
"grad_norm": 11.310667037963867,
"learning_rate": 0.0001989056047817567,
"loss": 1.6645,
"step": 75
},
{
"epoch": 0.014903421904108247,
"grad_norm": 6.586666107177734,
"learning_rate": 0.0001988717360050141,
"loss": 2.2651,
"step": 76
},
{
"epoch": 0.015099519560741249,
"grad_norm": 4.402716159820557,
"learning_rate": 0.00019883735407962846,
"loss": 1.3483,
"step": 77
},
{
"epoch": 0.015295617217374253,
"grad_norm": 9.384387016296387,
"learning_rate": 0.00019880245918404342,
"loss": 2.6391,
"step": 78
},
{
"epoch": 0.015491714874007256,
"grad_norm": 6.753894329071045,
"learning_rate": 0.000198767051499365,
"loss": 2.9391,
"step": 79
},
{
"epoch": 0.01568781253064026,
"grad_norm": 6.399787902832031,
"learning_rate": 0.00019873113120936074,
"loss": 3.7452,
"step": 80
},
{
"epoch": 0.01588391018727326,
"grad_norm": 8.880107879638672,
"learning_rate": 0.00019869469850045842,
"loss": 1.2771,
"step": 81
},
{
"epoch": 0.016080007843906265,
"grad_norm": 12.630661964416504,
"learning_rate": 0.00019865775356174545,
"loss": 2.2072,
"step": 82
},
{
"epoch": 0.01627610550053927,
"grad_norm": 7.974503993988037,
"learning_rate": 0.00019862029658496762,
"loss": 1.9795,
"step": 83
},
{
"epoch": 0.016472203157172273,
"grad_norm": 50.43594741821289,
"learning_rate": 0.00019858232776452837,
"loss": 1.5331,
"step": 84
},
{
"epoch": 0.016668300813805274,
"grad_norm": 7.273484230041504,
"learning_rate": 0.00019854384729748746,
"loss": 2.4005,
"step": 85
},
{
"epoch": 0.016864398470438278,
"grad_norm": 5.826492786407471,
"learning_rate": 0.00019850485538356027,
"loss": 2.1915,
"step": 86
},
{
"epoch": 0.017060496127071282,
"grad_norm": 9.881019592285156,
"learning_rate": 0.0001984653522251165,
"loss": 2.3309,
"step": 87
},
{
"epoch": 0.017256593783704286,
"grad_norm": 9.147713661193848,
"learning_rate": 0.00019842533802717923,
"loss": 1.1404,
"step": 88
},
{
"epoch": 0.017452691440337287,
"grad_norm": 13.98263931274414,
"learning_rate": 0.00019838481299742398,
"loss": 1.2166,
"step": 89
},
{
"epoch": 0.01764878909697029,
"grad_norm": 8.206791877746582,
"learning_rate": 0.0001983437773461774,
"loss": 2.6039,
"step": 90
},
{
"epoch": 0.017844886753603295,
"grad_norm": 10.445443153381348,
"learning_rate": 0.00019830223128641637,
"loss": 2.3554,
"step": 91
},
{
"epoch": 0.0180409844102363,
"grad_norm": 11.756292343139648,
"learning_rate": 0.00019826017503376666,
"loss": 1.7371,
"step": 92
},
{
"epoch": 0.0182370820668693,
"grad_norm": 7.509032249450684,
"learning_rate": 0.00019821760880650214,
"loss": 1.389,
"step": 93
},
{
"epoch": 0.018433179723502304,
"grad_norm": 8.619280815124512,
"learning_rate": 0.00019817453282554333,
"loss": 1.6818,
"step": 94
},
{
"epoch": 0.01862927738013531,
"grad_norm": 9.11640739440918,
"learning_rate": 0.00019813094731445654,
"loss": 1.631,
"step": 95
},
{
"epoch": 0.01882537503676831,
"grad_norm": 14.109521865844727,
"learning_rate": 0.00019808685249945245,
"loss": 2.0497,
"step": 96
},
{
"epoch": 0.019021472693401313,
"grad_norm": 10.804281234741211,
"learning_rate": 0.00019804224860938506,
"loss": 2.2364,
"step": 97
},
{
"epoch": 0.019217570350034317,
"grad_norm": 7.363731384277344,
"learning_rate": 0.0001979971358757505,
"loss": 1.0967,
"step": 98
},
{
"epoch": 0.01941366800666732,
"grad_norm": 15.269912719726562,
"learning_rate": 0.0001979515145326859,
"loss": 2.8752,
"step": 99
},
{
"epoch": 0.019609765663300322,
"grad_norm": 5.457535266876221,
"learning_rate": 0.000197905384816968,
"loss": 1.7098,
"step": 100
},
{
"epoch": 0.019805863319933326,
"grad_norm": 4.689967632293701,
"learning_rate": 0.00019785874696801202,
"loss": 2.2133,
"step": 101
},
{
"epoch": 0.02000196097656633,
"grad_norm": 10.993409156799316,
"learning_rate": 0.00019781160122787046,
"loss": 2.314,
"step": 102
},
{
"epoch": 0.020198058633199335,
"grad_norm": 8.199251174926758,
"learning_rate": 0.00019776394784123177,
"loss": 2.5164,
"step": 103
},
{
"epoch": 0.020394156289832335,
"grad_norm": 15.144885063171387,
"learning_rate": 0.00019771578705541916,
"loss": 2.0058,
"step": 104
},
{
"epoch": 0.02059025394646534,
"grad_norm": 5.252450466156006,
"learning_rate": 0.00019766711912038915,
"loss": 1.7012,
"step": 105
},
{
"epoch": 0.020786351603098344,
"grad_norm": 8.265049934387207,
"learning_rate": 0.0001976179442887305,
"loss": 1.8646,
"step": 106
},
{
"epoch": 0.020982449259731348,
"grad_norm": 8.365408897399902,
"learning_rate": 0.00019756826281566272,
"loss": 1.9615,
"step": 107
},
{
"epoch": 0.02117854691636435,
"grad_norm": 7.514213562011719,
"learning_rate": 0.00019751807495903484,
"loss": 1.4897,
"step": 108
},
{
"epoch": 0.021374644572997353,
"grad_norm": 15.234655380249023,
"learning_rate": 0.00019746738097932407,
"loss": 2.0467,
"step": 109
},
{
"epoch": 0.021570742229630357,
"grad_norm": 6.856448650360107,
"learning_rate": 0.0001974161811396343,
"loss": 1.4492,
"step": 110
},
{
"epoch": 0.021766839886263357,
"grad_norm": 7.893224716186523,
"learning_rate": 0.00019736447570569503,
"loss": 1.919,
"step": 111
},
{
"epoch": 0.02196293754289636,
"grad_norm": 8.966511726379395,
"learning_rate": 0.0001973122649458597,
"loss": 2.4484,
"step": 112
},
{
"epoch": 0.022159035199529366,
"grad_norm": 7.631579875946045,
"learning_rate": 0.00019725954913110442,
"loss": 1.4992,
"step": 113
},
{
"epoch": 0.02235513285616237,
"grad_norm": 7.418518543243408,
"learning_rate": 0.0001972063285350266,
"loss": 0.8401,
"step": 114
},
{
"epoch": 0.02255123051279537,
"grad_norm": 7.739930629730225,
"learning_rate": 0.00019715260343384347,
"loss": 2.0713,
"step": 115
},
{
"epoch": 0.022747328169428375,
"grad_norm": 6.441893100738525,
"learning_rate": 0.00019709837410639063,
"loss": 1.4438,
"step": 116
},
{
"epoch": 0.02294342582606138,
"grad_norm": 6.008083820343018,
"learning_rate": 0.0001970436408341207,
"loss": 1.3503,
"step": 117
},
{
"epoch": 0.023139523482694383,
"grad_norm": 7.100820541381836,
"learning_rate": 0.00019698840390110176,
"loss": 1.4726,
"step": 118
},
{
"epoch": 0.023335621139327384,
"grad_norm": 10.213756561279297,
"learning_rate": 0.0001969326635940159,
"loss": 0.8107,
"step": 119
},
{
"epoch": 0.023531718795960388,
"grad_norm": 5.251387119293213,
"learning_rate": 0.00019687642020215775,
"loss": 1.5542,
"step": 120
},
{
"epoch": 0.023727816452593392,
"grad_norm": 6.100740432739258,
"learning_rate": 0.00019681967401743297,
"loss": 1.2512,
"step": 121
},
{
"epoch": 0.023923914109226396,
"grad_norm": 7.356696128845215,
"learning_rate": 0.00019676242533435678,
"loss": 2.4725,
"step": 122
},
{
"epoch": 0.024120011765859397,
"grad_norm": 11.542431831359863,
"learning_rate": 0.00019670467445005233,
"loss": 3.0307,
"step": 123
},
{
"epoch": 0.0243161094224924,
"grad_norm": 12.166086196899414,
"learning_rate": 0.00019664642166424928,
"loss": 1.2784,
"step": 124
},
{
"epoch": 0.024512207079125405,
"grad_norm": 5.222433090209961,
"learning_rate": 0.00019658766727928206,
"loss": 1.1759,
"step": 125
},
{
"epoch": 0.024708304735758406,
"grad_norm": 4.77174711227417,
"learning_rate": 0.00019652841160008858,
"loss": 1.1041,
"step": 126
},
{
"epoch": 0.02490440239239141,
"grad_norm": 4.879274368286133,
"learning_rate": 0.0001964686549342084,
"loss": 2.6326,
"step": 127
},
{
"epoch": 0.025100500049024414,
"grad_norm": 14.171689987182617,
"learning_rate": 0.00019640839759178116,
"loss": 3.4144,
"step": 128
},
{
"epoch": 0.02529659770565742,
"grad_norm": 7.598373889923096,
"learning_rate": 0.00019634763988554522,
"loss": 2.0596,
"step": 129
},
{
"epoch": 0.02549269536229042,
"grad_norm": 6.88770866394043,
"learning_rate": 0.00019628638213083565,
"loss": 1.4691,
"step": 130
},
{
"epoch": 0.025688793018923423,
"grad_norm": 7.128096580505371,
"learning_rate": 0.00019622462464558295,
"loss": 1.3307,
"step": 131
},
{
"epoch": 0.025884890675556427,
"grad_norm": 6.430881500244141,
"learning_rate": 0.00019616236775031113,
"loss": 0.9491,
"step": 132
},
{
"epoch": 0.02608098833218943,
"grad_norm": 9.912070274353027,
"learning_rate": 0.00019609961176813624,
"loss": 2.5006,
"step": 133
},
{
"epoch": 0.026277085988822432,
"grad_norm": 8.550467491149902,
"learning_rate": 0.0001960363570247645,
"loss": 2.4952,
"step": 134
},
{
"epoch": 0.026473183645455436,
"grad_norm": 4.201476573944092,
"learning_rate": 0.0001959726038484909,
"loss": 0.9033,
"step": 135
},
{
"epoch": 0.02666928130208844,
"grad_norm": 5.774847984313965,
"learning_rate": 0.00019590835257019714,
"loss": 2.1291,
"step": 136
},
{
"epoch": 0.026865378958721445,
"grad_norm": 8.179195404052734,
"learning_rate": 0.00019584360352335023,
"loss": 2.7527,
"step": 137
},
{
"epoch": 0.027061476615354445,
"grad_norm": 15.658841133117676,
"learning_rate": 0.0001957783570440005,
"loss": 1.8304,
"step": 138
},
{
"epoch": 0.02725757427198745,
"grad_norm": 5.7399163246154785,
"learning_rate": 0.0001957126134707801,
"loss": 1.7071,
"step": 139
},
{
"epoch": 0.027453671928620454,
"grad_norm": 5.0817389488220215,
"learning_rate": 0.00019564637314490108,
"loss": 1.8933,
"step": 140
},
{
"epoch": 0.027649769585253458,
"grad_norm": 5.634946346282959,
"learning_rate": 0.0001955796364101535,
"loss": 1.7343,
"step": 141
},
{
"epoch": 0.02784586724188646,
"grad_norm": 6.406938552856445,
"learning_rate": 0.00019551240361290407,
"loss": 2.3013,
"step": 142
},
{
"epoch": 0.028041964898519463,
"grad_norm": 8.239458084106445,
"learning_rate": 0.00019544467510209388,
"loss": 1.2177,
"step": 143
},
{
"epoch": 0.028238062555152467,
"grad_norm": 11.887965202331543,
"learning_rate": 0.0001953764512292369,
"loss": 2.4312,
"step": 144
},
{
"epoch": 0.028434160211785468,
"grad_norm": 7.482359409332275,
"learning_rate": 0.00019530773234841803,
"loss": 1.1083,
"step": 145
},
{
"epoch": 0.028630257868418472,
"grad_norm": 8.86729621887207,
"learning_rate": 0.00019523851881629126,
"loss": 1.6451,
"step": 146
},
{
"epoch": 0.028826355525051476,
"grad_norm": 7.395509719848633,
"learning_rate": 0.0001951688109920778,
"loss": 1.31,
"step": 147
},
{
"epoch": 0.02902245318168448,
"grad_norm": 4.955163955688477,
"learning_rate": 0.00019509860923756442,
"loss": 2.5206,
"step": 148
},
{
"epoch": 0.02921855083831748,
"grad_norm": 5.034746170043945,
"learning_rate": 0.00019502791391710125,
"loss": 0.9336,
"step": 149
},
{
"epoch": 0.029414648494950485,
"grad_norm": 12.375234603881836,
"learning_rate": 0.00019495672539760007,
"loss": 2.1276,
"step": 150
},
{
"epoch": 0.02961074615158349,
"grad_norm": 5.832932949066162,
"learning_rate": 0.00019488504404853248,
"loss": 1.3252,
"step": 151
},
{
"epoch": 0.029806843808216493,
"grad_norm": 5.934417724609375,
"learning_rate": 0.00019481287024192775,
"loss": 1.5907,
"step": 152
},
{
"epoch": 0.030002941464849494,
"grad_norm": 9.238896369934082,
"learning_rate": 0.00019474020435237117,
"loss": 1.1184,
"step": 153
},
{
"epoch": 0.030199039121482498,
"grad_norm": 9.787931442260742,
"learning_rate": 0.00019466704675700185,
"loss": 1.4931,
"step": 154
},
{
"epoch": 0.030395136778115502,
"grad_norm": 7.260796070098877,
"learning_rate": 0.00019459339783551094,
"loss": 0.8924,
"step": 155
},
{
"epoch": 0.030591234434748506,
"grad_norm": 8.712836265563965,
"learning_rate": 0.00019451925797013954,
"loss": 1.586,
"step": 156
},
{
"epoch": 0.030787332091381507,
"grad_norm": 11.15104866027832,
"learning_rate": 0.00019444462754567682,
"loss": 1.5007,
"step": 157
},
{
"epoch": 0.03098342974801451,
"grad_norm": 7.158255100250244,
"learning_rate": 0.00019436950694945798,
"loss": 2.4118,
"step": 158
},
{
"epoch": 0.031179527404647515,
"grad_norm": 11.58385944366455,
"learning_rate": 0.00019429389657136213,
"loss": 2.1638,
"step": 159
},
{
"epoch": 0.03137562506128052,
"grad_norm": 7.469117641448975,
"learning_rate": 0.00019421779680381054,
"loss": 3.0682,
"step": 160
},
{
"epoch": 0.031571722717913524,
"grad_norm": 10.78966999053955,
"learning_rate": 0.00019414120804176426,
"loss": 1.1822,
"step": 161
},
{
"epoch": 0.03176782037454652,
"grad_norm": 9.68694019317627,
"learning_rate": 0.00019406413068272238,
"loss": 2.5351,
"step": 162
},
{
"epoch": 0.031963918031179525,
"grad_norm": 11.67428970336914,
"learning_rate": 0.00019398656512671972,
"loss": 1.9244,
"step": 163
},
{
"epoch": 0.03216001568781253,
"grad_norm": 12.72513198852539,
"learning_rate": 0.00019390851177632497,
"loss": 3.2138,
"step": 164
},
{
"epoch": 0.03235611334444553,
"grad_norm": 8.345921516418457,
"learning_rate": 0.00019382997103663838,
"loss": 2.6435,
"step": 165
},
{
"epoch": 0.03255221100107854,
"grad_norm": 7.740304470062256,
"learning_rate": 0.0001937509433152899,
"loss": 0.8189,
"step": 166
},
{
"epoch": 0.03274830865771154,
"grad_norm": 9.329862594604492,
"learning_rate": 0.0001936714290224368,
"loss": 1.4106,
"step": 167
},
{
"epoch": 0.032944406314344546,
"grad_norm": 7.179844379425049,
"learning_rate": 0.00019359142857076176,
"loss": 1.8125,
"step": 168
},
{
"epoch": 0.03314050397097755,
"grad_norm": 7.835447311401367,
"learning_rate": 0.00019351094237547066,
"loss": 1.6617,
"step": 169
},
{
"epoch": 0.03333660162761055,
"grad_norm": 6.018518924713135,
"learning_rate": 0.0001934299708542904,
"loss": 2.4333,
"step": 170
},
{
"epoch": 0.03353269928424355,
"grad_norm": 8.176468849182129,
"learning_rate": 0.00019334851442746664,
"loss": 2.5915,
"step": 171
},
{
"epoch": 0.033728796940876556,
"grad_norm": 8.241739273071289,
"learning_rate": 0.00019326657351776186,
"loss": 1.666,
"step": 172
},
{
"epoch": 0.03392489459750956,
"grad_norm": 8.064835548400879,
"learning_rate": 0.000193184148550453,
"loss": 1.477,
"step": 173
},
{
"epoch": 0.034120992254142564,
"grad_norm": 5.790217399597168,
"learning_rate": 0.00019310123995332917,
"loss": 0.7703,
"step": 174
},
{
"epoch": 0.03431708991077557,
"grad_norm": 9.38430118560791,
"learning_rate": 0.00019301784815668974,
"loss": 1.5785,
"step": 175
},
{
"epoch": 0.03451318756740857,
"grad_norm": 8.252826690673828,
"learning_rate": 0.00019293397359334167,
"loss": 2.1462,
"step": 176
},
{
"epoch": 0.03470928522404157,
"grad_norm": 12.65652847290039,
"learning_rate": 0.00019284961669859766,
"loss": 1.3009,
"step": 177
},
{
"epoch": 0.034905382880674574,
"grad_norm": 6.8490753173828125,
"learning_rate": 0.00019276477791027374,
"loss": 2.4905,
"step": 178
},
{
"epoch": 0.03510148053730758,
"grad_norm": 4.2581048011779785,
"learning_rate": 0.0001926794576686869,
"loss": 0.9042,
"step": 179
},
{
"epoch": 0.03529757819394058,
"grad_norm": 6.415445327758789,
"learning_rate": 0.0001925936564166529,
"loss": 2.238,
"step": 180
},
{
"epoch": 0.035493675850573586,
"grad_norm": 13.620756149291992,
"learning_rate": 0.00019250737459948405,
"loss": 1.5966,
"step": 181
},
{
"epoch": 0.03568977350720659,
"grad_norm": 10.609662055969238,
"learning_rate": 0.00019242061266498675,
"loss": 1.081,
"step": 182
},
{
"epoch": 0.035885871163839594,
"grad_norm": 8.404073715209961,
"learning_rate": 0.00019233337106345925,
"loss": 1.849,
"step": 183
},
{
"epoch": 0.0360819688204726,
"grad_norm": 5.560455322265625,
"learning_rate": 0.00019224565024768926,
"loss": 1.4533,
"step": 184
},
{
"epoch": 0.036278066477105596,
"grad_norm": 7.896220684051514,
"learning_rate": 0.00019215745067295169,
"loss": 2.482,
"step": 185
},
{
"epoch": 0.0364741641337386,
"grad_norm": 9.554024696350098,
"learning_rate": 0.00019206877279700612,
"loss": 1.9367,
"step": 186
},
{
"epoch": 0.036670261790371604,
"grad_norm": 3.333113193511963,
"learning_rate": 0.00019197961708009473,
"loss": 1.1477,
"step": 187
},
{
"epoch": 0.03686635944700461,
"grad_norm": 9.468240737915039,
"learning_rate": 0.00019188998398493953,
"loss": 1.0849,
"step": 188
},
{
"epoch": 0.03706245710363761,
"grad_norm": 10.807921409606934,
"learning_rate": 0.00019179987397674022,
"loss": 2.0192,
"step": 189
},
{
"epoch": 0.03725855476027062,
"grad_norm": 7.14724588394165,
"learning_rate": 0.0001917092875231717,
"loss": 2.1502,
"step": 190
},
{
"epoch": 0.03745465241690362,
"grad_norm": 12.262707710266113,
"learning_rate": 0.00019161822509438162,
"loss": 2.423,
"step": 191
},
{
"epoch": 0.03765075007353662,
"grad_norm": 35.0489387512207,
"learning_rate": 0.000191526687162988,
"loss": 2.5959,
"step": 192
},
{
"epoch": 0.03784684773016962,
"grad_norm": 6.615735054016113,
"learning_rate": 0.0001914346742040767,
"loss": 1.7733,
"step": 193
},
{
"epoch": 0.038042945386802626,
"grad_norm": 4.537426471710205,
"learning_rate": 0.00019134218669519896,
"loss": 1.0028,
"step": 194
},
{
"epoch": 0.03823904304343563,
"grad_norm": 5.247801303863525,
"learning_rate": 0.00019124922511636912,
"loss": 0.8412,
"step": 195
},
{
"epoch": 0.038435140700068635,
"grad_norm": 6.2183918952941895,
"learning_rate": 0.00019115578995006173,
"loss": 1.7212,
"step": 196
},
{
"epoch": 0.03863123835670164,
"grad_norm": 9.330825805664062,
"learning_rate": 0.00019106188168120948,
"loss": 1.5341,
"step": 197
},
{
"epoch": 0.03882733601333464,
"grad_norm": 9.86260986328125,
"learning_rate": 0.00019096750079720037,
"loss": 2.765,
"step": 198
},
{
"epoch": 0.03902343366996765,
"grad_norm": 10.341052055358887,
"learning_rate": 0.00019087264778787534,
"loss": 1.9024,
"step": 199
},
{
"epoch": 0.039219531326600644,
"grad_norm": 9.549159049987793,
"learning_rate": 0.00019077732314552566,
"loss": 1.2644,
"step": 200
},
{
"epoch": 0.03941562898323365,
"grad_norm": 5.25094747543335,
"learning_rate": 0.00019068152736489036,
"loss": 1.334,
"step": 201
},
{
"epoch": 0.03961172663986665,
"grad_norm": 7.197662830352783,
"learning_rate": 0.00019058526094315378,
"loss": 1.9093,
"step": 202
},
{
"epoch": 0.03980782429649966,
"grad_norm": 8.476766586303711,
"learning_rate": 0.0001904885243799429,
"loss": 1.477,
"step": 203
},
{
"epoch": 0.04000392195313266,
"grad_norm": 8.232537269592285,
"learning_rate": 0.00019039131817732462,
"loss": 1.4013,
"step": 204
},
{
"epoch": 0.040200019609765665,
"grad_norm": 15.687997817993164,
"learning_rate": 0.0001902936428398035,
"loss": 1.6772,
"step": 205
},
{
"epoch": 0.04039611726639867,
"grad_norm": 7.573246479034424,
"learning_rate": 0.00019019549887431877,
"loss": 1.5007,
"step": 206
},
{
"epoch": 0.040592214923031666,
"grad_norm": 11.531679153442383,
"learning_rate": 0.0001900968867902419,
"loss": 2.6798,
"step": 207
},
{
"epoch": 0.04078831257966467,
"grad_norm": 6.225399494171143,
"learning_rate": 0.00018999780709937398,
"loss": 1.3078,
"step": 208
},
{
"epoch": 0.040984410236297675,
"grad_norm": 10.358306884765625,
"learning_rate": 0.0001898982603159429,
"loss": 1.7353,
"step": 209
},
{
"epoch": 0.04118050789293068,
"grad_norm": 8.146821975708008,
"learning_rate": 0.00018979824695660087,
"loss": 1.415,
"step": 210
},
{
"epoch": 0.04137660554956368,
"grad_norm": 4.390834808349609,
"learning_rate": 0.00018969776754042156,
"loss": 1.7612,
"step": 211
},
{
"epoch": 0.04157270320619669,
"grad_norm": 7.958174228668213,
"learning_rate": 0.0001895968225888976,
"loss": 2.6614,
"step": 212
},
{
"epoch": 0.04176880086282969,
"grad_norm": 9.981225967407227,
"learning_rate": 0.00018949541262593762,
"loss": 2.0158,
"step": 213
},
{
"epoch": 0.041964898519462696,
"grad_norm": 4.456605911254883,
"learning_rate": 0.00018939353817786387,
"loss": 1.0621,
"step": 214
},
{
"epoch": 0.04216099617609569,
"grad_norm": 7.546274662017822,
"learning_rate": 0.00018929119977340917,
"loss": 1.7333,
"step": 215
},
{
"epoch": 0.0423570938327287,
"grad_norm": 11.629569053649902,
"learning_rate": 0.0001891883979437143,
"loss": 1.4268,
"step": 216
},
{
"epoch": 0.0425531914893617,
"grad_norm": 17.710948944091797,
"learning_rate": 0.00018908513322232528,
"loss": 2.8701,
"step": 217
},
{
"epoch": 0.042749289145994705,
"grad_norm": 6.267049789428711,
"learning_rate": 0.00018898140614519054,
"loss": 1.6313,
"step": 218
},
{
"epoch": 0.04294538680262771,
"grad_norm": 4.971591949462891,
"learning_rate": 0.00018887721725065814,
"loss": 2.0962,
"step": 219
},
{
"epoch": 0.043141484459260714,
"grad_norm": 5.603585243225098,
"learning_rate": 0.00018877256707947306,
"loss": 0.6683,
"step": 220
},
{
"epoch": 0.04333758211589372,
"grad_norm": 6.029137134552002,
"learning_rate": 0.00018866745617477423,
"loss": 1.5375,
"step": 221
},
{
"epoch": 0.043533679772526715,
"grad_norm": 7.4105143547058105,
"learning_rate": 0.00018856188508209183,
"loss": 1.9524,
"step": 222
},
{
"epoch": 0.04372977742915972,
"grad_norm": 8.321500778198242,
"learning_rate": 0.00018845585434934452,
"loss": 2.1109,
"step": 223
},
{
"epoch": 0.04392587508579272,
"grad_norm": 9.238992691040039,
"learning_rate": 0.00018834936452683638,
"loss": 1.4247,
"step": 224
},
{
"epoch": 0.04412197274242573,
"grad_norm": 5.125700950622559,
"learning_rate": 0.00018824241616725434,
"loss": 1.1266,
"step": 225
},
{
"epoch": 0.04431807039905873,
"grad_norm": 7.538069725036621,
"learning_rate": 0.000188135009825665,
"loss": 2.1554,
"step": 226
},
{
"epoch": 0.044514168055691736,
"grad_norm": 8.309137344360352,
"learning_rate": 0.00018802714605951199,
"loss": 1.1435,
"step": 227
},
{
"epoch": 0.04471026571232474,
"grad_norm": 22.02942657470703,
"learning_rate": 0.00018791882542861302,
"loss": 1.8154,
"step": 228
},
{
"epoch": 0.044906363368957744,
"grad_norm": 7.017299652099609,
"learning_rate": 0.0001878100484951569,
"loss": 1.4998,
"step": 229
},
{
"epoch": 0.04510246102559074,
"grad_norm": 18.39406394958496,
"learning_rate": 0.00018770081582370068,
"loss": 2.1662,
"step": 230
},
{
"epoch": 0.045298558682223745,
"grad_norm": 9.11802864074707,
"learning_rate": 0.0001875911279811667,
"loss": 0.7446,
"step": 231
},
{
"epoch": 0.04549465633885675,
"grad_norm": 7.193735122680664,
"learning_rate": 0.00018748098553683968,
"loss": 1.9472,
"step": 232
},
{
"epoch": 0.045690753995489754,
"grad_norm": 23.407245635986328,
"learning_rate": 0.0001873703890623637,
"loss": 2.1782,
"step": 233
},
{
"epoch": 0.04588685165212276,
"grad_norm": 6.547053813934326,
"learning_rate": 0.00018725933913173938,
"loss": 1.9687,
"step": 234
},
{
"epoch": 0.04608294930875576,
"grad_norm": 10.576699256896973,
"learning_rate": 0.00018714783632132068,
"loss": 1.8832,
"step": 235
},
{
"epoch": 0.046279046965388766,
"grad_norm": 5.852027416229248,
"learning_rate": 0.00018703588120981207,
"loss": 1.8932,
"step": 236
},
{
"epoch": 0.04647514462202176,
"grad_norm": 7.023755073547363,
"learning_rate": 0.00018692347437826548,
"loss": 3.7953,
"step": 237
},
{
"epoch": 0.04667124227865477,
"grad_norm": 13.61612606048584,
"learning_rate": 0.00018681061641007737,
"loss": 1.9077,
"step": 238
},
{
"epoch": 0.04686733993528777,
"grad_norm": 5.3344526290893555,
"learning_rate": 0.0001866973078909854,
"loss": 1.4342,
"step": 239
},
{
"epoch": 0.047063437591920776,
"grad_norm": 38.80408477783203,
"learning_rate": 0.00018658354940906586,
"loss": 2.3665,
"step": 240
},
{
"epoch": 0.04725953524855378,
"grad_norm": 9.670344352722168,
"learning_rate": 0.00018646934155473022,
"loss": 0.9006,
"step": 241
},
{
"epoch": 0.047455632905186784,
"grad_norm": 5.1102495193481445,
"learning_rate": 0.00018635468492072228,
"loss": 1.2289,
"step": 242
},
{
"epoch": 0.04765173056181979,
"grad_norm": 9.1209077835083,
"learning_rate": 0.00018623958010211493,
"loss": 1.6009,
"step": 243
},
{
"epoch": 0.04784782821845279,
"grad_norm": 16.793027877807617,
"learning_rate": 0.0001861240276963073,
"loss": 0.94,
"step": 244
},
{
"epoch": 0.04804392587508579,
"grad_norm": 6.90054988861084,
"learning_rate": 0.00018600802830302134,
"loss": 1.559,
"step": 245
},
{
"epoch": 0.048240023531718794,
"grad_norm": 13.111268043518066,
"learning_rate": 0.0001858915825242991,
"loss": 2.1186,
"step": 246
},
{
"epoch": 0.0484361211883518,
"grad_norm": 6.356579780578613,
"learning_rate": 0.00018577469096449925,
"loss": 1.6653,
"step": 247
},
{
"epoch": 0.0486322188449848,
"grad_norm": 9.505541801452637,
"learning_rate": 0.00018565735423029404,
"loss": 0.9774,
"step": 248
},
{
"epoch": 0.048828316501617806,
"grad_norm": 8.927581787109375,
"learning_rate": 0.00018553957293066632,
"loss": 2.6455,
"step": 249
},
{
"epoch": 0.04902441415825081,
"grad_norm": 7.568793773651123,
"learning_rate": 0.00018542134767690616,
"loss": 1.1464,
"step": 250
},
{
"epoch": 0.049220511814883815,
"grad_norm": 7.632232189178467,
"learning_rate": 0.00018530267908260784,
"loss": 1.2671,
"step": 251
},
{
"epoch": 0.04941660947151681,
"grad_norm": 4.4279561042785645,
"learning_rate": 0.00018518356776366657,
"loss": 2.0384,
"step": 252
},
{
"epoch": 0.049612707128149816,
"grad_norm": 10.818602561950684,
"learning_rate": 0.00018506401433827528,
"loss": 1.0559,
"step": 253
},
{
"epoch": 0.04980880478478282,
"grad_norm": 5.57148551940918,
"learning_rate": 0.00018494401942692153,
"loss": 0.9603,
"step": 254
},
{
"epoch": 0.050004902441415824,
"grad_norm": 11.1985502243042,
"learning_rate": 0.00018482358365238413,
"loss": 2.4928,
"step": 255
},
{
"epoch": 0.05020100009804883,
"grad_norm": 4.890799522399902,
"learning_rate": 0.00018470270763973004,
"loss": 1.4034,
"step": 256
},
{
"epoch": 0.05039709775468183,
"grad_norm": 6.2078680992126465,
"learning_rate": 0.00018458139201631108,
"loss": 1.782,
"step": 257
},
{
"epoch": 0.05059319541131484,
"grad_norm": 24.89278221130371,
"learning_rate": 0.00018445963741176065,
"loss": 3.7879,
"step": 258
},
{
"epoch": 0.05078929306794784,
"grad_norm": 5.363570213317871,
"learning_rate": 0.00018433744445799045,
"loss": 1.4292,
"step": 259
},
{
"epoch": 0.05098539072458084,
"grad_norm": 7.669764041900635,
"learning_rate": 0.0001842148137891873,
"loss": 2.0483,
"step": 260
},
{
"epoch": 0.05118148838121384,
"grad_norm": 5.229150295257568,
"learning_rate": 0.00018409174604180976,
"loss": 3.2863,
"step": 261
},
{
"epoch": 0.05137758603784685,
"grad_norm": 5.850373268127441,
"learning_rate": 0.0001839682418545848,
"loss": 1.8197,
"step": 262
},
{
"epoch": 0.05157368369447985,
"grad_norm": 7.138283729553223,
"learning_rate": 0.00018384430186850454,
"loss": 2.7101,
"step": 263
},
{
"epoch": 0.051769781351112855,
"grad_norm": 10.918169975280762,
"learning_rate": 0.000183719926726823,
"loss": 1.8243,
"step": 264
},
{
"epoch": 0.05196587900774586,
"grad_norm": 9.205517768859863,
"learning_rate": 0.00018359511707505258,
"loss": 1.4992,
"step": 265
},
{
"epoch": 0.05216197666437886,
"grad_norm": 8.567139625549316,
"learning_rate": 0.00018346987356096086,
"loss": 1.051,
"step": 266
},
{
"epoch": 0.05235807432101187,
"grad_norm": 10.313075065612793,
"learning_rate": 0.00018334419683456717,
"loss": 2.6062,
"step": 267
},
{
"epoch": 0.052554171977644865,
"grad_norm": 7.515801906585693,
"learning_rate": 0.0001832180875481392,
"loss": 1.266,
"step": 268
},
{
"epoch": 0.05275026963427787,
"grad_norm": 5.345809459686279,
"learning_rate": 0.00018309154635618965,
"loss": 1.2526,
"step": 269
},
{
"epoch": 0.05294636729091087,
"grad_norm": 13.568882942199707,
"learning_rate": 0.00018296457391547296,
"loss": 2.5183,
"step": 270
},
{
"epoch": 0.05314246494754388,
"grad_norm": 10.022235870361328,
"learning_rate": 0.00018283717088498155,
"loss": 2.2774,
"step": 271
},
{
"epoch": 0.05333856260417688,
"grad_norm": 6.537176132202148,
"learning_rate": 0.0001827093379259428,
"loss": 1.4989,
"step": 272
},
{
"epoch": 0.053534660260809885,
"grad_norm": 17.213987350463867,
"learning_rate": 0.00018258107570181533,
"loss": 2.4885,
"step": 273
},
{
"epoch": 0.05373075791744289,
"grad_norm": 6.48647403717041,
"learning_rate": 0.00018245238487828573,
"loss": 1.2309,
"step": 274
},
{
"epoch": 0.05392685557407589,
"grad_norm": 5.479822158813477,
"learning_rate": 0.000182323266123265,
"loss": 1.8959,
"step": 275
},
{
"epoch": 0.05412295323070889,
"grad_norm": 7.716124534606934,
"learning_rate": 0.00018219372010688515,
"loss": 1.8321,
"step": 276
},
{
"epoch": 0.054319050887341895,
"grad_norm": 9.968965530395508,
"learning_rate": 0.00018206374750149567,
"loss": 4.1652,
"step": 277
},
{
"epoch": 0.0545151485439749,
"grad_norm": 6.009235382080078,
"learning_rate": 0.00018193334898166007,
"loss": 0.8178,
"step": 278
},
{
"epoch": 0.0547112462006079,
"grad_norm": 8.031886100769043,
"learning_rate": 0.00018180252522415242,
"loss": 1.783,
"step": 279
},
{
"epoch": 0.05490734385724091,
"grad_norm": 5.5589680671691895,
"learning_rate": 0.00018167127690795368,
"loss": 1.3049,
"step": 280
},
{
"epoch": 0.05510344151387391,
"grad_norm": 5.04995059967041,
"learning_rate": 0.0001815396047142485,
"loss": 0.8962,
"step": 281
},
{
"epoch": 0.055299539170506916,
"grad_norm": 5.3526692390441895,
"learning_rate": 0.0001814075093264212,
"loss": 1.201,
"step": 282
},
{
"epoch": 0.05549563682713991,
"grad_norm": 11.980429649353027,
"learning_rate": 0.00018127499143005268,
"loss": 0.6955,
"step": 283
},
{
"epoch": 0.05569173448377292,
"grad_norm": 38.28229904174805,
"learning_rate": 0.00018114205171291663,
"loss": 1.7335,
"step": 284
},
{
"epoch": 0.05588783214040592,
"grad_norm": 6.15138053894043,
"learning_rate": 0.000181008690864976,
"loss": 1.2766,
"step": 285
},
{
"epoch": 0.056083929797038926,
"grad_norm": 7.846836566925049,
"learning_rate": 0.00018087490957837944,
"loss": 1.155,
"step": 286
},
{
"epoch": 0.05628002745367193,
"grad_norm": 7.675628185272217,
"learning_rate": 0.00018074070854745772,
"loss": 1.6129,
"step": 287
},
{
"epoch": 0.056476125110304934,
"grad_norm": 12.245649337768555,
"learning_rate": 0.00018060608846872005,
"loss": 1.7585,
"step": 288
},
{
"epoch": 0.05667222276693794,
"grad_norm": 10.520101547241211,
"learning_rate": 0.00018047105004085053,
"loss": 1.9265,
"step": 289
},
{
"epoch": 0.056868320423570935,
"grad_norm": 7.400151252746582,
"learning_rate": 0.00018033559396470454,
"loss": 1.4189,
"step": 290
},
{
"epoch": 0.05706441808020394,
"grad_norm": 12.058060646057129,
"learning_rate": 0.00018019972094330503,
"loss": 2.3312,
"step": 291
},
{
"epoch": 0.057260515736836944,
"grad_norm": 5.313794136047363,
"learning_rate": 0.00018006343168183893,
"loss": 2.0051,
"step": 292
},
{
"epoch": 0.05745661339346995,
"grad_norm": 11.182997703552246,
"learning_rate": 0.0001799267268876535,
"loss": 1.4779,
"step": 293
},
{
"epoch": 0.05765271105010295,
"grad_norm": 16.24866485595703,
"learning_rate": 0.0001797896072702526,
"loss": 2.4689,
"step": 294
},
{
"epoch": 0.057848808706735956,
"grad_norm": 7.471411228179932,
"learning_rate": 0.00017965207354129307,
"loss": 3.0599,
"step": 295
},
{
"epoch": 0.05804490636336896,
"grad_norm": 7.715878486633301,
"learning_rate": 0.00017951412641458098,
"loss": 0.8256,
"step": 296
},
{
"epoch": 0.058241004020001964,
"grad_norm": 22.084482192993164,
"learning_rate": 0.000179375766606068,
"loss": 2.457,
"step": 297
},
{
"epoch": 0.05843710167663496,
"grad_norm": 8.041847229003906,
"learning_rate": 0.00017923699483384753,
"loss": 1.5642,
"step": 298
},
{
"epoch": 0.058633199333267966,
"grad_norm": 12.814888000488281,
"learning_rate": 0.00017909781181815117,
"loss": 1.5129,
"step": 299
},
{
"epoch": 0.05882929698990097,
"grad_norm": 9.216371536254883,
"learning_rate": 0.0001789582182813449,
"loss": 2.0632,
"step": 300
},
{
"epoch": 0.059025394646533974,
"grad_norm": 12.80371379852295,
"learning_rate": 0.00017881821494792528,
"loss": 2.8705,
"step": 301
},
{
"epoch": 0.05922149230316698,
"grad_norm": 7.234943389892578,
"learning_rate": 0.00017867780254451576,
"loss": 2.6664,
"step": 302
},
{
"epoch": 0.05941758995979998,
"grad_norm": 11.168726921081543,
"learning_rate": 0.00017853698179986282,
"loss": 1.347,
"step": 303
},
{
"epoch": 0.059613687616432987,
"grad_norm": 19.369266510009766,
"learning_rate": 0.00017839575344483238,
"loss": 2.68,
"step": 304
},
{
"epoch": 0.059809785273065984,
"grad_norm": 7.1730570793151855,
"learning_rate": 0.0001782541182124057,
"loss": 2.3908,
"step": 305
},
{
"epoch": 0.06000588292969899,
"grad_norm": 7.243929862976074,
"learning_rate": 0.0001781120768376759,
"loss": 1.0056,
"step": 306
},
{
"epoch": 0.06020198058633199,
"grad_norm": 7.748988628387451,
"learning_rate": 0.00017796963005784394,
"loss": 2.1776,
"step": 307
},
{
"epoch": 0.060398078242964996,
"grad_norm": 13.446945190429688,
"learning_rate": 0.0001778267786122148,
"loss": 2.3275,
"step": 308
},
{
"epoch": 0.060594175899598,
"grad_norm": 10.720627784729004,
"learning_rate": 0.0001776835232421938,
"loss": 1.046,
"step": 309
},
{
"epoch": 0.060790273556231005,
"grad_norm": 11.274985313415527,
"learning_rate": 0.00017753986469128257,
"loss": 2.4269,
"step": 310
},
{
"epoch": 0.06098637121286401,
"grad_norm": 8.671335220336914,
"learning_rate": 0.00017739580370507532,
"loss": 2.1488,
"step": 311
},
{
"epoch": 0.06118246886949701,
"grad_norm": 8.375978469848633,
"learning_rate": 0.0001772513410312548,
"loss": 1.8458,
"step": 312
},
{
"epoch": 0.06137856652613001,
"grad_norm": 11.178112983703613,
"learning_rate": 0.00017710647741958868,
"loss": 2.7169,
"step": 313
},
{
"epoch": 0.061574664182763014,
"grad_norm": 8.29799747467041,
"learning_rate": 0.00017696121362192544,
"loss": 1.455,
"step": 314
},
{
"epoch": 0.06177076183939602,
"grad_norm": 6.712766647338867,
"learning_rate": 0.00017681555039219054,
"loss": 1.2604,
"step": 315
},
{
"epoch": 0.06196685949602902,
"grad_norm": 7.891608238220215,
"learning_rate": 0.00017666948848638257,
"loss": 2.1795,
"step": 316
},
{
"epoch": 0.06216295715266203,
"grad_norm": 5.039219379425049,
"learning_rate": 0.00017652302866256916,
"loss": 0.9069,
"step": 317
},
{
"epoch": 0.06235905480929503,
"grad_norm": 9.421103477478027,
"learning_rate": 0.00017637617168088325,
"loss": 2.4256,
"step": 318
},
{
"epoch": 0.06255515246592804,
"grad_norm": 4.435902118682861,
"learning_rate": 0.000176228918303519,
"loss": 1.9269,
"step": 319
},
{
"epoch": 0.06275125012256104,
"grad_norm": 10.938987731933594,
"learning_rate": 0.00017608126929472795,
"loss": 1.4649,
"step": 320
},
{
"epoch": 0.06294734777919404,
"grad_norm": 6.332970142364502,
"learning_rate": 0.00017593322542081485,
"loss": 2.0089,
"step": 321
},
{
"epoch": 0.06314344543582705,
"grad_norm": 6.731532573699951,
"learning_rate": 0.00017578478745013392,
"loss": 2.4046,
"step": 322
},
{
"epoch": 0.06333954309246005,
"grad_norm": 8.772012710571289,
"learning_rate": 0.00017563595615308474,
"loss": 1.4935,
"step": 323
},
{
"epoch": 0.06353564074909304,
"grad_norm": 5.693745136260986,
"learning_rate": 0.00017548673230210823,
"loss": 1.848,
"step": 324
},
{
"epoch": 0.06373173840572605,
"grad_norm": 15.056157112121582,
"learning_rate": 0.0001753371166716828,
"loss": 1.4598,
"step": 325
},
{
"epoch": 0.06392783606235905,
"grad_norm": 9.370506286621094,
"learning_rate": 0.00017518711003832002,
"loss": 1.4809,
"step": 326
},
{
"epoch": 0.06412393371899205,
"grad_norm": 19.398839950561523,
"learning_rate": 0.000175036713180561,
"loss": 1.0093,
"step": 327
},
{
"epoch": 0.06432003137562506,
"grad_norm": 4.393742084503174,
"learning_rate": 0.00017488592687897193,
"loss": 0.817,
"step": 328
},
{
"epoch": 0.06451612903225806,
"grad_norm": 6.7713799476623535,
"learning_rate": 0.00017473475191614037,
"loss": 2.1701,
"step": 329
},
{
"epoch": 0.06471222668889107,
"grad_norm": 5.920267581939697,
"learning_rate": 0.00017458318907667098,
"loss": 3.3491,
"step": 330
},
{
"epoch": 0.06490832434552407,
"grad_norm": 15.095996856689453,
"learning_rate": 0.0001744312391471816,
"loss": 1.7637,
"step": 331
},
{
"epoch": 0.06510442200215708,
"grad_norm": 9.470211029052734,
"learning_rate": 0.00017427890291629893,
"loss": 2.7744,
"step": 332
},
{
"epoch": 0.06530051965879008,
"grad_norm": 9.082067489624023,
"learning_rate": 0.00017412618117465477,
"loss": 3.1791,
"step": 333
},
{
"epoch": 0.06549661731542308,
"grad_norm": 5.174635410308838,
"learning_rate": 0.0001739730747148816,
"loss": 1.2189,
"step": 334
},
{
"epoch": 0.06569271497205609,
"grad_norm": 5.053405284881592,
"learning_rate": 0.00017381958433160865,
"loss": 1.7119,
"step": 335
},
{
"epoch": 0.06588881262868909,
"grad_norm": 5.771046161651611,
"learning_rate": 0.0001736657108214578,
"loss": 1.4188,
"step": 336
},
{
"epoch": 0.0660849102853221,
"grad_norm": 8.400517463684082,
"learning_rate": 0.00017351145498303925,
"loss": 2.3167,
"step": 337
},
{
"epoch": 0.0662810079419551,
"grad_norm": 4.6646728515625,
"learning_rate": 0.0001733568176169476,
"loss": 1.2102,
"step": 338
},
{
"epoch": 0.06647710559858809,
"grad_norm": 8.288646697998047,
"learning_rate": 0.0001732017995257575,
"loss": 2.4803,
"step": 339
},
{
"epoch": 0.0666732032552211,
"grad_norm": 10.970074653625488,
"learning_rate": 0.00017304640151401967,
"loss": 2.5839,
"step": 340
},
{
"epoch": 0.0668693009118541,
"grad_norm": 6.0125732421875,
"learning_rate": 0.00017289062438825665,
"loss": 1.5807,
"step": 341
},
{
"epoch": 0.0670653985684871,
"grad_norm": 5.844028472900391,
"learning_rate": 0.0001727344689569585,
"loss": 3.34,
"step": 342
},
{
"epoch": 0.06726149622512011,
"grad_norm": 7.1026387214660645,
"learning_rate": 0.00017257793603057871,
"loss": 1.4347,
"step": 343
},
{
"epoch": 0.06745759388175311,
"grad_norm": 9.198262214660645,
"learning_rate": 0.00017242102642153016,
"loss": 1.834,
"step": 344
},
{
"epoch": 0.06765369153838612,
"grad_norm": 5.76854133605957,
"learning_rate": 0.00017226374094418044,
"loss": 0.9294,
"step": 345
},
{
"epoch": 0.06784978919501912,
"grad_norm": 10.319186210632324,
"learning_rate": 0.0001721060804148482,
"loss": 2.0088,
"step": 346
},
{
"epoch": 0.06804588685165212,
"grad_norm": 22.298240661621094,
"learning_rate": 0.00017194804565179842,
"loss": 2.6901,
"step": 347
},
{
"epoch": 0.06824198450828513,
"grad_norm": 11.38401985168457,
"learning_rate": 0.00017178963747523847,
"loss": 2.6342,
"step": 348
},
{
"epoch": 0.06824198450828513,
"eval_loss": 0.4400941729545593,
"eval_runtime": 78.7276,
"eval_samples_per_second": 27.284,
"eval_steps_per_second": 13.642,
"step": 348
}
],
"logging_steps": 1,
"max_steps": 1389,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 348,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7.492671941640192e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}