Arabic-DeepSeek-R1-Distill-8B / trainer_state.json
Omartificial-Intelligence-Space's picture
upload 11 files
36367f7 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 288,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003472222222222222,
"grad_norm": 0.6381791234016418,
"learning_rate": 4e-05,
"loss": 1.664,
"step": 1
},
{
"epoch": 0.006944444444444444,
"grad_norm": 0.6412252187728882,
"learning_rate": 8e-05,
"loss": 1.6095,
"step": 2
},
{
"epoch": 0.010416666666666666,
"grad_norm": 0.6355312466621399,
"learning_rate": 0.00012,
"loss": 1.6431,
"step": 3
},
{
"epoch": 0.013888888888888888,
"grad_norm": 0.7827576994895935,
"learning_rate": 0.00016,
"loss": 1.7554,
"step": 4
},
{
"epoch": 0.017361111111111112,
"grad_norm": 0.7365180850028992,
"learning_rate": 0.0002,
"loss": 1.6866,
"step": 5
},
{
"epoch": 0.020833333333333332,
"grad_norm": 0.7607629299163818,
"learning_rate": 0.00019929328621908128,
"loss": 1.5447,
"step": 6
},
{
"epoch": 0.024305555555555556,
"grad_norm": 0.7718775272369385,
"learning_rate": 0.00019858657243816254,
"loss": 1.5777,
"step": 7
},
{
"epoch": 0.027777777777777776,
"grad_norm": 0.7301136255264282,
"learning_rate": 0.00019787985865724384,
"loss": 1.3402,
"step": 8
},
{
"epoch": 0.03125,
"grad_norm": 0.6953989863395691,
"learning_rate": 0.0001971731448763251,
"loss": 1.4443,
"step": 9
},
{
"epoch": 0.034722222222222224,
"grad_norm": 0.605755090713501,
"learning_rate": 0.00019646643109540637,
"loss": 1.3609,
"step": 10
},
{
"epoch": 0.03819444444444445,
"grad_norm": 0.5867183804512024,
"learning_rate": 0.00019575971731448764,
"loss": 1.2765,
"step": 11
},
{
"epoch": 0.041666666666666664,
"grad_norm": 0.5973581075668335,
"learning_rate": 0.00019505300353356894,
"loss": 1.2585,
"step": 12
},
{
"epoch": 0.04513888888888889,
"grad_norm": 0.7357228398323059,
"learning_rate": 0.0001943462897526502,
"loss": 1.2199,
"step": 13
},
{
"epoch": 0.04861111111111111,
"grad_norm": 0.7804207801818848,
"learning_rate": 0.00019363957597173144,
"loss": 1.1996,
"step": 14
},
{
"epoch": 0.052083333333333336,
"grad_norm": 0.813244640827179,
"learning_rate": 0.0001929328621908127,
"loss": 1.1735,
"step": 15
},
{
"epoch": 0.05555555555555555,
"grad_norm": 0.7277530431747437,
"learning_rate": 0.00019222614840989398,
"loss": 1.1094,
"step": 16
},
{
"epoch": 0.059027777777777776,
"grad_norm": 0.6414357423782349,
"learning_rate": 0.00019151943462897527,
"loss": 1.1786,
"step": 17
},
{
"epoch": 0.0625,
"grad_norm": 0.5542577505111694,
"learning_rate": 0.00019081272084805654,
"loss": 0.9343,
"step": 18
},
{
"epoch": 0.06597222222222222,
"grad_norm": 0.5263577699661255,
"learning_rate": 0.0001901060070671378,
"loss": 1.0607,
"step": 19
},
{
"epoch": 0.06944444444444445,
"grad_norm": 0.49260228872299194,
"learning_rate": 0.00018939929328621908,
"loss": 0.9934,
"step": 20
},
{
"epoch": 0.07291666666666667,
"grad_norm": 0.5072307586669922,
"learning_rate": 0.00018869257950530037,
"loss": 1.0165,
"step": 21
},
{
"epoch": 0.0763888888888889,
"grad_norm": 0.5154305696487427,
"learning_rate": 0.00018798586572438164,
"loss": 1.0159,
"step": 22
},
{
"epoch": 0.0798611111111111,
"grad_norm": 0.5033209919929504,
"learning_rate": 0.0001872791519434629,
"loss": 1.0102,
"step": 23
},
{
"epoch": 0.08333333333333333,
"grad_norm": 0.5032006502151489,
"learning_rate": 0.00018657243816254417,
"loss": 1.0215,
"step": 24
},
{
"epoch": 0.08680555555555555,
"grad_norm": 0.4746347665786743,
"learning_rate": 0.00018586572438162547,
"loss": 0.867,
"step": 25
},
{
"epoch": 0.09027777777777778,
"grad_norm": 0.4780479967594147,
"learning_rate": 0.00018515901060070673,
"loss": 0.951,
"step": 26
},
{
"epoch": 0.09375,
"grad_norm": 0.45668691396713257,
"learning_rate": 0.000184452296819788,
"loss": 0.9464,
"step": 27
},
{
"epoch": 0.09722222222222222,
"grad_norm": 0.43334755301475525,
"learning_rate": 0.00018374558303886927,
"loss": 0.8768,
"step": 28
},
{
"epoch": 0.10069444444444445,
"grad_norm": 0.4458893835544586,
"learning_rate": 0.00018303886925795054,
"loss": 0.8553,
"step": 29
},
{
"epoch": 0.10416666666666667,
"grad_norm": 0.4605867564678192,
"learning_rate": 0.00018233215547703183,
"loss": 0.948,
"step": 30
},
{
"epoch": 0.1076388888888889,
"grad_norm": 0.44694918394088745,
"learning_rate": 0.0001816254416961131,
"loss": 0.8755,
"step": 31
},
{
"epoch": 0.1111111111111111,
"grad_norm": 0.45669859647750854,
"learning_rate": 0.00018091872791519434,
"loss": 0.8554,
"step": 32
},
{
"epoch": 0.11458333333333333,
"grad_norm": 0.4380110502243042,
"learning_rate": 0.0001802120141342756,
"loss": 0.8699,
"step": 33
},
{
"epoch": 0.11805555555555555,
"grad_norm": 0.4494687020778656,
"learning_rate": 0.0001795053003533569,
"loss": 0.8463,
"step": 34
},
{
"epoch": 0.12152777777777778,
"grad_norm": 0.4385037422180176,
"learning_rate": 0.00017879858657243817,
"loss": 0.7964,
"step": 35
},
{
"epoch": 0.125,
"grad_norm": 0.42709654569625854,
"learning_rate": 0.00017809187279151944,
"loss": 0.8675,
"step": 36
},
{
"epoch": 0.1284722222222222,
"grad_norm": 0.462563157081604,
"learning_rate": 0.0001773851590106007,
"loss": 0.9125,
"step": 37
},
{
"epoch": 0.13194444444444445,
"grad_norm": 0.4372367560863495,
"learning_rate": 0.00017667844522968197,
"loss": 0.8352,
"step": 38
},
{
"epoch": 0.13541666666666666,
"grad_norm": 0.4251866340637207,
"learning_rate": 0.00017597173144876327,
"loss": 0.8695,
"step": 39
},
{
"epoch": 0.1388888888888889,
"grad_norm": 0.43755847215652466,
"learning_rate": 0.00017526501766784453,
"loss": 0.8544,
"step": 40
},
{
"epoch": 0.1423611111111111,
"grad_norm": 0.44858112931251526,
"learning_rate": 0.0001745583038869258,
"loss": 0.835,
"step": 41
},
{
"epoch": 0.14583333333333334,
"grad_norm": 0.4447510838508606,
"learning_rate": 0.00017385159010600707,
"loss": 0.8537,
"step": 42
},
{
"epoch": 0.14930555555555555,
"grad_norm": 0.42954930663108826,
"learning_rate": 0.00017314487632508836,
"loss": 0.7603,
"step": 43
},
{
"epoch": 0.1527777777777778,
"grad_norm": 0.4357329308986664,
"learning_rate": 0.00017243816254416963,
"loss": 0.7315,
"step": 44
},
{
"epoch": 0.15625,
"grad_norm": 0.4562698304653168,
"learning_rate": 0.0001717314487632509,
"loss": 0.9103,
"step": 45
},
{
"epoch": 0.1597222222222222,
"grad_norm": 0.46614545583724976,
"learning_rate": 0.00017102473498233216,
"loss": 0.832,
"step": 46
},
{
"epoch": 0.16319444444444445,
"grad_norm": 0.42011672258377075,
"learning_rate": 0.00017031802120141343,
"loss": 0.7172,
"step": 47
},
{
"epoch": 0.16666666666666666,
"grad_norm": 0.40585461258888245,
"learning_rate": 0.00016961130742049473,
"loss": 0.7087,
"step": 48
},
{
"epoch": 0.1701388888888889,
"grad_norm": 0.4490915834903717,
"learning_rate": 0.000168904593639576,
"loss": 0.7956,
"step": 49
},
{
"epoch": 0.1736111111111111,
"grad_norm": 0.4251860976219177,
"learning_rate": 0.00016819787985865723,
"loss": 0.7763,
"step": 50
},
{
"epoch": 0.17708333333333334,
"grad_norm": 0.4503968060016632,
"learning_rate": 0.0001674911660777385,
"loss": 0.833,
"step": 51
},
{
"epoch": 0.18055555555555555,
"grad_norm": 0.452923059463501,
"learning_rate": 0.0001667844522968198,
"loss": 0.729,
"step": 52
},
{
"epoch": 0.1840277777777778,
"grad_norm": 0.45421820878982544,
"learning_rate": 0.00016607773851590106,
"loss": 0.8678,
"step": 53
},
{
"epoch": 0.1875,
"grad_norm": 0.47255003452301025,
"learning_rate": 0.00016537102473498233,
"loss": 0.7856,
"step": 54
},
{
"epoch": 0.1909722222222222,
"grad_norm": 0.49460238218307495,
"learning_rate": 0.0001646643109540636,
"loss": 0.8103,
"step": 55
},
{
"epoch": 0.19444444444444445,
"grad_norm": 0.47668716311454773,
"learning_rate": 0.0001639575971731449,
"loss": 0.6687,
"step": 56
},
{
"epoch": 0.19791666666666666,
"grad_norm": 0.4631917178630829,
"learning_rate": 0.00016325088339222616,
"loss": 0.7362,
"step": 57
},
{
"epoch": 0.2013888888888889,
"grad_norm": 0.4908764064311981,
"learning_rate": 0.00016254416961130743,
"loss": 0.7734,
"step": 58
},
{
"epoch": 0.2048611111111111,
"grad_norm": 0.47517484426498413,
"learning_rate": 0.0001618374558303887,
"loss": 0.8048,
"step": 59
},
{
"epoch": 0.20833333333333334,
"grad_norm": 0.4693131744861603,
"learning_rate": 0.00016113074204946996,
"loss": 0.8148,
"step": 60
},
{
"epoch": 0.21180555555555555,
"grad_norm": 0.48101750016212463,
"learning_rate": 0.00016042402826855126,
"loss": 0.7089,
"step": 61
},
{
"epoch": 0.2152777777777778,
"grad_norm": 0.45942020416259766,
"learning_rate": 0.00015971731448763253,
"loss": 0.7649,
"step": 62
},
{
"epoch": 0.21875,
"grad_norm": 0.46566373109817505,
"learning_rate": 0.0001590106007067138,
"loss": 0.7081,
"step": 63
},
{
"epoch": 0.2222222222222222,
"grad_norm": 0.4782656133174896,
"learning_rate": 0.00015830388692579506,
"loss": 0.6578,
"step": 64
},
{
"epoch": 0.22569444444444445,
"grad_norm": 0.4600412845611572,
"learning_rate": 0.00015759717314487635,
"loss": 0.6907,
"step": 65
},
{
"epoch": 0.22916666666666666,
"grad_norm": 0.47648143768310547,
"learning_rate": 0.00015689045936395762,
"loss": 0.7159,
"step": 66
},
{
"epoch": 0.2326388888888889,
"grad_norm": 0.4845995008945465,
"learning_rate": 0.00015618374558303886,
"loss": 0.6186,
"step": 67
},
{
"epoch": 0.2361111111111111,
"grad_norm": 0.5142258405685425,
"learning_rate": 0.00015547703180212013,
"loss": 0.7643,
"step": 68
},
{
"epoch": 0.23958333333333334,
"grad_norm": 0.5000512599945068,
"learning_rate": 0.0001547703180212014,
"loss": 0.734,
"step": 69
},
{
"epoch": 0.24305555555555555,
"grad_norm": 0.473004549741745,
"learning_rate": 0.0001540636042402827,
"loss": 0.689,
"step": 70
},
{
"epoch": 0.2465277777777778,
"grad_norm": 0.49982285499572754,
"learning_rate": 0.00015335689045936396,
"loss": 0.7162,
"step": 71
},
{
"epoch": 0.25,
"grad_norm": 0.48723116517066956,
"learning_rate": 0.00015265017667844523,
"loss": 0.605,
"step": 72
},
{
"epoch": 0.2534722222222222,
"grad_norm": 0.4668457508087158,
"learning_rate": 0.0001519434628975265,
"loss": 0.7956,
"step": 73
},
{
"epoch": 0.2569444444444444,
"grad_norm": 0.5139826536178589,
"learning_rate": 0.0001512367491166078,
"loss": 0.6877,
"step": 74
},
{
"epoch": 0.2604166666666667,
"grad_norm": 0.48903441429138184,
"learning_rate": 0.00015053003533568906,
"loss": 0.7341,
"step": 75
},
{
"epoch": 0.2638888888888889,
"grad_norm": 0.4892962872982025,
"learning_rate": 0.00014982332155477032,
"loss": 0.6833,
"step": 76
},
{
"epoch": 0.2673611111111111,
"grad_norm": 0.510981023311615,
"learning_rate": 0.0001491166077738516,
"loss": 0.6567,
"step": 77
},
{
"epoch": 0.2708333333333333,
"grad_norm": 0.48602378368377686,
"learning_rate": 0.00014840989399293289,
"loss": 0.6592,
"step": 78
},
{
"epoch": 0.2743055555555556,
"grad_norm": 0.5149515271186829,
"learning_rate": 0.00014770318021201415,
"loss": 0.7435,
"step": 79
},
{
"epoch": 0.2777777777777778,
"grad_norm": 0.4999159872531891,
"learning_rate": 0.00014699646643109542,
"loss": 0.7397,
"step": 80
},
{
"epoch": 0.28125,
"grad_norm": 0.5036943554878235,
"learning_rate": 0.0001462897526501767,
"loss": 0.6791,
"step": 81
},
{
"epoch": 0.2847222222222222,
"grad_norm": 0.4811699688434601,
"learning_rate": 0.00014558303886925796,
"loss": 0.6376,
"step": 82
},
{
"epoch": 0.2881944444444444,
"grad_norm": 0.5215135812759399,
"learning_rate": 0.00014487632508833925,
"loss": 0.6547,
"step": 83
},
{
"epoch": 0.2916666666666667,
"grad_norm": 0.5180763006210327,
"learning_rate": 0.00014416961130742052,
"loss": 0.6667,
"step": 84
},
{
"epoch": 0.2951388888888889,
"grad_norm": 0.5010943412780762,
"learning_rate": 0.00014346289752650176,
"loss": 0.6496,
"step": 85
},
{
"epoch": 0.2986111111111111,
"grad_norm": 0.47832784056663513,
"learning_rate": 0.00014275618374558303,
"loss": 0.6516,
"step": 86
},
{
"epoch": 0.3020833333333333,
"grad_norm": 0.5183102488517761,
"learning_rate": 0.00014204946996466432,
"loss": 0.6822,
"step": 87
},
{
"epoch": 0.3055555555555556,
"grad_norm": 0.5250104069709778,
"learning_rate": 0.0001413427561837456,
"loss": 0.6837,
"step": 88
},
{
"epoch": 0.3090277777777778,
"grad_norm": 0.4821203052997589,
"learning_rate": 0.00014063604240282686,
"loss": 0.6734,
"step": 89
},
{
"epoch": 0.3125,
"grad_norm": 0.5341935157775879,
"learning_rate": 0.00013992932862190812,
"loss": 0.7098,
"step": 90
},
{
"epoch": 0.3159722222222222,
"grad_norm": 0.5351920127868652,
"learning_rate": 0.0001392226148409894,
"loss": 0.6806,
"step": 91
},
{
"epoch": 0.3194444444444444,
"grad_norm": 0.5185657143592834,
"learning_rate": 0.00013851590106007068,
"loss": 0.6212,
"step": 92
},
{
"epoch": 0.3229166666666667,
"grad_norm": 0.4946865439414978,
"learning_rate": 0.00013780918727915195,
"loss": 0.7244,
"step": 93
},
{
"epoch": 0.3263888888888889,
"grad_norm": 0.51689213514328,
"learning_rate": 0.00013710247349823322,
"loss": 0.532,
"step": 94
},
{
"epoch": 0.3298611111111111,
"grad_norm": 0.5387365221977234,
"learning_rate": 0.0001363957597173145,
"loss": 0.6225,
"step": 95
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.5210251212120056,
"learning_rate": 0.00013568904593639578,
"loss": 0.6565,
"step": 96
},
{
"epoch": 0.3368055555555556,
"grad_norm": 0.5470120310783386,
"learning_rate": 0.00013498233215547705,
"loss": 0.6181,
"step": 97
},
{
"epoch": 0.3402777777777778,
"grad_norm": 0.5134613513946533,
"learning_rate": 0.00013427561837455832,
"loss": 0.5943,
"step": 98
},
{
"epoch": 0.34375,
"grad_norm": 0.5114628076553345,
"learning_rate": 0.00013356890459363958,
"loss": 0.7112,
"step": 99
},
{
"epoch": 0.3472222222222222,
"grad_norm": 0.5393586158752441,
"learning_rate": 0.00013286219081272085,
"loss": 0.6513,
"step": 100
},
{
"epoch": 0.3506944444444444,
"grad_norm": 0.5121696591377258,
"learning_rate": 0.00013215547703180215,
"loss": 0.6112,
"step": 101
},
{
"epoch": 0.3541666666666667,
"grad_norm": 0.5529887676239014,
"learning_rate": 0.0001314487632508834,
"loss": 0.704,
"step": 102
},
{
"epoch": 0.3576388888888889,
"grad_norm": 0.5504875779151917,
"learning_rate": 0.00013074204946996465,
"loss": 0.6367,
"step": 103
},
{
"epoch": 0.3611111111111111,
"grad_norm": 0.5555490255355835,
"learning_rate": 0.00013003533568904592,
"loss": 0.6988,
"step": 104
},
{
"epoch": 0.3645833333333333,
"grad_norm": 0.5071545839309692,
"learning_rate": 0.00012932862190812722,
"loss": 0.6598,
"step": 105
},
{
"epoch": 0.3680555555555556,
"grad_norm": 0.5332069396972656,
"learning_rate": 0.00012862190812720848,
"loss": 0.65,
"step": 106
},
{
"epoch": 0.3715277777777778,
"grad_norm": 0.5629200339317322,
"learning_rate": 0.00012791519434628975,
"loss": 0.699,
"step": 107
},
{
"epoch": 0.375,
"grad_norm": 0.5515280961990356,
"learning_rate": 0.00012720848056537102,
"loss": 0.6418,
"step": 108
},
{
"epoch": 0.3784722222222222,
"grad_norm": 0.527538001537323,
"learning_rate": 0.0001265017667844523,
"loss": 0.7302,
"step": 109
},
{
"epoch": 0.3819444444444444,
"grad_norm": 0.5716756582260132,
"learning_rate": 0.00012579505300353358,
"loss": 0.7443,
"step": 110
},
{
"epoch": 0.3854166666666667,
"grad_norm": 0.5540614724159241,
"learning_rate": 0.00012508833922261485,
"loss": 0.6564,
"step": 111
},
{
"epoch": 0.3888888888888889,
"grad_norm": 0.5193327069282532,
"learning_rate": 0.00012438162544169612,
"loss": 0.6783,
"step": 112
},
{
"epoch": 0.3923611111111111,
"grad_norm": 0.5474222302436829,
"learning_rate": 0.00012367491166077738,
"loss": 0.6519,
"step": 113
},
{
"epoch": 0.3958333333333333,
"grad_norm": 0.5058162808418274,
"learning_rate": 0.00012296819787985868,
"loss": 0.5785,
"step": 114
},
{
"epoch": 0.3993055555555556,
"grad_norm": 0.5280646085739136,
"learning_rate": 0.00012226148409893994,
"loss": 0.6308,
"step": 115
},
{
"epoch": 0.4027777777777778,
"grad_norm": 0.536086916923523,
"learning_rate": 0.0001215547703180212,
"loss": 0.5696,
"step": 116
},
{
"epoch": 0.40625,
"grad_norm": 0.5223795771598816,
"learning_rate": 0.00012084805653710247,
"loss": 0.5861,
"step": 117
},
{
"epoch": 0.4097222222222222,
"grad_norm": 0.5218377113342285,
"learning_rate": 0.00012014134275618376,
"loss": 0.6151,
"step": 118
},
{
"epoch": 0.4131944444444444,
"grad_norm": 0.5461903214454651,
"learning_rate": 0.00011943462897526503,
"loss": 0.585,
"step": 119
},
{
"epoch": 0.4166666666666667,
"grad_norm": 0.5451020002365112,
"learning_rate": 0.0001187279151943463,
"loss": 0.6571,
"step": 120
},
{
"epoch": 0.4201388888888889,
"grad_norm": 0.5320534706115723,
"learning_rate": 0.00011802120141342756,
"loss": 0.6191,
"step": 121
},
{
"epoch": 0.4236111111111111,
"grad_norm": 0.5509814023971558,
"learning_rate": 0.00011731448763250883,
"loss": 0.542,
"step": 122
},
{
"epoch": 0.4270833333333333,
"grad_norm": 0.5494756102561951,
"learning_rate": 0.00011660777385159012,
"loss": 0.7145,
"step": 123
},
{
"epoch": 0.4305555555555556,
"grad_norm": 0.5778016448020935,
"learning_rate": 0.00011590106007067139,
"loss": 0.6787,
"step": 124
},
{
"epoch": 0.4340277777777778,
"grad_norm": 0.5546441674232483,
"learning_rate": 0.00011519434628975265,
"loss": 0.5897,
"step": 125
},
{
"epoch": 0.4375,
"grad_norm": 0.5590822100639343,
"learning_rate": 0.00011448763250883391,
"loss": 0.6331,
"step": 126
},
{
"epoch": 0.4409722222222222,
"grad_norm": 0.5571174025535583,
"learning_rate": 0.00011378091872791521,
"loss": 0.6624,
"step": 127
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.579007625579834,
"learning_rate": 0.00011307420494699648,
"loss": 0.6377,
"step": 128
},
{
"epoch": 0.4479166666666667,
"grad_norm": 0.5956988334655762,
"learning_rate": 0.00011236749116607774,
"loss": 0.6828,
"step": 129
},
{
"epoch": 0.4513888888888889,
"grad_norm": 0.5952633023262024,
"learning_rate": 0.00011166077738515901,
"loss": 0.5603,
"step": 130
},
{
"epoch": 0.4548611111111111,
"grad_norm": 0.561600387096405,
"learning_rate": 0.00011095406360424028,
"loss": 0.6311,
"step": 131
},
{
"epoch": 0.4583333333333333,
"grad_norm": 0.5515158772468567,
"learning_rate": 0.00011024734982332157,
"loss": 0.7293,
"step": 132
},
{
"epoch": 0.4618055555555556,
"grad_norm": 0.5889860987663269,
"learning_rate": 0.00010954063604240284,
"loss": 0.6693,
"step": 133
},
{
"epoch": 0.4652777777777778,
"grad_norm": 0.5815720558166504,
"learning_rate": 0.0001088339222614841,
"loss": 0.6881,
"step": 134
},
{
"epoch": 0.46875,
"grad_norm": 0.6205659508705139,
"learning_rate": 0.00010812720848056536,
"loss": 0.6732,
"step": 135
},
{
"epoch": 0.4722222222222222,
"grad_norm": 0.5689711570739746,
"learning_rate": 0.00010742049469964666,
"loss": 0.6206,
"step": 136
},
{
"epoch": 0.4756944444444444,
"grad_norm": 0.5453592538833618,
"learning_rate": 0.00010671378091872792,
"loss": 0.6257,
"step": 137
},
{
"epoch": 0.4791666666666667,
"grad_norm": 0.5330798625946045,
"learning_rate": 0.00010600706713780919,
"loss": 0.4674,
"step": 138
},
{
"epoch": 0.4826388888888889,
"grad_norm": 0.5798875093460083,
"learning_rate": 0.00010530035335689046,
"loss": 0.6017,
"step": 139
},
{
"epoch": 0.4861111111111111,
"grad_norm": 0.5574885010719299,
"learning_rate": 0.00010459363957597175,
"loss": 0.6087,
"step": 140
},
{
"epoch": 0.4895833333333333,
"grad_norm": 0.5381941199302673,
"learning_rate": 0.00010388692579505302,
"loss": 0.6543,
"step": 141
},
{
"epoch": 0.4930555555555556,
"grad_norm": 0.5750802755355835,
"learning_rate": 0.00010318021201413429,
"loss": 0.6662,
"step": 142
},
{
"epoch": 0.4965277777777778,
"grad_norm": 0.6241365671157837,
"learning_rate": 0.00010247349823321554,
"loss": 0.5976,
"step": 143
},
{
"epoch": 0.5,
"grad_norm": 0.5461493134498596,
"learning_rate": 0.00010176678445229681,
"loss": 0.6796,
"step": 144
},
{
"epoch": 0.5034722222222222,
"grad_norm": 0.5892328023910522,
"learning_rate": 0.0001010600706713781,
"loss": 0.6166,
"step": 145
},
{
"epoch": 0.5069444444444444,
"grad_norm": 0.5670603513717651,
"learning_rate": 0.00010035335689045937,
"loss": 0.651,
"step": 146
},
{
"epoch": 0.5104166666666666,
"grad_norm": 0.5573473572731018,
"learning_rate": 9.964664310954064e-05,
"loss": 0.6268,
"step": 147
},
{
"epoch": 0.5138888888888888,
"grad_norm": 0.5641173720359802,
"learning_rate": 9.893992932862192e-05,
"loss": 0.6515,
"step": 148
},
{
"epoch": 0.5173611111111112,
"grad_norm": 0.5903286337852478,
"learning_rate": 9.823321554770319e-05,
"loss": 0.6705,
"step": 149
},
{
"epoch": 0.5208333333333334,
"grad_norm": 0.5907771587371826,
"learning_rate": 9.752650176678447e-05,
"loss": 0.6184,
"step": 150
},
{
"epoch": 0.5243055555555556,
"grad_norm": 0.5515082478523254,
"learning_rate": 9.681978798586572e-05,
"loss": 0.661,
"step": 151
},
{
"epoch": 0.5277777777777778,
"grad_norm": 0.5268746614456177,
"learning_rate": 9.611307420494699e-05,
"loss": 0.6082,
"step": 152
},
{
"epoch": 0.53125,
"grad_norm": 0.5877247452735901,
"learning_rate": 9.540636042402827e-05,
"loss": 0.5856,
"step": 153
},
{
"epoch": 0.5347222222222222,
"grad_norm": 0.5402259826660156,
"learning_rate": 9.469964664310954e-05,
"loss": 0.574,
"step": 154
},
{
"epoch": 0.5381944444444444,
"grad_norm": 0.5235267281532288,
"learning_rate": 9.399293286219082e-05,
"loss": 0.5598,
"step": 155
},
{
"epoch": 0.5416666666666666,
"grad_norm": 0.5417196750640869,
"learning_rate": 9.328621908127209e-05,
"loss": 0.6447,
"step": 156
},
{
"epoch": 0.5451388888888888,
"grad_norm": 0.6259889006614685,
"learning_rate": 9.257950530035337e-05,
"loss": 0.6227,
"step": 157
},
{
"epoch": 0.5486111111111112,
"grad_norm": 0.567189633846283,
"learning_rate": 9.187279151943463e-05,
"loss": 0.5788,
"step": 158
},
{
"epoch": 0.5520833333333334,
"grad_norm": 0.5860530138015747,
"learning_rate": 9.116607773851592e-05,
"loss": 0.5026,
"step": 159
},
{
"epoch": 0.5555555555555556,
"grad_norm": 0.5970867872238159,
"learning_rate": 9.045936395759717e-05,
"loss": 0.65,
"step": 160
},
{
"epoch": 0.5590277777777778,
"grad_norm": 0.5621108412742615,
"learning_rate": 8.975265017667845e-05,
"loss": 0.6118,
"step": 161
},
{
"epoch": 0.5625,
"grad_norm": 0.5695009827613831,
"learning_rate": 8.904593639575972e-05,
"loss": 0.5119,
"step": 162
},
{
"epoch": 0.5659722222222222,
"grad_norm": 0.5741239786148071,
"learning_rate": 8.833922261484099e-05,
"loss": 0.5273,
"step": 163
},
{
"epoch": 0.5694444444444444,
"grad_norm": 0.5961455702781677,
"learning_rate": 8.763250883392227e-05,
"loss": 0.6197,
"step": 164
},
{
"epoch": 0.5729166666666666,
"grad_norm": 0.5358597636222839,
"learning_rate": 8.692579505300353e-05,
"loss": 0.5629,
"step": 165
},
{
"epoch": 0.5763888888888888,
"grad_norm": 0.5704313516616821,
"learning_rate": 8.621908127208482e-05,
"loss": 0.5838,
"step": 166
},
{
"epoch": 0.5798611111111112,
"grad_norm": 0.6421118378639221,
"learning_rate": 8.551236749116608e-05,
"loss": 0.5964,
"step": 167
},
{
"epoch": 0.5833333333333334,
"grad_norm": 0.5588890314102173,
"learning_rate": 8.480565371024736e-05,
"loss": 0.4957,
"step": 168
},
{
"epoch": 0.5868055555555556,
"grad_norm": 0.5954290628433228,
"learning_rate": 8.409893992932862e-05,
"loss": 0.6364,
"step": 169
},
{
"epoch": 0.5902777777777778,
"grad_norm": 0.5738086700439453,
"learning_rate": 8.33922261484099e-05,
"loss": 0.5661,
"step": 170
},
{
"epoch": 0.59375,
"grad_norm": 0.6033700704574585,
"learning_rate": 8.268551236749117e-05,
"loss": 0.6128,
"step": 171
},
{
"epoch": 0.5972222222222222,
"grad_norm": 0.6137632727622986,
"learning_rate": 8.197879858657245e-05,
"loss": 0.6271,
"step": 172
},
{
"epoch": 0.6006944444444444,
"grad_norm": 0.5929612517356873,
"learning_rate": 8.127208480565371e-05,
"loss": 0.6328,
"step": 173
},
{
"epoch": 0.6041666666666666,
"grad_norm": 0.5914517641067505,
"learning_rate": 8.056537102473498e-05,
"loss": 0.6406,
"step": 174
},
{
"epoch": 0.6076388888888888,
"grad_norm": 0.5911087989807129,
"learning_rate": 7.985865724381626e-05,
"loss": 0.5432,
"step": 175
},
{
"epoch": 0.6111111111111112,
"grad_norm": 0.6051996350288391,
"learning_rate": 7.915194346289753e-05,
"loss": 0.576,
"step": 176
},
{
"epoch": 0.6145833333333334,
"grad_norm": 0.5913227200508118,
"learning_rate": 7.844522968197881e-05,
"loss": 0.6012,
"step": 177
},
{
"epoch": 0.6180555555555556,
"grad_norm": 0.5444892048835754,
"learning_rate": 7.773851590106007e-05,
"loss": 0.5324,
"step": 178
},
{
"epoch": 0.6215277777777778,
"grad_norm": 0.5951313972473145,
"learning_rate": 7.703180212014135e-05,
"loss": 0.6932,
"step": 179
},
{
"epoch": 0.625,
"grad_norm": 0.5791239142417908,
"learning_rate": 7.632508833922261e-05,
"loss": 0.6144,
"step": 180
},
{
"epoch": 0.6284722222222222,
"grad_norm": 0.5847236514091492,
"learning_rate": 7.56183745583039e-05,
"loss": 0.5894,
"step": 181
},
{
"epoch": 0.6319444444444444,
"grad_norm": 0.6364589929580688,
"learning_rate": 7.491166077738516e-05,
"loss": 0.5187,
"step": 182
},
{
"epoch": 0.6354166666666666,
"grad_norm": 0.572428822517395,
"learning_rate": 7.420494699646644e-05,
"loss": 0.5653,
"step": 183
},
{
"epoch": 0.6388888888888888,
"grad_norm": 0.5649181008338928,
"learning_rate": 7.349823321554771e-05,
"loss": 0.5586,
"step": 184
},
{
"epoch": 0.6423611111111112,
"grad_norm": 0.5847340226173401,
"learning_rate": 7.279151943462898e-05,
"loss": 0.6389,
"step": 185
},
{
"epoch": 0.6458333333333334,
"grad_norm": 0.5655092597007751,
"learning_rate": 7.208480565371026e-05,
"loss": 0.5809,
"step": 186
},
{
"epoch": 0.6493055555555556,
"grad_norm": 0.5791696310043335,
"learning_rate": 7.137809187279151e-05,
"loss": 0.6295,
"step": 187
},
{
"epoch": 0.6527777777777778,
"grad_norm": 0.6031994223594666,
"learning_rate": 7.06713780918728e-05,
"loss": 0.5962,
"step": 188
},
{
"epoch": 0.65625,
"grad_norm": 0.6025336980819702,
"learning_rate": 6.996466431095406e-05,
"loss": 0.6424,
"step": 189
},
{
"epoch": 0.6597222222222222,
"grad_norm": 0.6521866917610168,
"learning_rate": 6.925795053003534e-05,
"loss": 0.5857,
"step": 190
},
{
"epoch": 0.6631944444444444,
"grad_norm": 0.5835008025169373,
"learning_rate": 6.855123674911661e-05,
"loss": 0.5807,
"step": 191
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.5866998434066772,
"learning_rate": 6.784452296819789e-05,
"loss": 0.5653,
"step": 192
},
{
"epoch": 0.6701388888888888,
"grad_norm": 0.5725639462471008,
"learning_rate": 6.713780918727916e-05,
"loss": 0.4654,
"step": 193
},
{
"epoch": 0.6736111111111112,
"grad_norm": 0.6040775775909424,
"learning_rate": 6.643109540636043e-05,
"loss": 0.6563,
"step": 194
},
{
"epoch": 0.6770833333333334,
"grad_norm": 0.6573815941810608,
"learning_rate": 6.57243816254417e-05,
"loss": 0.5836,
"step": 195
},
{
"epoch": 0.6805555555555556,
"grad_norm": 0.5753028392791748,
"learning_rate": 6.501766784452296e-05,
"loss": 0.6142,
"step": 196
},
{
"epoch": 0.6840277777777778,
"grad_norm": 0.6084535717964172,
"learning_rate": 6.431095406360424e-05,
"loss": 0.5592,
"step": 197
},
{
"epoch": 0.6875,
"grad_norm": 0.6668769717216492,
"learning_rate": 6.360424028268551e-05,
"loss": 0.563,
"step": 198
},
{
"epoch": 0.6909722222222222,
"grad_norm": 0.620083212852478,
"learning_rate": 6.289752650176679e-05,
"loss": 0.4906,
"step": 199
},
{
"epoch": 0.6944444444444444,
"grad_norm": 0.6372251510620117,
"learning_rate": 6.219081272084806e-05,
"loss": 0.5754,
"step": 200
},
{
"epoch": 0.6979166666666666,
"grad_norm": 0.602571964263916,
"learning_rate": 6.148409893992934e-05,
"loss": 0.5631,
"step": 201
},
{
"epoch": 0.7013888888888888,
"grad_norm": 0.610701322555542,
"learning_rate": 6.07773851590106e-05,
"loss": 0.6417,
"step": 202
},
{
"epoch": 0.7048611111111112,
"grad_norm": 0.6244217753410339,
"learning_rate": 6.007067137809188e-05,
"loss": 0.6194,
"step": 203
},
{
"epoch": 0.7083333333333334,
"grad_norm": 0.6154200434684753,
"learning_rate": 5.936395759717315e-05,
"loss": 0.5724,
"step": 204
},
{
"epoch": 0.7118055555555556,
"grad_norm": 0.5975745916366577,
"learning_rate": 5.8657243816254415e-05,
"loss": 0.6117,
"step": 205
},
{
"epoch": 0.7152777777777778,
"grad_norm": 0.6204216480255127,
"learning_rate": 5.7950530035335696e-05,
"loss": 0.6153,
"step": 206
},
{
"epoch": 0.71875,
"grad_norm": 0.5886074900627136,
"learning_rate": 5.724381625441696e-05,
"loss": 0.4968,
"step": 207
},
{
"epoch": 0.7222222222222222,
"grad_norm": 0.5940162539482117,
"learning_rate": 5.653710247349824e-05,
"loss": 0.6544,
"step": 208
},
{
"epoch": 0.7256944444444444,
"grad_norm": 0.596792459487915,
"learning_rate": 5.5830388692579505e-05,
"loss": 0.5685,
"step": 209
},
{
"epoch": 0.7291666666666666,
"grad_norm": 0.5985198020935059,
"learning_rate": 5.5123674911660786e-05,
"loss": 0.5795,
"step": 210
},
{
"epoch": 0.7326388888888888,
"grad_norm": 0.5998839735984802,
"learning_rate": 5.441696113074205e-05,
"loss": 0.5229,
"step": 211
},
{
"epoch": 0.7361111111111112,
"grad_norm": 0.5900275707244873,
"learning_rate": 5.371024734982333e-05,
"loss": 0.5882,
"step": 212
},
{
"epoch": 0.7395833333333334,
"grad_norm": 0.5826575756072998,
"learning_rate": 5.3003533568904595e-05,
"loss": 0.4709,
"step": 213
},
{
"epoch": 0.7430555555555556,
"grad_norm": 0.5909337997436523,
"learning_rate": 5.2296819787985876e-05,
"loss": 0.5528,
"step": 214
},
{
"epoch": 0.7465277777777778,
"grad_norm": 0.5588683485984802,
"learning_rate": 5.1590106007067144e-05,
"loss": 0.5373,
"step": 215
},
{
"epoch": 0.75,
"grad_norm": 0.5555421710014343,
"learning_rate": 5.0883392226148405e-05,
"loss": 0.5779,
"step": 216
},
{
"epoch": 0.7534722222222222,
"grad_norm": 0.5925477743148804,
"learning_rate": 5.0176678445229686e-05,
"loss": 0.5896,
"step": 217
},
{
"epoch": 0.7569444444444444,
"grad_norm": 0.6183862090110779,
"learning_rate": 4.946996466431096e-05,
"loss": 0.6245,
"step": 218
},
{
"epoch": 0.7604166666666666,
"grad_norm": 0.5807591080665588,
"learning_rate": 4.8763250883392234e-05,
"loss": 0.5509,
"step": 219
},
{
"epoch": 0.7638888888888888,
"grad_norm": 0.5723608136177063,
"learning_rate": 4.8056537102473495e-05,
"loss": 0.5927,
"step": 220
},
{
"epoch": 0.7673611111111112,
"grad_norm": 0.5748648643493652,
"learning_rate": 4.734982332155477e-05,
"loss": 0.5009,
"step": 221
},
{
"epoch": 0.7708333333333334,
"grad_norm": 0.5595970749855042,
"learning_rate": 4.664310954063604e-05,
"loss": 0.5225,
"step": 222
},
{
"epoch": 0.7743055555555556,
"grad_norm": 0.5699341893196106,
"learning_rate": 4.593639575971732e-05,
"loss": 0.5945,
"step": 223
},
{
"epoch": 0.7777777777777778,
"grad_norm": 0.5965125560760498,
"learning_rate": 4.5229681978798585e-05,
"loss": 0.6039,
"step": 224
},
{
"epoch": 0.78125,
"grad_norm": 0.601673424243927,
"learning_rate": 4.452296819787986e-05,
"loss": 0.6116,
"step": 225
},
{
"epoch": 0.7847222222222222,
"grad_norm": 0.6311600804328918,
"learning_rate": 4.381625441696113e-05,
"loss": 0.6356,
"step": 226
},
{
"epoch": 0.7881944444444444,
"grad_norm": 0.6146146059036255,
"learning_rate": 4.310954063604241e-05,
"loss": 0.5925,
"step": 227
},
{
"epoch": 0.7916666666666666,
"grad_norm": 0.5608177781105042,
"learning_rate": 4.240282685512368e-05,
"loss": 0.4546,
"step": 228
},
{
"epoch": 0.7951388888888888,
"grad_norm": 0.6077985167503357,
"learning_rate": 4.169611307420495e-05,
"loss": 0.6111,
"step": 229
},
{
"epoch": 0.7986111111111112,
"grad_norm": 0.5868244171142578,
"learning_rate": 4.0989399293286223e-05,
"loss": 0.5945,
"step": 230
},
{
"epoch": 0.8020833333333334,
"grad_norm": 0.587386965751648,
"learning_rate": 4.028268551236749e-05,
"loss": 0.5294,
"step": 231
},
{
"epoch": 0.8055555555555556,
"grad_norm": 0.5953108668327332,
"learning_rate": 3.9575971731448765e-05,
"loss": 0.5621,
"step": 232
},
{
"epoch": 0.8090277777777778,
"grad_norm": 0.6050392985343933,
"learning_rate": 3.886925795053003e-05,
"loss": 0.5374,
"step": 233
},
{
"epoch": 0.8125,
"grad_norm": 0.6083176732063293,
"learning_rate": 3.816254416961131e-05,
"loss": 0.5843,
"step": 234
},
{
"epoch": 0.8159722222222222,
"grad_norm": 0.5946184992790222,
"learning_rate": 3.745583038869258e-05,
"loss": 0.5049,
"step": 235
},
{
"epoch": 0.8194444444444444,
"grad_norm": 0.5910969972610474,
"learning_rate": 3.6749116607773855e-05,
"loss": 0.5749,
"step": 236
},
{
"epoch": 0.8229166666666666,
"grad_norm": 0.5982712507247925,
"learning_rate": 3.604240282685513e-05,
"loss": 0.5882,
"step": 237
},
{
"epoch": 0.8263888888888888,
"grad_norm": 0.6084917783737183,
"learning_rate": 3.53356890459364e-05,
"loss": 0.5169,
"step": 238
},
{
"epoch": 0.8298611111111112,
"grad_norm": 0.6236574649810791,
"learning_rate": 3.462897526501767e-05,
"loss": 0.6085,
"step": 239
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.592654824256897,
"learning_rate": 3.3922261484098945e-05,
"loss": 0.5525,
"step": 240
},
{
"epoch": 0.8368055555555556,
"grad_norm": 0.6017850041389465,
"learning_rate": 3.321554770318021e-05,
"loss": 0.636,
"step": 241
},
{
"epoch": 0.8402777777777778,
"grad_norm": 0.5793926119804382,
"learning_rate": 3.250883392226148e-05,
"loss": 0.6003,
"step": 242
},
{
"epoch": 0.84375,
"grad_norm": 0.5898053050041199,
"learning_rate": 3.1802120141342755e-05,
"loss": 0.6925,
"step": 243
},
{
"epoch": 0.8472222222222222,
"grad_norm": 0.6041194796562195,
"learning_rate": 3.109540636042403e-05,
"loss": 0.5509,
"step": 244
},
{
"epoch": 0.8506944444444444,
"grad_norm": 0.5889320373535156,
"learning_rate": 3.03886925795053e-05,
"loss": 0.5377,
"step": 245
},
{
"epoch": 0.8541666666666666,
"grad_norm": 0.65882408618927,
"learning_rate": 2.9681978798586574e-05,
"loss": 0.6909,
"step": 246
},
{
"epoch": 0.8576388888888888,
"grad_norm": 0.5652550458908081,
"learning_rate": 2.8975265017667848e-05,
"loss": 0.6553,
"step": 247
},
{
"epoch": 0.8611111111111112,
"grad_norm": 0.6159343123435974,
"learning_rate": 2.826855123674912e-05,
"loss": 0.4816,
"step": 248
},
{
"epoch": 0.8645833333333334,
"grad_norm": 0.5917626023292542,
"learning_rate": 2.7561837455830393e-05,
"loss": 0.6042,
"step": 249
},
{
"epoch": 0.8680555555555556,
"grad_norm": 0.5773218870162964,
"learning_rate": 2.6855123674911664e-05,
"loss": 0.5723,
"step": 250
},
{
"epoch": 0.8715277777777778,
"grad_norm": 0.5936452150344849,
"learning_rate": 2.6148409893992938e-05,
"loss": 0.6071,
"step": 251
},
{
"epoch": 0.875,
"grad_norm": 0.6094207763671875,
"learning_rate": 2.5441696113074202e-05,
"loss": 0.5406,
"step": 252
},
{
"epoch": 0.8784722222222222,
"grad_norm": 0.6131781935691833,
"learning_rate": 2.473498233215548e-05,
"loss": 0.562,
"step": 253
},
{
"epoch": 0.8819444444444444,
"grad_norm": 0.5903745889663696,
"learning_rate": 2.4028268551236747e-05,
"loss": 0.5841,
"step": 254
},
{
"epoch": 0.8854166666666666,
"grad_norm": 0.5778252482414246,
"learning_rate": 2.332155477031802e-05,
"loss": 0.6071,
"step": 255
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.6057181358337402,
"learning_rate": 2.2614840989399292e-05,
"loss": 0.5794,
"step": 256
},
{
"epoch": 0.8923611111111112,
"grad_norm": 0.5691949129104614,
"learning_rate": 2.1908127208480567e-05,
"loss": 0.5634,
"step": 257
},
{
"epoch": 0.8958333333333334,
"grad_norm": 0.6154379844665527,
"learning_rate": 2.120141342756184e-05,
"loss": 0.5389,
"step": 258
},
{
"epoch": 0.8993055555555556,
"grad_norm": 0.5671730637550354,
"learning_rate": 2.0494699646643112e-05,
"loss": 0.6045,
"step": 259
},
{
"epoch": 0.9027777777777778,
"grad_norm": 0.6344681978225708,
"learning_rate": 1.9787985865724383e-05,
"loss": 0.4922,
"step": 260
},
{
"epoch": 0.90625,
"grad_norm": 0.584071934223175,
"learning_rate": 1.9081272084805653e-05,
"loss": 0.5283,
"step": 261
},
{
"epoch": 0.9097222222222222,
"grad_norm": 0.5744527578353882,
"learning_rate": 1.8374558303886928e-05,
"loss": 0.51,
"step": 262
},
{
"epoch": 0.9131944444444444,
"grad_norm": 0.5983773469924927,
"learning_rate": 1.76678445229682e-05,
"loss": 0.619,
"step": 263
},
{
"epoch": 0.9166666666666666,
"grad_norm": 0.5973845720291138,
"learning_rate": 1.6961130742049473e-05,
"loss": 0.5131,
"step": 264
},
{
"epoch": 0.9201388888888888,
"grad_norm": 0.6193533539772034,
"learning_rate": 1.625441696113074e-05,
"loss": 0.577,
"step": 265
},
{
"epoch": 0.9236111111111112,
"grad_norm": 0.6047850251197815,
"learning_rate": 1.5547703180212014e-05,
"loss": 0.6277,
"step": 266
},
{
"epoch": 0.9270833333333334,
"grad_norm": 0.6007250547409058,
"learning_rate": 1.4840989399293287e-05,
"loss": 0.6026,
"step": 267
},
{
"epoch": 0.9305555555555556,
"grad_norm": 0.5738986730575562,
"learning_rate": 1.413427561837456e-05,
"loss": 0.5842,
"step": 268
},
{
"epoch": 0.9340277777777778,
"grad_norm": 0.5671294927597046,
"learning_rate": 1.3427561837455832e-05,
"loss": 0.4875,
"step": 269
},
{
"epoch": 0.9375,
"grad_norm": 0.5837607383728027,
"learning_rate": 1.2720848056537101e-05,
"loss": 0.6032,
"step": 270
},
{
"epoch": 0.9409722222222222,
"grad_norm": 0.6210164427757263,
"learning_rate": 1.2014134275618374e-05,
"loss": 0.6143,
"step": 271
},
{
"epoch": 0.9444444444444444,
"grad_norm": 0.5613874197006226,
"learning_rate": 1.1307420494699646e-05,
"loss": 0.5193,
"step": 272
},
{
"epoch": 0.9479166666666666,
"grad_norm": 0.5788601636886597,
"learning_rate": 1.060070671378092e-05,
"loss": 0.5993,
"step": 273
},
{
"epoch": 0.9513888888888888,
"grad_norm": 0.5956071615219116,
"learning_rate": 9.893992932862191e-06,
"loss": 0.6244,
"step": 274
},
{
"epoch": 0.9548611111111112,
"grad_norm": 0.5741413235664368,
"learning_rate": 9.187279151943464e-06,
"loss": 0.5912,
"step": 275
},
{
"epoch": 0.9583333333333334,
"grad_norm": 0.5746176242828369,
"learning_rate": 8.480565371024736e-06,
"loss": 0.4959,
"step": 276
},
{
"epoch": 0.9618055555555556,
"grad_norm": 0.6048396825790405,
"learning_rate": 7.773851590106007e-06,
"loss": 0.5358,
"step": 277
},
{
"epoch": 0.9652777777777778,
"grad_norm": 0.5789562463760376,
"learning_rate": 7.06713780918728e-06,
"loss": 0.526,
"step": 278
},
{
"epoch": 0.96875,
"grad_norm": 0.623224139213562,
"learning_rate": 6.3604240282685506e-06,
"loss": 0.6296,
"step": 279
},
{
"epoch": 0.9722222222222222,
"grad_norm": 0.5628058910369873,
"learning_rate": 5.653710247349823e-06,
"loss": 0.519,
"step": 280
},
{
"epoch": 0.9756944444444444,
"grad_norm": 0.574334442615509,
"learning_rate": 4.946996466431096e-06,
"loss": 0.5241,
"step": 281
},
{
"epoch": 0.9791666666666666,
"grad_norm": 0.6212002038955688,
"learning_rate": 4.240282685512368e-06,
"loss": 0.6307,
"step": 282
},
{
"epoch": 0.9826388888888888,
"grad_norm": 0.6209303140640259,
"learning_rate": 3.53356890459364e-06,
"loss": 0.5419,
"step": 283
},
{
"epoch": 0.9861111111111112,
"grad_norm": 0.6056732535362244,
"learning_rate": 2.8268551236749116e-06,
"loss": 0.6091,
"step": 284
},
{
"epoch": 0.9895833333333334,
"grad_norm": 0.632924497127533,
"learning_rate": 2.120141342756184e-06,
"loss": 0.6159,
"step": 285
},
{
"epoch": 0.9930555555555556,
"grad_norm": 0.6058059334754944,
"learning_rate": 1.4134275618374558e-06,
"loss": 0.5805,
"step": 286
},
{
"epoch": 0.9965277777777778,
"grad_norm": 0.6084165573120117,
"learning_rate": 7.067137809187279e-07,
"loss": 0.6208,
"step": 287
},
{
"epoch": 1.0,
"grad_norm": 0.6049368977546692,
"learning_rate": 0.0,
"loss": 0.4819,
"step": 288
}
],
"logging_steps": 1,
"max_steps": 288,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.5964978416533504e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}