lesso's picture
Training in progress, step 150, checkpoint
8de4256 verified
raw
history blame
27.9 kB
{
"best_metric": 0.2286633849143982,
"best_model_checkpoint": "miner_id_24/checkpoint-100",
"epoch": 0.21770682148040638,
"eval_steps": 50,
"global_step": 150,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001451378809869376,
"grad_norm": 0.11486738920211792,
"learning_rate": 1.0110000000000001e-05,
"loss": 0.2249,
"step": 1
},
{
"epoch": 0.001451378809869376,
"eval_loss": 0.27678582072257996,
"eval_runtime": 142.974,
"eval_samples_per_second": 2.035,
"eval_steps_per_second": 0.511,
"step": 1
},
{
"epoch": 0.002902757619738752,
"grad_norm": 0.11387992650270462,
"learning_rate": 2.0220000000000003e-05,
"loss": 0.4195,
"step": 2
},
{
"epoch": 0.0043541364296081275,
"grad_norm": 0.1770201325416565,
"learning_rate": 3.033e-05,
"loss": 0.3746,
"step": 3
},
{
"epoch": 0.005805515239477504,
"grad_norm": 0.15089277923107147,
"learning_rate": 4.0440000000000006e-05,
"loss": 0.3461,
"step": 4
},
{
"epoch": 0.00725689404934688,
"grad_norm": 0.11748471856117249,
"learning_rate": 5.055e-05,
"loss": 0.1969,
"step": 5
},
{
"epoch": 0.008708272859216255,
"grad_norm": 0.1325271725654602,
"learning_rate": 6.066e-05,
"loss": 0.2835,
"step": 6
},
{
"epoch": 0.010159651669085631,
"grad_norm": 0.14989273250102997,
"learning_rate": 7.077e-05,
"loss": 0.44,
"step": 7
},
{
"epoch": 0.011611030478955007,
"grad_norm": 0.1532517969608307,
"learning_rate": 8.088000000000001e-05,
"loss": 0.4155,
"step": 8
},
{
"epoch": 0.013062409288824383,
"grad_norm": 0.12538368999958038,
"learning_rate": 9.099000000000001e-05,
"loss": 0.3613,
"step": 9
},
{
"epoch": 0.01451378809869376,
"grad_norm": 0.13840965926647186,
"learning_rate": 0.0001011,
"loss": 0.7261,
"step": 10
},
{
"epoch": 0.015965166908563134,
"grad_norm": 0.1670045405626297,
"learning_rate": 0.00010056789473684211,
"loss": 0.5663,
"step": 11
},
{
"epoch": 0.01741654571843251,
"grad_norm": 0.13095274567604065,
"learning_rate": 0.00010003578947368421,
"loss": 0.3288,
"step": 12
},
{
"epoch": 0.018867924528301886,
"grad_norm": 0.19772979617118835,
"learning_rate": 9.950368421052632e-05,
"loss": 0.613,
"step": 13
},
{
"epoch": 0.020319303338171262,
"grad_norm": 0.16937138140201569,
"learning_rate": 9.897157894736842e-05,
"loss": 0.4953,
"step": 14
},
{
"epoch": 0.02177068214804064,
"grad_norm": 0.1333189755678177,
"learning_rate": 9.843947368421053e-05,
"loss": 0.2195,
"step": 15
},
{
"epoch": 0.023222060957910014,
"grad_norm": 0.1428212672472,
"learning_rate": 9.790736842105264e-05,
"loss": 0.2334,
"step": 16
},
{
"epoch": 0.02467343976777939,
"grad_norm": 0.14330728352069855,
"learning_rate": 9.737526315789474e-05,
"loss": 0.3142,
"step": 17
},
{
"epoch": 0.026124818577648767,
"grad_norm": 0.1493055373430252,
"learning_rate": 9.684315789473684e-05,
"loss": 0.2551,
"step": 18
},
{
"epoch": 0.027576197387518143,
"grad_norm": 0.16706344485282898,
"learning_rate": 9.631105263157895e-05,
"loss": 0.5574,
"step": 19
},
{
"epoch": 0.02902757619738752,
"grad_norm": 0.11606133729219437,
"learning_rate": 9.577894736842105e-05,
"loss": 0.1948,
"step": 20
},
{
"epoch": 0.030478955007256895,
"grad_norm": 0.0992949903011322,
"learning_rate": 9.524684210526317e-05,
"loss": 0.1042,
"step": 21
},
{
"epoch": 0.03193033381712627,
"grad_norm": 0.15178389847278595,
"learning_rate": 9.471473684210526e-05,
"loss": 0.4452,
"step": 22
},
{
"epoch": 0.033381712626995644,
"grad_norm": 0.15305830538272858,
"learning_rate": 9.418263157894737e-05,
"loss": 0.4336,
"step": 23
},
{
"epoch": 0.03483309143686502,
"grad_norm": 0.21439702808856964,
"learning_rate": 9.365052631578948e-05,
"loss": 0.346,
"step": 24
},
{
"epoch": 0.036284470246734396,
"grad_norm": 0.1342068761587143,
"learning_rate": 9.311842105263157e-05,
"loss": 0.3454,
"step": 25
},
{
"epoch": 0.03773584905660377,
"grad_norm": 0.1589202582836151,
"learning_rate": 9.258631578947368e-05,
"loss": 0.4931,
"step": 26
},
{
"epoch": 0.03918722786647315,
"grad_norm": 0.16286569833755493,
"learning_rate": 9.20542105263158e-05,
"loss": 0.5843,
"step": 27
},
{
"epoch": 0.040638606676342524,
"grad_norm": 0.11265375465154648,
"learning_rate": 9.15221052631579e-05,
"loss": 0.2569,
"step": 28
},
{
"epoch": 0.0420899854862119,
"grad_norm": 0.18966278433799744,
"learning_rate": 9.099000000000001e-05,
"loss": 0.4922,
"step": 29
},
{
"epoch": 0.04354136429608128,
"grad_norm": 0.2602640688419342,
"learning_rate": 9.045789473684212e-05,
"loss": 0.3959,
"step": 30
},
{
"epoch": 0.04499274310595065,
"grad_norm": 0.12830835580825806,
"learning_rate": 8.992578947368421e-05,
"loss": 0.3903,
"step": 31
},
{
"epoch": 0.04644412191582003,
"grad_norm": 0.10974445939064026,
"learning_rate": 8.939368421052632e-05,
"loss": 0.2795,
"step": 32
},
{
"epoch": 0.047895500725689405,
"grad_norm": 0.16427350044250488,
"learning_rate": 8.886157894736841e-05,
"loss": 0.4776,
"step": 33
},
{
"epoch": 0.04934687953555878,
"grad_norm": 0.12569822371006012,
"learning_rate": 8.832947368421054e-05,
"loss": 0.2615,
"step": 34
},
{
"epoch": 0.05079825834542816,
"grad_norm": 0.10434585809707642,
"learning_rate": 8.779736842105264e-05,
"loss": 0.2248,
"step": 35
},
{
"epoch": 0.05224963715529753,
"grad_norm": 0.1004101112484932,
"learning_rate": 8.726526315789474e-05,
"loss": 0.2055,
"step": 36
},
{
"epoch": 0.05370101596516691,
"grad_norm": 0.12491065263748169,
"learning_rate": 8.673315789473685e-05,
"loss": 0.2482,
"step": 37
},
{
"epoch": 0.055152394775036286,
"grad_norm": 0.11319632083177567,
"learning_rate": 8.620105263157896e-05,
"loss": 0.2185,
"step": 38
},
{
"epoch": 0.05660377358490566,
"grad_norm": 0.15422259271144867,
"learning_rate": 8.566894736842105e-05,
"loss": 0.3448,
"step": 39
},
{
"epoch": 0.05805515239477504,
"grad_norm": 0.15866440534591675,
"learning_rate": 8.513684210526316e-05,
"loss": 0.6739,
"step": 40
},
{
"epoch": 0.059506531204644414,
"grad_norm": 0.1435173898935318,
"learning_rate": 8.460473684210527e-05,
"loss": 0.2159,
"step": 41
},
{
"epoch": 0.06095791001451379,
"grad_norm": 0.21297410130500793,
"learning_rate": 8.407263157894738e-05,
"loss": 0.3077,
"step": 42
},
{
"epoch": 0.062409288824383166,
"grad_norm": 0.1251886934041977,
"learning_rate": 8.354052631578948e-05,
"loss": 0.2113,
"step": 43
},
{
"epoch": 0.06386066763425254,
"grad_norm": 0.1815638542175293,
"learning_rate": 8.300842105263158e-05,
"loss": 0.6564,
"step": 44
},
{
"epoch": 0.06531204644412192,
"grad_norm": 0.2576989531517029,
"learning_rate": 8.247631578947369e-05,
"loss": 0.5205,
"step": 45
},
{
"epoch": 0.06676342525399129,
"grad_norm": 0.27275457978248596,
"learning_rate": 8.19442105263158e-05,
"loss": 0.9175,
"step": 46
},
{
"epoch": 0.06821480406386067,
"grad_norm": 0.26641345024108887,
"learning_rate": 8.141210526315789e-05,
"loss": 0.812,
"step": 47
},
{
"epoch": 0.06966618287373004,
"grad_norm": 0.2708298861980438,
"learning_rate": 8.088000000000001e-05,
"loss": 0.6076,
"step": 48
},
{
"epoch": 0.07111756168359942,
"grad_norm": 0.5135082602500916,
"learning_rate": 8.03478947368421e-05,
"loss": 0.9538,
"step": 49
},
{
"epoch": 0.07256894049346879,
"grad_norm": 0.46205997467041016,
"learning_rate": 7.981578947368421e-05,
"loss": 1.1091,
"step": 50
},
{
"epoch": 0.07256894049346879,
"eval_loss": 0.24128209054470062,
"eval_runtime": 146.2789,
"eval_samples_per_second": 1.989,
"eval_steps_per_second": 0.499,
"step": 50
},
{
"epoch": 0.07402031930333818,
"grad_norm": 0.15042653679847717,
"learning_rate": 7.928368421052632e-05,
"loss": 0.5749,
"step": 51
},
{
"epoch": 0.07547169811320754,
"grad_norm": 0.14070290327072144,
"learning_rate": 7.875157894736842e-05,
"loss": 0.3815,
"step": 52
},
{
"epoch": 0.07692307692307693,
"grad_norm": 0.12616680562496185,
"learning_rate": 7.821947368421053e-05,
"loss": 0.3485,
"step": 53
},
{
"epoch": 0.0783744557329463,
"grad_norm": 0.1807371824979782,
"learning_rate": 7.768736842105263e-05,
"loss": 0.6332,
"step": 54
},
{
"epoch": 0.07982583454281568,
"grad_norm": 0.1610403060913086,
"learning_rate": 7.715526315789474e-05,
"loss": 0.4334,
"step": 55
},
{
"epoch": 0.08127721335268505,
"grad_norm": 0.16740575432777405,
"learning_rate": 7.662315789473685e-05,
"loss": 0.5169,
"step": 56
},
{
"epoch": 0.08272859216255443,
"grad_norm": 0.13583853840827942,
"learning_rate": 7.609105263157895e-05,
"loss": 0.3515,
"step": 57
},
{
"epoch": 0.0841799709724238,
"grad_norm": 0.20852766931056976,
"learning_rate": 7.555894736842105e-05,
"loss": 0.8417,
"step": 58
},
{
"epoch": 0.08563134978229318,
"grad_norm": 0.12840092182159424,
"learning_rate": 7.502684210526316e-05,
"loss": 0.3527,
"step": 59
},
{
"epoch": 0.08708272859216255,
"grad_norm": 0.20794260501861572,
"learning_rate": 7.449473684210526e-05,
"loss": 0.8209,
"step": 60
},
{
"epoch": 0.08853410740203194,
"grad_norm": 0.15332041680812836,
"learning_rate": 7.396263157894738e-05,
"loss": 0.5123,
"step": 61
},
{
"epoch": 0.0899854862119013,
"grad_norm": 0.12152067571878433,
"learning_rate": 7.343052631578949e-05,
"loss": 0.5442,
"step": 62
},
{
"epoch": 0.09143686502177069,
"grad_norm": 0.15734325349330902,
"learning_rate": 7.289842105263158e-05,
"loss": 0.4926,
"step": 63
},
{
"epoch": 0.09288824383164006,
"grad_norm": 0.10759459435939789,
"learning_rate": 7.236631578947369e-05,
"loss": 0.2285,
"step": 64
},
{
"epoch": 0.09433962264150944,
"grad_norm": 0.11365488916635513,
"learning_rate": 7.183421052631579e-05,
"loss": 0.3032,
"step": 65
},
{
"epoch": 0.09579100145137881,
"grad_norm": 0.13057614862918854,
"learning_rate": 7.13021052631579e-05,
"loss": 0.4789,
"step": 66
},
{
"epoch": 0.09724238026124818,
"grad_norm": 0.1509314924478531,
"learning_rate": 7.077e-05,
"loss": 0.5571,
"step": 67
},
{
"epoch": 0.09869375907111756,
"grad_norm": 0.16016532480716705,
"learning_rate": 7.023789473684211e-05,
"loss": 0.2708,
"step": 68
},
{
"epoch": 0.10014513788098693,
"grad_norm": 0.2185145616531372,
"learning_rate": 6.970578947368422e-05,
"loss": 0.4046,
"step": 69
},
{
"epoch": 0.10159651669085631,
"grad_norm": 0.14823275804519653,
"learning_rate": 6.917368421052633e-05,
"loss": 0.5717,
"step": 70
},
{
"epoch": 0.10304789550072568,
"grad_norm": 0.16261740028858185,
"learning_rate": 6.864157894736842e-05,
"loss": 0.3313,
"step": 71
},
{
"epoch": 0.10449927431059507,
"grad_norm": 0.1639692485332489,
"learning_rate": 6.810947368421053e-05,
"loss": 0.7474,
"step": 72
},
{
"epoch": 0.10595065312046444,
"grad_norm": 0.11137572675943375,
"learning_rate": 6.757736842105264e-05,
"loss": 0.3237,
"step": 73
},
{
"epoch": 0.10740203193033382,
"grad_norm": 0.09403427690267563,
"learning_rate": 6.704526315789473e-05,
"loss": 0.213,
"step": 74
},
{
"epoch": 0.10885341074020319,
"grad_norm": 0.1198626309633255,
"learning_rate": 6.651315789473685e-05,
"loss": 0.3147,
"step": 75
},
{
"epoch": 0.11030478955007257,
"grad_norm": 0.18822187185287476,
"learning_rate": 6.598105263157895e-05,
"loss": 0.2479,
"step": 76
},
{
"epoch": 0.11175616835994194,
"grad_norm": 0.08338137716054916,
"learning_rate": 6.544894736842106e-05,
"loss": 0.1284,
"step": 77
},
{
"epoch": 0.11320754716981132,
"grad_norm": 0.11395663768053055,
"learning_rate": 6.491684210526317e-05,
"loss": 0.304,
"step": 78
},
{
"epoch": 0.11465892597968069,
"grad_norm": 0.12172822654247284,
"learning_rate": 6.438473684210526e-05,
"loss": 0.5225,
"step": 79
},
{
"epoch": 0.11611030478955008,
"grad_norm": 0.18436570465564728,
"learning_rate": 6.385263157894737e-05,
"loss": 1.1369,
"step": 80
},
{
"epoch": 0.11756168359941944,
"grad_norm": 0.08676780015230179,
"learning_rate": 6.332052631578948e-05,
"loss": 0.2241,
"step": 81
},
{
"epoch": 0.11901306240928883,
"grad_norm": 0.14665937423706055,
"learning_rate": 6.278842105263159e-05,
"loss": 0.4818,
"step": 82
},
{
"epoch": 0.1204644412191582,
"grad_norm": 0.12407947331666946,
"learning_rate": 6.22563157894737e-05,
"loss": 0.4061,
"step": 83
},
{
"epoch": 0.12191582002902758,
"grad_norm": 0.12156087160110474,
"learning_rate": 6.172421052631579e-05,
"loss": 0.4842,
"step": 84
},
{
"epoch": 0.12336719883889695,
"grad_norm": 0.164754256606102,
"learning_rate": 6.11921052631579e-05,
"loss": 0.6039,
"step": 85
},
{
"epoch": 0.12481857764876633,
"grad_norm": 0.13035471737384796,
"learning_rate": 6.066e-05,
"loss": 0.5963,
"step": 86
},
{
"epoch": 0.1262699564586357,
"grad_norm": 0.08667127043008804,
"learning_rate": 6.012789473684211e-05,
"loss": 0.104,
"step": 87
},
{
"epoch": 0.12772133526850507,
"grad_norm": 0.12500996887683868,
"learning_rate": 5.959578947368421e-05,
"loss": 0.1519,
"step": 88
},
{
"epoch": 0.12917271407837447,
"grad_norm": 0.1052846610546112,
"learning_rate": 5.9063684210526324e-05,
"loss": 0.168,
"step": 89
},
{
"epoch": 0.13062409288824384,
"grad_norm": 0.16326619684696198,
"learning_rate": 5.8531578947368425e-05,
"loss": 1.0778,
"step": 90
},
{
"epoch": 0.1320754716981132,
"grad_norm": 0.18031665682792664,
"learning_rate": 5.7999473684210527e-05,
"loss": 0.8578,
"step": 91
},
{
"epoch": 0.13352685050798258,
"grad_norm": 0.14327001571655273,
"learning_rate": 5.7467368421052635e-05,
"loss": 0.8367,
"step": 92
},
{
"epoch": 0.13497822931785197,
"grad_norm": 0.12158166617155075,
"learning_rate": 5.6935263157894736e-05,
"loss": 0.2978,
"step": 93
},
{
"epoch": 0.13642960812772134,
"grad_norm": 0.17898587882518768,
"learning_rate": 5.640315789473684e-05,
"loss": 0.7248,
"step": 94
},
{
"epoch": 0.1378809869375907,
"grad_norm": 0.1620013415813446,
"learning_rate": 5.587105263157895e-05,
"loss": 0.5909,
"step": 95
},
{
"epoch": 0.13933236574746008,
"grad_norm": 0.24596630036830902,
"learning_rate": 5.533894736842106e-05,
"loss": 1.3,
"step": 96
},
{
"epoch": 0.14078374455732948,
"grad_norm": 0.28974536061286926,
"learning_rate": 5.480684210526316e-05,
"loss": 1.0699,
"step": 97
},
{
"epoch": 0.14223512336719885,
"grad_norm": 0.2533659338951111,
"learning_rate": 5.4274736842105264e-05,
"loss": 0.6266,
"step": 98
},
{
"epoch": 0.14368650217706821,
"grad_norm": 0.36984333395957947,
"learning_rate": 5.374263157894737e-05,
"loss": 1.3249,
"step": 99
},
{
"epoch": 0.14513788098693758,
"grad_norm": 0.4221994876861572,
"learning_rate": 5.3210526315789474e-05,
"loss": 0.9726,
"step": 100
},
{
"epoch": 0.14513788098693758,
"eval_loss": 0.2286633849143982,
"eval_runtime": 145.162,
"eval_samples_per_second": 2.005,
"eval_steps_per_second": 0.503,
"step": 100
},
{
"epoch": 0.14658925979680695,
"grad_norm": 0.13860496878623962,
"learning_rate": 5.2678421052631576e-05,
"loss": 0.7931,
"step": 101
},
{
"epoch": 0.14804063860667635,
"grad_norm": 0.1530522257089615,
"learning_rate": 5.214631578947369e-05,
"loss": 0.4883,
"step": 102
},
{
"epoch": 0.14949201741654572,
"grad_norm": 0.09421262890100479,
"learning_rate": 5.161421052631579e-05,
"loss": 0.1757,
"step": 103
},
{
"epoch": 0.1509433962264151,
"grad_norm": 0.11253266036510468,
"learning_rate": 5.10821052631579e-05,
"loss": 0.4355,
"step": 104
},
{
"epoch": 0.15239477503628446,
"grad_norm": 0.11226090788841248,
"learning_rate": 5.055e-05,
"loss": 0.3406,
"step": 105
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.09969495981931686,
"learning_rate": 5.0017894736842104e-05,
"loss": 0.33,
"step": 106
},
{
"epoch": 0.15529753265602322,
"grad_norm": 0.14994554221630096,
"learning_rate": 4.948578947368421e-05,
"loss": 0.7712,
"step": 107
},
{
"epoch": 0.1567489114658926,
"grad_norm": 0.16747573018074036,
"learning_rate": 4.895368421052632e-05,
"loss": 0.4805,
"step": 108
},
{
"epoch": 0.15820029027576196,
"grad_norm": 0.15101346373558044,
"learning_rate": 4.842157894736842e-05,
"loss": 0.6745,
"step": 109
},
{
"epoch": 0.15965166908563136,
"grad_norm": 0.14325077831745148,
"learning_rate": 4.7889473684210523e-05,
"loss": 0.2481,
"step": 110
},
{
"epoch": 0.16110304789550073,
"grad_norm": 0.6589614152908325,
"learning_rate": 4.735736842105263e-05,
"loss": 0.385,
"step": 111
},
{
"epoch": 0.1625544267053701,
"grad_norm": 0.14817237854003906,
"learning_rate": 4.682526315789474e-05,
"loss": 0.4519,
"step": 112
},
{
"epoch": 0.16400580551523947,
"grad_norm": 0.12075196951627731,
"learning_rate": 4.629315789473684e-05,
"loss": 0.32,
"step": 113
},
{
"epoch": 0.16545718432510886,
"grad_norm": 0.10896284133195877,
"learning_rate": 4.576105263157895e-05,
"loss": 0.252,
"step": 114
},
{
"epoch": 0.16690856313497823,
"grad_norm": 0.15797024965286255,
"learning_rate": 4.522894736842106e-05,
"loss": 0.7193,
"step": 115
},
{
"epoch": 0.1683599419448476,
"grad_norm": 0.12811946868896484,
"learning_rate": 4.469684210526316e-05,
"loss": 0.5991,
"step": 116
},
{
"epoch": 0.16981132075471697,
"grad_norm": 0.17125706374645233,
"learning_rate": 4.416473684210527e-05,
"loss": 0.6612,
"step": 117
},
{
"epoch": 0.17126269956458637,
"grad_norm": 0.21033529937267303,
"learning_rate": 4.363263157894737e-05,
"loss": 0.637,
"step": 118
},
{
"epoch": 0.17271407837445574,
"grad_norm": 0.14343442022800446,
"learning_rate": 4.310052631578948e-05,
"loss": 0.3919,
"step": 119
},
{
"epoch": 0.1741654571843251,
"grad_norm": 0.09223167598247528,
"learning_rate": 4.256842105263158e-05,
"loss": 0.2206,
"step": 120
},
{
"epoch": 0.17561683599419448,
"grad_norm": 0.10201866924762726,
"learning_rate": 4.203631578947369e-05,
"loss": 0.2657,
"step": 121
},
{
"epoch": 0.17706821480406387,
"grad_norm": 0.09830533713102341,
"learning_rate": 4.150421052631579e-05,
"loss": 0.1767,
"step": 122
},
{
"epoch": 0.17851959361393324,
"grad_norm": 0.11305135488510132,
"learning_rate": 4.09721052631579e-05,
"loss": 0.3528,
"step": 123
},
{
"epoch": 0.1799709724238026,
"grad_norm": 0.11795450747013092,
"learning_rate": 4.0440000000000006e-05,
"loss": 0.2826,
"step": 124
},
{
"epoch": 0.18142235123367198,
"grad_norm": 0.10810278356075287,
"learning_rate": 3.990789473684211e-05,
"loss": 0.3728,
"step": 125
},
{
"epoch": 0.18287373004354138,
"grad_norm": 0.07889194041490555,
"learning_rate": 3.937578947368421e-05,
"loss": 0.1485,
"step": 126
},
{
"epoch": 0.18432510885341075,
"grad_norm": 0.12167462706565857,
"learning_rate": 3.884368421052632e-05,
"loss": 0.3191,
"step": 127
},
{
"epoch": 0.18577648766328012,
"grad_norm": 0.13652844727039337,
"learning_rate": 3.8311578947368426e-05,
"loss": 0.3272,
"step": 128
},
{
"epoch": 0.18722786647314948,
"grad_norm": 0.11355964839458466,
"learning_rate": 3.777947368421053e-05,
"loss": 0.2216,
"step": 129
},
{
"epoch": 0.18867924528301888,
"grad_norm": 0.12072530388832092,
"learning_rate": 3.724736842105263e-05,
"loss": 0.2511,
"step": 130
},
{
"epoch": 0.19013062409288825,
"grad_norm": 0.11075890809297562,
"learning_rate": 3.6715263157894744e-05,
"loss": 0.1769,
"step": 131
},
{
"epoch": 0.19158200290275762,
"grad_norm": 0.09322332590818405,
"learning_rate": 3.6183157894736845e-05,
"loss": 0.2052,
"step": 132
},
{
"epoch": 0.193033381712627,
"grad_norm": 0.12875616550445557,
"learning_rate": 3.565105263157895e-05,
"loss": 0.1961,
"step": 133
},
{
"epoch": 0.19448476052249636,
"grad_norm": 0.11314037442207336,
"learning_rate": 3.5118947368421055e-05,
"loss": 0.2227,
"step": 134
},
{
"epoch": 0.19593613933236576,
"grad_norm": 0.10332886129617691,
"learning_rate": 3.458684210526316e-05,
"loss": 0.339,
"step": 135
},
{
"epoch": 0.19738751814223512,
"grad_norm": 0.17729254066944122,
"learning_rate": 3.4054736842105265e-05,
"loss": 0.2802,
"step": 136
},
{
"epoch": 0.1988388969521045,
"grad_norm": 0.08777690678834915,
"learning_rate": 3.3522631578947366e-05,
"loss": 0.1391,
"step": 137
},
{
"epoch": 0.20029027576197386,
"grad_norm": 0.1536087989807129,
"learning_rate": 3.2990526315789475e-05,
"loss": 0.3643,
"step": 138
},
{
"epoch": 0.20174165457184326,
"grad_norm": 0.11051812022924423,
"learning_rate": 3.245842105263158e-05,
"loss": 0.1991,
"step": 139
},
{
"epoch": 0.20319303338171263,
"grad_norm": 0.09295262396335602,
"learning_rate": 3.1926315789473685e-05,
"loss": 0.2913,
"step": 140
},
{
"epoch": 0.204644412191582,
"grad_norm": 0.13505689799785614,
"learning_rate": 3.139421052631579e-05,
"loss": 0.8371,
"step": 141
},
{
"epoch": 0.20609579100145137,
"grad_norm": 0.1267169862985611,
"learning_rate": 3.0862105263157894e-05,
"loss": 0.6024,
"step": 142
},
{
"epoch": 0.20754716981132076,
"grad_norm": 0.13407814502716064,
"learning_rate": 3.033e-05,
"loss": 0.4159,
"step": 143
},
{
"epoch": 0.20899854862119013,
"grad_norm": 0.11469519138336182,
"learning_rate": 2.9797894736842104e-05,
"loss": 0.2402,
"step": 144
},
{
"epoch": 0.2104499274310595,
"grad_norm": 0.169059157371521,
"learning_rate": 2.9265789473684213e-05,
"loss": 0.3534,
"step": 145
},
{
"epoch": 0.21190130624092887,
"grad_norm": 0.19773763418197632,
"learning_rate": 2.8733684210526317e-05,
"loss": 0.5001,
"step": 146
},
{
"epoch": 0.21335268505079827,
"grad_norm": 0.4277418553829193,
"learning_rate": 2.820157894736842e-05,
"loss": 0.7581,
"step": 147
},
{
"epoch": 0.21480406386066764,
"grad_norm": 0.2791711688041687,
"learning_rate": 2.766947368421053e-05,
"loss": 0.8721,
"step": 148
},
{
"epoch": 0.216255442670537,
"grad_norm": 0.40901464223861694,
"learning_rate": 2.7137368421052632e-05,
"loss": 0.8766,
"step": 149
},
{
"epoch": 0.21770682148040638,
"grad_norm": 0.39813101291656494,
"learning_rate": 2.6605263157894737e-05,
"loss": 0.5074,
"step": 150
},
{
"epoch": 0.21770682148040638,
"eval_loss": 0.23037858307361603,
"eval_runtime": 143.626,
"eval_samples_per_second": 2.026,
"eval_steps_per_second": 0.508,
"step": 150
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 1
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.8064531192283136e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}