lesso's picture
Training in progress, step 200, checkpoint
2641c55 verified
{
"best_metric": 1.1349619626998901,
"best_model_checkpoint": "miner_id_24/checkpoint-200",
"epoch": 0.03266372693124286,
"eval_steps": 50,
"global_step": 200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00016331863465621427,
"grad_norm": 1.671673059463501,
"learning_rate": 1.001e-05,
"loss": 1.0081,
"step": 1
},
{
"epoch": 0.00016331863465621427,
"eval_loss": 1.9000940322875977,
"eval_runtime": 345.3153,
"eval_samples_per_second": 7.466,
"eval_steps_per_second": 1.868,
"step": 1
},
{
"epoch": 0.00032663726931242854,
"grad_norm": 2.1853973865509033,
"learning_rate": 2.002e-05,
"loss": 1.0249,
"step": 2
},
{
"epoch": 0.0004899559039686428,
"grad_norm": 2.061541795730591,
"learning_rate": 3.0029999999999995e-05,
"loss": 1.0868,
"step": 3
},
{
"epoch": 0.0006532745386248571,
"grad_norm": 2.0026113986968994,
"learning_rate": 4.004e-05,
"loss": 1.2093,
"step": 4
},
{
"epoch": 0.0008165931732810714,
"grad_norm": 1.3478747606277466,
"learning_rate": 5.005e-05,
"loss": 0.9586,
"step": 5
},
{
"epoch": 0.0009799118079372856,
"grad_norm": 1.2198477983474731,
"learning_rate": 6.005999999999999e-05,
"loss": 1.001,
"step": 6
},
{
"epoch": 0.0011432304425935,
"grad_norm": 1.0008301734924316,
"learning_rate": 7.006999999999998e-05,
"loss": 1.0274,
"step": 7
},
{
"epoch": 0.0013065490772497142,
"grad_norm": 0.9373151063919067,
"learning_rate": 8.008e-05,
"loss": 1.0383,
"step": 8
},
{
"epoch": 0.0014698677119059284,
"grad_norm": 0.984042763710022,
"learning_rate": 9.009e-05,
"loss": 0.9748,
"step": 9
},
{
"epoch": 0.0016331863465621427,
"grad_norm": 0.9123459458351135,
"learning_rate": 0.0001001,
"loss": 1.0147,
"step": 10
},
{
"epoch": 0.001796504981218357,
"grad_norm": 0.941821277141571,
"learning_rate": 9.957315789473684e-05,
"loss": 0.8924,
"step": 11
},
{
"epoch": 0.0019598236158745713,
"grad_norm": 0.8732932209968567,
"learning_rate": 9.904631578947367e-05,
"loss": 0.9978,
"step": 12
},
{
"epoch": 0.0021231422505307855,
"grad_norm": 0.8224309086799622,
"learning_rate": 9.851947368421052e-05,
"loss": 1.064,
"step": 13
},
{
"epoch": 0.002286460885187,
"grad_norm": 0.9332011342048645,
"learning_rate": 9.799263157894736e-05,
"loss": 1.0589,
"step": 14
},
{
"epoch": 0.002449779519843214,
"grad_norm": 0.9349180459976196,
"learning_rate": 9.746578947368421e-05,
"loss": 1.1028,
"step": 15
},
{
"epoch": 0.0026130981544994283,
"grad_norm": 0.8387799263000488,
"learning_rate": 9.693894736842104e-05,
"loss": 1.0083,
"step": 16
},
{
"epoch": 0.0027764167891556426,
"grad_norm": 0.8592500686645508,
"learning_rate": 9.641210526315789e-05,
"loss": 1.0184,
"step": 17
},
{
"epoch": 0.002939735423811857,
"grad_norm": 0.8980984687805176,
"learning_rate": 9.588526315789473e-05,
"loss": 0.937,
"step": 18
},
{
"epoch": 0.003103054058468071,
"grad_norm": 0.8308787941932678,
"learning_rate": 9.535842105263157e-05,
"loss": 0.9237,
"step": 19
},
{
"epoch": 0.0032663726931242854,
"grad_norm": 1.0007582902908325,
"learning_rate": 9.483157894736841e-05,
"loss": 1.0406,
"step": 20
},
{
"epoch": 0.0034296913277804997,
"grad_norm": 1.081726312637329,
"learning_rate": 9.430473684210526e-05,
"loss": 1.007,
"step": 21
},
{
"epoch": 0.003593009962436714,
"grad_norm": 0.9985299110412598,
"learning_rate": 9.37778947368421e-05,
"loss": 1.0914,
"step": 22
},
{
"epoch": 0.0037563285970929282,
"grad_norm": 1.0678647756576538,
"learning_rate": 9.325105263157894e-05,
"loss": 1.2774,
"step": 23
},
{
"epoch": 0.0039196472317491425,
"grad_norm": 1.0457429885864258,
"learning_rate": 9.272421052631578e-05,
"loss": 0.9798,
"step": 24
},
{
"epoch": 0.004082965866405357,
"grad_norm": 1.1700959205627441,
"learning_rate": 9.219736842105263e-05,
"loss": 1.2144,
"step": 25
},
{
"epoch": 0.004246284501061571,
"grad_norm": 1.3236618041992188,
"learning_rate": 9.167052631578946e-05,
"loss": 1.1766,
"step": 26
},
{
"epoch": 0.004409603135717786,
"grad_norm": 1.2073756456375122,
"learning_rate": 9.114368421052632e-05,
"loss": 0.8459,
"step": 27
},
{
"epoch": 0.004572921770374,
"grad_norm": 1.3933292627334595,
"learning_rate": 9.061684210526315e-05,
"loss": 1.0467,
"step": 28
},
{
"epoch": 0.004736240405030214,
"grad_norm": 1.4482253789901733,
"learning_rate": 9.009e-05,
"loss": 1.1401,
"step": 29
},
{
"epoch": 0.004899559039686428,
"grad_norm": 1.365537405014038,
"learning_rate": 8.956315789473683e-05,
"loss": 0.9932,
"step": 30
},
{
"epoch": 0.005062877674342643,
"grad_norm": 1.475220799446106,
"learning_rate": 8.903631578947368e-05,
"loss": 1.2499,
"step": 31
},
{
"epoch": 0.005226196308998857,
"grad_norm": 1.531992793083191,
"learning_rate": 8.850947368421052e-05,
"loss": 1.0115,
"step": 32
},
{
"epoch": 0.005389514943655071,
"grad_norm": 2.305574655532837,
"learning_rate": 8.798263157894736e-05,
"loss": 0.9368,
"step": 33
},
{
"epoch": 0.005552833578311285,
"grad_norm": 1.7354161739349365,
"learning_rate": 8.745578947368422e-05,
"loss": 1.049,
"step": 34
},
{
"epoch": 0.0057161522129675,
"grad_norm": 2.7026562690734863,
"learning_rate": 8.692894736842105e-05,
"loss": 1.3359,
"step": 35
},
{
"epoch": 0.005879470847623714,
"grad_norm": 2.1211535930633545,
"learning_rate": 8.64021052631579e-05,
"loss": 1.2523,
"step": 36
},
{
"epoch": 0.0060427894822799285,
"grad_norm": 2.5288071632385254,
"learning_rate": 8.587526315789473e-05,
"loss": 1.43,
"step": 37
},
{
"epoch": 0.006206108116936142,
"grad_norm": 2.433478832244873,
"learning_rate": 8.534842105263157e-05,
"loss": 1.214,
"step": 38
},
{
"epoch": 0.006369426751592357,
"grad_norm": 2.750303268432617,
"learning_rate": 8.482157894736842e-05,
"loss": 1.0099,
"step": 39
},
{
"epoch": 0.006532745386248571,
"grad_norm": 2.981351375579834,
"learning_rate": 8.429473684210525e-05,
"loss": 1.6271,
"step": 40
},
{
"epoch": 0.0066960640209047856,
"grad_norm": 3.7224624156951904,
"learning_rate": 8.376789473684211e-05,
"loss": 1.733,
"step": 41
},
{
"epoch": 0.006859382655560999,
"grad_norm": 3.8278751373291016,
"learning_rate": 8.324105263157894e-05,
"loss": 1.5315,
"step": 42
},
{
"epoch": 0.007022701290217214,
"grad_norm": 2.7715137004852295,
"learning_rate": 8.271421052631579e-05,
"loss": 1.103,
"step": 43
},
{
"epoch": 0.007186019924873428,
"grad_norm": 4.1672892570495605,
"learning_rate": 8.218736842105262e-05,
"loss": 1.6296,
"step": 44
},
{
"epoch": 0.007349338559529643,
"grad_norm": 4.840080261230469,
"learning_rate": 8.166052631578947e-05,
"loss": 1.3456,
"step": 45
},
{
"epoch": 0.0075126571941858565,
"grad_norm": 4.377506256103516,
"learning_rate": 8.113368421052631e-05,
"loss": 1.6744,
"step": 46
},
{
"epoch": 0.007675975828842071,
"grad_norm": 4.180212497711182,
"learning_rate": 8.060684210526315e-05,
"loss": 1.4903,
"step": 47
},
{
"epoch": 0.007839294463498285,
"grad_norm": 4.473518371582031,
"learning_rate": 8.008e-05,
"loss": 2.0731,
"step": 48
},
{
"epoch": 0.008002613098154499,
"grad_norm": 7.101559162139893,
"learning_rate": 7.955315789473684e-05,
"loss": 2.0784,
"step": 49
},
{
"epoch": 0.008165931732810714,
"grad_norm": 12.615863800048828,
"learning_rate": 7.902631578947368e-05,
"loss": 3.4556,
"step": 50
},
{
"epoch": 0.008165931732810714,
"eval_loss": 1.2871240377426147,
"eval_runtime": 345.3086,
"eval_samples_per_second": 7.466,
"eval_steps_per_second": 1.868,
"step": 50
},
{
"epoch": 0.008329250367466928,
"grad_norm": 1.8585017919540405,
"learning_rate": 7.849947368421052e-05,
"loss": 0.9102,
"step": 51
},
{
"epoch": 0.008492569002123142,
"grad_norm": 1.3500086069107056,
"learning_rate": 7.797263157894736e-05,
"loss": 1.079,
"step": 52
},
{
"epoch": 0.008655887636779356,
"grad_norm": 0.9420222043991089,
"learning_rate": 7.744578947368421e-05,
"loss": 0.9128,
"step": 53
},
{
"epoch": 0.008819206271435572,
"grad_norm": 0.7498682141304016,
"learning_rate": 7.691894736842104e-05,
"loss": 0.9181,
"step": 54
},
{
"epoch": 0.008982524906091785,
"grad_norm": 0.6257831454277039,
"learning_rate": 7.63921052631579e-05,
"loss": 0.8515,
"step": 55
},
{
"epoch": 0.009145843540748,
"grad_norm": 0.6922722458839417,
"learning_rate": 7.586526315789473e-05,
"loss": 1.0347,
"step": 56
},
{
"epoch": 0.009309162175404213,
"grad_norm": 0.6588111519813538,
"learning_rate": 7.533842105263158e-05,
"loss": 0.9266,
"step": 57
},
{
"epoch": 0.009472480810060429,
"grad_norm": 0.6781211495399475,
"learning_rate": 7.481157894736841e-05,
"loss": 0.8475,
"step": 58
},
{
"epoch": 0.009635799444716642,
"grad_norm": 0.7402060627937317,
"learning_rate": 7.428473684210526e-05,
"loss": 0.9287,
"step": 59
},
{
"epoch": 0.009799118079372856,
"grad_norm": 0.7259023189544678,
"learning_rate": 7.375789473684209e-05,
"loss": 0.9206,
"step": 60
},
{
"epoch": 0.00996243671402907,
"grad_norm": 0.7616604566574097,
"learning_rate": 7.323105263157895e-05,
"loss": 1.0049,
"step": 61
},
{
"epoch": 0.010125755348685286,
"grad_norm": 0.7781885862350464,
"learning_rate": 7.270421052631578e-05,
"loss": 0.9962,
"step": 62
},
{
"epoch": 0.0102890739833415,
"grad_norm": 0.7381305694580078,
"learning_rate": 7.217736842105263e-05,
"loss": 0.9049,
"step": 63
},
{
"epoch": 0.010452392617997713,
"grad_norm": 0.6955409646034241,
"learning_rate": 7.165052631578947e-05,
"loss": 0.8963,
"step": 64
},
{
"epoch": 0.010615711252653927,
"grad_norm": 0.8050999045372009,
"learning_rate": 7.11236842105263e-05,
"loss": 1.0838,
"step": 65
},
{
"epoch": 0.010779029887310143,
"grad_norm": 0.7544150352478027,
"learning_rate": 7.059684210526315e-05,
"loss": 0.9493,
"step": 66
},
{
"epoch": 0.010942348521966357,
"grad_norm": 0.8155741095542908,
"learning_rate": 7.006999999999998e-05,
"loss": 1.1459,
"step": 67
},
{
"epoch": 0.01110566715662257,
"grad_norm": 0.7876653671264648,
"learning_rate": 6.954315789473684e-05,
"loss": 1.1572,
"step": 68
},
{
"epoch": 0.011268985791278784,
"grad_norm": 0.7840471267700195,
"learning_rate": 6.901631578947368e-05,
"loss": 0.9516,
"step": 69
},
{
"epoch": 0.011432304425935,
"grad_norm": 0.8437424302101135,
"learning_rate": 6.848947368421052e-05,
"loss": 1.0117,
"step": 70
},
{
"epoch": 0.011595623060591214,
"grad_norm": 1.0028568506240845,
"learning_rate": 6.796263157894737e-05,
"loss": 1.0057,
"step": 71
},
{
"epoch": 0.011758941695247428,
"grad_norm": 0.9504867792129517,
"learning_rate": 6.74357894736842e-05,
"loss": 0.7451,
"step": 72
},
{
"epoch": 0.011922260329903641,
"grad_norm": 0.9289828538894653,
"learning_rate": 6.690894736842105e-05,
"loss": 0.9866,
"step": 73
},
{
"epoch": 0.012085578964559857,
"grad_norm": 0.9018694758415222,
"learning_rate": 6.638210526315788e-05,
"loss": 0.9381,
"step": 74
},
{
"epoch": 0.01224889759921607,
"grad_norm": 1.0793242454528809,
"learning_rate": 6.585526315789474e-05,
"loss": 0.8213,
"step": 75
},
{
"epoch": 0.012412216233872285,
"grad_norm": 1.2196921110153198,
"learning_rate": 6.532842105263157e-05,
"loss": 1.2063,
"step": 76
},
{
"epoch": 0.012575534868528498,
"grad_norm": 1.1026328802108765,
"learning_rate": 6.480157894736842e-05,
"loss": 0.9335,
"step": 77
},
{
"epoch": 0.012738853503184714,
"grad_norm": 1.1513912677764893,
"learning_rate": 6.427473684210526e-05,
"loss": 0.8602,
"step": 78
},
{
"epoch": 0.012902172137840928,
"grad_norm": 1.6112369298934937,
"learning_rate": 6.37478947368421e-05,
"loss": 0.9575,
"step": 79
},
{
"epoch": 0.013065490772497142,
"grad_norm": 1.4818161725997925,
"learning_rate": 6.322105263157894e-05,
"loss": 1.0907,
"step": 80
},
{
"epoch": 0.013228809407153356,
"grad_norm": 1.3476381301879883,
"learning_rate": 6.269421052631577e-05,
"loss": 1.1235,
"step": 81
},
{
"epoch": 0.013392128041809571,
"grad_norm": 2.3844518661499023,
"learning_rate": 6.216736842105263e-05,
"loss": 1.5409,
"step": 82
},
{
"epoch": 0.013555446676465785,
"grad_norm": 3.1227309703826904,
"learning_rate": 6.164052631578947e-05,
"loss": 1.0545,
"step": 83
},
{
"epoch": 0.013718765311121999,
"grad_norm": 1.8032991886138916,
"learning_rate": 6.111368421052631e-05,
"loss": 1.458,
"step": 84
},
{
"epoch": 0.013882083945778213,
"grad_norm": 3.0967934131622314,
"learning_rate": 6.058684210526315e-05,
"loss": 1.4376,
"step": 85
},
{
"epoch": 0.014045402580434428,
"grad_norm": 1.9893653392791748,
"learning_rate": 6.005999999999999e-05,
"loss": 1.2981,
"step": 86
},
{
"epoch": 0.014208721215090642,
"grad_norm": 1.9208163022994995,
"learning_rate": 5.953315789473684e-05,
"loss": 0.9923,
"step": 87
},
{
"epoch": 0.014372039849746856,
"grad_norm": 2.3992538452148438,
"learning_rate": 5.9006315789473676e-05,
"loss": 1.1079,
"step": 88
},
{
"epoch": 0.01453535848440307,
"grad_norm": 2.272313117980957,
"learning_rate": 5.847947368421053e-05,
"loss": 1.1979,
"step": 89
},
{
"epoch": 0.014698677119059285,
"grad_norm": 2.8902785778045654,
"learning_rate": 5.795263157894737e-05,
"loss": 1.3671,
"step": 90
},
{
"epoch": 0.014861995753715499,
"grad_norm": 2.847869634628296,
"learning_rate": 5.742578947368421e-05,
"loss": 0.7691,
"step": 91
},
{
"epoch": 0.015025314388371713,
"grad_norm": 2.585005044937134,
"learning_rate": 5.6898947368421046e-05,
"loss": 1.2341,
"step": 92
},
{
"epoch": 0.015188633023027927,
"grad_norm": 3.3812036514282227,
"learning_rate": 5.6372105263157886e-05,
"loss": 1.3677,
"step": 93
},
{
"epoch": 0.015351951657684142,
"grad_norm": 4.347215175628662,
"learning_rate": 5.584526315789473e-05,
"loss": 1.5338,
"step": 94
},
{
"epoch": 0.015515270292340356,
"grad_norm": 2.856776237487793,
"learning_rate": 5.531842105263158e-05,
"loss": 1.4113,
"step": 95
},
{
"epoch": 0.01567858892699657,
"grad_norm": 3.887981414794922,
"learning_rate": 5.4791578947368424e-05,
"loss": 1.4782,
"step": 96
},
{
"epoch": 0.015841907561652786,
"grad_norm": 3.390988826751709,
"learning_rate": 5.426473684210526e-05,
"loss": 2.0514,
"step": 97
},
{
"epoch": 0.016005226196308998,
"grad_norm": 90.9446792602539,
"learning_rate": 5.37378947368421e-05,
"loss": 1.5698,
"step": 98
},
{
"epoch": 0.016168544830965213,
"grad_norm": 4.8651556968688965,
"learning_rate": 5.321105263157894e-05,
"loss": 1.732,
"step": 99
},
{
"epoch": 0.01633186346562143,
"grad_norm": 8.150487899780273,
"learning_rate": 5.268421052631578e-05,
"loss": 2.5082,
"step": 100
},
{
"epoch": 0.01633186346562143,
"eval_loss": 1.1979554891586304,
"eval_runtime": 345.8196,
"eval_samples_per_second": 7.455,
"eval_steps_per_second": 1.865,
"step": 100
},
{
"epoch": 0.01649518210027764,
"grad_norm": 0.7058354020118713,
"learning_rate": 5.2157368421052626e-05,
"loss": 0.7955,
"step": 101
},
{
"epoch": 0.016658500734933857,
"grad_norm": 0.7873735427856445,
"learning_rate": 5.163052631578947e-05,
"loss": 0.9298,
"step": 102
},
{
"epoch": 0.01682181936959007,
"grad_norm": 0.6918278336524963,
"learning_rate": 5.110368421052632e-05,
"loss": 0.941,
"step": 103
},
{
"epoch": 0.016985138004246284,
"grad_norm": 0.6960136294364929,
"learning_rate": 5.057684210526316e-05,
"loss": 0.8583,
"step": 104
},
{
"epoch": 0.0171484566389025,
"grad_norm": 0.6780745983123779,
"learning_rate": 5.005e-05,
"loss": 0.8733,
"step": 105
},
{
"epoch": 0.017311775273558712,
"grad_norm": 0.6429307460784912,
"learning_rate": 4.9523157894736836e-05,
"loss": 0.9417,
"step": 106
},
{
"epoch": 0.017475093908214927,
"grad_norm": 0.6558470726013184,
"learning_rate": 4.899631578947368e-05,
"loss": 0.8916,
"step": 107
},
{
"epoch": 0.017638412542871143,
"grad_norm": 0.6242132186889648,
"learning_rate": 4.846947368421052e-05,
"loss": 1.0048,
"step": 108
},
{
"epoch": 0.017801731177527355,
"grad_norm": 0.5471253991127014,
"learning_rate": 4.794263157894737e-05,
"loss": 0.8926,
"step": 109
},
{
"epoch": 0.01796504981218357,
"grad_norm": 0.5996319055557251,
"learning_rate": 4.7415789473684206e-05,
"loss": 1.032,
"step": 110
},
{
"epoch": 0.018128368446839783,
"grad_norm": 0.6278873682022095,
"learning_rate": 4.688894736842105e-05,
"loss": 1.0255,
"step": 111
},
{
"epoch": 0.018291687081496,
"grad_norm": 0.622478723526001,
"learning_rate": 4.636210526315789e-05,
"loss": 0.9705,
"step": 112
},
{
"epoch": 0.018455005716152214,
"grad_norm": 0.6352484226226807,
"learning_rate": 4.583526315789473e-05,
"loss": 0.7911,
"step": 113
},
{
"epoch": 0.018618324350808426,
"grad_norm": 0.681956946849823,
"learning_rate": 4.530842105263158e-05,
"loss": 1.0143,
"step": 114
},
{
"epoch": 0.01878164298546464,
"grad_norm": 0.7090697288513184,
"learning_rate": 4.4781578947368416e-05,
"loss": 0.9628,
"step": 115
},
{
"epoch": 0.018944961620120857,
"grad_norm": 0.7363274097442627,
"learning_rate": 4.425473684210526e-05,
"loss": 1.0271,
"step": 116
},
{
"epoch": 0.01910828025477707,
"grad_norm": 0.7926649451255798,
"learning_rate": 4.372789473684211e-05,
"loss": 1.124,
"step": 117
},
{
"epoch": 0.019271598889433285,
"grad_norm": 0.7625167369842529,
"learning_rate": 4.320105263157895e-05,
"loss": 1.1374,
"step": 118
},
{
"epoch": 0.019434917524089497,
"grad_norm": 0.8049649000167847,
"learning_rate": 4.2674210526315786e-05,
"loss": 0.9706,
"step": 119
},
{
"epoch": 0.019598236158745713,
"grad_norm": 0.7378389239311218,
"learning_rate": 4.2147368421052626e-05,
"loss": 0.8136,
"step": 120
},
{
"epoch": 0.019761554793401928,
"grad_norm": 0.8343645930290222,
"learning_rate": 4.162052631578947e-05,
"loss": 0.9652,
"step": 121
},
{
"epoch": 0.01992487342805814,
"grad_norm": 1.196692705154419,
"learning_rate": 4.109368421052631e-05,
"loss": 0.9455,
"step": 122
},
{
"epoch": 0.020088192062714356,
"grad_norm": 0.9792311787605286,
"learning_rate": 4.056684210526316e-05,
"loss": 1.0111,
"step": 123
},
{
"epoch": 0.02025151069737057,
"grad_norm": 0.9197056889533997,
"learning_rate": 4.004e-05,
"loss": 0.975,
"step": 124
},
{
"epoch": 0.020414829332026783,
"grad_norm": 1.0094839334487915,
"learning_rate": 3.951315789473684e-05,
"loss": 1.1968,
"step": 125
},
{
"epoch": 0.020578147966683,
"grad_norm": 1.0122300386428833,
"learning_rate": 3.898631578947368e-05,
"loss": 0.9602,
"step": 126
},
{
"epoch": 0.02074146660133921,
"grad_norm": 1.0537972450256348,
"learning_rate": 3.845947368421052e-05,
"loss": 0.9842,
"step": 127
},
{
"epoch": 0.020904785235995427,
"grad_norm": 1.1098566055297852,
"learning_rate": 3.7932631578947367e-05,
"loss": 0.6929,
"step": 128
},
{
"epoch": 0.021068103870651642,
"grad_norm": 1.246183156967163,
"learning_rate": 3.7405789473684206e-05,
"loss": 1.0037,
"step": 129
},
{
"epoch": 0.021231422505307854,
"grad_norm": 1.2114588022232056,
"learning_rate": 3.6878947368421045e-05,
"loss": 0.9154,
"step": 130
},
{
"epoch": 0.02139474113996407,
"grad_norm": 1.2241871356964111,
"learning_rate": 3.635210526315789e-05,
"loss": 1.0221,
"step": 131
},
{
"epoch": 0.021558059774620286,
"grad_norm": 1.3580658435821533,
"learning_rate": 3.582526315789474e-05,
"loss": 1.0275,
"step": 132
},
{
"epoch": 0.021721378409276498,
"grad_norm": 1.6600242853164673,
"learning_rate": 3.5298421052631576e-05,
"loss": 1.1471,
"step": 133
},
{
"epoch": 0.021884697043932713,
"grad_norm": 1.9517477750778198,
"learning_rate": 3.477157894736842e-05,
"loss": 1.21,
"step": 134
},
{
"epoch": 0.022048015678588925,
"grad_norm": 2.107109308242798,
"learning_rate": 3.424473684210526e-05,
"loss": 1.4034,
"step": 135
},
{
"epoch": 0.02221133431324514,
"grad_norm": 3.42795991897583,
"learning_rate": 3.37178947368421e-05,
"loss": 1.2696,
"step": 136
},
{
"epoch": 0.022374652947901356,
"grad_norm": 2.0616705417633057,
"learning_rate": 3.319105263157894e-05,
"loss": 1.1452,
"step": 137
},
{
"epoch": 0.02253797158255757,
"grad_norm": 2.5637259483337402,
"learning_rate": 3.2664210526315786e-05,
"loss": 1.1689,
"step": 138
},
{
"epoch": 0.022701290217213784,
"grad_norm": 3.1429243087768555,
"learning_rate": 3.213736842105263e-05,
"loss": 1.4649,
"step": 139
},
{
"epoch": 0.02286460885187,
"grad_norm": 3.0663323402404785,
"learning_rate": 3.161052631578947e-05,
"loss": 1.4323,
"step": 140
},
{
"epoch": 0.023027927486526212,
"grad_norm": 3.1458401679992676,
"learning_rate": 3.108368421052632e-05,
"loss": 0.9673,
"step": 141
},
{
"epoch": 0.023191246121182427,
"grad_norm": 3.3782215118408203,
"learning_rate": 3.0556842105263156e-05,
"loss": 1.3108,
"step": 142
},
{
"epoch": 0.02335456475583864,
"grad_norm": 3.1077187061309814,
"learning_rate": 3.0029999999999995e-05,
"loss": 1.6808,
"step": 143
},
{
"epoch": 0.023517883390494855,
"grad_norm": 3.3202567100524902,
"learning_rate": 2.9503157894736838e-05,
"loss": 1.6597,
"step": 144
},
{
"epoch": 0.02368120202515107,
"grad_norm": 2.8775930404663086,
"learning_rate": 2.8976315789473684e-05,
"loss": 0.9313,
"step": 145
},
{
"epoch": 0.023844520659807283,
"grad_norm": 3.3522984981536865,
"learning_rate": 2.8449473684210523e-05,
"loss": 1.8898,
"step": 146
},
{
"epoch": 0.0240078392944635,
"grad_norm": 4.523774147033691,
"learning_rate": 2.7922631578947366e-05,
"loss": 1.1704,
"step": 147
},
{
"epoch": 0.024171157929119714,
"grad_norm": 5.291178226470947,
"learning_rate": 2.7395789473684212e-05,
"loss": 2.2631,
"step": 148
},
{
"epoch": 0.024334476563775926,
"grad_norm": 5.7989959716796875,
"learning_rate": 2.686894736842105e-05,
"loss": 2.0457,
"step": 149
},
{
"epoch": 0.02449779519843214,
"grad_norm": 10.409539222717285,
"learning_rate": 2.634210526315789e-05,
"loss": 1.8767,
"step": 150
},
{
"epoch": 0.02449779519843214,
"eval_loss": 1.1476330757141113,
"eval_runtime": 345.8578,
"eval_samples_per_second": 7.454,
"eval_steps_per_second": 1.865,
"step": 150
},
{
"epoch": 0.024661113833088357,
"grad_norm": 0.5229954719543457,
"learning_rate": 2.5815263157894736e-05,
"loss": 0.9078,
"step": 151
},
{
"epoch": 0.02482443246774457,
"grad_norm": 0.4165619909763336,
"learning_rate": 2.528842105263158e-05,
"loss": 0.7202,
"step": 152
},
{
"epoch": 0.024987751102400785,
"grad_norm": 0.4411524832248688,
"learning_rate": 2.4761578947368418e-05,
"loss": 0.7657,
"step": 153
},
{
"epoch": 0.025151069737056997,
"grad_norm": 0.5870716571807861,
"learning_rate": 2.423473684210526e-05,
"loss": 0.9064,
"step": 154
},
{
"epoch": 0.025314388371713212,
"grad_norm": 0.5716325044631958,
"learning_rate": 2.3707894736842103e-05,
"loss": 0.7211,
"step": 155
},
{
"epoch": 0.025477707006369428,
"grad_norm": 0.5912550091743469,
"learning_rate": 2.3181052631578946e-05,
"loss": 0.9137,
"step": 156
},
{
"epoch": 0.02564102564102564,
"grad_norm": 0.49093204736709595,
"learning_rate": 2.265421052631579e-05,
"loss": 0.9175,
"step": 157
},
{
"epoch": 0.025804344275681856,
"grad_norm": 0.5379444360733032,
"learning_rate": 2.212736842105263e-05,
"loss": 1.0109,
"step": 158
},
{
"epoch": 0.02596766291033807,
"grad_norm": 0.5582550764083862,
"learning_rate": 2.1600526315789474e-05,
"loss": 0.9187,
"step": 159
},
{
"epoch": 0.026130981544994283,
"grad_norm": 0.5776020288467407,
"learning_rate": 2.1073684210526313e-05,
"loss": 0.8756,
"step": 160
},
{
"epoch": 0.0262943001796505,
"grad_norm": 0.6320648789405823,
"learning_rate": 2.0546842105263155e-05,
"loss": 1.002,
"step": 161
},
{
"epoch": 0.02645761881430671,
"grad_norm": 0.5884015560150146,
"learning_rate": 2.002e-05,
"loss": 0.82,
"step": 162
},
{
"epoch": 0.026620937448962927,
"grad_norm": 0.6072445511817932,
"learning_rate": 1.949315789473684e-05,
"loss": 0.9141,
"step": 163
},
{
"epoch": 0.026784256083619142,
"grad_norm": 0.578770637512207,
"learning_rate": 1.8966315789473683e-05,
"loss": 0.8936,
"step": 164
},
{
"epoch": 0.026947574718275354,
"grad_norm": 0.7762899994850159,
"learning_rate": 1.8439473684210522e-05,
"loss": 0.67,
"step": 165
},
{
"epoch": 0.02711089335293157,
"grad_norm": 0.7052147388458252,
"learning_rate": 1.791263157894737e-05,
"loss": 0.7984,
"step": 166
},
{
"epoch": 0.027274211987587785,
"grad_norm": 0.6481052041053772,
"learning_rate": 1.738578947368421e-05,
"loss": 0.9779,
"step": 167
},
{
"epoch": 0.027437530622243998,
"grad_norm": 0.6853712201118469,
"learning_rate": 1.685894736842105e-05,
"loss": 0.8923,
"step": 168
},
{
"epoch": 0.027600849256900213,
"grad_norm": 0.7015615105628967,
"learning_rate": 1.6332105263157893e-05,
"loss": 0.8661,
"step": 169
},
{
"epoch": 0.027764167891556425,
"grad_norm": 0.8164605498313904,
"learning_rate": 1.5805263157894735e-05,
"loss": 0.8789,
"step": 170
},
{
"epoch": 0.02792748652621264,
"grad_norm": 0.8073320984840393,
"learning_rate": 1.5278421052631578e-05,
"loss": 1.0431,
"step": 171
},
{
"epoch": 0.028090805160868856,
"grad_norm": 0.9141411185264587,
"learning_rate": 1.4751578947368419e-05,
"loss": 1.069,
"step": 172
},
{
"epoch": 0.02825412379552507,
"grad_norm": 1.0805171728134155,
"learning_rate": 1.4224736842105262e-05,
"loss": 1.0579,
"step": 173
},
{
"epoch": 0.028417442430181284,
"grad_norm": 0.8766102194786072,
"learning_rate": 1.3697894736842106e-05,
"loss": 0.7755,
"step": 174
},
{
"epoch": 0.0285807610648375,
"grad_norm": 0.9655460715293884,
"learning_rate": 1.3171052631578945e-05,
"loss": 0.967,
"step": 175
},
{
"epoch": 0.02874407969949371,
"grad_norm": 0.9592732787132263,
"learning_rate": 1.264421052631579e-05,
"loss": 0.9174,
"step": 176
},
{
"epoch": 0.028907398334149927,
"grad_norm": 1.3226499557495117,
"learning_rate": 1.211736842105263e-05,
"loss": 1.2512,
"step": 177
},
{
"epoch": 0.02907071696880614,
"grad_norm": 1.1680970191955566,
"learning_rate": 1.1590526315789473e-05,
"loss": 0.948,
"step": 178
},
{
"epoch": 0.029234035603462355,
"grad_norm": 1.0348116159439087,
"learning_rate": 1.1063684210526316e-05,
"loss": 0.8562,
"step": 179
},
{
"epoch": 0.02939735423811857,
"grad_norm": 1.6529384851455688,
"learning_rate": 1.0536842105263156e-05,
"loss": 1.1671,
"step": 180
},
{
"epoch": 0.029560672872774783,
"grad_norm": 1.5153594017028809,
"learning_rate": 1.001e-05,
"loss": 1.0921,
"step": 181
},
{
"epoch": 0.029723991507430998,
"grad_norm": 1.3825584650039673,
"learning_rate": 9.483157894736842e-06,
"loss": 1.1131,
"step": 182
},
{
"epoch": 0.029887310142087214,
"grad_norm": 2.2654531002044678,
"learning_rate": 8.956315789473684e-06,
"loss": 1.129,
"step": 183
},
{
"epoch": 0.030050628776743426,
"grad_norm": 1.8482561111450195,
"learning_rate": 8.429473684210525e-06,
"loss": 1.1118,
"step": 184
},
{
"epoch": 0.03021394741139964,
"grad_norm": 2.0934503078460693,
"learning_rate": 7.902631578947368e-06,
"loss": 1.0566,
"step": 185
},
{
"epoch": 0.030377266046055854,
"grad_norm": 2.0610568523406982,
"learning_rate": 7.3757894736842095e-06,
"loss": 1.5909,
"step": 186
},
{
"epoch": 0.03054058468071207,
"grad_norm": 2.510016918182373,
"learning_rate": 6.848947368421053e-06,
"loss": 1.1805,
"step": 187
},
{
"epoch": 0.030703903315368285,
"grad_norm": 2.6841964721679688,
"learning_rate": 6.322105263157895e-06,
"loss": 1.2795,
"step": 188
},
{
"epoch": 0.030867221950024497,
"grad_norm": 2.3344948291778564,
"learning_rate": 5.7952631578947365e-06,
"loss": 1.2345,
"step": 189
},
{
"epoch": 0.031030540584680712,
"grad_norm": 2.7954304218292236,
"learning_rate": 5.268421052631578e-06,
"loss": 1.2223,
"step": 190
},
{
"epoch": 0.031193859219336928,
"grad_norm": 2.858825922012329,
"learning_rate": 4.741578947368421e-06,
"loss": 1.7338,
"step": 191
},
{
"epoch": 0.03135717785399314,
"grad_norm": 3.322049856185913,
"learning_rate": 4.2147368421052626e-06,
"loss": 1.4427,
"step": 192
},
{
"epoch": 0.03152049648864935,
"grad_norm": 2.922865152359009,
"learning_rate": 3.6878947368421047e-06,
"loss": 1.5083,
"step": 193
},
{
"epoch": 0.03168381512330557,
"grad_norm": 4.792754173278809,
"learning_rate": 3.1610526315789474e-06,
"loss": 1.614,
"step": 194
},
{
"epoch": 0.03184713375796178,
"grad_norm": 2.5056612491607666,
"learning_rate": 2.634210526315789e-06,
"loss": 1.0081,
"step": 195
},
{
"epoch": 0.032010452392617995,
"grad_norm": 3.5326714515686035,
"learning_rate": 2.1073684210526313e-06,
"loss": 0.9609,
"step": 196
},
{
"epoch": 0.032173771027274214,
"grad_norm": 5.468371868133545,
"learning_rate": 1.5805263157894737e-06,
"loss": 2.8999,
"step": 197
},
{
"epoch": 0.03233708966193043,
"grad_norm": 5.357850074768066,
"learning_rate": 1.0536842105263156e-06,
"loss": 2.4673,
"step": 198
},
{
"epoch": 0.03250040829658664,
"grad_norm": 5.674871444702148,
"learning_rate": 5.268421052631578e-07,
"loss": 2.0768,
"step": 199
},
{
"epoch": 0.03266372693124286,
"grad_norm": 5.8404860496521,
"learning_rate": 0.0,
"loss": 1.6252,
"step": 200
},
{
"epoch": 0.03266372693124286,
"eval_loss": 1.1349619626998901,
"eval_runtime": 346.0024,
"eval_samples_per_second": 7.451,
"eval_steps_per_second": 1.864,
"step": 200
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.65769979075625e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}