|
{ |
|
"best_metric": 1.1956850290298462, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-200", |
|
"epoch": 0.09519276534983341, |
|
"eval_steps": 50, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00047596382674916705, |
|
"grad_norm": 18.821937561035156, |
|
"learning_rate": 1.0170000000000001e-05, |
|
"loss": 3.7617, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00047596382674916705, |
|
"eval_loss": 2.0851309299468994, |
|
"eval_runtime": 94.1092, |
|
"eval_samples_per_second": 9.404, |
|
"eval_steps_per_second": 2.359, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0009519276534983341, |
|
"grad_norm": 20.83098602294922, |
|
"learning_rate": 2.0340000000000002e-05, |
|
"loss": 3.8855, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0014278914802475012, |
|
"grad_norm": 18.45833969116211, |
|
"learning_rate": 3.051e-05, |
|
"loss": 4.0341, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0019038553069966682, |
|
"grad_norm": 15.618721961975098, |
|
"learning_rate": 4.0680000000000004e-05, |
|
"loss": 3.4497, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.002379819133745835, |
|
"grad_norm": 14.598923683166504, |
|
"learning_rate": 5.085e-05, |
|
"loss": 3.2374, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0028557829604950024, |
|
"grad_norm": 16.081514358520508, |
|
"learning_rate": 6.102e-05, |
|
"loss": 3.4533, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0033317467872441696, |
|
"grad_norm": 14.391841888427734, |
|
"learning_rate": 7.119e-05, |
|
"loss": 3.2119, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0038077106139933364, |
|
"grad_norm": 13.710631370544434, |
|
"learning_rate": 8.136000000000001e-05, |
|
"loss": 3.3193, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.004283674440742504, |
|
"grad_norm": 12.184770584106445, |
|
"learning_rate": 9.153000000000001e-05, |
|
"loss": 3.0283, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.00475963826749167, |
|
"grad_norm": 11.110732078552246, |
|
"learning_rate": 0.0001017, |
|
"loss": 3.0447, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.005235602094240838, |
|
"grad_norm": 13.03386402130127, |
|
"learning_rate": 0.00010116473684210527, |
|
"loss": 3.0503, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.005711565920990005, |
|
"grad_norm": 11.393197059631348, |
|
"learning_rate": 0.00010062947368421052, |
|
"loss": 3.1731, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.006187529747739172, |
|
"grad_norm": 11.392594337463379, |
|
"learning_rate": 0.00010009421052631579, |
|
"loss": 2.7808, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.006663493574488339, |
|
"grad_norm": 11.060273170471191, |
|
"learning_rate": 9.955894736842107e-05, |
|
"loss": 2.9735, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.007139457401237506, |
|
"grad_norm": 10.616848945617676, |
|
"learning_rate": 9.902368421052632e-05, |
|
"loss": 2.8027, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.007615421227986673, |
|
"grad_norm": 11.916224479675293, |
|
"learning_rate": 9.848842105263159e-05, |
|
"loss": 3.3891, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.00809138505473584, |
|
"grad_norm": 11.447843551635742, |
|
"learning_rate": 9.795315789473685e-05, |
|
"loss": 3.1294, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.008567348881485007, |
|
"grad_norm": 9.17795467376709, |
|
"learning_rate": 9.74178947368421e-05, |
|
"loss": 2.7995, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.009043312708234174, |
|
"grad_norm": 10.779789924621582, |
|
"learning_rate": 9.688263157894737e-05, |
|
"loss": 3.2716, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.00951927653498334, |
|
"grad_norm": 10.012481689453125, |
|
"learning_rate": 9.634736842105264e-05, |
|
"loss": 2.9491, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.009995240361732508, |
|
"grad_norm": 10.166115760803223, |
|
"learning_rate": 9.58121052631579e-05, |
|
"loss": 3.024, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.010471204188481676, |
|
"grad_norm": 9.625226020812988, |
|
"learning_rate": 9.527684210526317e-05, |
|
"loss": 2.9964, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.010947168015230843, |
|
"grad_norm": 11.936861991882324, |
|
"learning_rate": 9.474157894736843e-05, |
|
"loss": 3.0226, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.01142313184198001, |
|
"grad_norm": 9.899079322814941, |
|
"learning_rate": 9.420631578947368e-05, |
|
"loss": 3.0418, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.011899095668729176, |
|
"grad_norm": 9.566582679748535, |
|
"learning_rate": 9.367105263157895e-05, |
|
"loss": 2.5971, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.012375059495478343, |
|
"grad_norm": 9.53864860534668, |
|
"learning_rate": 9.313578947368422e-05, |
|
"loss": 3.0019, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.01285102332222751, |
|
"grad_norm": 13.411975860595703, |
|
"learning_rate": 9.260052631578948e-05, |
|
"loss": 2.7033, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.013326987148976678, |
|
"grad_norm": 26.838424682617188, |
|
"learning_rate": 9.206526315789475e-05, |
|
"loss": 2.5882, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.013802950975725845, |
|
"grad_norm": 11.076753616333008, |
|
"learning_rate": 9.153000000000001e-05, |
|
"loss": 2.9107, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.014278914802475012, |
|
"grad_norm": 10.66457748413086, |
|
"learning_rate": 9.099473684210527e-05, |
|
"loss": 3.1288, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.014754878629224179, |
|
"grad_norm": 11.16288948059082, |
|
"learning_rate": 9.045947368421053e-05, |
|
"loss": 2.913, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.015230842455973346, |
|
"grad_norm": 9.841339111328125, |
|
"learning_rate": 8.99242105263158e-05, |
|
"loss": 2.8651, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.015706806282722512, |
|
"grad_norm": 9.972783088684082, |
|
"learning_rate": 8.938894736842105e-05, |
|
"loss": 2.6382, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.01618277010947168, |
|
"grad_norm": 8.628253936767578, |
|
"learning_rate": 8.885368421052633e-05, |
|
"loss": 2.8631, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.016658733936220846, |
|
"grad_norm": 9.764812469482422, |
|
"learning_rate": 8.83184210526316e-05, |
|
"loss": 3.1602, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.017134697762970014, |
|
"grad_norm": 9.357998847961426, |
|
"learning_rate": 8.778315789473685e-05, |
|
"loss": 3.232, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.017610661589719183, |
|
"grad_norm": 8.898050308227539, |
|
"learning_rate": 8.724789473684211e-05, |
|
"loss": 3.064, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.018086625416468348, |
|
"grad_norm": 8.951642990112305, |
|
"learning_rate": 8.671263157894738e-05, |
|
"loss": 2.6545, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.018562589243217516, |
|
"grad_norm": 8.89450454711914, |
|
"learning_rate": 8.617736842105263e-05, |
|
"loss": 2.8535, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.01903855306996668, |
|
"grad_norm": 8.89578628540039, |
|
"learning_rate": 8.56421052631579e-05, |
|
"loss": 2.775, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01951451689671585, |
|
"grad_norm": 11.88039493560791, |
|
"learning_rate": 8.510684210526316e-05, |
|
"loss": 2.9885, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.019990480723465015, |
|
"grad_norm": 9.176898956298828, |
|
"learning_rate": 8.457157894736843e-05, |
|
"loss": 2.7476, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.020466444550214184, |
|
"grad_norm": 10.42405891418457, |
|
"learning_rate": 8.403631578947369e-05, |
|
"loss": 3.1671, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.020942408376963352, |
|
"grad_norm": 9.580604553222656, |
|
"learning_rate": 8.350105263157896e-05, |
|
"loss": 2.7088, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.021418372203712517, |
|
"grad_norm": 11.169540405273438, |
|
"learning_rate": 8.296578947368421e-05, |
|
"loss": 3.0075, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.021894336030461686, |
|
"grad_norm": 8.268192291259766, |
|
"learning_rate": 8.243052631578948e-05, |
|
"loss": 2.5109, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.02237029985721085, |
|
"grad_norm": 9.292203903198242, |
|
"learning_rate": 8.189526315789474e-05, |
|
"loss": 2.9182, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.02284626368396002, |
|
"grad_norm": 10.59662914276123, |
|
"learning_rate": 8.136000000000001e-05, |
|
"loss": 3.2872, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.023322227510709188, |
|
"grad_norm": 10.883740425109863, |
|
"learning_rate": 8.082473684210527e-05, |
|
"loss": 3.0321, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.023798191337458353, |
|
"grad_norm": 8.746244430541992, |
|
"learning_rate": 8.028947368421054e-05, |
|
"loss": 2.9888, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.023798191337458353, |
|
"eval_loss": 1.4506478309631348, |
|
"eval_runtime": 94.223, |
|
"eval_samples_per_second": 9.393, |
|
"eval_steps_per_second": 2.356, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02427415516420752, |
|
"grad_norm": 8.132503509521484, |
|
"learning_rate": 7.975421052631579e-05, |
|
"loss": 2.8369, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.024750118990956686, |
|
"grad_norm": 7.467820644378662, |
|
"learning_rate": 7.921894736842106e-05, |
|
"loss": 3.1423, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.025226082817705855, |
|
"grad_norm": 6.4784369468688965, |
|
"learning_rate": 7.868368421052632e-05, |
|
"loss": 2.7428, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.02570204664445502, |
|
"grad_norm": 6.803327560424805, |
|
"learning_rate": 7.814842105263157e-05, |
|
"loss": 2.8238, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.02617801047120419, |
|
"grad_norm": 9.669547080993652, |
|
"learning_rate": 7.761315789473685e-05, |
|
"loss": 2.7933, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.026653974297953357, |
|
"grad_norm": 9.685455322265625, |
|
"learning_rate": 7.70778947368421e-05, |
|
"loss": 3.0942, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.027129938124702522, |
|
"grad_norm": 8.498844146728516, |
|
"learning_rate": 7.654263157894737e-05, |
|
"loss": 3.0961, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.02760590195145169, |
|
"grad_norm": 7.0263776779174805, |
|
"learning_rate": 7.600736842105264e-05, |
|
"loss": 2.9287, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.028081865778200855, |
|
"grad_norm": 7.7478742599487305, |
|
"learning_rate": 7.54721052631579e-05, |
|
"loss": 2.7414, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.028557829604950024, |
|
"grad_norm": 6.691257953643799, |
|
"learning_rate": 7.493684210526315e-05, |
|
"loss": 2.6123, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.029033793431699192, |
|
"grad_norm": 7.181465148925781, |
|
"learning_rate": 7.440157894736843e-05, |
|
"loss": 2.6851, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.029509757258448358, |
|
"grad_norm": 6.531435012817383, |
|
"learning_rate": 7.386631578947369e-05, |
|
"loss": 2.6416, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.029985721085197526, |
|
"grad_norm": 9.599217414855957, |
|
"learning_rate": 7.333105263157895e-05, |
|
"loss": 2.7074, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.03046168491194669, |
|
"grad_norm": 7.4359846115112305, |
|
"learning_rate": 7.279578947368422e-05, |
|
"loss": 2.8352, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.03093764873869586, |
|
"grad_norm": 6.91318941116333, |
|
"learning_rate": 7.226052631578947e-05, |
|
"loss": 2.5786, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.031413612565445025, |
|
"grad_norm": 8.484053611755371, |
|
"learning_rate": 7.172526315789474e-05, |
|
"loss": 3.0093, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.0318895763921942, |
|
"grad_norm": 7.760731220245361, |
|
"learning_rate": 7.119e-05, |
|
"loss": 2.8831, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.03236554021894336, |
|
"grad_norm": 7.7634100914001465, |
|
"learning_rate": 7.065473684210527e-05, |
|
"loss": 2.8134, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.03284150404569253, |
|
"grad_norm": 15.038714408874512, |
|
"learning_rate": 7.011947368421053e-05, |
|
"loss": 2.446, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.03331746787244169, |
|
"grad_norm": 6.84593391418457, |
|
"learning_rate": 6.95842105263158e-05, |
|
"loss": 2.8314, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.033793431699190864, |
|
"grad_norm": 6.8339667320251465, |
|
"learning_rate": 6.904894736842105e-05, |
|
"loss": 2.507, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.03426939552594003, |
|
"grad_norm": 8.750052452087402, |
|
"learning_rate": 6.851368421052632e-05, |
|
"loss": 2.637, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.034745359352689194, |
|
"grad_norm": 7.736267566680908, |
|
"learning_rate": 6.797842105263158e-05, |
|
"loss": 2.748, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.035221323179438366, |
|
"grad_norm": 7.89774227142334, |
|
"learning_rate": 6.744315789473685e-05, |
|
"loss": 2.7948, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.03569728700618753, |
|
"grad_norm": 7.224119663238525, |
|
"learning_rate": 6.690789473684211e-05, |
|
"loss": 2.7275, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.036173250832936696, |
|
"grad_norm": 7.180510520935059, |
|
"learning_rate": 6.637263157894738e-05, |
|
"loss": 2.5776, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.03664921465968586, |
|
"grad_norm": 6.440933704376221, |
|
"learning_rate": 6.583736842105263e-05, |
|
"loss": 2.241, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.03712517848643503, |
|
"grad_norm": 8.913047790527344, |
|
"learning_rate": 6.53021052631579e-05, |
|
"loss": 2.9477, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.0376011423131842, |
|
"grad_norm": 7.736593723297119, |
|
"learning_rate": 6.476684210526316e-05, |
|
"loss": 2.8696, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.03807710613993336, |
|
"grad_norm": 7.847418308258057, |
|
"learning_rate": 6.423157894736841e-05, |
|
"loss": 2.3813, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.038553069966682535, |
|
"grad_norm": 9.366430282592773, |
|
"learning_rate": 6.369631578947368e-05, |
|
"loss": 2.8983, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.0390290337934317, |
|
"grad_norm": 6.681727886199951, |
|
"learning_rate": 6.316105263157896e-05, |
|
"loss": 2.4428, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.039504997620180865, |
|
"grad_norm": 8.800130844116211, |
|
"learning_rate": 6.262578947368421e-05, |
|
"loss": 2.9551, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.03998096144693003, |
|
"grad_norm": 8.04470157623291, |
|
"learning_rate": 6.209052631578948e-05, |
|
"loss": 2.7357, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.0404569252736792, |
|
"grad_norm": 6.852024555206299, |
|
"learning_rate": 6.155526315789474e-05, |
|
"loss": 2.6536, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.04093288910042837, |
|
"grad_norm": 7.370736598968506, |
|
"learning_rate": 6.102e-05, |
|
"loss": 2.6235, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.04140885292717753, |
|
"grad_norm": 7.6142401695251465, |
|
"learning_rate": 6.048473684210526e-05, |
|
"loss": 2.7034, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.041884816753926704, |
|
"grad_norm": 7.17495059967041, |
|
"learning_rate": 5.9949473684210527e-05, |
|
"loss": 2.7207, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.04236078058067587, |
|
"grad_norm": 7.216758728027344, |
|
"learning_rate": 5.94142105263158e-05, |
|
"loss": 2.4448, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.042836744407425034, |
|
"grad_norm": 7.9468817710876465, |
|
"learning_rate": 5.887894736842106e-05, |
|
"loss": 2.9818, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.043312708234174206, |
|
"grad_norm": 8.539376258850098, |
|
"learning_rate": 5.834368421052632e-05, |
|
"loss": 2.4128, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.04378867206092337, |
|
"grad_norm": 8.345818519592285, |
|
"learning_rate": 5.780842105263158e-05, |
|
"loss": 2.7606, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.044264635887672536, |
|
"grad_norm": 8.508038520812988, |
|
"learning_rate": 5.727315789473684e-05, |
|
"loss": 2.6035, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.0447405997144217, |
|
"grad_norm": 8.71206283569336, |
|
"learning_rate": 5.673789473684211e-05, |
|
"loss": 2.9024, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.04521656354117087, |
|
"grad_norm": 7.256693363189697, |
|
"learning_rate": 5.620263157894738e-05, |
|
"loss": 2.5654, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.04569252736792004, |
|
"grad_norm": 6.628811359405518, |
|
"learning_rate": 5.566736842105264e-05, |
|
"loss": 2.0765, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.0461684911946692, |
|
"grad_norm": 9.192995071411133, |
|
"learning_rate": 5.51321052631579e-05, |
|
"loss": 3.0441, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.046644455021418375, |
|
"grad_norm": 9.181817054748535, |
|
"learning_rate": 5.459684210526316e-05, |
|
"loss": 2.8811, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.04712041884816754, |
|
"grad_norm": 9.44265079498291, |
|
"learning_rate": 5.406157894736842e-05, |
|
"loss": 2.6583, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.047596382674916705, |
|
"grad_norm": 9.560362815856934, |
|
"learning_rate": 5.352631578947368e-05, |
|
"loss": 2.5805, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.047596382674916705, |
|
"eval_loss": 1.34207284450531, |
|
"eval_runtime": 94.197, |
|
"eval_samples_per_second": 9.395, |
|
"eval_steps_per_second": 2.357, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04807234650166587, |
|
"grad_norm": 7.516977787017822, |
|
"learning_rate": 5.299105263157895e-05, |
|
"loss": 2.5181, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.04854831032841504, |
|
"grad_norm": 7.009253025054932, |
|
"learning_rate": 5.245578947368422e-05, |
|
"loss": 2.6719, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.04902427415516421, |
|
"grad_norm": 6.846374988555908, |
|
"learning_rate": 5.192052631578948e-05, |
|
"loss": 2.5637, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.04950023798191337, |
|
"grad_norm": 5.941534519195557, |
|
"learning_rate": 5.1385263157894744e-05, |
|
"loss": 2.2845, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.049976201808662545, |
|
"grad_norm": 7.417452812194824, |
|
"learning_rate": 5.085e-05, |
|
"loss": 3.1752, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.05045216563541171, |
|
"grad_norm": 6.396870136260986, |
|
"learning_rate": 5.031473684210526e-05, |
|
"loss": 2.5322, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.050928129462160875, |
|
"grad_norm": 6.419276237487793, |
|
"learning_rate": 4.9779473684210534e-05, |
|
"loss": 2.5846, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.05140409328891004, |
|
"grad_norm": 9.035964012145996, |
|
"learning_rate": 4.924421052631579e-05, |
|
"loss": 2.8623, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.05188005711565921, |
|
"grad_norm": 6.224599361419678, |
|
"learning_rate": 4.870894736842105e-05, |
|
"loss": 2.3845, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.05235602094240838, |
|
"grad_norm": 7.3065409660339355, |
|
"learning_rate": 4.817368421052632e-05, |
|
"loss": 2.7384, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.05283198476915754, |
|
"grad_norm": 6.494760036468506, |
|
"learning_rate": 4.763842105263158e-05, |
|
"loss": 2.3574, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.053307948595906714, |
|
"grad_norm": 5.581643104553223, |
|
"learning_rate": 4.710315789473684e-05, |
|
"loss": 2.2699, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.05378391242265588, |
|
"grad_norm": 6.040269374847412, |
|
"learning_rate": 4.656789473684211e-05, |
|
"loss": 2.5859, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.054259876249405044, |
|
"grad_norm": 6.788252353668213, |
|
"learning_rate": 4.6032631578947374e-05, |
|
"loss": 2.4009, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.05473584007615421, |
|
"grad_norm": 7.658879280090332, |
|
"learning_rate": 4.549736842105263e-05, |
|
"loss": 2.0763, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.05521180390290338, |
|
"grad_norm": 9.23805046081543, |
|
"learning_rate": 4.49621052631579e-05, |
|
"loss": 2.2703, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.055687767729652546, |
|
"grad_norm": 7.436956882476807, |
|
"learning_rate": 4.4426842105263164e-05, |
|
"loss": 2.7847, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.05616373155640171, |
|
"grad_norm": 8.176994323730469, |
|
"learning_rate": 4.389157894736842e-05, |
|
"loss": 2.5287, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.05663969538315088, |
|
"grad_norm": 7.547792434692383, |
|
"learning_rate": 4.335631578947369e-05, |
|
"loss": 2.81, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.05711565920990005, |
|
"grad_norm": 7.150010108947754, |
|
"learning_rate": 4.282105263157895e-05, |
|
"loss": 2.7525, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.05759162303664921, |
|
"grad_norm": 7.117188930511475, |
|
"learning_rate": 4.228578947368421e-05, |
|
"loss": 2.5581, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.058067586863398385, |
|
"grad_norm": 6.114779472351074, |
|
"learning_rate": 4.175052631578948e-05, |
|
"loss": 2.5203, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.05854355069014755, |
|
"grad_norm": 6.867531776428223, |
|
"learning_rate": 4.121526315789474e-05, |
|
"loss": 2.6818, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.059019514516896715, |
|
"grad_norm": 7.646029949188232, |
|
"learning_rate": 4.0680000000000004e-05, |
|
"loss": 2.6518, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.05949547834364588, |
|
"grad_norm": 8.080942153930664, |
|
"learning_rate": 4.014473684210527e-05, |
|
"loss": 2.8059, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.05997144217039505, |
|
"grad_norm": 6.926398277282715, |
|
"learning_rate": 3.960947368421053e-05, |
|
"loss": 2.5314, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.06044740599714422, |
|
"grad_norm": 7.169958114624023, |
|
"learning_rate": 3.907421052631579e-05, |
|
"loss": 2.5107, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.06092336982389338, |
|
"grad_norm": 7.661369323730469, |
|
"learning_rate": 3.853894736842105e-05, |
|
"loss": 2.3918, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.061399333650642554, |
|
"grad_norm": 7.439793586730957, |
|
"learning_rate": 3.800368421052632e-05, |
|
"loss": 2.8541, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.06187529747739172, |
|
"grad_norm": 6.659045696258545, |
|
"learning_rate": 3.746842105263158e-05, |
|
"loss": 2.5837, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.062351261304140884, |
|
"grad_norm": 8.68233585357666, |
|
"learning_rate": 3.693315789473684e-05, |
|
"loss": 2.5465, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.06282722513089005, |
|
"grad_norm": 7.132856845855713, |
|
"learning_rate": 3.639789473684211e-05, |
|
"loss": 2.4912, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.06330318895763921, |
|
"grad_norm": 7.528649806976318, |
|
"learning_rate": 3.586263157894737e-05, |
|
"loss": 2.7235, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.0637791527843884, |
|
"grad_norm": 6.1327667236328125, |
|
"learning_rate": 3.5327368421052633e-05, |
|
"loss": 2.4542, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.06425511661113756, |
|
"grad_norm": 6.09030294418335, |
|
"learning_rate": 3.47921052631579e-05, |
|
"loss": 2.3305, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.06473108043788672, |
|
"grad_norm": 6.725249290466309, |
|
"learning_rate": 3.425684210526316e-05, |
|
"loss": 2.4557, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.06520704426463589, |
|
"grad_norm": 6.322257041931152, |
|
"learning_rate": 3.3721578947368424e-05, |
|
"loss": 2.2034, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.06568300809138505, |
|
"grad_norm": 5.7170305252075195, |
|
"learning_rate": 3.318631578947369e-05, |
|
"loss": 2.1993, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.06615897191813422, |
|
"grad_norm": 7.491881370544434, |
|
"learning_rate": 3.265105263157895e-05, |
|
"loss": 2.6061, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.06663493574488338, |
|
"grad_norm": 7.218158721923828, |
|
"learning_rate": 3.211578947368421e-05, |
|
"loss": 2.622, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.06711089957163256, |
|
"grad_norm": 10.26785659790039, |
|
"learning_rate": 3.158052631578948e-05, |
|
"loss": 2.6993, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.06758686339838173, |
|
"grad_norm": 7.6761393547058105, |
|
"learning_rate": 3.104526315789474e-05, |
|
"loss": 2.6623, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.06806282722513089, |
|
"grad_norm": 7.0337395668029785, |
|
"learning_rate": 3.051e-05, |
|
"loss": 2.632, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.06853879105188006, |
|
"grad_norm": 7.8988189697265625, |
|
"learning_rate": 2.9974736842105263e-05, |
|
"loss": 3.1195, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.06901475487862922, |
|
"grad_norm": 8.312713623046875, |
|
"learning_rate": 2.943947368421053e-05, |
|
"loss": 2.5157, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.06949071870537839, |
|
"grad_norm": 7.518918514251709, |
|
"learning_rate": 2.890421052631579e-05, |
|
"loss": 2.7354, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.06996668253212755, |
|
"grad_norm": 7.5307841300964355, |
|
"learning_rate": 2.8368947368421054e-05, |
|
"loss": 2.5474, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.07044264635887673, |
|
"grad_norm": 8.257736206054688, |
|
"learning_rate": 2.783368421052632e-05, |
|
"loss": 2.8553, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.0709186101856259, |
|
"grad_norm": 6.292891979217529, |
|
"learning_rate": 2.729842105263158e-05, |
|
"loss": 2.3596, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.07139457401237506, |
|
"grad_norm": 9.312799453735352, |
|
"learning_rate": 2.676315789473684e-05, |
|
"loss": 2.5738, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07139457401237506, |
|
"eval_loss": 1.2638894319534302, |
|
"eval_runtime": 94.1331, |
|
"eval_samples_per_second": 9.402, |
|
"eval_steps_per_second": 2.358, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07187053783912423, |
|
"grad_norm": 5.808437824249268, |
|
"learning_rate": 2.622789473684211e-05, |
|
"loss": 2.4866, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.07234650166587339, |
|
"grad_norm": 6.367058753967285, |
|
"learning_rate": 2.5692631578947372e-05, |
|
"loss": 2.5468, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.07282246549262256, |
|
"grad_norm": 7.049677848815918, |
|
"learning_rate": 2.515736842105263e-05, |
|
"loss": 2.6916, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.07329842931937172, |
|
"grad_norm": 6.486673831939697, |
|
"learning_rate": 2.4622105263157897e-05, |
|
"loss": 2.5986, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.0737743931461209, |
|
"grad_norm": 6.483848571777344, |
|
"learning_rate": 2.408684210526316e-05, |
|
"loss": 2.6228, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.07425035697287007, |
|
"grad_norm": 8.12568473815918, |
|
"learning_rate": 2.355157894736842e-05, |
|
"loss": 3.0707, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.07472632079961923, |
|
"grad_norm": 6.405067443847656, |
|
"learning_rate": 2.3016315789473687e-05, |
|
"loss": 2.654, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.0752022846263684, |
|
"grad_norm": 5.833227634429932, |
|
"learning_rate": 2.248105263157895e-05, |
|
"loss": 2.4285, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.07567824845311756, |
|
"grad_norm": 5.303393840789795, |
|
"learning_rate": 2.194578947368421e-05, |
|
"loss": 2.2755, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.07615421227986673, |
|
"grad_norm": 6.173555850982666, |
|
"learning_rate": 2.1410526315789474e-05, |
|
"loss": 2.4719, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.07663017610661589, |
|
"grad_norm": 5.758234977722168, |
|
"learning_rate": 2.087526315789474e-05, |
|
"loss": 2.6648, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.07710613993336507, |
|
"grad_norm": 5.576732158660889, |
|
"learning_rate": 2.0340000000000002e-05, |
|
"loss": 2.5195, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.07758210376011423, |
|
"grad_norm": 6.40316104888916, |
|
"learning_rate": 1.9804736842105264e-05, |
|
"loss": 2.6616, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.0780580675868634, |
|
"grad_norm": 5.3443427085876465, |
|
"learning_rate": 1.9269473684210526e-05, |
|
"loss": 2.1508, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.07853403141361257, |
|
"grad_norm": 7.0513014793396, |
|
"learning_rate": 1.873421052631579e-05, |
|
"loss": 2.3545, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.07900999524036173, |
|
"grad_norm": 6.783944129943848, |
|
"learning_rate": 1.8198947368421054e-05, |
|
"loss": 2.361, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.0794859590671109, |
|
"grad_norm": 7.000679016113281, |
|
"learning_rate": 1.7663684210526317e-05, |
|
"loss": 2.6073, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.07996192289386006, |
|
"grad_norm": 5.986188888549805, |
|
"learning_rate": 1.712842105263158e-05, |
|
"loss": 2.3954, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.08043788672060924, |
|
"grad_norm": 6.529272556304932, |
|
"learning_rate": 1.6593157894736845e-05, |
|
"loss": 2.7285, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.0809138505473584, |
|
"grad_norm": 5.4884772300720215, |
|
"learning_rate": 1.6057894736842104e-05, |
|
"loss": 2.1842, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.08138981437410757, |
|
"grad_norm": 6.433114528656006, |
|
"learning_rate": 1.552263157894737e-05, |
|
"loss": 2.3042, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.08186577820085673, |
|
"grad_norm": 6.97841215133667, |
|
"learning_rate": 1.4987368421052632e-05, |
|
"loss": 2.2691, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.0823417420276059, |
|
"grad_norm": 6.401273727416992, |
|
"learning_rate": 1.4452105263157896e-05, |
|
"loss": 2.5178, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.08281770585435506, |
|
"grad_norm": 6.540492534637451, |
|
"learning_rate": 1.391684210526316e-05, |
|
"loss": 2.2453, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.08329366968110423, |
|
"grad_norm": 6.713190078735352, |
|
"learning_rate": 1.338157894736842e-05, |
|
"loss": 2.2148, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.08376963350785341, |
|
"grad_norm": 7.125333309173584, |
|
"learning_rate": 1.2846315789473686e-05, |
|
"loss": 2.635, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.08424559733460257, |
|
"grad_norm": 7.104244709014893, |
|
"learning_rate": 1.2311052631578948e-05, |
|
"loss": 2.5499, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.08472156116135174, |
|
"grad_norm": 7.630082130432129, |
|
"learning_rate": 1.177578947368421e-05, |
|
"loss": 2.3113, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.0851975249881009, |
|
"grad_norm": 7.205958366394043, |
|
"learning_rate": 1.1240526315789475e-05, |
|
"loss": 2.3001, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.08567348881485007, |
|
"grad_norm": 6.465893268585205, |
|
"learning_rate": 1.0705263157894737e-05, |
|
"loss": 2.4646, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.08614945264159923, |
|
"grad_norm": 6.484694957733154, |
|
"learning_rate": 1.0170000000000001e-05, |
|
"loss": 2.3538, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.08662541646834841, |
|
"grad_norm": 6.740143775939941, |
|
"learning_rate": 9.634736842105263e-06, |
|
"loss": 2.1774, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.08710138029509758, |
|
"grad_norm": 6.847339153289795, |
|
"learning_rate": 9.099473684210527e-06, |
|
"loss": 2.5101, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.08757734412184674, |
|
"grad_norm": 7.767678260803223, |
|
"learning_rate": 8.56421052631579e-06, |
|
"loss": 2.5202, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.08805330794859591, |
|
"grad_norm": 8.59636116027832, |
|
"learning_rate": 8.028947368421052e-06, |
|
"loss": 2.7538, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.08852927177534507, |
|
"grad_norm": 8.195631980895996, |
|
"learning_rate": 7.493684210526316e-06, |
|
"loss": 2.4046, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.08900523560209424, |
|
"grad_norm": 6.393590927124023, |
|
"learning_rate": 6.95842105263158e-06, |
|
"loss": 2.2026, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.0894811994288434, |
|
"grad_norm": 6.6723551750183105, |
|
"learning_rate": 6.423157894736843e-06, |
|
"loss": 2.1685, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.08995716325559258, |
|
"grad_norm": 7.914825439453125, |
|
"learning_rate": 5.887894736842105e-06, |
|
"loss": 2.5318, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.09043312708234175, |
|
"grad_norm": 5.79133415222168, |
|
"learning_rate": 5.3526315789473684e-06, |
|
"loss": 2.1259, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 8.203691482543945, |
|
"learning_rate": 4.817368421052632e-06, |
|
"loss": 2.6837, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.09138505473584008, |
|
"grad_norm": 8.923236846923828, |
|
"learning_rate": 4.282105263157895e-06, |
|
"loss": 2.6346, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.09186101856258924, |
|
"grad_norm": 7.227407932281494, |
|
"learning_rate": 3.746842105263158e-06, |
|
"loss": 2.5861, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.0923369823893384, |
|
"grad_norm": 6.976375102996826, |
|
"learning_rate": 3.2115789473684215e-06, |
|
"loss": 2.2964, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.09281294621608757, |
|
"grad_norm": 6.509735584259033, |
|
"learning_rate": 2.6763157894736842e-06, |
|
"loss": 2.341, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.09328891004283675, |
|
"grad_norm": 8.12691879272461, |
|
"learning_rate": 2.1410526315789474e-06, |
|
"loss": 2.4133, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.09376487386958592, |
|
"grad_norm": 7.400885581970215, |
|
"learning_rate": 1.6057894736842107e-06, |
|
"loss": 2.6121, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.09424083769633508, |
|
"grad_norm": 6.809811592102051, |
|
"learning_rate": 1.0705263157894737e-06, |
|
"loss": 2.4394, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.09471680152308425, |
|
"grad_norm": 7.949410915374756, |
|
"learning_rate": 5.352631578947368e-07, |
|
"loss": 2.7608, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.09519276534983341, |
|
"grad_norm": 8.497697830200195, |
|
"learning_rate": 0.0, |
|
"loss": 2.583, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.09519276534983341, |
|
"eval_loss": 1.1956850290298462, |
|
"eval_runtime": 94.1756, |
|
"eval_samples_per_second": 9.397, |
|
"eval_steps_per_second": 2.357, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7.15499809800192e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|