|
{ |
|
"best_metric": 11.5, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-50", |
|
"epoch": 3.007518796992481, |
|
"eval_steps": 50, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03007518796992481, |
|
"grad_norm": 9.266636334359646e-05, |
|
"learning_rate": 1.0100000000000002e-05, |
|
"loss": 46.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03007518796992481, |
|
"eval_loss": 11.5, |
|
"eval_runtime": 0.3095, |
|
"eval_samples_per_second": 1437.638, |
|
"eval_steps_per_second": 45.229, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.06015037593984962, |
|
"grad_norm": 9.364843572257087e-05, |
|
"learning_rate": 2.0200000000000003e-05, |
|
"loss": 46.0, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.09022556390977443, |
|
"grad_norm": 0.00010662364365998656, |
|
"learning_rate": 3.0299999999999998e-05, |
|
"loss": 46.0, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.12030075187969924, |
|
"grad_norm": 9.021838195621967e-05, |
|
"learning_rate": 4.0400000000000006e-05, |
|
"loss": 46.0, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.15037593984962405, |
|
"grad_norm": 7.925021054688841e-05, |
|
"learning_rate": 5.05e-05, |
|
"loss": 46.0, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.18045112781954886, |
|
"grad_norm": 8.351929864147678e-05, |
|
"learning_rate": 6.0599999999999996e-05, |
|
"loss": 46.0, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 9.711394523037598e-05, |
|
"learning_rate": 7.07e-05, |
|
"loss": 46.0, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.24060150375939848, |
|
"grad_norm": 9.60054385359399e-05, |
|
"learning_rate": 8.080000000000001e-05, |
|
"loss": 46.0, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.2706766917293233, |
|
"grad_norm": 0.00012731400784105062, |
|
"learning_rate": 9.09e-05, |
|
"loss": 46.0, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.3007518796992481, |
|
"grad_norm": 0.00011434268526500091, |
|
"learning_rate": 0.000101, |
|
"loss": 46.0, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.3308270676691729, |
|
"grad_norm": 9.765291906660423e-05, |
|
"learning_rate": 9.987777777777779e-05, |
|
"loss": 46.0, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.3609022556390977, |
|
"grad_norm": 9.876976400846615e-05, |
|
"learning_rate": 9.875555555555555e-05, |
|
"loss": 46.0, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.39097744360902253, |
|
"grad_norm": 0.00011370878928573802, |
|
"learning_rate": 9.763333333333334e-05, |
|
"loss": 46.0, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 0.0001137641811510548, |
|
"learning_rate": 9.651111111111111e-05, |
|
"loss": 46.0, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.45112781954887216, |
|
"grad_norm": 0.0001710877986624837, |
|
"learning_rate": 9.538888888888889e-05, |
|
"loss": 46.0, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.48120300751879697, |
|
"grad_norm": 0.0001575473725097254, |
|
"learning_rate": 9.426666666666668e-05, |
|
"loss": 46.0, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.5112781954887218, |
|
"grad_norm": 0.0001343897165497765, |
|
"learning_rate": 9.314444444444445e-05, |
|
"loss": 46.0, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.5413533834586466, |
|
"grad_norm": 0.00015038810670375824, |
|
"learning_rate": 9.202222222222223e-05, |
|
"loss": 46.0, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 0.00013464110088534653, |
|
"learning_rate": 9.09e-05, |
|
"loss": 46.0, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.6015037593984962, |
|
"grad_norm": 0.0001557513460284099, |
|
"learning_rate": 8.977777777777778e-05, |
|
"loss": 46.0, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 0.00022612961765844375, |
|
"learning_rate": 8.865555555555555e-05, |
|
"loss": 46.0, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.6616541353383458, |
|
"grad_norm": 0.00020132240024395287, |
|
"learning_rate": 8.753333333333334e-05, |
|
"loss": 46.0, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.6917293233082706, |
|
"grad_norm": 0.00018423757865093648, |
|
"learning_rate": 8.641111111111111e-05, |
|
"loss": 46.0, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.7218045112781954, |
|
"grad_norm": 0.00016721716383472085, |
|
"learning_rate": 8.528888888888889e-05, |
|
"loss": 46.0, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.7518796992481203, |
|
"grad_norm": 0.0001902187941595912, |
|
"learning_rate": 8.416666666666668e-05, |
|
"loss": 46.0, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.7819548872180451, |
|
"grad_norm": 0.00020407498232088983, |
|
"learning_rate": 8.304444444444444e-05, |
|
"loss": 46.0, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.8120300751879699, |
|
"grad_norm": 0.0002489977050572634, |
|
"learning_rate": 8.192222222222222e-05, |
|
"loss": 46.0, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 0.00030909531051293015, |
|
"learning_rate": 8.080000000000001e-05, |
|
"loss": 46.0, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.8721804511278195, |
|
"grad_norm": 0.000234145627473481, |
|
"learning_rate": 7.967777777777777e-05, |
|
"loss": 46.0, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.9022556390977443, |
|
"grad_norm": 0.0002202718605985865, |
|
"learning_rate": 7.855555555555556e-05, |
|
"loss": 46.0, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.9323308270676691, |
|
"grad_norm": 0.0002214343985542655, |
|
"learning_rate": 7.743333333333334e-05, |
|
"loss": 46.0, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.9624060150375939, |
|
"grad_norm": 0.00025516573805361986, |
|
"learning_rate": 7.631111111111111e-05, |
|
"loss": 46.0, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.9924812030075187, |
|
"grad_norm": 0.00025323766749352217, |
|
"learning_rate": 7.51888888888889e-05, |
|
"loss": 46.0, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.0225563909774436, |
|
"grad_norm": 0.00027011558995582163, |
|
"learning_rate": 7.406666666666666e-05, |
|
"loss": 46.0, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 0.0002727561804931611, |
|
"learning_rate": 7.294444444444445e-05, |
|
"loss": 46.0, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.0827067669172932, |
|
"grad_norm": 0.00034800541470758617, |
|
"learning_rate": 7.182222222222222e-05, |
|
"loss": 46.0, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.112781954887218, |
|
"grad_norm": 0.0003243626852054149, |
|
"learning_rate": 7.07e-05, |
|
"loss": 46.0, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 0.0002872659242711961, |
|
"learning_rate": 6.957777777777777e-05, |
|
"loss": 46.0, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.1729323308270676, |
|
"grad_norm": 0.00030295943724922836, |
|
"learning_rate": 6.845555555555556e-05, |
|
"loss": 46.0, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.2030075187969924, |
|
"grad_norm": 0.00025122929946519434, |
|
"learning_rate": 6.733333333333333e-05, |
|
"loss": 46.0, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.2330827067669172, |
|
"grad_norm": 0.00027522369055077434, |
|
"learning_rate": 6.621111111111111e-05, |
|
"loss": 46.0, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.263157894736842, |
|
"grad_norm": 0.00043568338151089847, |
|
"learning_rate": 6.50888888888889e-05, |
|
"loss": 46.0, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.2932330827067668, |
|
"grad_norm": 0.0003823314909823239, |
|
"learning_rate": 6.396666666666666e-05, |
|
"loss": 46.0, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.3233082706766917, |
|
"grad_norm": 0.00034535420127213, |
|
"learning_rate": 6.284444444444445e-05, |
|
"loss": 46.0, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.3533834586466165, |
|
"grad_norm": 0.0003590960695873946, |
|
"learning_rate": 6.172222222222223e-05, |
|
"loss": 46.0, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.3834586466165413, |
|
"grad_norm": 0.00029228764469735324, |
|
"learning_rate": 6.0599999999999996e-05, |
|
"loss": 46.0, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.413533834586466, |
|
"grad_norm": 0.0002954736119136214, |
|
"learning_rate": 5.9477777777777784e-05, |
|
"loss": 46.0, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.443609022556391, |
|
"grad_norm": 0.0005244979402050376, |
|
"learning_rate": 5.835555555555555e-05, |
|
"loss": 46.0, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.4736842105263157, |
|
"grad_norm": 0.00042245551594533026, |
|
"learning_rate": 5.723333333333333e-05, |
|
"loss": 46.0, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.5037593984962405, |
|
"grad_norm": 0.00039055367233231664, |
|
"learning_rate": 5.6111111111111114e-05, |
|
"loss": 46.0, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.5037593984962405, |
|
"eval_loss": 11.5, |
|
"eval_runtime": 0.305, |
|
"eval_samples_per_second": 1458.84, |
|
"eval_steps_per_second": 45.896, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.5338345864661656, |
|
"grad_norm": 0.0003437866980675608, |
|
"learning_rate": 5.498888888888888e-05, |
|
"loss": 46.0, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.5639097744360901, |
|
"grad_norm": 0.00032857232145033777, |
|
"learning_rate": 5.3866666666666664e-05, |
|
"loss": 46.0, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.5939849624060152, |
|
"grad_norm": 0.00038689709617756307, |
|
"learning_rate": 5.274444444444445e-05, |
|
"loss": 46.0, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.6240601503759398, |
|
"grad_norm": 0.0004355922865215689, |
|
"learning_rate": 5.162222222222222e-05, |
|
"loss": 46.0, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.6541353383458648, |
|
"grad_norm": 0.0005018658121116459, |
|
"learning_rate": 5.05e-05, |
|
"loss": 46.0, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.6842105263157894, |
|
"grad_norm": 0.00045422802213579416, |
|
"learning_rate": 4.9377777777777776e-05, |
|
"loss": 46.0, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 0.0003774865763261914, |
|
"learning_rate": 4.825555555555556e-05, |
|
"loss": 46.0, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.744360902255639, |
|
"grad_norm": 0.00042634535930119455, |
|
"learning_rate": 4.713333333333334e-05, |
|
"loss": 46.0, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.774436090225564, |
|
"grad_norm": 0.000391370733268559, |
|
"learning_rate": 4.601111111111111e-05, |
|
"loss": 46.0, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.8045112781954886, |
|
"grad_norm": 0.00042282594949938357, |
|
"learning_rate": 4.488888888888889e-05, |
|
"loss": 46.0, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.8345864661654137, |
|
"grad_norm": 0.0005342218209989369, |
|
"learning_rate": 4.376666666666667e-05, |
|
"loss": 46.0, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.8646616541353382, |
|
"grad_norm": 0.0004987895372323692, |
|
"learning_rate": 4.2644444444444443e-05, |
|
"loss": 46.0, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.8947368421052633, |
|
"grad_norm": 0.00042961034341715276, |
|
"learning_rate": 4.152222222222222e-05, |
|
"loss": 46.0, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.9248120300751879, |
|
"grad_norm": 0.00040588382398709655, |
|
"learning_rate": 4.0400000000000006e-05, |
|
"loss": 46.0, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.954887218045113, |
|
"grad_norm": 0.00041778653394430876, |
|
"learning_rate": 3.927777777777778e-05, |
|
"loss": 46.0, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.9849624060150375, |
|
"grad_norm": 0.00044860667549073696, |
|
"learning_rate": 3.8155555555555555e-05, |
|
"loss": 46.0, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 2.0150375939849625, |
|
"grad_norm": 0.00040975239244289696, |
|
"learning_rate": 3.703333333333333e-05, |
|
"loss": 46.0, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 2.045112781954887, |
|
"grad_norm": 0.00044060847721993923, |
|
"learning_rate": 3.591111111111111e-05, |
|
"loss": 46.0, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 2.075187969924812, |
|
"grad_norm": 0.0006466583581641316, |
|
"learning_rate": 3.4788888888888886e-05, |
|
"loss": 46.0, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 0.000526844582054764, |
|
"learning_rate": 3.366666666666667e-05, |
|
"loss": 46.0, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.1353383458646618, |
|
"grad_norm": 0.0005004553822800517, |
|
"learning_rate": 3.254444444444445e-05, |
|
"loss": 46.0, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 2.1654135338345863, |
|
"grad_norm": 0.0004915460012853146, |
|
"learning_rate": 3.142222222222222e-05, |
|
"loss": 46.0, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.1954887218045114, |
|
"grad_norm": 0.0004236290114931762, |
|
"learning_rate": 3.0299999999999998e-05, |
|
"loss": 46.0, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 2.225563909774436, |
|
"grad_norm": 0.000418194686062634, |
|
"learning_rate": 2.9177777777777776e-05, |
|
"loss": 46.0, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.255639097744361, |
|
"grad_norm": 0.0006017014384269714, |
|
"learning_rate": 2.8055555555555557e-05, |
|
"loss": 46.0, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 0.0006093091797083616, |
|
"learning_rate": 2.6933333333333332e-05, |
|
"loss": 46.0, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.3157894736842106, |
|
"grad_norm": 0.0005279547767713666, |
|
"learning_rate": 2.581111111111111e-05, |
|
"loss": 46.0, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 2.345864661654135, |
|
"grad_norm": 0.000503562914673239, |
|
"learning_rate": 2.4688888888888888e-05, |
|
"loss": 46.0, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.3759398496240602, |
|
"grad_norm": 0.0005086955497972667, |
|
"learning_rate": 2.356666666666667e-05, |
|
"loss": 46.0, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 2.406015037593985, |
|
"grad_norm": 0.0005278786411508918, |
|
"learning_rate": 2.2444444444444444e-05, |
|
"loss": 46.0, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.43609022556391, |
|
"grad_norm": 0.0005769961280748248, |
|
"learning_rate": 2.1322222222222222e-05, |
|
"loss": 46.0, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.4661654135338344, |
|
"grad_norm": 0.0006405999301932752, |
|
"learning_rate": 2.0200000000000003e-05, |
|
"loss": 46.0, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.4962406015037595, |
|
"grad_norm": 0.000521977839525789, |
|
"learning_rate": 1.9077777777777778e-05, |
|
"loss": 46.0, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 2.526315789473684, |
|
"grad_norm": 0.0004871116252616048, |
|
"learning_rate": 1.7955555555555556e-05, |
|
"loss": 46.0, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.556390977443609, |
|
"grad_norm": 0.0005196544225327671, |
|
"learning_rate": 1.6833333333333334e-05, |
|
"loss": 46.0, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 2.5864661654135337, |
|
"grad_norm": 0.00043869740329682827, |
|
"learning_rate": 1.571111111111111e-05, |
|
"loss": 46.0, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.6165413533834587, |
|
"grad_norm": 0.0005383821553550661, |
|
"learning_rate": 1.4588888888888888e-05, |
|
"loss": 46.0, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.6466165413533833, |
|
"grad_norm": 0.0007268090848810971, |
|
"learning_rate": 1.3466666666666666e-05, |
|
"loss": 46.0, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.6766917293233083, |
|
"grad_norm": 0.0006505560595542192, |
|
"learning_rate": 1.2344444444444444e-05, |
|
"loss": 46.0, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 2.706766917293233, |
|
"grad_norm": 0.0005259292083792388, |
|
"learning_rate": 1.1222222222222222e-05, |
|
"loss": 46.0, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.736842105263158, |
|
"grad_norm": 0.0004948357818648219, |
|
"learning_rate": 1.0100000000000002e-05, |
|
"loss": 46.0, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 2.7669172932330826, |
|
"grad_norm": 0.00042854080675169826, |
|
"learning_rate": 8.977777777777778e-06, |
|
"loss": 46.0, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.7969924812030076, |
|
"grad_norm": 0.00048204101040028036, |
|
"learning_rate": 7.855555555555556e-06, |
|
"loss": 46.0, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.827067669172932, |
|
"grad_norm": 0.0006985705113038421, |
|
"learning_rate": 6.733333333333333e-06, |
|
"loss": 46.0, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.0005821044323965907, |
|
"learning_rate": 5.611111111111111e-06, |
|
"loss": 46.0, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.887218045112782, |
|
"grad_norm": 0.0005263919592835009, |
|
"learning_rate": 4.488888888888889e-06, |
|
"loss": 46.0, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.917293233082707, |
|
"grad_norm": 0.0005317186005413532, |
|
"learning_rate": 3.3666666666666665e-06, |
|
"loss": 46.0, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 2.9473684210526314, |
|
"grad_norm": 0.00046762413694523275, |
|
"learning_rate": 2.2444444444444445e-06, |
|
"loss": 46.0, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.9774436090225564, |
|
"grad_norm": 0.0006510710809379816, |
|
"learning_rate": 1.1222222222222222e-06, |
|
"loss": 46.0, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 3.007518796992481, |
|
"grad_norm": 0.00044690861250273883, |
|
"learning_rate": 0.0, |
|
"loss": 46.0, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.007518796992481, |
|
"eval_loss": 11.5, |
|
"eval_runtime": 0.3049, |
|
"eval_samples_per_second": 1459.491, |
|
"eval_steps_per_second": 45.917, |
|
"step": 100 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 1 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 128633851084800.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|