|
{ |
|
"best_metric": 0.7936509251594543, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-200", |
|
"epoch": 0.5376344086021505, |
|
"eval_steps": 50, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002688172043010753, |
|
"grad_norm": 2.552518367767334, |
|
"learning_rate": 1.008e-05, |
|
"loss": 4.2159, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.002688172043010753, |
|
"eval_loss": 2.7429797649383545, |
|
"eval_runtime": 20.0523, |
|
"eval_samples_per_second": 7.83, |
|
"eval_steps_per_second": 1.995, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.005376344086021506, |
|
"grad_norm": 4.294527053833008, |
|
"learning_rate": 2.016e-05, |
|
"loss": 5.2935, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.008064516129032258, |
|
"grad_norm": 3.8511745929718018, |
|
"learning_rate": 3.024e-05, |
|
"loss": 4.435, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.010752688172043012, |
|
"grad_norm": 4.876188278198242, |
|
"learning_rate": 4.032e-05, |
|
"loss": 5.0044, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.013440860215053764, |
|
"grad_norm": 4.7782392501831055, |
|
"learning_rate": 5.04e-05, |
|
"loss": 4.3481, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.016129032258064516, |
|
"grad_norm": 4.02572774887085, |
|
"learning_rate": 6.048e-05, |
|
"loss": 3.9645, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01881720430107527, |
|
"grad_norm": 4.530522346496582, |
|
"learning_rate": 7.055999999999999e-05, |
|
"loss": 3.7968, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.021505376344086023, |
|
"grad_norm": 3.8177340030670166, |
|
"learning_rate": 8.064e-05, |
|
"loss": 3.9752, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.024193548387096774, |
|
"grad_norm": 3.42366361618042, |
|
"learning_rate": 9.072e-05, |
|
"loss": 3.7669, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.026881720430107527, |
|
"grad_norm": 3.45857572555542, |
|
"learning_rate": 0.0001008, |
|
"loss": 3.1495, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02956989247311828, |
|
"grad_norm": 5.064671516418457, |
|
"learning_rate": 0.00010026947368421052, |
|
"loss": 2.8932, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.03225806451612903, |
|
"grad_norm": 7.331480979919434, |
|
"learning_rate": 9.973894736842104e-05, |
|
"loss": 2.1098, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.03494623655913978, |
|
"grad_norm": 6.972163677215576, |
|
"learning_rate": 9.920842105263157e-05, |
|
"loss": 3.326, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03763440860215054, |
|
"grad_norm": 4.03656005859375, |
|
"learning_rate": 9.86778947368421e-05, |
|
"loss": 2.4259, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.04032258064516129, |
|
"grad_norm": 6.760096073150635, |
|
"learning_rate": 9.814736842105264e-05, |
|
"loss": 1.9487, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.043010752688172046, |
|
"grad_norm": 4.662262439727783, |
|
"learning_rate": 9.761684210526316e-05, |
|
"loss": 2.489, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0456989247311828, |
|
"grad_norm": 5.427197456359863, |
|
"learning_rate": 9.708631578947368e-05, |
|
"loss": 2.9028, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.04838709677419355, |
|
"grad_norm": 6.550124168395996, |
|
"learning_rate": 9.655578947368421e-05, |
|
"loss": 2.5639, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.051075268817204304, |
|
"grad_norm": 5.742612361907959, |
|
"learning_rate": 9.602526315789473e-05, |
|
"loss": 1.5072, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.053763440860215055, |
|
"grad_norm": 4.321170330047607, |
|
"learning_rate": 9.549473684210525e-05, |
|
"loss": 1.1754, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.056451612903225805, |
|
"grad_norm": 3.095183849334717, |
|
"learning_rate": 9.496421052631579e-05, |
|
"loss": 1.5671, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.05913978494623656, |
|
"grad_norm": 2.853109121322632, |
|
"learning_rate": 9.443368421052631e-05, |
|
"loss": 1.2742, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.06182795698924731, |
|
"grad_norm": 3.149440050125122, |
|
"learning_rate": 9.390315789473683e-05, |
|
"loss": 2.0117, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.06451612903225806, |
|
"grad_norm": 3.0655312538146973, |
|
"learning_rate": 9.337263157894737e-05, |
|
"loss": 2.2624, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.06720430107526881, |
|
"grad_norm": 3.4676096439361572, |
|
"learning_rate": 9.28421052631579e-05, |
|
"loss": 2.3864, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06989247311827956, |
|
"grad_norm": 3.4489779472351074, |
|
"learning_rate": 9.231157894736842e-05, |
|
"loss": 1.5537, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.07258064516129033, |
|
"grad_norm": 3.5075297355651855, |
|
"learning_rate": 9.178105263157895e-05, |
|
"loss": 1.6461, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.07526881720430108, |
|
"grad_norm": 2.4398374557495117, |
|
"learning_rate": 9.125052631578948e-05, |
|
"loss": 1.4466, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.07795698924731183, |
|
"grad_norm": 3.644092321395874, |
|
"learning_rate": 9.072e-05, |
|
"loss": 2.507, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.08064516129032258, |
|
"grad_norm": 3.263528823852539, |
|
"learning_rate": 9.018947368421052e-05, |
|
"loss": 1.2544, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08333333333333333, |
|
"grad_norm": 3.7490763664245605, |
|
"learning_rate": 8.965894736842104e-05, |
|
"loss": 1.2655, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.08602150537634409, |
|
"grad_norm": 3.2782695293426514, |
|
"learning_rate": 8.912842105263157e-05, |
|
"loss": 1.7594, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.08870967741935484, |
|
"grad_norm": 2.6544880867004395, |
|
"learning_rate": 8.85978947368421e-05, |
|
"loss": 1.6763, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0913978494623656, |
|
"grad_norm": 2.461487054824829, |
|
"learning_rate": 8.806736842105264e-05, |
|
"loss": 1.4153, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.09408602150537634, |
|
"grad_norm": 2.037763833999634, |
|
"learning_rate": 8.753684210526316e-05, |
|
"loss": 1.4954, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0967741935483871, |
|
"grad_norm": 2.2501564025878906, |
|
"learning_rate": 8.700631578947369e-05, |
|
"loss": 1.8277, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.09946236559139784, |
|
"grad_norm": 2.666536331176758, |
|
"learning_rate": 8.647578947368421e-05, |
|
"loss": 2.2322, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.10215053763440861, |
|
"grad_norm": 3.9293577671051025, |
|
"learning_rate": 8.594526315789473e-05, |
|
"loss": 2.394, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.10483870967741936, |
|
"grad_norm": 5.047905921936035, |
|
"learning_rate": 8.541473684210525e-05, |
|
"loss": 2.2877, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.10752688172043011, |
|
"grad_norm": 3.5374960899353027, |
|
"learning_rate": 8.488421052631578e-05, |
|
"loss": 1.552, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.11021505376344086, |
|
"grad_norm": 4.972512722015381, |
|
"learning_rate": 8.435368421052631e-05, |
|
"loss": 2.5782, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.11290322580645161, |
|
"grad_norm": 4.151662349700928, |
|
"learning_rate": 8.382315789473684e-05, |
|
"loss": 1.3923, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.11559139784946236, |
|
"grad_norm": 3.7074525356292725, |
|
"learning_rate": 8.329263157894737e-05, |
|
"loss": 1.3753, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.11827956989247312, |
|
"grad_norm": 3.257704973220825, |
|
"learning_rate": 8.27621052631579e-05, |
|
"loss": 1.5227, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.12096774193548387, |
|
"grad_norm": 2.8200066089630127, |
|
"learning_rate": 8.223157894736842e-05, |
|
"loss": 1.0166, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12365591397849462, |
|
"grad_norm": 9.245245933532715, |
|
"learning_rate": 8.170105263157894e-05, |
|
"loss": 3.0691, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.12634408602150538, |
|
"grad_norm": 6.105884075164795, |
|
"learning_rate": 8.117052631578946e-05, |
|
"loss": 2.7272, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.12903225806451613, |
|
"grad_norm": 9.19589900970459, |
|
"learning_rate": 8.064e-05, |
|
"loss": 2.4159, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.13172043010752688, |
|
"grad_norm": 6.029378890991211, |
|
"learning_rate": 8.010947368421052e-05, |
|
"loss": 1.4225, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.13440860215053763, |
|
"grad_norm": 4.892293930053711, |
|
"learning_rate": 7.957894736842105e-05, |
|
"loss": 0.7743, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.13440860215053763, |
|
"eval_loss": 1.1576098203659058, |
|
"eval_runtime": 20.0295, |
|
"eval_samples_per_second": 7.838, |
|
"eval_steps_per_second": 1.997, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.13709677419354838, |
|
"grad_norm": 3.6493349075317383, |
|
"learning_rate": 7.904842105263158e-05, |
|
"loss": 3.4341, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.13978494623655913, |
|
"grad_norm": 3.946737051010132, |
|
"learning_rate": 7.85178947368421e-05, |
|
"loss": 4.1712, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.1424731182795699, |
|
"grad_norm": 3.0488572120666504, |
|
"learning_rate": 7.798736842105263e-05, |
|
"loss": 2.7678, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.14516129032258066, |
|
"grad_norm": 2.641152858734131, |
|
"learning_rate": 7.745684210526315e-05, |
|
"loss": 3.4299, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1478494623655914, |
|
"grad_norm": 2.6064324378967285, |
|
"learning_rate": 7.692631578947369e-05, |
|
"loss": 2.8889, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.15053763440860216, |
|
"grad_norm": 2.853044271469116, |
|
"learning_rate": 7.639578947368421e-05, |
|
"loss": 3.9092, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.1532258064516129, |
|
"grad_norm": 3.7264606952667236, |
|
"learning_rate": 7.586526315789473e-05, |
|
"loss": 3.1663, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.15591397849462366, |
|
"grad_norm": 2.3590335845947266, |
|
"learning_rate": 7.533473684210526e-05, |
|
"loss": 2.5423, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1586021505376344, |
|
"grad_norm": 3.6274800300598145, |
|
"learning_rate": 7.480421052631578e-05, |
|
"loss": 2.7857, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.16129032258064516, |
|
"grad_norm": 3.466926097869873, |
|
"learning_rate": 7.427368421052632e-05, |
|
"loss": 2.1407, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1639784946236559, |
|
"grad_norm": 2.135305166244507, |
|
"learning_rate": 7.374315789473685e-05, |
|
"loss": 2.0501, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.16666666666666666, |
|
"grad_norm": 2.145434617996216, |
|
"learning_rate": 7.321263157894737e-05, |
|
"loss": 1.8671, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.1693548387096774, |
|
"grad_norm": 2.250318765640259, |
|
"learning_rate": 7.26821052631579e-05, |
|
"loss": 2.8158, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.17204301075268819, |
|
"grad_norm": 2.0222816467285156, |
|
"learning_rate": 7.215157894736842e-05, |
|
"loss": 1.9672, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.17473118279569894, |
|
"grad_norm": 2.0015830993652344, |
|
"learning_rate": 7.162105263157894e-05, |
|
"loss": 1.7105, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.1774193548387097, |
|
"grad_norm": 2.1358511447906494, |
|
"learning_rate": 7.109052631578947e-05, |
|
"loss": 2.0722, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.18010752688172044, |
|
"grad_norm": 2.1164755821228027, |
|
"learning_rate": 7.055999999999999e-05, |
|
"loss": 1.3778, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.1827956989247312, |
|
"grad_norm": 2.6545348167419434, |
|
"learning_rate": 7.002947368421052e-05, |
|
"loss": 1.5793, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.18548387096774194, |
|
"grad_norm": 2.278327703475952, |
|
"learning_rate": 6.949894736842105e-05, |
|
"loss": 1.3986, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1881720430107527, |
|
"grad_norm": 2.1539969444274902, |
|
"learning_rate": 6.896842105263158e-05, |
|
"loss": 0.8514, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.19086021505376344, |
|
"grad_norm": 2.815715789794922, |
|
"learning_rate": 6.843789473684211e-05, |
|
"loss": 1.9194, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1935483870967742, |
|
"grad_norm": 1.7789136171340942, |
|
"learning_rate": 6.790736842105263e-05, |
|
"loss": 1.1168, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.19623655913978494, |
|
"grad_norm": 2.7908670902252197, |
|
"learning_rate": 6.737684210526315e-05, |
|
"loss": 1.6579, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1989247311827957, |
|
"grad_norm": 3.0436408519744873, |
|
"learning_rate": 6.684631578947368e-05, |
|
"loss": 2.1317, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.20161290322580644, |
|
"grad_norm": 2.6586692333221436, |
|
"learning_rate": 6.631578947368421e-05, |
|
"loss": 1.5078, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.20430107526881722, |
|
"grad_norm": 2.3957104682922363, |
|
"learning_rate": 6.578526315789473e-05, |
|
"loss": 1.4137, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.20698924731182797, |
|
"grad_norm": 2.467710256576538, |
|
"learning_rate": 6.525473684210526e-05, |
|
"loss": 1.5379, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.20967741935483872, |
|
"grad_norm": 2.3089401721954346, |
|
"learning_rate": 6.47242105263158e-05, |
|
"loss": 1.3105, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.21236559139784947, |
|
"grad_norm": 3.3448803424835205, |
|
"learning_rate": 6.419368421052632e-05, |
|
"loss": 2.9907, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.21505376344086022, |
|
"grad_norm": 1.918912649154663, |
|
"learning_rate": 6.366315789473684e-05, |
|
"loss": 1.0538, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.21774193548387097, |
|
"grad_norm": 2.1565945148468018, |
|
"learning_rate": 6.313263157894736e-05, |
|
"loss": 0.8427, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.22043010752688172, |
|
"grad_norm": 1.9118677377700806, |
|
"learning_rate": 6.26021052631579e-05, |
|
"loss": 0.8702, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.22311827956989247, |
|
"grad_norm": 2.161041498184204, |
|
"learning_rate": 6.207157894736842e-05, |
|
"loss": 1.5612, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.22580645161290322, |
|
"grad_norm": 1.8799899816513062, |
|
"learning_rate": 6.154105263157894e-05, |
|
"loss": 1.0472, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.22849462365591397, |
|
"grad_norm": 3.115447521209717, |
|
"learning_rate": 6.1010526315789474e-05, |
|
"loss": 1.3093, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.23118279569892472, |
|
"grad_norm": 2.5335817337036133, |
|
"learning_rate": 6.048e-05, |
|
"loss": 1.5528, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.23387096774193547, |
|
"grad_norm": 2.8771631717681885, |
|
"learning_rate": 5.994947368421052e-05, |
|
"loss": 1.6662, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.23655913978494625, |
|
"grad_norm": 3.9417519569396973, |
|
"learning_rate": 5.941894736842104e-05, |
|
"loss": 1.8012, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.239247311827957, |
|
"grad_norm": 3.2698731422424316, |
|
"learning_rate": 5.888842105263158e-05, |
|
"loss": 2.3075, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.24193548387096775, |
|
"grad_norm": 2.6435606479644775, |
|
"learning_rate": 5.835789473684211e-05, |
|
"loss": 1.3781, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.2446236559139785, |
|
"grad_norm": 3.099759817123413, |
|
"learning_rate": 5.782736842105263e-05, |
|
"loss": 2.1561, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.24731182795698925, |
|
"grad_norm": 3.145350217819214, |
|
"learning_rate": 5.7296842105263154e-05, |
|
"loss": 1.4506, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 4.324268341064453, |
|
"learning_rate": 5.676631578947368e-05, |
|
"loss": 0.9423, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.25268817204301075, |
|
"grad_norm": 2.9348642826080322, |
|
"learning_rate": 5.623578947368421e-05, |
|
"loss": 1.7886, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.2553763440860215, |
|
"grad_norm": 1.738901138305664, |
|
"learning_rate": 5.570526315789474e-05, |
|
"loss": 0.676, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.25806451612903225, |
|
"grad_norm": 2.8710925579071045, |
|
"learning_rate": 5.5174736842105266e-05, |
|
"loss": 0.9851, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.260752688172043, |
|
"grad_norm": 3.99045729637146, |
|
"learning_rate": 5.464421052631579e-05, |
|
"loss": 1.8844, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.26344086021505375, |
|
"grad_norm": 4.602737903594971, |
|
"learning_rate": 5.411368421052631e-05, |
|
"loss": 1.3643, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.2661290322580645, |
|
"grad_norm": 3.7166576385498047, |
|
"learning_rate": 5.358315789473684e-05, |
|
"loss": 0.6935, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.26881720430107525, |
|
"grad_norm": 3.1870436668395996, |
|
"learning_rate": 5.3052631578947364e-05, |
|
"loss": 1.2186, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.26881720430107525, |
|
"eval_loss": 0.9393484592437744, |
|
"eval_runtime": 20.0614, |
|
"eval_samples_per_second": 7.826, |
|
"eval_steps_per_second": 1.994, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.271505376344086, |
|
"grad_norm": 2.3323299884796143, |
|
"learning_rate": 5.252210526315789e-05, |
|
"loss": 3.5211, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.27419354838709675, |
|
"grad_norm": 2.551894187927246, |
|
"learning_rate": 5.199157894736842e-05, |
|
"loss": 4.4608, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2768817204301075, |
|
"grad_norm": 2.600679397583008, |
|
"learning_rate": 5.1461052631578946e-05, |
|
"loss": 4.0707, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.27956989247311825, |
|
"grad_norm": 1.9256941080093384, |
|
"learning_rate": 5.0930526315789476e-05, |
|
"loss": 2.119, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.28225806451612906, |
|
"grad_norm": 2.2405953407287598, |
|
"learning_rate": 5.04e-05, |
|
"loss": 3.1069, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.2849462365591398, |
|
"grad_norm": 1.9663243293762207, |
|
"learning_rate": 4.986947368421052e-05, |
|
"loss": 2.7355, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.28763440860215056, |
|
"grad_norm": 2.2502665519714355, |
|
"learning_rate": 4.933894736842105e-05, |
|
"loss": 3.8927, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2903225806451613, |
|
"grad_norm": 2.449840784072876, |
|
"learning_rate": 4.880842105263158e-05, |
|
"loss": 2.5857, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.29301075268817206, |
|
"grad_norm": 2.5077898502349854, |
|
"learning_rate": 4.8277894736842103e-05, |
|
"loss": 2.0671, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.2956989247311828, |
|
"grad_norm": 2.9152989387512207, |
|
"learning_rate": 4.7747368421052626e-05, |
|
"loss": 2.771, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.29838709677419356, |
|
"grad_norm": 2.127004861831665, |
|
"learning_rate": 4.7216842105263156e-05, |
|
"loss": 2.009, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.3010752688172043, |
|
"grad_norm": 2.5167322158813477, |
|
"learning_rate": 4.6686315789473686e-05, |
|
"loss": 2.798, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.30376344086021506, |
|
"grad_norm": 1.6835650205612183, |
|
"learning_rate": 4.615578947368421e-05, |
|
"loss": 2.0009, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.3064516129032258, |
|
"grad_norm": 1.6886358261108398, |
|
"learning_rate": 4.562526315789474e-05, |
|
"loss": 1.8121, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.30913978494623656, |
|
"grad_norm": 2.1738483905792236, |
|
"learning_rate": 4.509473684210526e-05, |
|
"loss": 2.1961, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.3118279569892473, |
|
"grad_norm": 1.9415655136108398, |
|
"learning_rate": 4.4564210526315784e-05, |
|
"loss": 1.3629, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.31451612903225806, |
|
"grad_norm": 2.252319812774658, |
|
"learning_rate": 4.403368421052632e-05, |
|
"loss": 1.5139, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.3172043010752688, |
|
"grad_norm": 2.353317975997925, |
|
"learning_rate": 4.350315789473684e-05, |
|
"loss": 1.9432, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.31989247311827956, |
|
"grad_norm": 2.4038631916046143, |
|
"learning_rate": 4.2972631578947366e-05, |
|
"loss": 1.7128, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.3225806451612903, |
|
"grad_norm": 1.8164989948272705, |
|
"learning_rate": 4.244210526315789e-05, |
|
"loss": 1.0845, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.32526881720430106, |
|
"grad_norm": 1.1192851066589355, |
|
"learning_rate": 4.191157894736842e-05, |
|
"loss": 0.4165, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.3279569892473118, |
|
"grad_norm": 1.6266660690307617, |
|
"learning_rate": 4.138105263157895e-05, |
|
"loss": 1.181, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.33064516129032256, |
|
"grad_norm": 1.9494553804397583, |
|
"learning_rate": 4.085052631578947e-05, |
|
"loss": 1.3595, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 2.290536403656006, |
|
"learning_rate": 4.032e-05, |
|
"loss": 2.1863, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.33602150537634407, |
|
"grad_norm": 1.9435406923294067, |
|
"learning_rate": 3.978947368421052e-05, |
|
"loss": 0.785, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.3387096774193548, |
|
"grad_norm": 1.6345041990280151, |
|
"learning_rate": 3.925894736842105e-05, |
|
"loss": 0.6604, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.34139784946236557, |
|
"grad_norm": 2.3190295696258545, |
|
"learning_rate": 3.8728421052631575e-05, |
|
"loss": 1.5344, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.34408602150537637, |
|
"grad_norm": 2.137402296066284, |
|
"learning_rate": 3.8197894736842105e-05, |
|
"loss": 1.1624, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.3467741935483871, |
|
"grad_norm": 3.097121238708496, |
|
"learning_rate": 3.766736842105263e-05, |
|
"loss": 1.1613, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.34946236559139787, |
|
"grad_norm": 1.4628736972808838, |
|
"learning_rate": 3.713684210526316e-05, |
|
"loss": 0.7341, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.3521505376344086, |
|
"grad_norm": 2.344586133956909, |
|
"learning_rate": 3.660631578947369e-05, |
|
"loss": 1.0426, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.3548387096774194, |
|
"grad_norm": 2.10186505317688, |
|
"learning_rate": 3.607578947368421e-05, |
|
"loss": 0.6319, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.3575268817204301, |
|
"grad_norm": 3.2302911281585693, |
|
"learning_rate": 3.554526315789473e-05, |
|
"loss": 1.7709, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.3602150537634409, |
|
"grad_norm": 2.4297268390655518, |
|
"learning_rate": 3.501473684210526e-05, |
|
"loss": 1.4642, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.3629032258064516, |
|
"grad_norm": 1.544806718826294, |
|
"learning_rate": 3.448421052631579e-05, |
|
"loss": 0.5347, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.3655913978494624, |
|
"grad_norm": 1.8195679187774658, |
|
"learning_rate": 3.3953684210526315e-05, |
|
"loss": 0.9418, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.3682795698924731, |
|
"grad_norm": 2.9363420009613037, |
|
"learning_rate": 3.342315789473684e-05, |
|
"loss": 1.5149, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.3709677419354839, |
|
"grad_norm": 1.6591293811798096, |
|
"learning_rate": 3.289263157894737e-05, |
|
"loss": 0.8025, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.3736559139784946, |
|
"grad_norm": 1.8703725337982178, |
|
"learning_rate": 3.23621052631579e-05, |
|
"loss": 0.9463, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.3763440860215054, |
|
"grad_norm": 1.9521825313568115, |
|
"learning_rate": 3.183157894736842e-05, |
|
"loss": 0.8371, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.3790322580645161, |
|
"grad_norm": 2.850534200668335, |
|
"learning_rate": 3.130105263157895e-05, |
|
"loss": 0.9155, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.3817204301075269, |
|
"grad_norm": 2.7304813861846924, |
|
"learning_rate": 3.077052631578947e-05, |
|
"loss": 1.3678, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.3844086021505376, |
|
"grad_norm": 5.073098659515381, |
|
"learning_rate": 3.024e-05, |
|
"loss": 1.4571, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.3870967741935484, |
|
"grad_norm": 4.6329498291015625, |
|
"learning_rate": 2.970947368421052e-05, |
|
"loss": 2.6228, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.3897849462365591, |
|
"grad_norm": 2.2699081897735596, |
|
"learning_rate": 2.9178947368421054e-05, |
|
"loss": 1.0457, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3924731182795699, |
|
"grad_norm": 2.2033205032348633, |
|
"learning_rate": 2.8648421052631577e-05, |
|
"loss": 1.2035, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.3951612903225806, |
|
"grad_norm": 4.73455810546875, |
|
"learning_rate": 2.8117894736842103e-05, |
|
"loss": 1.7207, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.3978494623655914, |
|
"grad_norm": 3.5149385929107666, |
|
"learning_rate": 2.7587368421052633e-05, |
|
"loss": 0.7439, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.40053763440860213, |
|
"grad_norm": 2.338346004486084, |
|
"learning_rate": 2.7056842105263156e-05, |
|
"loss": 0.5313, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.4032258064516129, |
|
"grad_norm": 1.7660259008407593, |
|
"learning_rate": 2.6526315789473682e-05, |
|
"loss": 0.3979, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4032258064516129, |
|
"eval_loss": 0.8605657815933228, |
|
"eval_runtime": 20.0392, |
|
"eval_samples_per_second": 7.835, |
|
"eval_steps_per_second": 1.996, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.40591397849462363, |
|
"grad_norm": 2.3321237564086914, |
|
"learning_rate": 2.599578947368421e-05, |
|
"loss": 3.6574, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.40860215053763443, |
|
"grad_norm": 2.897630214691162, |
|
"learning_rate": 2.5465263157894738e-05, |
|
"loss": 2.9582, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.4112903225806452, |
|
"grad_norm": 2.3918817043304443, |
|
"learning_rate": 2.493473684210526e-05, |
|
"loss": 2.7901, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.41397849462365593, |
|
"grad_norm": 2.5373759269714355, |
|
"learning_rate": 2.440421052631579e-05, |
|
"loss": 3.1311, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 2.5285072326660156, |
|
"learning_rate": 2.3873684210526313e-05, |
|
"loss": 2.5189, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.41935483870967744, |
|
"grad_norm": 2.504533052444458, |
|
"learning_rate": 2.3343157894736843e-05, |
|
"loss": 3.2569, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.4220430107526882, |
|
"grad_norm": 2.3029768466949463, |
|
"learning_rate": 2.281263157894737e-05, |
|
"loss": 2.6988, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.42473118279569894, |
|
"grad_norm": 2.2616817951202393, |
|
"learning_rate": 2.2282105263157892e-05, |
|
"loss": 2.3272, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.4274193548387097, |
|
"grad_norm": 2.2895076274871826, |
|
"learning_rate": 2.175157894736842e-05, |
|
"loss": 2.3862, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.43010752688172044, |
|
"grad_norm": 1.9649949073791504, |
|
"learning_rate": 2.1221052631578944e-05, |
|
"loss": 2.1478, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.4327956989247312, |
|
"grad_norm": 2.0496561527252197, |
|
"learning_rate": 2.0690526315789474e-05, |
|
"loss": 1.3665, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.43548387096774194, |
|
"grad_norm": 2.4721364974975586, |
|
"learning_rate": 2.016e-05, |
|
"loss": 3.1437, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.4381720430107527, |
|
"grad_norm": 2.5533034801483154, |
|
"learning_rate": 1.9629473684210526e-05, |
|
"loss": 2.2656, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.44086021505376344, |
|
"grad_norm": 1.7975753545761108, |
|
"learning_rate": 1.9098947368421053e-05, |
|
"loss": 1.4366, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.4435483870967742, |
|
"grad_norm": 2.068775177001953, |
|
"learning_rate": 1.856842105263158e-05, |
|
"loss": 1.3049, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.44623655913978494, |
|
"grad_norm": 2.0522570610046387, |
|
"learning_rate": 1.8037894736842105e-05, |
|
"loss": 1.8766, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.4489247311827957, |
|
"grad_norm": 1.8050427436828613, |
|
"learning_rate": 1.750736842105263e-05, |
|
"loss": 1.501, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.45161290322580644, |
|
"grad_norm": 2.184504508972168, |
|
"learning_rate": 1.6976842105263157e-05, |
|
"loss": 1.7882, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.4543010752688172, |
|
"grad_norm": 1.4424117803573608, |
|
"learning_rate": 1.6446315789473684e-05, |
|
"loss": 0.6933, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.45698924731182794, |
|
"grad_norm": 2.0629289150238037, |
|
"learning_rate": 1.591578947368421e-05, |
|
"loss": 1.1983, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.4596774193548387, |
|
"grad_norm": 2.0509307384490967, |
|
"learning_rate": 1.5385263157894736e-05, |
|
"loss": 1.709, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.46236559139784944, |
|
"grad_norm": 1.423912525177002, |
|
"learning_rate": 1.485473684210526e-05, |
|
"loss": 0.6537, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.4650537634408602, |
|
"grad_norm": 2.0477561950683594, |
|
"learning_rate": 1.4324210526315789e-05, |
|
"loss": 0.791, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.46774193548387094, |
|
"grad_norm": 2.9739744663238525, |
|
"learning_rate": 1.3793684210526316e-05, |
|
"loss": 1.7767, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.47043010752688175, |
|
"grad_norm": 2.3965930938720703, |
|
"learning_rate": 1.3263157894736841e-05, |
|
"loss": 1.8458, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.4731182795698925, |
|
"grad_norm": 2.0706305503845215, |
|
"learning_rate": 1.2732631578947369e-05, |
|
"loss": 0.8644, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.47580645161290325, |
|
"grad_norm": 1.389144778251648, |
|
"learning_rate": 1.2202105263157895e-05, |
|
"loss": 0.5712, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.478494623655914, |
|
"grad_norm": 2.4105353355407715, |
|
"learning_rate": 1.1671578947368421e-05, |
|
"loss": 1.7319, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.48118279569892475, |
|
"grad_norm": 1.4054794311523438, |
|
"learning_rate": 1.1141052631578946e-05, |
|
"loss": 0.594, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.4838709677419355, |
|
"grad_norm": 1.2603883743286133, |
|
"learning_rate": 1.0610526315789472e-05, |
|
"loss": 0.521, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.48655913978494625, |
|
"grad_norm": 2.3499302864074707, |
|
"learning_rate": 1.008e-05, |
|
"loss": 1.2467, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.489247311827957, |
|
"grad_norm": 1.8965387344360352, |
|
"learning_rate": 9.549473684210526e-06, |
|
"loss": 1.0413, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.49193548387096775, |
|
"grad_norm": 1.9322240352630615, |
|
"learning_rate": 9.018947368421052e-06, |
|
"loss": 1.1569, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.4946236559139785, |
|
"grad_norm": 2.103156089782715, |
|
"learning_rate": 8.488421052631579e-06, |
|
"loss": 1.1426, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.49731182795698925, |
|
"grad_norm": 1.795997142791748, |
|
"learning_rate": 7.957894736842105e-06, |
|
"loss": 1.1144, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 2.546483278274536, |
|
"learning_rate": 7.42736842105263e-06, |
|
"loss": 1.2147, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.5026881720430108, |
|
"grad_norm": 2.0869460105895996, |
|
"learning_rate": 6.896842105263158e-06, |
|
"loss": 1.4491, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.5053763440860215, |
|
"grad_norm": 2.946209192276001, |
|
"learning_rate": 6.3663157894736845e-06, |
|
"loss": 1.5029, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.5080645161290323, |
|
"grad_norm": 1.2845457792282104, |
|
"learning_rate": 5.835789473684211e-06, |
|
"loss": 0.5538, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.510752688172043, |
|
"grad_norm": 1.6108278036117554, |
|
"learning_rate": 5.305263157894736e-06, |
|
"loss": 0.5395, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.5134408602150538, |
|
"grad_norm": 2.4547786712646484, |
|
"learning_rate": 4.774736842105263e-06, |
|
"loss": 0.7314, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.5161290322580645, |
|
"grad_norm": 4.319262981414795, |
|
"learning_rate": 4.244210526315789e-06, |
|
"loss": 2.1648, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.5188172043010753, |
|
"grad_norm": 1.8788155317306519, |
|
"learning_rate": 3.713684210526315e-06, |
|
"loss": 0.9098, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.521505376344086, |
|
"grad_norm": 2.349867820739746, |
|
"learning_rate": 3.1831578947368422e-06, |
|
"loss": 1.2202, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.5241935483870968, |
|
"grad_norm": 2.8968505859375, |
|
"learning_rate": 2.652631578947368e-06, |
|
"loss": 1.8995, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.5268817204301075, |
|
"grad_norm": 4.167923927307129, |
|
"learning_rate": 2.1221052631578947e-06, |
|
"loss": 2.6088, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.5295698924731183, |
|
"grad_norm": 4.212091445922852, |
|
"learning_rate": 1.5915789473684211e-06, |
|
"loss": 1.8908, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.532258064516129, |
|
"grad_norm": 1.670013666152954, |
|
"learning_rate": 1.0610526315789473e-06, |
|
"loss": 0.529, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.5349462365591398, |
|
"grad_norm": 2.4547181129455566, |
|
"learning_rate": 5.305263157894737e-07, |
|
"loss": 0.7957, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.5376344086021505, |
|
"grad_norm": 2.100339651107788, |
|
"learning_rate": 0.0, |
|
"loss": 0.3624, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5376344086021505, |
|
"eval_loss": 0.7936509251594543, |
|
"eval_runtime": 20.0443, |
|
"eval_samples_per_second": 7.833, |
|
"eval_steps_per_second": 1.996, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6.65254329581568e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|