|
{ |
|
"best_metric": 0.22681911289691925, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-200", |
|
"epoch": 0.29027576197387517, |
|
"eval_steps": 50, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.001451378809869376, |
|
"grad_norm": 0.11486738920211792, |
|
"learning_rate": 1.0110000000000001e-05, |
|
"loss": 0.2249, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.001451378809869376, |
|
"eval_loss": 0.27678582072257996, |
|
"eval_runtime": 142.974, |
|
"eval_samples_per_second": 2.035, |
|
"eval_steps_per_second": 0.511, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.002902757619738752, |
|
"grad_norm": 0.11387992650270462, |
|
"learning_rate": 2.0220000000000003e-05, |
|
"loss": 0.4195, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0043541364296081275, |
|
"grad_norm": 0.1770201325416565, |
|
"learning_rate": 3.033e-05, |
|
"loss": 0.3746, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.005805515239477504, |
|
"grad_norm": 0.15089277923107147, |
|
"learning_rate": 4.0440000000000006e-05, |
|
"loss": 0.3461, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.00725689404934688, |
|
"grad_norm": 0.11748471856117249, |
|
"learning_rate": 5.055e-05, |
|
"loss": 0.1969, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.008708272859216255, |
|
"grad_norm": 0.1325271725654602, |
|
"learning_rate": 6.066e-05, |
|
"loss": 0.2835, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.010159651669085631, |
|
"grad_norm": 0.14989273250102997, |
|
"learning_rate": 7.077e-05, |
|
"loss": 0.44, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.011611030478955007, |
|
"grad_norm": 0.1532517969608307, |
|
"learning_rate": 8.088000000000001e-05, |
|
"loss": 0.4155, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.013062409288824383, |
|
"grad_norm": 0.12538368999958038, |
|
"learning_rate": 9.099000000000001e-05, |
|
"loss": 0.3613, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01451378809869376, |
|
"grad_norm": 0.13840965926647186, |
|
"learning_rate": 0.0001011, |
|
"loss": 0.7261, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.015965166908563134, |
|
"grad_norm": 0.1670045405626297, |
|
"learning_rate": 0.00010056789473684211, |
|
"loss": 0.5663, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.01741654571843251, |
|
"grad_norm": 0.13095274567604065, |
|
"learning_rate": 0.00010003578947368421, |
|
"loss": 0.3288, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.018867924528301886, |
|
"grad_norm": 0.19772979617118835, |
|
"learning_rate": 9.950368421052632e-05, |
|
"loss": 0.613, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.020319303338171262, |
|
"grad_norm": 0.16937138140201569, |
|
"learning_rate": 9.897157894736842e-05, |
|
"loss": 0.4953, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.02177068214804064, |
|
"grad_norm": 0.1333189755678177, |
|
"learning_rate": 9.843947368421053e-05, |
|
"loss": 0.2195, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.023222060957910014, |
|
"grad_norm": 0.1428212672472, |
|
"learning_rate": 9.790736842105264e-05, |
|
"loss": 0.2334, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02467343976777939, |
|
"grad_norm": 0.14330728352069855, |
|
"learning_rate": 9.737526315789474e-05, |
|
"loss": 0.3142, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.026124818577648767, |
|
"grad_norm": 0.1493055373430252, |
|
"learning_rate": 9.684315789473684e-05, |
|
"loss": 0.2551, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.027576197387518143, |
|
"grad_norm": 0.16706344485282898, |
|
"learning_rate": 9.631105263157895e-05, |
|
"loss": 0.5574, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.02902757619738752, |
|
"grad_norm": 0.11606133729219437, |
|
"learning_rate": 9.577894736842105e-05, |
|
"loss": 0.1948, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.030478955007256895, |
|
"grad_norm": 0.0992949903011322, |
|
"learning_rate": 9.524684210526317e-05, |
|
"loss": 0.1042, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03193033381712627, |
|
"grad_norm": 0.15178389847278595, |
|
"learning_rate": 9.471473684210526e-05, |
|
"loss": 0.4452, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.033381712626995644, |
|
"grad_norm": 0.15305830538272858, |
|
"learning_rate": 9.418263157894737e-05, |
|
"loss": 0.4336, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.03483309143686502, |
|
"grad_norm": 0.21439702808856964, |
|
"learning_rate": 9.365052631578948e-05, |
|
"loss": 0.346, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.036284470246734396, |
|
"grad_norm": 0.1342068761587143, |
|
"learning_rate": 9.311842105263157e-05, |
|
"loss": 0.3454, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03773584905660377, |
|
"grad_norm": 0.1589202582836151, |
|
"learning_rate": 9.258631578947368e-05, |
|
"loss": 0.4931, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.03918722786647315, |
|
"grad_norm": 0.16286569833755493, |
|
"learning_rate": 9.20542105263158e-05, |
|
"loss": 0.5843, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.040638606676342524, |
|
"grad_norm": 0.11265375465154648, |
|
"learning_rate": 9.15221052631579e-05, |
|
"loss": 0.2569, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.0420899854862119, |
|
"grad_norm": 0.18966278433799744, |
|
"learning_rate": 9.099000000000001e-05, |
|
"loss": 0.4922, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.04354136429608128, |
|
"grad_norm": 0.2602640688419342, |
|
"learning_rate": 9.045789473684212e-05, |
|
"loss": 0.3959, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04499274310595065, |
|
"grad_norm": 0.12830835580825806, |
|
"learning_rate": 8.992578947368421e-05, |
|
"loss": 0.3903, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.04644412191582003, |
|
"grad_norm": 0.10974445939064026, |
|
"learning_rate": 8.939368421052632e-05, |
|
"loss": 0.2795, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.047895500725689405, |
|
"grad_norm": 0.16427350044250488, |
|
"learning_rate": 8.886157894736841e-05, |
|
"loss": 0.4776, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.04934687953555878, |
|
"grad_norm": 0.12569822371006012, |
|
"learning_rate": 8.832947368421054e-05, |
|
"loss": 0.2615, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.05079825834542816, |
|
"grad_norm": 0.10434585809707642, |
|
"learning_rate": 8.779736842105264e-05, |
|
"loss": 0.2248, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.05224963715529753, |
|
"grad_norm": 0.1004101112484932, |
|
"learning_rate": 8.726526315789474e-05, |
|
"loss": 0.2055, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.05370101596516691, |
|
"grad_norm": 0.12491065263748169, |
|
"learning_rate": 8.673315789473685e-05, |
|
"loss": 0.2482, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.055152394775036286, |
|
"grad_norm": 0.11319632083177567, |
|
"learning_rate": 8.620105263157896e-05, |
|
"loss": 0.2185, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.05660377358490566, |
|
"grad_norm": 0.15422259271144867, |
|
"learning_rate": 8.566894736842105e-05, |
|
"loss": 0.3448, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.05805515239477504, |
|
"grad_norm": 0.15866440534591675, |
|
"learning_rate": 8.513684210526316e-05, |
|
"loss": 0.6739, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.059506531204644414, |
|
"grad_norm": 0.1435173898935318, |
|
"learning_rate": 8.460473684210527e-05, |
|
"loss": 0.2159, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.06095791001451379, |
|
"grad_norm": 0.21297410130500793, |
|
"learning_rate": 8.407263157894738e-05, |
|
"loss": 0.3077, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.062409288824383166, |
|
"grad_norm": 0.1251886934041977, |
|
"learning_rate": 8.354052631578948e-05, |
|
"loss": 0.2113, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.06386066763425254, |
|
"grad_norm": 0.1815638542175293, |
|
"learning_rate": 8.300842105263158e-05, |
|
"loss": 0.6564, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.06531204644412192, |
|
"grad_norm": 0.2576989531517029, |
|
"learning_rate": 8.247631578947369e-05, |
|
"loss": 0.5205, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.06676342525399129, |
|
"grad_norm": 0.27275457978248596, |
|
"learning_rate": 8.19442105263158e-05, |
|
"loss": 0.9175, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.06821480406386067, |
|
"grad_norm": 0.26641345024108887, |
|
"learning_rate": 8.141210526315789e-05, |
|
"loss": 0.812, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.06966618287373004, |
|
"grad_norm": 0.2708298861980438, |
|
"learning_rate": 8.088000000000001e-05, |
|
"loss": 0.6076, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.07111756168359942, |
|
"grad_norm": 0.5135082602500916, |
|
"learning_rate": 8.03478947368421e-05, |
|
"loss": 0.9538, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.07256894049346879, |
|
"grad_norm": 0.46205997467041016, |
|
"learning_rate": 7.981578947368421e-05, |
|
"loss": 1.1091, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.07256894049346879, |
|
"eval_loss": 0.24128209054470062, |
|
"eval_runtime": 146.2789, |
|
"eval_samples_per_second": 1.989, |
|
"eval_steps_per_second": 0.499, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.07402031930333818, |
|
"grad_norm": 0.15042653679847717, |
|
"learning_rate": 7.928368421052632e-05, |
|
"loss": 0.5749, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.07547169811320754, |
|
"grad_norm": 0.14070290327072144, |
|
"learning_rate": 7.875157894736842e-05, |
|
"loss": 0.3815, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.07692307692307693, |
|
"grad_norm": 0.12616680562496185, |
|
"learning_rate": 7.821947368421053e-05, |
|
"loss": 0.3485, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.0783744557329463, |
|
"grad_norm": 0.1807371824979782, |
|
"learning_rate": 7.768736842105263e-05, |
|
"loss": 0.6332, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.07982583454281568, |
|
"grad_norm": 0.1610403060913086, |
|
"learning_rate": 7.715526315789474e-05, |
|
"loss": 0.4334, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.08127721335268505, |
|
"grad_norm": 0.16740575432777405, |
|
"learning_rate": 7.662315789473685e-05, |
|
"loss": 0.5169, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.08272859216255443, |
|
"grad_norm": 0.13583853840827942, |
|
"learning_rate": 7.609105263157895e-05, |
|
"loss": 0.3515, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.0841799709724238, |
|
"grad_norm": 0.20852766931056976, |
|
"learning_rate": 7.555894736842105e-05, |
|
"loss": 0.8417, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.08563134978229318, |
|
"grad_norm": 0.12840092182159424, |
|
"learning_rate": 7.502684210526316e-05, |
|
"loss": 0.3527, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.08708272859216255, |
|
"grad_norm": 0.20794260501861572, |
|
"learning_rate": 7.449473684210526e-05, |
|
"loss": 0.8209, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.08853410740203194, |
|
"grad_norm": 0.15332041680812836, |
|
"learning_rate": 7.396263157894738e-05, |
|
"loss": 0.5123, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.0899854862119013, |
|
"grad_norm": 0.12152067571878433, |
|
"learning_rate": 7.343052631578949e-05, |
|
"loss": 0.5442, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.09143686502177069, |
|
"grad_norm": 0.15734325349330902, |
|
"learning_rate": 7.289842105263158e-05, |
|
"loss": 0.4926, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.09288824383164006, |
|
"grad_norm": 0.10759459435939789, |
|
"learning_rate": 7.236631578947369e-05, |
|
"loss": 0.2285, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.09433962264150944, |
|
"grad_norm": 0.11365488916635513, |
|
"learning_rate": 7.183421052631579e-05, |
|
"loss": 0.3032, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.09579100145137881, |
|
"grad_norm": 0.13057614862918854, |
|
"learning_rate": 7.13021052631579e-05, |
|
"loss": 0.4789, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.09724238026124818, |
|
"grad_norm": 0.1509314924478531, |
|
"learning_rate": 7.077e-05, |
|
"loss": 0.5571, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.09869375907111756, |
|
"grad_norm": 0.16016532480716705, |
|
"learning_rate": 7.023789473684211e-05, |
|
"loss": 0.2708, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.10014513788098693, |
|
"grad_norm": 0.2185145616531372, |
|
"learning_rate": 6.970578947368422e-05, |
|
"loss": 0.4046, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.10159651669085631, |
|
"grad_norm": 0.14823275804519653, |
|
"learning_rate": 6.917368421052633e-05, |
|
"loss": 0.5717, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.10304789550072568, |
|
"grad_norm": 0.16261740028858185, |
|
"learning_rate": 6.864157894736842e-05, |
|
"loss": 0.3313, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.10449927431059507, |
|
"grad_norm": 0.1639692485332489, |
|
"learning_rate": 6.810947368421053e-05, |
|
"loss": 0.7474, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.10595065312046444, |
|
"grad_norm": 0.11137572675943375, |
|
"learning_rate": 6.757736842105264e-05, |
|
"loss": 0.3237, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.10740203193033382, |
|
"grad_norm": 0.09403427690267563, |
|
"learning_rate": 6.704526315789473e-05, |
|
"loss": 0.213, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.10885341074020319, |
|
"grad_norm": 0.1198626309633255, |
|
"learning_rate": 6.651315789473685e-05, |
|
"loss": 0.3147, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.11030478955007257, |
|
"grad_norm": 0.18822187185287476, |
|
"learning_rate": 6.598105263157895e-05, |
|
"loss": 0.2479, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.11175616835994194, |
|
"grad_norm": 0.08338137716054916, |
|
"learning_rate": 6.544894736842106e-05, |
|
"loss": 0.1284, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.11320754716981132, |
|
"grad_norm": 0.11395663768053055, |
|
"learning_rate": 6.491684210526317e-05, |
|
"loss": 0.304, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.11465892597968069, |
|
"grad_norm": 0.12172822654247284, |
|
"learning_rate": 6.438473684210526e-05, |
|
"loss": 0.5225, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.11611030478955008, |
|
"grad_norm": 0.18436570465564728, |
|
"learning_rate": 6.385263157894737e-05, |
|
"loss": 1.1369, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.11756168359941944, |
|
"grad_norm": 0.08676780015230179, |
|
"learning_rate": 6.332052631578948e-05, |
|
"loss": 0.2241, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.11901306240928883, |
|
"grad_norm": 0.14665937423706055, |
|
"learning_rate": 6.278842105263159e-05, |
|
"loss": 0.4818, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.1204644412191582, |
|
"grad_norm": 0.12407947331666946, |
|
"learning_rate": 6.22563157894737e-05, |
|
"loss": 0.4061, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.12191582002902758, |
|
"grad_norm": 0.12156087160110474, |
|
"learning_rate": 6.172421052631579e-05, |
|
"loss": 0.4842, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.12336719883889695, |
|
"grad_norm": 0.164754256606102, |
|
"learning_rate": 6.11921052631579e-05, |
|
"loss": 0.6039, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.12481857764876633, |
|
"grad_norm": 0.13035471737384796, |
|
"learning_rate": 6.066e-05, |
|
"loss": 0.5963, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.1262699564586357, |
|
"grad_norm": 0.08667127043008804, |
|
"learning_rate": 6.012789473684211e-05, |
|
"loss": 0.104, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.12772133526850507, |
|
"grad_norm": 0.12500996887683868, |
|
"learning_rate": 5.959578947368421e-05, |
|
"loss": 0.1519, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.12917271407837447, |
|
"grad_norm": 0.1052846610546112, |
|
"learning_rate": 5.9063684210526324e-05, |
|
"loss": 0.168, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.13062409288824384, |
|
"grad_norm": 0.16326619684696198, |
|
"learning_rate": 5.8531578947368425e-05, |
|
"loss": 1.0778, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1320754716981132, |
|
"grad_norm": 0.18031665682792664, |
|
"learning_rate": 5.7999473684210527e-05, |
|
"loss": 0.8578, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.13352685050798258, |
|
"grad_norm": 0.14327001571655273, |
|
"learning_rate": 5.7467368421052635e-05, |
|
"loss": 0.8367, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.13497822931785197, |
|
"grad_norm": 0.12158166617155075, |
|
"learning_rate": 5.6935263157894736e-05, |
|
"loss": 0.2978, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.13642960812772134, |
|
"grad_norm": 0.17898587882518768, |
|
"learning_rate": 5.640315789473684e-05, |
|
"loss": 0.7248, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.1378809869375907, |
|
"grad_norm": 0.1620013415813446, |
|
"learning_rate": 5.587105263157895e-05, |
|
"loss": 0.5909, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.13933236574746008, |
|
"grad_norm": 0.24596630036830902, |
|
"learning_rate": 5.533894736842106e-05, |
|
"loss": 1.3, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.14078374455732948, |
|
"grad_norm": 0.28974536061286926, |
|
"learning_rate": 5.480684210526316e-05, |
|
"loss": 1.0699, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.14223512336719885, |
|
"grad_norm": 0.2533659338951111, |
|
"learning_rate": 5.4274736842105264e-05, |
|
"loss": 0.6266, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.14368650217706821, |
|
"grad_norm": 0.36984333395957947, |
|
"learning_rate": 5.374263157894737e-05, |
|
"loss": 1.3249, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.14513788098693758, |
|
"grad_norm": 0.4221994876861572, |
|
"learning_rate": 5.3210526315789474e-05, |
|
"loss": 0.9726, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14513788098693758, |
|
"eval_loss": 0.2286633849143982, |
|
"eval_runtime": 145.162, |
|
"eval_samples_per_second": 2.005, |
|
"eval_steps_per_second": 0.503, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14658925979680695, |
|
"grad_norm": 0.13860496878623962, |
|
"learning_rate": 5.2678421052631576e-05, |
|
"loss": 0.7931, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.14804063860667635, |
|
"grad_norm": 0.1530522257089615, |
|
"learning_rate": 5.214631578947369e-05, |
|
"loss": 0.4883, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.14949201741654572, |
|
"grad_norm": 0.09421262890100479, |
|
"learning_rate": 5.161421052631579e-05, |
|
"loss": 0.1757, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.1509433962264151, |
|
"grad_norm": 0.11253266036510468, |
|
"learning_rate": 5.10821052631579e-05, |
|
"loss": 0.4355, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.15239477503628446, |
|
"grad_norm": 0.11226090788841248, |
|
"learning_rate": 5.055e-05, |
|
"loss": 0.3406, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.15384615384615385, |
|
"grad_norm": 0.09969495981931686, |
|
"learning_rate": 5.0017894736842104e-05, |
|
"loss": 0.33, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.15529753265602322, |
|
"grad_norm": 0.14994554221630096, |
|
"learning_rate": 4.948578947368421e-05, |
|
"loss": 0.7712, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.1567489114658926, |
|
"grad_norm": 0.16747573018074036, |
|
"learning_rate": 4.895368421052632e-05, |
|
"loss": 0.4805, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.15820029027576196, |
|
"grad_norm": 0.15101346373558044, |
|
"learning_rate": 4.842157894736842e-05, |
|
"loss": 0.6745, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.15965166908563136, |
|
"grad_norm": 0.14325077831745148, |
|
"learning_rate": 4.7889473684210523e-05, |
|
"loss": 0.2481, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.16110304789550073, |
|
"grad_norm": 0.6589614152908325, |
|
"learning_rate": 4.735736842105263e-05, |
|
"loss": 0.385, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.1625544267053701, |
|
"grad_norm": 0.14817237854003906, |
|
"learning_rate": 4.682526315789474e-05, |
|
"loss": 0.4519, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.16400580551523947, |
|
"grad_norm": 0.12075196951627731, |
|
"learning_rate": 4.629315789473684e-05, |
|
"loss": 0.32, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.16545718432510886, |
|
"grad_norm": 0.10896284133195877, |
|
"learning_rate": 4.576105263157895e-05, |
|
"loss": 0.252, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.16690856313497823, |
|
"grad_norm": 0.15797024965286255, |
|
"learning_rate": 4.522894736842106e-05, |
|
"loss": 0.7193, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.1683599419448476, |
|
"grad_norm": 0.12811946868896484, |
|
"learning_rate": 4.469684210526316e-05, |
|
"loss": 0.5991, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.16981132075471697, |
|
"grad_norm": 0.17125706374645233, |
|
"learning_rate": 4.416473684210527e-05, |
|
"loss": 0.6612, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.17126269956458637, |
|
"grad_norm": 0.21033529937267303, |
|
"learning_rate": 4.363263157894737e-05, |
|
"loss": 0.637, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.17271407837445574, |
|
"grad_norm": 0.14343442022800446, |
|
"learning_rate": 4.310052631578948e-05, |
|
"loss": 0.3919, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.1741654571843251, |
|
"grad_norm": 0.09223167598247528, |
|
"learning_rate": 4.256842105263158e-05, |
|
"loss": 0.2206, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.17561683599419448, |
|
"grad_norm": 0.10201866924762726, |
|
"learning_rate": 4.203631578947369e-05, |
|
"loss": 0.2657, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.17706821480406387, |
|
"grad_norm": 0.09830533713102341, |
|
"learning_rate": 4.150421052631579e-05, |
|
"loss": 0.1767, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.17851959361393324, |
|
"grad_norm": 0.11305135488510132, |
|
"learning_rate": 4.09721052631579e-05, |
|
"loss": 0.3528, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.1799709724238026, |
|
"grad_norm": 0.11795450747013092, |
|
"learning_rate": 4.0440000000000006e-05, |
|
"loss": 0.2826, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.18142235123367198, |
|
"grad_norm": 0.10810278356075287, |
|
"learning_rate": 3.990789473684211e-05, |
|
"loss": 0.3728, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.18287373004354138, |
|
"grad_norm": 0.07889194041490555, |
|
"learning_rate": 3.937578947368421e-05, |
|
"loss": 0.1485, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.18432510885341075, |
|
"grad_norm": 0.12167462706565857, |
|
"learning_rate": 3.884368421052632e-05, |
|
"loss": 0.3191, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.18577648766328012, |
|
"grad_norm": 0.13652844727039337, |
|
"learning_rate": 3.8311578947368426e-05, |
|
"loss": 0.3272, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.18722786647314948, |
|
"grad_norm": 0.11355964839458466, |
|
"learning_rate": 3.777947368421053e-05, |
|
"loss": 0.2216, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.18867924528301888, |
|
"grad_norm": 0.12072530388832092, |
|
"learning_rate": 3.724736842105263e-05, |
|
"loss": 0.2511, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.19013062409288825, |
|
"grad_norm": 0.11075890809297562, |
|
"learning_rate": 3.6715263157894744e-05, |
|
"loss": 0.1769, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.19158200290275762, |
|
"grad_norm": 0.09322332590818405, |
|
"learning_rate": 3.6183157894736845e-05, |
|
"loss": 0.2052, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.193033381712627, |
|
"grad_norm": 0.12875616550445557, |
|
"learning_rate": 3.565105263157895e-05, |
|
"loss": 0.1961, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.19448476052249636, |
|
"grad_norm": 0.11314037442207336, |
|
"learning_rate": 3.5118947368421055e-05, |
|
"loss": 0.2227, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.19593613933236576, |
|
"grad_norm": 0.10332886129617691, |
|
"learning_rate": 3.458684210526316e-05, |
|
"loss": 0.339, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.19738751814223512, |
|
"grad_norm": 0.17729254066944122, |
|
"learning_rate": 3.4054736842105265e-05, |
|
"loss": 0.2802, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.1988388969521045, |
|
"grad_norm": 0.08777690678834915, |
|
"learning_rate": 3.3522631578947366e-05, |
|
"loss": 0.1391, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.20029027576197386, |
|
"grad_norm": 0.1536087989807129, |
|
"learning_rate": 3.2990526315789475e-05, |
|
"loss": 0.3643, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.20174165457184326, |
|
"grad_norm": 0.11051812022924423, |
|
"learning_rate": 3.245842105263158e-05, |
|
"loss": 0.1991, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.20319303338171263, |
|
"grad_norm": 0.09295262396335602, |
|
"learning_rate": 3.1926315789473685e-05, |
|
"loss": 0.2913, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.204644412191582, |
|
"grad_norm": 0.13505689799785614, |
|
"learning_rate": 3.139421052631579e-05, |
|
"loss": 0.8371, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.20609579100145137, |
|
"grad_norm": 0.1267169862985611, |
|
"learning_rate": 3.0862105263157894e-05, |
|
"loss": 0.6024, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.20754716981132076, |
|
"grad_norm": 0.13407814502716064, |
|
"learning_rate": 3.033e-05, |
|
"loss": 0.4159, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.20899854862119013, |
|
"grad_norm": 0.11469519138336182, |
|
"learning_rate": 2.9797894736842104e-05, |
|
"loss": 0.2402, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2104499274310595, |
|
"grad_norm": 0.169059157371521, |
|
"learning_rate": 2.9265789473684213e-05, |
|
"loss": 0.3534, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.21190130624092887, |
|
"grad_norm": 0.19773763418197632, |
|
"learning_rate": 2.8733684210526317e-05, |
|
"loss": 0.5001, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.21335268505079827, |
|
"grad_norm": 0.4277418553829193, |
|
"learning_rate": 2.820157894736842e-05, |
|
"loss": 0.7581, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.21480406386066764, |
|
"grad_norm": 0.2791711688041687, |
|
"learning_rate": 2.766947368421053e-05, |
|
"loss": 0.8721, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.216255442670537, |
|
"grad_norm": 0.40901464223861694, |
|
"learning_rate": 2.7137368421052632e-05, |
|
"loss": 0.8766, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.21770682148040638, |
|
"grad_norm": 0.39813101291656494, |
|
"learning_rate": 2.6605263157894737e-05, |
|
"loss": 0.5074, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.21770682148040638, |
|
"eval_loss": 0.23037858307361603, |
|
"eval_runtime": 143.626, |
|
"eval_samples_per_second": 2.026, |
|
"eval_steps_per_second": 0.508, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.21915820029027577, |
|
"grad_norm": 0.16082116961479187, |
|
"learning_rate": 2.6073157894736845e-05, |
|
"loss": 1.008, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.22060957910014514, |
|
"grad_norm": 0.17968745529651642, |
|
"learning_rate": 2.554105263157895e-05, |
|
"loss": 0.8331, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.2220609579100145, |
|
"grad_norm": 0.17362633347511292, |
|
"learning_rate": 2.5008947368421052e-05, |
|
"loss": 0.5327, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.22351233671988388, |
|
"grad_norm": 0.1643386036157608, |
|
"learning_rate": 2.447684210526316e-05, |
|
"loss": 1.0695, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.22496371552975328, |
|
"grad_norm": 0.12344465404748917, |
|
"learning_rate": 2.3944736842105262e-05, |
|
"loss": 0.5125, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.22641509433962265, |
|
"grad_norm": 0.2044529914855957, |
|
"learning_rate": 2.341263157894737e-05, |
|
"loss": 1.0597, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.22786647314949202, |
|
"grad_norm": 0.1536094695329666, |
|
"learning_rate": 2.2880526315789475e-05, |
|
"loss": 0.2413, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.22931785195936139, |
|
"grad_norm": 0.16920588910579681, |
|
"learning_rate": 2.234842105263158e-05, |
|
"loss": 0.7497, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.23076923076923078, |
|
"grad_norm": 0.12118630111217499, |
|
"learning_rate": 2.1816315789473685e-05, |
|
"loss": 0.3208, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.23222060957910015, |
|
"grad_norm": 0.14840605854988098, |
|
"learning_rate": 2.128421052631579e-05, |
|
"loss": 0.387, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.23367198838896952, |
|
"grad_norm": 0.10502210259437561, |
|
"learning_rate": 2.0752105263157895e-05, |
|
"loss": 0.4105, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.2351233671988389, |
|
"grad_norm": 0.23360522091388702, |
|
"learning_rate": 2.0220000000000003e-05, |
|
"loss": 0.4883, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.2365747460087083, |
|
"grad_norm": 0.1557024121284485, |
|
"learning_rate": 1.9687894736842104e-05, |
|
"loss": 0.3857, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.23802612481857766, |
|
"grad_norm": 0.15826496481895447, |
|
"learning_rate": 1.9155789473684213e-05, |
|
"loss": 0.8257, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.23947750362844702, |
|
"grad_norm": 0.12281662970781326, |
|
"learning_rate": 1.8623684210526314e-05, |
|
"loss": 0.398, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.2409288824383164, |
|
"grad_norm": 0.17840005457401276, |
|
"learning_rate": 1.8091578947368423e-05, |
|
"loss": 1.0399, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.24238026124818576, |
|
"grad_norm": 0.2307438999414444, |
|
"learning_rate": 1.7559473684210528e-05, |
|
"loss": 0.6128, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.24383164005805516, |
|
"grad_norm": 0.12274490296840668, |
|
"learning_rate": 1.7027368421052632e-05, |
|
"loss": 0.3564, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.24528301886792453, |
|
"grad_norm": 0.12504906952381134, |
|
"learning_rate": 1.6495263157894737e-05, |
|
"loss": 0.5111, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.2467343976777939, |
|
"grad_norm": 0.10843324661254883, |
|
"learning_rate": 1.5963157894736842e-05, |
|
"loss": 0.3324, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.24818577648766327, |
|
"grad_norm": 0.12919478118419647, |
|
"learning_rate": 1.5431052631578947e-05, |
|
"loss": 0.3152, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.24963715529753266, |
|
"grad_norm": 0.19524158537387848, |
|
"learning_rate": 1.4898947368421052e-05, |
|
"loss": 0.4494, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.251088534107402, |
|
"grad_norm": 0.13108614087104797, |
|
"learning_rate": 1.4366842105263159e-05, |
|
"loss": 0.2326, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.2525399129172714, |
|
"grad_norm": 0.09462467581033707, |
|
"learning_rate": 1.3834736842105265e-05, |
|
"loss": 0.3285, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.2539912917271408, |
|
"grad_norm": 0.11037827283143997, |
|
"learning_rate": 1.3302631578947369e-05, |
|
"loss": 0.3239, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.25544267053701014, |
|
"grad_norm": 0.13191723823547363, |
|
"learning_rate": 1.2770526315789475e-05, |
|
"loss": 0.9917, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.25689404934687954, |
|
"grad_norm": 0.11104253679513931, |
|
"learning_rate": 1.223842105263158e-05, |
|
"loss": 0.3113, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.25834542815674894, |
|
"grad_norm": 0.16667191684246063, |
|
"learning_rate": 1.1706315789473685e-05, |
|
"loss": 0.5103, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.2597968069666183, |
|
"grad_norm": 0.13359755277633667, |
|
"learning_rate": 1.117421052631579e-05, |
|
"loss": 0.5022, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.2612481857764877, |
|
"grad_norm": 0.11774896085262299, |
|
"learning_rate": 1.0642105263157895e-05, |
|
"loss": 0.2854, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.262699564586357, |
|
"grad_norm": 0.14967772364616394, |
|
"learning_rate": 1.0110000000000001e-05, |
|
"loss": 0.3724, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.2641509433962264, |
|
"grad_norm": 0.17375238239765167, |
|
"learning_rate": 9.577894736842106e-06, |
|
"loss": 0.5366, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.2656023222060958, |
|
"grad_norm": 0.12037857621908188, |
|
"learning_rate": 9.045789473684211e-06, |
|
"loss": 0.2978, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.26705370101596515, |
|
"grad_norm": 0.09681455045938492, |
|
"learning_rate": 8.513684210526316e-06, |
|
"loss": 0.2166, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.26850507982583455, |
|
"grad_norm": 0.09586142748594284, |
|
"learning_rate": 7.981578947368421e-06, |
|
"loss": 0.655, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.26995645863570394, |
|
"grad_norm": 0.09345651417970657, |
|
"learning_rate": 7.449473684210526e-06, |
|
"loss": 0.2769, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.2714078374455733, |
|
"grad_norm": 0.09641759097576141, |
|
"learning_rate": 6.917368421052633e-06, |
|
"loss": 0.2374, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.2728592162554427, |
|
"grad_norm": 0.12996430695056915, |
|
"learning_rate": 6.385263157894738e-06, |
|
"loss": 0.2477, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.274310595065312, |
|
"grad_norm": 0.14311741292476654, |
|
"learning_rate": 5.8531578947368425e-06, |
|
"loss": 0.2815, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.2757619738751814, |
|
"grad_norm": 0.12616676092147827, |
|
"learning_rate": 5.321052631578947e-06, |
|
"loss": 0.2024, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2772133526850508, |
|
"grad_norm": 0.10585512220859528, |
|
"learning_rate": 4.788947368421053e-06, |
|
"loss": 0.2822, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.27866473149492016, |
|
"grad_norm": 0.1168060377240181, |
|
"learning_rate": 4.256842105263158e-06, |
|
"loss": 0.1666, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.28011611030478956, |
|
"grad_norm": 0.12357722222805023, |
|
"learning_rate": 3.724736842105263e-06, |
|
"loss": 0.4234, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.28156748911465895, |
|
"grad_norm": 0.1898827701807022, |
|
"learning_rate": 3.192631578947369e-06, |
|
"loss": 0.426, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.2830188679245283, |
|
"grad_norm": 0.2734042704105377, |
|
"learning_rate": 2.6605263157894737e-06, |
|
"loss": 0.8311, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.2844702467343977, |
|
"grad_norm": 0.23327644169330597, |
|
"learning_rate": 2.128421052631579e-06, |
|
"loss": 0.8963, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.28592162554426703, |
|
"grad_norm": 0.23755201697349548, |
|
"learning_rate": 1.5963157894736844e-06, |
|
"loss": 0.6194, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.28737300435413643, |
|
"grad_norm": 0.3003562390804291, |
|
"learning_rate": 1.0642105263157895e-06, |
|
"loss": 0.6794, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.2888243831640058, |
|
"grad_norm": 0.45043861865997314, |
|
"learning_rate": 5.321052631578948e-07, |
|
"loss": 0.6781, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.29027576197387517, |
|
"grad_norm": 0.43113571405410767, |
|
"learning_rate": 0.0, |
|
"loss": 1.0902, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.29027576197387517, |
|
"eval_loss": 0.22681911289691925, |
|
"eval_runtime": 144.9273, |
|
"eval_samples_per_second": 2.008, |
|
"eval_steps_per_second": 0.504, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.4053840999350272e+17, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|