|
{ |
|
"best_metric": 11.5, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-150", |
|
"epoch": 1.0, |
|
"eval_steps": 150, |
|
"global_step": 169, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.005917159763313609, |
|
"grad_norm": 3.417472544242628e-05, |
|
"learning_rate": 5e-06, |
|
"loss": 46.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.005917159763313609, |
|
"eval_loss": 11.5, |
|
"eval_runtime": 1.4564, |
|
"eval_samples_per_second": 195.691, |
|
"eval_steps_per_second": 49.438, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.011834319526627219, |
|
"grad_norm": 3.082885086769238e-05, |
|
"learning_rate": 1e-05, |
|
"loss": 46.0, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01775147928994083, |
|
"grad_norm": 1.8164773791795596e-05, |
|
"learning_rate": 1.5e-05, |
|
"loss": 46.0, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.023668639053254437, |
|
"grad_norm": 2.2182997781783342e-05, |
|
"learning_rate": 2e-05, |
|
"loss": 46.0, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.029585798816568046, |
|
"grad_norm": 3.892195672960952e-05, |
|
"learning_rate": 2.5e-05, |
|
"loss": 46.0, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.03550295857988166, |
|
"grad_norm": 3.778233440243639e-05, |
|
"learning_rate": 3e-05, |
|
"loss": 46.0, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.04142011834319527, |
|
"grad_norm": 3.2646468753227964e-05, |
|
"learning_rate": 3.5e-05, |
|
"loss": 46.0, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.047337278106508875, |
|
"grad_norm": 2.2486128727905452e-05, |
|
"learning_rate": 4e-05, |
|
"loss": 46.0, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.05325443786982249, |
|
"grad_norm": 4.4991043978370726e-05, |
|
"learning_rate": 4.5e-05, |
|
"loss": 46.0, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.05917159763313609, |
|
"grad_norm": 3.6158784496365115e-05, |
|
"learning_rate": 5e-05, |
|
"loss": 46.0, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0650887573964497, |
|
"grad_norm": 3.0189292374416254e-05, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 46.0, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.07100591715976332, |
|
"grad_norm": 2.9184000595705584e-05, |
|
"learning_rate": 6e-05, |
|
"loss": 46.0, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.07692307692307693, |
|
"grad_norm": 2.1351072064135224e-05, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 46.0, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.08284023668639054, |
|
"grad_norm": 3.9580500015290454e-05, |
|
"learning_rate": 7e-05, |
|
"loss": 46.0, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.08875739644970414, |
|
"grad_norm": 3.6071560316486284e-05, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 46.0, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.09467455621301775, |
|
"grad_norm": 2.757324546109885e-05, |
|
"learning_rate": 8e-05, |
|
"loss": 46.0, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.10059171597633136, |
|
"grad_norm": 4.150742097408511e-05, |
|
"learning_rate": 8.5e-05, |
|
"loss": 46.0, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.10650887573964497, |
|
"grad_norm": 3.0252529541030526e-05, |
|
"learning_rate": 9e-05, |
|
"loss": 46.0, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.11242603550295859, |
|
"grad_norm": 4.47008824266959e-05, |
|
"learning_rate": 9.5e-05, |
|
"loss": 46.0, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.11834319526627218, |
|
"grad_norm": 4.676431854022667e-05, |
|
"learning_rate": 0.0001, |
|
"loss": 46.0, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.1242603550295858, |
|
"grad_norm": 3.7658752262359485e-05, |
|
"learning_rate": 9.99888864929809e-05, |
|
"loss": 46.0, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.1301775147928994, |
|
"grad_norm": 5.5233460443560034e-05, |
|
"learning_rate": 9.995555091232516e-05, |
|
"loss": 46.0, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.13609467455621302, |
|
"grad_norm": 5.739947300753556e-05, |
|
"learning_rate": 9.990000807704114e-05, |
|
"loss": 46.0, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.14201183431952663, |
|
"grad_norm": 4.8324771341867745e-05, |
|
"learning_rate": 9.982228267815643e-05, |
|
"loss": 46.0, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.14792899408284024, |
|
"grad_norm": 4.9891212256625295e-05, |
|
"learning_rate": 9.972240926774168e-05, |
|
"loss": 46.0, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.15384615384615385, |
|
"grad_norm": 3.683907198137604e-05, |
|
"learning_rate": 9.96004322435508e-05, |
|
"loss": 46.0, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.15976331360946747, |
|
"grad_norm": 5.67148053960409e-05, |
|
"learning_rate": 9.945640582928437e-05, |
|
"loss": 46.0, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.16568047337278108, |
|
"grad_norm": 6.611295975744724e-05, |
|
"learning_rate": 9.929039405048501e-05, |
|
"loss": 46.0, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.17159763313609466, |
|
"grad_norm": 4.908446135232225e-05, |
|
"learning_rate": 9.910247070607552e-05, |
|
"loss": 46.0, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.17751479289940827, |
|
"grad_norm": 6.706344720441848e-05, |
|
"learning_rate": 9.889271933555213e-05, |
|
"loss": 46.0, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.1834319526627219, |
|
"grad_norm": 8.02269860287197e-05, |
|
"learning_rate": 9.866123318184803e-05, |
|
"loss": 46.0, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.1893491124260355, |
|
"grad_norm": 6.0282156482571736e-05, |
|
"learning_rate": 9.840811514988294e-05, |
|
"loss": 46.0, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.1952662721893491, |
|
"grad_norm": 7.383022602880374e-05, |
|
"learning_rate": 9.813347776081789e-05, |
|
"loss": 46.0, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.20118343195266272, |
|
"grad_norm": 6.196251342771575e-05, |
|
"learning_rate": 9.783744310203491e-05, |
|
"loss": 46.0, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.20710059171597633, |
|
"grad_norm": 8.574208186473697e-05, |
|
"learning_rate": 9.752014277286432e-05, |
|
"loss": 46.0, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.21301775147928995, |
|
"grad_norm": 0.00013027463865000755, |
|
"learning_rate": 9.718171782608356e-05, |
|
"loss": 46.0, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.21893491124260356, |
|
"grad_norm": 0.00014059020031709224, |
|
"learning_rate": 9.682231870521347e-05, |
|
"loss": 46.0, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.22485207100591717, |
|
"grad_norm": 9.982455958379433e-05, |
|
"learning_rate": 9.644210517764014e-05, |
|
"loss": 46.0, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.23076923076923078, |
|
"grad_norm": 8.106425229925662e-05, |
|
"learning_rate": 9.60412462635919e-05, |
|
"loss": 46.0, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.23668639053254437, |
|
"grad_norm": 7.230762275867164e-05, |
|
"learning_rate": 9.561992016100293e-05, |
|
"loss": 46.0, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.24260355029585798, |
|
"grad_norm": 7.070993888191879e-05, |
|
"learning_rate": 9.517831416629716e-05, |
|
"loss": 46.0, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.2485207100591716, |
|
"grad_norm": 0.0001705414615571499, |
|
"learning_rate": 9.471662459112747e-05, |
|
"loss": 46.0, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.25443786982248523, |
|
"grad_norm": 0.0001224745938088745, |
|
"learning_rate": 9.423505667510724e-05, |
|
"loss": 46.0, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.2603550295857988, |
|
"grad_norm": 7.8432189184241e-05, |
|
"learning_rate": 9.373382449457304e-05, |
|
"loss": 46.0, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.26627218934911245, |
|
"grad_norm": 6.436308467527851e-05, |
|
"learning_rate": 9.321315086741916e-05, |
|
"loss": 46.0, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.27218934911242604, |
|
"grad_norm": 9.919815784087405e-05, |
|
"learning_rate": 9.267326725404599e-05, |
|
"loss": 46.0, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.2781065088757396, |
|
"grad_norm": 7.329176878556609e-05, |
|
"learning_rate": 9.21144136544666e-05, |
|
"loss": 46.0, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.28402366863905326, |
|
"grad_norm": 0.00013015043805353343, |
|
"learning_rate": 9.153683850161706e-05, |
|
"loss": 46.0, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.28994082840236685, |
|
"grad_norm": 0.000146089427289553, |
|
"learning_rate": 9.094079855091797e-05, |
|
"loss": 46.0, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.2958579881656805, |
|
"grad_norm": 0.00011953832290600985, |
|
"learning_rate": 9.032655876613636e-05, |
|
"loss": 46.0, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.30177514792899407, |
|
"grad_norm": 0.0001868321414804086, |
|
"learning_rate": 8.96943922015986e-05, |
|
"loss": 46.0, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.3076923076923077, |
|
"grad_norm": 9.340223914477974e-05, |
|
"learning_rate": 8.904457988080681e-05, |
|
"loss": 46.0, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.3136094674556213, |
|
"grad_norm": 0.00012810462794732302, |
|
"learning_rate": 8.83774106715125e-05, |
|
"loss": 46.0, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.31952662721893493, |
|
"grad_norm": 0.0001306675694650039, |
|
"learning_rate": 8.76931811573033e-05, |
|
"loss": 46.0, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.3254437869822485, |
|
"grad_norm": 0.00019835562852676958, |
|
"learning_rate": 8.699219550575953e-05, |
|
"loss": 46.0, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.33136094674556216, |
|
"grad_norm": 0.00010900765482801944, |
|
"learning_rate": 8.627476533323957e-05, |
|
"loss": 46.0, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.33727810650887574, |
|
"grad_norm": 0.00016158001380972564, |
|
"learning_rate": 8.554120956635375e-05, |
|
"loss": 46.0, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.3431952662721893, |
|
"grad_norm": 0.00012027497723465785, |
|
"learning_rate": 8.479185430018858e-05, |
|
"loss": 46.0, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.34911242603550297, |
|
"grad_norm": 0.00011706927034538239, |
|
"learning_rate": 8.402703265334455e-05, |
|
"loss": 46.0, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.35502958579881655, |
|
"grad_norm": 0.00020923654665239155, |
|
"learning_rate": 8.324708461985124e-05, |
|
"loss": 46.0, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.3609467455621302, |
|
"grad_norm": 0.00013141462113708258, |
|
"learning_rate": 8.245235691802644e-05, |
|
"loss": 46.0, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.3668639053254438, |
|
"grad_norm": 0.0001231282512890175, |
|
"learning_rate": 8.164320283634585e-05, |
|
"loss": 46.0, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.3727810650887574, |
|
"grad_norm": 0.00014354191080201417, |
|
"learning_rate": 8.081998207639212e-05, |
|
"loss": 46.0, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.378698224852071, |
|
"grad_norm": 0.00018254865426570177, |
|
"learning_rate": 7.998306059295301e-05, |
|
"loss": 46.0, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 0.00015154361608438194, |
|
"learning_rate": 7.913281043133978e-05, |
|
"loss": 46.0, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.3905325443786982, |
|
"grad_norm": 0.00014137514517642558, |
|
"learning_rate": 7.826960956199794e-05, |
|
"loss": 46.0, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.39644970414201186, |
|
"grad_norm": 0.000186969613423571, |
|
"learning_rate": 7.739384171248435e-05, |
|
"loss": 46.0, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.40236686390532544, |
|
"grad_norm": 0.00019048571994062513, |
|
"learning_rate": 7.650589619688469e-05, |
|
"loss": 46.0, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.40828402366863903, |
|
"grad_norm": 0.000217212233110331, |
|
"learning_rate": 7.560616774274775e-05, |
|
"loss": 46.0, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.41420118343195267, |
|
"grad_norm": 0.00014128351176623255, |
|
"learning_rate": 7.469505631561317e-05, |
|
"loss": 46.0, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.42011834319526625, |
|
"grad_norm": 0.000143827244755812, |
|
"learning_rate": 7.377296694121058e-05, |
|
"loss": 46.0, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.4260355029585799, |
|
"grad_norm": 0.00021143256162758917, |
|
"learning_rate": 7.284030952540937e-05, |
|
"loss": 46.0, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.4319526627218935, |
|
"grad_norm": 0.0002473762142471969, |
|
"learning_rate": 7.189749867199899e-05, |
|
"loss": 46.0, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.4378698224852071, |
|
"grad_norm": 0.00019267069001216441, |
|
"learning_rate": 7.094495349838092e-05, |
|
"loss": 46.0, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.4437869822485207, |
|
"grad_norm": 0.0002110818022629246, |
|
"learning_rate": 6.998309744925411e-05, |
|
"loss": 46.0, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.44970414201183434, |
|
"grad_norm": 0.00015903066378086805, |
|
"learning_rate": 6.901235810837669e-05, |
|
"loss": 46.0, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.4556213017751479, |
|
"grad_norm": 0.0001764715852914378, |
|
"learning_rate": 6.803316700848779e-05, |
|
"loss": 46.0, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.46153846153846156, |
|
"grad_norm": 0.00025015155551955104, |
|
"learning_rate": 6.704595943947385e-05, |
|
"loss": 46.0, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.46745562130177515, |
|
"grad_norm": 0.00027083579334430397, |
|
"learning_rate": 6.605117425486482e-05, |
|
"loss": 46.0, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.47337278106508873, |
|
"grad_norm": 0.00024672580184414983, |
|
"learning_rate": 6.504925367674594e-05, |
|
"loss": 46.0, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.47928994082840237, |
|
"grad_norm": 0.0003929804952349514, |
|
"learning_rate": 6.404064309917231e-05, |
|
"loss": 46.0, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.48520710059171596, |
|
"grad_norm": 0.0004307347990106791, |
|
"learning_rate": 6.302579089017327e-05, |
|
"loss": 46.0, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.4911242603550296, |
|
"grad_norm": 0.00020378813496790826, |
|
"learning_rate": 6.200514819243476e-05, |
|
"loss": 46.0, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.4970414201183432, |
|
"grad_norm": 0.00033084748429246247, |
|
"learning_rate": 6.097916872274815e-05, |
|
"loss": 46.0, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.5029585798816568, |
|
"grad_norm": 0.0002085510641336441, |
|
"learning_rate": 5.994830857031499e-05, |
|
"loss": 46.0, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.5088757396449705, |
|
"grad_norm": 0.00018850974447559565, |
|
"learning_rate": 5.891302599399685e-05, |
|
"loss": 46.0, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.514792899408284, |
|
"grad_norm": 0.00025538477348163724, |
|
"learning_rate": 5.78737812186009e-05, |
|
"loss": 46.0, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.5207100591715976, |
|
"grad_norm": 0.00026542809791862965, |
|
"learning_rate": 5.683103623029135e-05, |
|
"loss": 46.0, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.5266272189349113, |
|
"grad_norm": 0.00019196125504095107, |
|
"learning_rate": 5.578525457121807e-05, |
|
"loss": 46.0, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.5325443786982249, |
|
"grad_norm": 0.00026723742485046387, |
|
"learning_rate": 5.473690113345342e-05, |
|
"loss": 46.0, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.5384615384615384, |
|
"grad_norm": 0.0002167681377613917, |
|
"learning_rate": 5.368644195232896e-05, |
|
"loss": 46.0, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.5443786982248521, |
|
"grad_norm": 0.00020223991305101663, |
|
"learning_rate": 5.263434399926398e-05, |
|
"loss": 46.0, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.5502958579881657, |
|
"grad_norm": 0.00031044651404954493, |
|
"learning_rate": 5.158107497417795e-05, |
|
"loss": 46.0, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.5562130177514792, |
|
"grad_norm": 0.00022696513042319566, |
|
"learning_rate": 5.052710309757899e-05, |
|
"loss": 46.0, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.5621301775147929, |
|
"grad_norm": 0.0003360291011631489, |
|
"learning_rate": 4.947289690242102e-05, |
|
"loss": 46.0, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.5680473372781065, |
|
"grad_norm": 0.0002260725013911724, |
|
"learning_rate": 4.841892502582206e-05, |
|
"loss": 46.0, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.5739644970414202, |
|
"grad_norm": 0.0002723989891819656, |
|
"learning_rate": 4.736565600073602e-05, |
|
"loss": 46.0, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.5798816568047337, |
|
"grad_norm": 0.00029556764638982713, |
|
"learning_rate": 4.631355804767105e-05, |
|
"loss": 46.0, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.5857988165680473, |
|
"grad_norm": 0.0002435845381114632, |
|
"learning_rate": 4.5263098866546586e-05, |
|
"loss": 46.0, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.591715976331361, |
|
"grad_norm": 0.0002843434049282223, |
|
"learning_rate": 4.421474542878195e-05, |
|
"loss": 46.0, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5976331360946746, |
|
"grad_norm": 0.0002432662295177579, |
|
"learning_rate": 4.316896376970866e-05, |
|
"loss": 46.0, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.6035502958579881, |
|
"grad_norm": 0.0004484949167817831, |
|
"learning_rate": 4.212621878139912e-05, |
|
"loss": 46.0, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.6094674556213018, |
|
"grad_norm": 0.00032589331385679543, |
|
"learning_rate": 4.108697400600316e-05, |
|
"loss": 46.0, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.6153846153846154, |
|
"grad_norm": 0.0002852332836482674, |
|
"learning_rate": 4.005169142968503e-05, |
|
"loss": 46.0, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.621301775147929, |
|
"grad_norm": 0.0004301300796214491, |
|
"learning_rate": 3.9020831277251863e-05, |
|
"loss": 46.0, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.6272189349112426, |
|
"grad_norm": 0.0003021268348675221, |
|
"learning_rate": 3.7994851807565254e-05, |
|
"loss": 46.0, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.6331360946745562, |
|
"grad_norm": 0.0005913956556469202, |
|
"learning_rate": 3.6974209109826726e-05, |
|
"loss": 46.0, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.6390532544378699, |
|
"grad_norm": 0.0003400585555937141, |
|
"learning_rate": 3.595935690082769e-05, |
|
"loss": 46.0, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.6449704142011834, |
|
"grad_norm": 0.00034251168835908175, |
|
"learning_rate": 3.495074632325407e-05, |
|
"loss": 46.0, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.650887573964497, |
|
"grad_norm": 0.0003791009949054569, |
|
"learning_rate": 3.394882574513519e-05, |
|
"loss": 46.0, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.6568047337278107, |
|
"grad_norm": 0.00034353407681919634, |
|
"learning_rate": 3.295404056052616e-05, |
|
"loss": 46.0, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.6627218934911243, |
|
"grad_norm": 0.00026513394550420344, |
|
"learning_rate": 3.196683299151223e-05, |
|
"loss": 46.0, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.6686390532544378, |
|
"grad_norm": 0.00021634704899042845, |
|
"learning_rate": 3.098764189162332e-05, |
|
"loss": 46.0, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.6745562130177515, |
|
"grad_norm": 0.0004164211277384311, |
|
"learning_rate": 3.0016902550745897e-05, |
|
"loss": 46.0, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.6804733727810651, |
|
"grad_norm": 0.00032345083309337497, |
|
"learning_rate": 2.905504650161909e-05, |
|
"loss": 46.0, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.6863905325443787, |
|
"grad_norm": 0.00039014124195091426, |
|
"learning_rate": 2.810250132800103e-05, |
|
"loss": 46.0, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.6923076923076923, |
|
"grad_norm": 0.00032619069679640234, |
|
"learning_rate": 2.715969047459066e-05, |
|
"loss": 46.0, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.6982248520710059, |
|
"grad_norm": 0.0004408117674756795, |
|
"learning_rate": 2.6227033058789408e-05, |
|
"loss": 46.0, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.7041420118343196, |
|
"grad_norm": 0.0005531953065656126, |
|
"learning_rate": 2.530494368438683e-05, |
|
"loss": 46.0, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.7100591715976331, |
|
"grad_norm": 0.00048261991469189525, |
|
"learning_rate": 2.4393832257252252e-05, |
|
"loss": 46.0, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.7159763313609467, |
|
"grad_norm": 0.00041955066262744367, |
|
"learning_rate": 2.349410380311532e-05, |
|
"loss": 46.0, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.7218934911242604, |
|
"grad_norm": 0.0004092879535164684, |
|
"learning_rate": 2.260615828751566e-05, |
|
"loss": 46.0, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.727810650887574, |
|
"grad_norm": 0.0005536783719435334, |
|
"learning_rate": 2.173039043800206e-05, |
|
"loss": 46.0, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.7337278106508875, |
|
"grad_norm": 0.00047029706183820963, |
|
"learning_rate": 2.086718956866024e-05, |
|
"loss": 46.0, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.7396449704142012, |
|
"grad_norm": 0.00032897520577535033, |
|
"learning_rate": 2.0016939407046987e-05, |
|
"loss": 46.0, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.7455621301775148, |
|
"grad_norm": 0.00024629110703244805, |
|
"learning_rate": 1.9180017923607886e-05, |
|
"loss": 46.0, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.7514792899408284, |
|
"grad_norm": 0.00034898053854703903, |
|
"learning_rate": 1.835679716365417e-05, |
|
"loss": 46.0, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.757396449704142, |
|
"grad_norm": 0.00032217547413893044, |
|
"learning_rate": 1.754764308197358e-05, |
|
"loss": 46.0, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.7633136094674556, |
|
"grad_norm": 0.00029810157138854265, |
|
"learning_rate": 1.675291538014877e-05, |
|
"loss": 46.0, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 0.000341045088134706, |
|
"learning_rate": 1.5972967346655448e-05, |
|
"loss": 46.0, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.7751479289940828, |
|
"grad_norm": 0.0003495199780445546, |
|
"learning_rate": 1.5208145699811415e-05, |
|
"loss": 46.0, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.7810650887573964, |
|
"grad_norm": 0.0003159924817737192, |
|
"learning_rate": 1.4458790433646263e-05, |
|
"loss": 46.0, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.7869822485207101, |
|
"grad_norm": 0.0002985662722494453, |
|
"learning_rate": 1.3725234666760428e-05, |
|
"loss": 46.0, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.7928994082840237, |
|
"grad_norm": 0.0003588286635931581, |
|
"learning_rate": 1.3007804494240478e-05, |
|
"loss": 46.0, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.7988165680473372, |
|
"grad_norm": 0.00040958283352665603, |
|
"learning_rate": 1.2306818842696716e-05, |
|
"loss": 46.0, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.8047337278106509, |
|
"grad_norm": 0.00039786347770132124, |
|
"learning_rate": 1.1622589328487504e-05, |
|
"loss": 46.0, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.8106508875739645, |
|
"grad_norm": 0.00025457568699494004, |
|
"learning_rate": 1.0955420119193199e-05, |
|
"loss": 46.0, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.8165680473372781, |
|
"grad_norm": 0.000526306452229619, |
|
"learning_rate": 1.03056077984014e-05, |
|
"loss": 46.0, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.8224852071005917, |
|
"grad_norm": 0.000289287039777264, |
|
"learning_rate": 9.673441233863662e-06, |
|
"loss": 46.0, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.8284023668639053, |
|
"grad_norm": 0.0002234600979136303, |
|
"learning_rate": 9.059201449082045e-06, |
|
"loss": 46.0, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.834319526627219, |
|
"grad_norm": 0.00036921919672749937, |
|
"learning_rate": 8.463161498382948e-06, |
|
"loss": 46.0, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.8402366863905325, |
|
"grad_norm": 0.00024386531731579453, |
|
"learning_rate": 7.885586345533397e-06, |
|
"loss": 46.0, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.8461538461538461, |
|
"grad_norm": 0.0003299729141872376, |
|
"learning_rate": 7.3267327459540015e-06, |
|
"loss": 46.0, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.8520710059171598, |
|
"grad_norm": 0.00021204911172389984, |
|
"learning_rate": 6.786849132580842e-06, |
|
"loss": 46.0, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.8579881656804734, |
|
"grad_norm": 0.00035832199500873685, |
|
"learning_rate": 6.266175505426958e-06, |
|
"loss": 46.0, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.863905325443787, |
|
"grad_norm": 0.0005066677113063633, |
|
"learning_rate": 5.76494332489278e-06, |
|
"loss": 46.0, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.8698224852071006, |
|
"grad_norm": 0.0002925771113950759, |
|
"learning_rate": 5.283375408872537e-06, |
|
"loss": 46.0, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.8757396449704142, |
|
"grad_norm": 0.00039165234193205833, |
|
"learning_rate": 4.821685833702849e-06, |
|
"loss": 46.0, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.8816568047337278, |
|
"grad_norm": 0.0003404826857149601, |
|
"learning_rate": 4.380079838997086e-06, |
|
"loss": 46.0, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.8875739644970414, |
|
"grad_norm": 0.0005336723406799138, |
|
"learning_rate": 3.958753736408105e-06, |
|
"loss": 46.0, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.8875739644970414, |
|
"eval_loss": 11.5, |
|
"eval_runtime": 1.4273, |
|
"eval_samples_per_second": 199.677, |
|
"eval_steps_per_second": 50.445, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.893491124260355, |
|
"grad_norm": 0.0005075408262200654, |
|
"learning_rate": 3.557894822359864e-06, |
|
"loss": 46.0, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.8994082840236687, |
|
"grad_norm": 0.00039445646689273417, |
|
"learning_rate": 3.1776812947865385e-06, |
|
"loss": 46.0, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.9053254437869822, |
|
"grad_norm": 0.00035393863799981773, |
|
"learning_rate": 2.8182821739164534e-06, |
|
"loss": 46.0, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.9112426035502958, |
|
"grad_norm": 0.0005202327738516033, |
|
"learning_rate": 2.4798572271356846e-06, |
|
"loss": 46.0, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.9171597633136095, |
|
"grad_norm": 0.0002673721464816481, |
|
"learning_rate": 2.1625568979651014e-06, |
|
"loss": 46.0, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.9230769230769231, |
|
"grad_norm": 0.0005175816477276385, |
|
"learning_rate": 1.8665222391821169e-06, |
|
"loss": 46.0, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.9289940828402367, |
|
"grad_norm": 0.0004950053407810628, |
|
"learning_rate": 1.5918848501170647e-06, |
|
"loss": 46.0, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.9349112426035503, |
|
"grad_norm": 0.0003163626533932984, |
|
"learning_rate": 1.338766818151982e-06, |
|
"loss": 46.0, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.9408284023668639, |
|
"grad_norm": 0.0004616921942215413, |
|
"learning_rate": 1.1072806644478739e-06, |
|
"loss": 46.0, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.9467455621301775, |
|
"grad_norm": 0.0003856797411572188, |
|
"learning_rate": 8.975292939244928e-07, |
|
"loss": 46.0, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.9526627218934911, |
|
"grad_norm": 0.0004804205091204494, |
|
"learning_rate": 7.096059495149854e-07, |
|
"loss": 46.0, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.9585798816568047, |
|
"grad_norm": 0.0006341671105474234, |
|
"learning_rate": 5.435941707156389e-07, |
|
"loss": 46.0, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.9644970414201184, |
|
"grad_norm": 0.0005707453237846494, |
|
"learning_rate": 3.9956775644920395e-07, |
|
"loss": 46.0, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.9704142011834319, |
|
"grad_norm": 0.0003669565194286406, |
|
"learning_rate": 2.77590732258326e-07, |
|
"loss": 46.0, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.9763313609467456, |
|
"grad_norm": 0.0006033536046743393, |
|
"learning_rate": 1.7771732184357904e-07, |
|
"loss": 46.0, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.9822485207100592, |
|
"grad_norm": 0.00046527519589290023, |
|
"learning_rate": 9.999192295886972e-08, |
|
"loss": 46.0, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.9881656804733728, |
|
"grad_norm": 0.0005233949050307274, |
|
"learning_rate": 4.4449087674847125e-08, |
|
"loss": 46.0, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.9940828402366864, |
|
"grad_norm": 0.0007509227143600583, |
|
"learning_rate": 1.111350701909486e-08, |
|
"loss": 46.0, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.00042899660184048116, |
|
"learning_rate": 0.0, |
|
"loss": 46.0, |
|
"step": 169 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 169, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 150, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 2, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 26545806065664.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|