{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.46077289571201696, "eval_steps": 50, "global_step": 136, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0033880359978824773, "grad_norm": 0.04680836945772171, "learning_rate": 7.499999999999999e-06, "loss": 10.3766, "step": 1 }, { "epoch": 0.0033880359978824773, "eval_loss": 10.376468658447266, "eval_runtime": 3.878, "eval_samples_per_second": 256.575, "eval_steps_per_second": 128.416, "step": 1 }, { "epoch": 0.006776071995764955, "grad_norm": 0.04365997761487961, "learning_rate": 1.4999999999999999e-05, "loss": 10.3771, "step": 2 }, { "epoch": 0.010164107993647432, "grad_norm": 0.041304562240839005, "learning_rate": 2.2499999999999998e-05, "loss": 10.3758, "step": 3 }, { "epoch": 0.01355214399152991, "grad_norm": 0.043574266135692596, "learning_rate": 2.9999999999999997e-05, "loss": 10.3764, "step": 4 }, { "epoch": 0.01694017998941239, "grad_norm": 0.045331861823797226, "learning_rate": 3.75e-05, "loss": 10.3757, "step": 5 }, { "epoch": 0.020328215987294864, "grad_norm": 0.04029256850481033, "learning_rate": 4.4999999999999996e-05, "loss": 10.3764, "step": 6 }, { "epoch": 0.023716251985177343, "grad_norm": 0.046235088258981705, "learning_rate": 5.2499999999999995e-05, "loss": 10.3752, "step": 7 }, { "epoch": 0.02710428798305982, "grad_norm": 0.04704303666949272, "learning_rate": 5.9999999999999995e-05, "loss": 10.3756, "step": 8 }, { "epoch": 0.030492323980942298, "grad_norm": 0.04604668170213699, "learning_rate": 6.75e-05, "loss": 10.3763, "step": 9 }, { "epoch": 0.03388035997882478, "grad_norm": 0.04410148411989212, "learning_rate": 7.5e-05, "loss": 10.3751, "step": 10 }, { "epoch": 0.03726839597670725, "grad_norm": 0.045992445200681686, "learning_rate": 8.25e-05, "loss": 10.3755, "step": 11 }, { "epoch": 0.04065643197458973, "grad_norm": 0.04755518212914467, "learning_rate": 8.999999999999999e-05, "loss": 10.374, "step": 12 }, { "epoch": 0.04404446797247221, "grad_norm": 0.0471210703253746, "learning_rate": 9.75e-05, "loss": 10.375, "step": 13 }, { "epoch": 0.047432503970354686, "grad_norm": 0.04655275121331215, "learning_rate": 0.00010499999999999999, "loss": 10.3753, "step": 14 }, { "epoch": 0.05082053996823716, "grad_norm": 0.04771905019879341, "learning_rate": 0.0001125, "loss": 10.3748, "step": 15 }, { "epoch": 0.05420857596611964, "grad_norm": 0.051401205360889435, "learning_rate": 0.00011999999999999999, "loss": 10.3749, "step": 16 }, { "epoch": 0.05759661196400212, "grad_norm": 0.045204389840364456, "learning_rate": 0.00012749999999999998, "loss": 10.3736, "step": 17 }, { "epoch": 0.060984647961884596, "grad_norm": 0.04979519173502922, "learning_rate": 0.000135, "loss": 10.3729, "step": 18 }, { "epoch": 0.06437268395976707, "grad_norm": 0.04958589747548103, "learning_rate": 0.0001425, "loss": 10.3735, "step": 19 }, { "epoch": 0.06776071995764955, "grad_norm": 0.054519593715667725, "learning_rate": 0.00015, "loss": 10.3723, "step": 20 }, { "epoch": 0.07114875595553202, "grad_norm": 0.060721371322870255, "learning_rate": 0.00014998857713672935, "loss": 10.372, "step": 21 }, { "epoch": 0.0745367919534145, "grad_norm": 0.06067592278122902, "learning_rate": 0.00014995431202643217, "loss": 10.3718, "step": 22 }, { "epoch": 0.07792482795129699, "grad_norm": 0.06362656503915787, "learning_rate": 0.000149897215106593, "loss": 10.3713, "step": 23 }, { "epoch": 0.08131286394917946, "grad_norm": 0.067110076546669, "learning_rate": 0.0001498173037694868, "loss": 10.3714, "step": 24 }, { "epoch": 0.08470089994706194, "grad_norm": 0.07370075583457947, "learning_rate": 0.0001497146023568809, "loss": 10.3712, "step": 25 }, { "epoch": 0.08808893594494442, "grad_norm": 0.0760367140173912, "learning_rate": 0.00014958914215262048, "loss": 10.371, "step": 26 }, { "epoch": 0.09147697194282689, "grad_norm": 0.08042097836732864, "learning_rate": 0.00014944096137309914, "loss": 10.37, "step": 27 }, { "epoch": 0.09486500794070937, "grad_norm": 0.08724083006381989, "learning_rate": 0.00014927010515561776, "loss": 10.3697, "step": 28 }, { "epoch": 0.09825304393859184, "grad_norm": 0.08779574185609818, "learning_rate": 0.00014907662554463532, "loss": 10.3689, "step": 29 }, { "epoch": 0.10164107993647432, "grad_norm": 0.09096319228410721, "learning_rate": 0.0001488605814759156, "loss": 10.3677, "step": 30 }, { "epoch": 0.1050291159343568, "grad_norm": 0.10570746660232544, "learning_rate": 0.00014862203875857477, "loss": 10.3666, "step": 31 }, { "epoch": 0.10841715193223928, "grad_norm": 0.1051798090338707, "learning_rate": 0.0001483610700550354, "loss": 10.3678, "step": 32 }, { "epoch": 0.11180518793012176, "grad_norm": 0.11076612770557404, "learning_rate": 0.00014807775485889264, "loss": 10.366, "step": 33 }, { "epoch": 0.11519322392800424, "grad_norm": 0.11073443293571472, "learning_rate": 0.0001477721794706997, "loss": 10.3658, "step": 34 }, { "epoch": 0.11858125992588671, "grad_norm": 0.11666611582040787, "learning_rate": 0.0001474444369716801, "loss": 10.3642, "step": 35 }, { "epoch": 0.12196929592376919, "grad_norm": 0.12058889120817184, "learning_rate": 0.0001470946271953739, "loss": 10.3624, "step": 36 }, { "epoch": 0.12535733192165166, "grad_norm": 0.13473331928253174, "learning_rate": 0.00014672285669722765, "loss": 10.3619, "step": 37 }, { "epoch": 0.12874536791953414, "grad_norm": 0.12408842891454697, "learning_rate": 0.00014632923872213652, "loss": 10.3616, "step": 38 }, { "epoch": 0.13213340391741663, "grad_norm": 0.12850341200828552, "learning_rate": 0.00014591389316994876, "loss": 10.3609, "step": 39 }, { "epoch": 0.1355214399152991, "grad_norm": 0.12435595691204071, "learning_rate": 0.0001454769465589431, "loss": 10.3597, "step": 40 }, { "epoch": 0.1389094759131816, "grad_norm": 0.11300837248563766, "learning_rate": 0.00014501853198729012, "loss": 10.3583, "step": 41 }, { "epoch": 0.14229751191106405, "grad_norm": 0.11374777555465698, "learning_rate": 0.00014453878909250904, "loss": 10.357, "step": 42 }, { "epoch": 0.14568554790894653, "grad_norm": 0.11273212730884552, "learning_rate": 0.00014403786400893302, "loss": 10.3559, "step": 43 }, { "epoch": 0.149073583906829, "grad_norm": 0.10369884222745895, "learning_rate": 0.00014351590932319504, "loss": 10.3545, "step": 44 }, { "epoch": 0.1524616199047115, "grad_norm": 0.09857185184955597, "learning_rate": 0.00014297308402774875, "loss": 10.3544, "step": 45 }, { "epoch": 0.15584965590259398, "grad_norm": 0.09635099023580551, "learning_rate": 0.0001424095534724375, "loss": 10.3537, "step": 46 }, { "epoch": 0.15923769190047643, "grad_norm": 0.08793843537569046, "learning_rate": 0.00014182548931412757, "loss": 10.3528, "step": 47 }, { "epoch": 0.1626257278983589, "grad_norm": 0.07799001038074493, "learning_rate": 0.0001412210694644195, "loss": 10.3516, "step": 48 }, { "epoch": 0.1660137638962414, "grad_norm": 0.07476358115673065, "learning_rate": 0.00014059647803545467, "loss": 10.3515, "step": 49 }, { "epoch": 0.16940179989412388, "grad_norm": 0.07342197000980377, "learning_rate": 0.0001399519052838329, "loss": 10.3523, "step": 50 }, { "epoch": 0.16940179989412388, "eval_loss": 10.351042747497559, "eval_runtime": 3.8684, "eval_samples_per_second": 257.211, "eval_steps_per_second": 128.735, "step": 50 }, { "epoch": 0.17278983589200636, "grad_norm": 0.06936267018318176, "learning_rate": 0.00013928754755265842, "loss": 10.3514, "step": 51 }, { "epoch": 0.17617787188988884, "grad_norm": 0.06112508103251457, "learning_rate": 0.00013860360721173193, "loss": 10.3511, "step": 52 }, { "epoch": 0.1795659078877713, "grad_norm": 0.06208924576640129, "learning_rate": 0.0001379002925959068, "loss": 10.3505, "step": 53 }, { "epoch": 0.18295394388565378, "grad_norm": 0.057701293379068375, "learning_rate": 0.0001371778179416281, "loss": 10.3505, "step": 54 }, { "epoch": 0.18634197988353626, "grad_norm": 0.050447553396224976, "learning_rate": 0.00013643640332167438, "loss": 10.349, "step": 55 }, { "epoch": 0.18973001588141875, "grad_norm": 0.05121852830052376, "learning_rate": 0.00013567627457812106, "loss": 10.3496, "step": 56 }, { "epoch": 0.19311805187930123, "grad_norm": 0.04376755282282829, "learning_rate": 0.00013489766325354695, "loss": 10.3488, "step": 57 }, { "epoch": 0.19650608787718368, "grad_norm": 0.04570882394909859, "learning_rate": 0.00013410080652050412, "loss": 10.3493, "step": 58 }, { "epoch": 0.19989412387506617, "grad_norm": 0.047347478568553925, "learning_rate": 0.0001332859471092728, "loss": 10.35, "step": 59 }, { "epoch": 0.20328215987294865, "grad_norm": 0.048713624477386475, "learning_rate": 0.00013245333323392333, "loss": 10.3484, "step": 60 }, { "epoch": 0.20667019587083113, "grad_norm": 0.04044219106435776, "learning_rate": 0.0001316032185167079, "loss": 10.3485, "step": 61 }, { "epoch": 0.2100582318687136, "grad_norm": 0.03352762386202812, "learning_rate": 0.00013073586191080457, "loss": 10.3494, "step": 62 }, { "epoch": 0.2134462678665961, "grad_norm": 0.03678755834698677, "learning_rate": 0.00012985152762143778, "loss": 10.3492, "step": 63 }, { "epoch": 0.21683430386447855, "grad_norm": 0.035951532423496246, "learning_rate": 0.00012895048502539882, "loss": 10.3475, "step": 64 }, { "epoch": 0.22022233986236103, "grad_norm": 0.04304710775613785, "learning_rate": 0.00012803300858899104, "loss": 10.349, "step": 65 }, { "epoch": 0.22361037586024352, "grad_norm": 0.035194698721170425, "learning_rate": 0.0001270993777844248, "loss": 10.3468, "step": 66 }, { "epoch": 0.226998411858126, "grad_norm": 0.029510466381907463, "learning_rate": 0.0001261498770046874, "loss": 10.3484, "step": 67 }, { "epoch": 0.23038644785600848, "grad_norm": 0.02603858895599842, "learning_rate": 0.00012518479547691435, "loss": 10.3485, "step": 68 }, { "epoch": 0.23377448385389094, "grad_norm": 0.028004931285977364, "learning_rate": 0.00012420442717428804, "loss": 10.3488, "step": 69 }, { "epoch": 0.23716251985177342, "grad_norm": 0.030328616499900818, "learning_rate": 0.00012320907072649044, "loss": 10.3479, "step": 70 }, { "epoch": 0.2405505558496559, "grad_norm": 0.027720727026462555, "learning_rate": 0.0001221990293287378, "loss": 10.3477, "step": 71 }, { "epoch": 0.24393859184753838, "grad_norm": 0.02903708443045616, "learning_rate": 0.00012117461064942435, "loss": 10.3479, "step": 72 }, { "epoch": 0.24732662784542087, "grad_norm": 0.026270205155014992, "learning_rate": 0.00012013612673640363, "loss": 10.3473, "step": 73 }, { "epoch": 0.2507146638433033, "grad_norm": 0.022023234516382217, "learning_rate": 0.00011908389392193547, "loss": 10.3462, "step": 74 }, { "epoch": 0.25410269984118583, "grad_norm": 0.018086234107613564, "learning_rate": 0.00011801823272632844, "loss": 10.3471, "step": 75 }, { "epoch": 0.2574907358390683, "grad_norm": 0.017115121707320213, "learning_rate": 0.00011693946776030599, "loss": 10.3481, "step": 76 }, { "epoch": 0.26087877183695074, "grad_norm": 0.022446399554610252, "learning_rate": 0.00011584792762612703, "loss": 10.3469, "step": 77 }, { "epoch": 0.26426680783483325, "grad_norm": 0.023726455867290497, "learning_rate": 0.00011474394481749035, "loss": 10.3475, "step": 78 }, { "epoch": 0.2676548438327157, "grad_norm": 0.02203121967613697, "learning_rate": 0.00011362785561825406, "loss": 10.3484, "step": 79 }, { "epoch": 0.2710428798305982, "grad_norm": 0.016930948942899704, "learning_rate": 0.0001125, "loss": 10.3476, "step": 80 }, { "epoch": 0.27443091582848067, "grad_norm": 0.017117898911237717, "learning_rate": 0.00011136072151847529, "loss": 10.3474, "step": 81 }, { "epoch": 0.2778189518263632, "grad_norm": 0.02053241617977619, "learning_rate": 0.00011021036720894179, "loss": 10.3467, "step": 82 }, { "epoch": 0.28120698782424564, "grad_norm": 0.022768596187233925, "learning_rate": 0.00010904928748046599, "loss": 10.3479, "step": 83 }, { "epoch": 0.2845950238221281, "grad_norm": 0.016698352992534637, "learning_rate": 0.0001078778360091808, "loss": 10.3468, "step": 84 }, { "epoch": 0.2879830598200106, "grad_norm": 0.020199725404381752, "learning_rate": 0.00010669636963055245, "loss": 10.3473, "step": 85 }, { "epoch": 0.29137109581789306, "grad_norm": 0.014243320561945438, "learning_rate": 0.00010550524823068502, "loss": 10.3472, "step": 86 }, { "epoch": 0.29475913181577557, "grad_norm": 0.019790155813097954, "learning_rate": 0.00010430483463669551, "loss": 10.3472, "step": 87 }, { "epoch": 0.298147167813658, "grad_norm": 0.01632939651608467, "learning_rate": 0.0001030954945061934, "loss": 10.3474, "step": 88 }, { "epoch": 0.3015352038115405, "grad_norm": 0.015530891716480255, "learning_rate": 0.0001018775962158975, "loss": 10.347, "step": 89 }, { "epoch": 0.304923239809423, "grad_norm": 0.015950841829180717, "learning_rate": 0.00010065151074942516, "loss": 10.3475, "step": 90 }, { "epoch": 0.30831127580730544, "grad_norm": 0.020643549039959908, "learning_rate": 9.941761158428674e-05, "loss": 10.347, "step": 91 }, { "epoch": 0.31169931180518795, "grad_norm": 0.019012747332453728, "learning_rate": 9.817627457812105e-05, "loss": 10.3465, "step": 92 }, { "epoch": 0.3150873478030704, "grad_norm": 0.014910330064594746, "learning_rate": 9.692787785420525e-05, "loss": 10.3482, "step": 93 }, { "epoch": 0.31847538380095286, "grad_norm": 0.01823591999709606, "learning_rate": 9.567280168627493e-05, "loss": 10.3467, "step": 94 }, { "epoch": 0.32186341979883537, "grad_norm": 0.01734175719320774, "learning_rate": 9.441142838268905e-05, "loss": 10.3468, "step": 95 }, { "epoch": 0.3252514557967178, "grad_norm": 0.021044116467237473, "learning_rate": 9.314414216997507e-05, "loss": 10.3477, "step": 96 }, { "epoch": 0.32863949179460034, "grad_norm": 0.01741754449903965, "learning_rate": 9.187132907578987e-05, "loss": 10.3468, "step": 97 }, { "epoch": 0.3320275277924828, "grad_norm": 0.018304290249943733, "learning_rate": 9.059337681133192e-05, "loss": 10.3482, "step": 98 }, { "epoch": 0.33541556379036525, "grad_norm": 0.017597520723938942, "learning_rate": 8.931067465324085e-05, "loss": 10.3468, "step": 99 }, { "epoch": 0.33880359978824776, "grad_norm": 0.01578284054994583, "learning_rate": 8.802361332501978e-05, "loss": 10.3476, "step": 100 }, { "epoch": 0.33880359978824776, "eval_loss": 10.34636116027832, "eval_runtime": 3.8912, "eval_samples_per_second": 255.704, "eval_steps_per_second": 127.981, "step": 100 }, { "epoch": 0.3421916357861302, "grad_norm": 0.02010318450629711, "learning_rate": 8.673258487801731e-05, "loss": 10.3469, "step": 101 }, { "epoch": 0.3455796717840127, "grad_norm": 0.02163493074476719, "learning_rate": 8.54379825720049e-05, "loss": 10.346, "step": 102 }, { "epoch": 0.3489677077818952, "grad_norm": 0.02064266800880432, "learning_rate": 8.414020075538605e-05, "loss": 10.3461, "step": 103 }, { "epoch": 0.3523557437797777, "grad_norm": 0.023022258654236794, "learning_rate": 8.2839634745074e-05, "loss": 10.3463, "step": 104 }, { "epoch": 0.35574377977766014, "grad_norm": 0.017432276159524918, "learning_rate": 8.153668070607437e-05, "loss": 10.3465, "step": 105 }, { "epoch": 0.3591318157755426, "grad_norm": 0.01911778748035431, "learning_rate": 8.023173553080938e-05, "loss": 10.3459, "step": 106 }, { "epoch": 0.3625198517734251, "grad_norm": 0.01831880584359169, "learning_rate": 7.89251967182208e-05, "loss": 10.3455, "step": 107 }, { "epoch": 0.36590788777130756, "grad_norm": 0.018690194934606552, "learning_rate": 7.761746225268758e-05, "loss": 10.3471, "step": 108 }, { "epoch": 0.36929592376919007, "grad_norm": 0.021015219390392303, "learning_rate": 7.630893048279627e-05, "loss": 10.3453, "step": 109 }, { "epoch": 0.3726839597670725, "grad_norm": 0.01983080990612507, "learning_rate": 7.5e-05, "loss": 10.3462, "step": 110 }, { "epoch": 0.376071995764955, "grad_norm": 0.021141625940799713, "learning_rate": 7.369106951720373e-05, "loss": 10.3457, "step": 111 }, { "epoch": 0.3794600317628375, "grad_norm": 0.021793803200125694, "learning_rate": 7.238253774731244e-05, "loss": 10.3457, "step": 112 }, { "epoch": 0.38284806776071995, "grad_norm": 0.019963612779974937, "learning_rate": 7.10748032817792e-05, "loss": 10.3463, "step": 113 }, { "epoch": 0.38623610375860246, "grad_norm": 0.020090965554118156, "learning_rate": 6.976826446919059e-05, "loss": 10.3458, "step": 114 }, { "epoch": 0.3896241397564849, "grad_norm": 0.02083776332437992, "learning_rate": 6.846331929392562e-05, "loss": 10.3469, "step": 115 }, { "epoch": 0.39301217575436737, "grad_norm": 0.01785002276301384, "learning_rate": 6.7160365254926e-05, "loss": 10.3465, "step": 116 }, { "epoch": 0.3964002117522499, "grad_norm": 0.02254386991262436, "learning_rate": 6.585979924461394e-05, "loss": 10.3452, "step": 117 }, { "epoch": 0.39978824775013233, "grad_norm": 0.023428700864315033, "learning_rate": 6.45620174279951e-05, "loss": 10.3475, "step": 118 }, { "epoch": 0.40317628374801484, "grad_norm": 0.01672559231519699, "learning_rate": 6.326741512198266e-05, "loss": 10.346, "step": 119 }, { "epoch": 0.4065643197458973, "grad_norm": 0.020479971542954445, "learning_rate": 6.197638667498022e-05, "loss": 10.3456, "step": 120 }, { "epoch": 0.40995235574377975, "grad_norm": 0.023177366703748703, "learning_rate": 6.068932534675913e-05, "loss": 10.3449, "step": 121 }, { "epoch": 0.41334039174166226, "grad_norm": 0.01870677061378956, "learning_rate": 5.9406623188668055e-05, "loss": 10.346, "step": 122 }, { "epoch": 0.4167284277395447, "grad_norm": 0.0195186547935009, "learning_rate": 5.812867092421013e-05, "loss": 10.3459, "step": 123 }, { "epoch": 0.4201164637374272, "grad_norm": 0.02066197618842125, "learning_rate": 5.685585783002493e-05, "loss": 10.3451, "step": 124 }, { "epoch": 0.4235044997353097, "grad_norm": 0.018614448606967926, "learning_rate": 5.558857161731093e-05, "loss": 10.3456, "step": 125 }, { "epoch": 0.4268925357331922, "grad_norm": 0.02485392615199089, "learning_rate": 5.4327198313725064e-05, "loss": 10.3456, "step": 126 }, { "epoch": 0.43028057173107465, "grad_norm": 0.019080353900790215, "learning_rate": 5.307212214579474e-05, "loss": 10.3463, "step": 127 }, { "epoch": 0.4336686077289571, "grad_norm": 0.02017894946038723, "learning_rate": 5.182372542187895e-05, "loss": 10.3456, "step": 128 }, { "epoch": 0.4370566437268396, "grad_norm": 0.02289474382996559, "learning_rate": 5.058238841571326e-05, "loss": 10.3443, "step": 129 }, { "epoch": 0.44044467972472207, "grad_norm": 0.02067600190639496, "learning_rate": 4.934848925057484e-05, "loss": 10.3459, "step": 130 }, { "epoch": 0.4438327157226046, "grad_norm": 0.02071257308125496, "learning_rate": 4.812240378410248e-05, "loss": 10.3452, "step": 131 }, { "epoch": 0.44722075172048703, "grad_norm": 0.02436411753296852, "learning_rate": 4.690450549380659e-05, "loss": 10.3449, "step": 132 }, { "epoch": 0.4506087877183695, "grad_norm": 0.0237566027790308, "learning_rate": 4.569516536330447e-05, "loss": 10.3462, "step": 133 }, { "epoch": 0.453996823716252, "grad_norm": 0.029039116576313972, "learning_rate": 4.449475176931499e-05, "loss": 10.3455, "step": 134 }, { "epoch": 0.45738485971413445, "grad_norm": 0.021290864795446396, "learning_rate": 4.3303630369447554e-05, "loss": 10.3446, "step": 135 }, { "epoch": 0.46077289571201696, "grad_norm": 0.01688864268362522, "learning_rate": 4.212216399081918e-05, "loss": 10.3439, "step": 136 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 17, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 58258116182016.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }