{ "best_metric": 0.770029604434967, "best_model_checkpoint": "miner_id_24/checkpoint-200", "epoch": 0.12168410805548795, "eval_steps": 25, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006084205402774398, "grad_norm": 3.7732460498809814, "learning_rate": 2.173913043478261e-06, "loss": 3.2093, "step": 1 }, { "epoch": 0.0006084205402774398, "eval_loss": 1.9844963550567627, "eval_runtime": 0.0679, "eval_samples_per_second": 736.566, "eval_steps_per_second": 44.194, "step": 1 }, { "epoch": 0.0012168410805548796, "grad_norm": 4.146840572357178, "learning_rate": 4.347826086956522e-06, "loss": 2.2685, "step": 2 }, { "epoch": 0.0018252616208323193, "grad_norm": 3.805527448654175, "learning_rate": 6.521739130434783e-06, "loss": 2.0894, "step": 3 }, { "epoch": 0.002433682161109759, "grad_norm": 3.5490667819976807, "learning_rate": 8.695652173913044e-06, "loss": 1.9026, "step": 4 }, { "epoch": 0.003042102701387199, "grad_norm": 3.1814687252044678, "learning_rate": 1.0869565217391305e-05, "loss": 1.9351, "step": 5 }, { "epoch": 0.0036505232416646385, "grad_norm": 3.0306947231292725, "learning_rate": 1.3043478260869566e-05, "loss": 1.9283, "step": 6 }, { "epoch": 0.004258943781942078, "grad_norm": 3.2333295345306396, "learning_rate": 1.5217391304347828e-05, "loss": 1.9199, "step": 7 }, { "epoch": 0.004867364322219518, "grad_norm": 2.786118745803833, "learning_rate": 1.739130434782609e-05, "loss": 1.9464, "step": 8 }, { "epoch": 0.005475784862496958, "grad_norm": 3.1137781143188477, "learning_rate": 1.956521739130435e-05, "loss": 1.8952, "step": 9 }, { "epoch": 0.006084205402774398, "grad_norm": 3.2737996578216553, "learning_rate": 2.173913043478261e-05, "loss": 1.8745, "step": 10 }, { "epoch": 0.006692625943051838, "grad_norm": 3.218707799911499, "learning_rate": 2.391304347826087e-05, "loss": 1.9421, "step": 11 }, { "epoch": 0.007301046483329277, "grad_norm": 3.402679204940796, "learning_rate": 2.608695652173913e-05, "loss": 2.0349, "step": 12 }, { "epoch": 0.007909467023606717, "grad_norm": 2.8429970741271973, "learning_rate": 2.826086956521739e-05, "loss": 4.0381, "step": 13 }, { "epoch": 0.008517887563884156, "grad_norm": 4.92161750793457, "learning_rate": 3.0434782608695656e-05, "loss": 2.315, "step": 14 }, { "epoch": 0.009126308104161596, "grad_norm": 4.043981075286865, "learning_rate": 3.260869565217392e-05, "loss": 2.028, "step": 15 }, { "epoch": 0.009734728644439037, "grad_norm": 3.5905041694641113, "learning_rate": 3.478260869565218e-05, "loss": 2.034, "step": 16 }, { "epoch": 0.010343149184716476, "grad_norm": 3.121873378753662, "learning_rate": 3.695652173913043e-05, "loss": 1.9274, "step": 17 }, { "epoch": 0.010951569724993915, "grad_norm": 3.123685836791992, "learning_rate": 3.91304347826087e-05, "loss": 1.9217, "step": 18 }, { "epoch": 0.011559990265271356, "grad_norm": 2.93074107170105, "learning_rate": 4.130434782608696e-05, "loss": 1.6898, "step": 19 }, { "epoch": 0.012168410805548795, "grad_norm": 3.0180726051330566, "learning_rate": 4.347826086956522e-05, "loss": 1.6951, "step": 20 }, { "epoch": 0.012776831345826235, "grad_norm": 2.9282548427581787, "learning_rate": 4.565217391304348e-05, "loss": 1.7592, "step": 21 }, { "epoch": 0.013385251886103676, "grad_norm": 2.5757176876068115, "learning_rate": 4.782608695652174e-05, "loss": 1.753, "step": 22 }, { "epoch": 0.013993672426381115, "grad_norm": 2.7244386672973633, "learning_rate": 5e-05, "loss": 1.7478, "step": 23 }, { "epoch": 0.014602092966658554, "grad_norm": 2.9556708335876465, "learning_rate": 5.217391304347826e-05, "loss": 1.7354, "step": 24 }, { "epoch": 0.015210513506935993, "grad_norm": 2.826549530029297, "learning_rate": 5.4347826086956524e-05, "loss": 1.8359, "step": 25 }, { "epoch": 0.015210513506935993, "eval_loss": 1.7007235288619995, "eval_runtime": 0.0651, "eval_samples_per_second": 767.794, "eval_steps_per_second": 46.068, "step": 25 }, { "epoch": 0.015818934047213434, "grad_norm": 4.59504508972168, "learning_rate": 5.652173913043478e-05, "loss": 2.9356, "step": 26 }, { "epoch": 0.016427354587490874, "grad_norm": 4.431028842926025, "learning_rate": 5.869565217391305e-05, "loss": 1.7887, "step": 27 }, { "epoch": 0.017035775127768313, "grad_norm": 3.4355099201202393, "learning_rate": 6.086956521739131e-05, "loss": 1.6548, "step": 28 }, { "epoch": 0.017644195668045752, "grad_norm": 2.5302505493164062, "learning_rate": 6.304347826086957e-05, "loss": 1.6289, "step": 29 }, { "epoch": 0.01825261620832319, "grad_norm": 1.6945921182632446, "learning_rate": 6.521739130434783e-05, "loss": 1.4231, "step": 30 }, { "epoch": 0.018861036748600634, "grad_norm": 1.6188069581985474, "learning_rate": 6.73913043478261e-05, "loss": 1.4109, "step": 31 }, { "epoch": 0.019469457288878073, "grad_norm": 1.2194888591766357, "learning_rate": 6.956521739130436e-05, "loss": 1.4766, "step": 32 }, { "epoch": 0.020077877829155513, "grad_norm": 1.0816359519958496, "learning_rate": 7.17391304347826e-05, "loss": 1.5166, "step": 33 }, { "epoch": 0.020686298369432952, "grad_norm": 1.0028973817825317, "learning_rate": 7.391304347826086e-05, "loss": 1.5929, "step": 34 }, { "epoch": 0.02129471890971039, "grad_norm": 0.9615126848220825, "learning_rate": 7.608695652173914e-05, "loss": 1.3961, "step": 35 }, { "epoch": 0.02190313944998783, "grad_norm": 1.0823183059692383, "learning_rate": 7.82608695652174e-05, "loss": 1.4874, "step": 36 }, { "epoch": 0.022511559990265273, "grad_norm": 1.1442559957504272, "learning_rate": 8.043478260869566e-05, "loss": 1.5171, "step": 37 }, { "epoch": 0.023119980530542712, "grad_norm": 3.612239360809326, "learning_rate": 8.260869565217392e-05, "loss": 3.3537, "step": 38 }, { "epoch": 0.02372840107082015, "grad_norm": 3.0585246086120605, "learning_rate": 8.478260869565218e-05, "loss": 1.571, "step": 39 }, { "epoch": 0.02433682161109759, "grad_norm": 2.00331711769104, "learning_rate": 8.695652173913044e-05, "loss": 1.5134, "step": 40 }, { "epoch": 0.02494524215137503, "grad_norm": 1.4210355281829834, "learning_rate": 8.91304347826087e-05, "loss": 1.2967, "step": 41 }, { "epoch": 0.02555366269165247, "grad_norm": 1.2288633584976196, "learning_rate": 9.130434782608696e-05, "loss": 1.4348, "step": 42 }, { "epoch": 0.02616208323192991, "grad_norm": 1.157564640045166, "learning_rate": 9.347826086956522e-05, "loss": 1.3196, "step": 43 }, { "epoch": 0.02677050377220735, "grad_norm": 1.2825106382369995, "learning_rate": 9.565217391304348e-05, "loss": 1.349, "step": 44 }, { "epoch": 0.02737892431248479, "grad_norm": 1.2434223890304565, "learning_rate": 9.782608695652174e-05, "loss": 1.3314, "step": 45 }, { "epoch": 0.02798734485276223, "grad_norm": 1.429315447807312, "learning_rate": 0.0001, "loss": 1.3863, "step": 46 }, { "epoch": 0.02859576539303967, "grad_norm": 0.9383629560470581, "learning_rate": 9.999972027814122e-05, "loss": 1.3054, "step": 47 }, { "epoch": 0.029204185933317108, "grad_norm": 0.7935044765472412, "learning_rate": 9.999888111604245e-05, "loss": 1.298, "step": 48 }, { "epoch": 0.029812606473594547, "grad_norm": 0.859842836856842, "learning_rate": 9.999748252413618e-05, "loss": 1.3098, "step": 49 }, { "epoch": 0.030421027013871987, "grad_norm": 1.2138994932174683, "learning_rate": 9.999552451980984e-05, "loss": 1.3987, "step": 50 }, { "epoch": 0.030421027013871987, "eval_loss": 1.2663160562515259, "eval_runtime": 0.064, "eval_samples_per_second": 781.103, "eval_steps_per_second": 46.866, "step": 50 }, { "epoch": 0.03102944755414943, "grad_norm": 3.1896111965179443, "learning_rate": 9.999300712740551e-05, "loss": 2.1177, "step": 51 }, { "epoch": 0.03163786809442687, "grad_norm": 1.8031097650527954, "learning_rate": 9.998993037821958e-05, "loss": 1.281, "step": 52 }, { "epoch": 0.03224628863470431, "grad_norm": 1.3828022480010986, "learning_rate": 9.998629431050251e-05, "loss": 1.286, "step": 53 }, { "epoch": 0.03285470917498175, "grad_norm": 1.2959613800048828, "learning_rate": 9.998209896945815e-05, "loss": 1.3098, "step": 54 }, { "epoch": 0.033463129715259186, "grad_norm": 0.9320961236953735, "learning_rate": 9.997734440724333e-05, "loss": 1.2237, "step": 55 }, { "epoch": 0.034071550255536626, "grad_norm": 0.7311733961105347, "learning_rate": 9.997203068296719e-05, "loss": 1.1421, "step": 56 }, { "epoch": 0.034679970795814065, "grad_norm": 0.7125067710876465, "learning_rate": 9.996615786269035e-05, "loss": 1.2081, "step": 57 }, { "epoch": 0.035288391336091504, "grad_norm": 0.6830151677131653, "learning_rate": 9.995972601942424e-05, "loss": 1.2123, "step": 58 }, { "epoch": 0.03589681187636894, "grad_norm": 0.6575762629508972, "learning_rate": 9.995273523313003e-05, "loss": 1.1829, "step": 59 }, { "epoch": 0.03650523241664638, "grad_norm": 0.8578693866729736, "learning_rate": 9.994518559071775e-05, "loss": 1.1763, "step": 60 }, { "epoch": 0.03711365295692383, "grad_norm": 0.8111445307731628, "learning_rate": 9.993707718604522e-05, "loss": 1.1614, "step": 61 }, { "epoch": 0.03772207349720127, "grad_norm": 0.9371657967567444, "learning_rate": 9.992841011991676e-05, "loss": 1.2245, "step": 62 }, { "epoch": 0.03833049403747871, "grad_norm": 2.6930816173553467, "learning_rate": 9.991918450008206e-05, "loss": 2.285, "step": 63 }, { "epoch": 0.03893891457775615, "grad_norm": 1.746653437614441, "learning_rate": 9.99094004412348e-05, "loss": 1.301, "step": 64 }, { "epoch": 0.039547335118033586, "grad_norm": 1.1019763946533203, "learning_rate": 9.98990580650112e-05, "loss": 1.1406, "step": 65 }, { "epoch": 0.040155755658311025, "grad_norm": 1.0221314430236816, "learning_rate": 9.988815749998852e-05, "loss": 1.0963, "step": 66 }, { "epoch": 0.040764176198588464, "grad_norm": 0.7163240313529968, "learning_rate": 9.987669888168351e-05, "loss": 1.0621, "step": 67 }, { "epoch": 0.041372596738865904, "grad_norm": 1.0129799842834473, "learning_rate": 9.986468235255065e-05, "loss": 1.0655, "step": 68 }, { "epoch": 0.04198101727914334, "grad_norm": 1.0162873268127441, "learning_rate": 9.985210806198042e-05, "loss": 1.1478, "step": 69 }, { "epoch": 0.04258943781942078, "grad_norm": 0.6798834204673767, "learning_rate": 9.983897616629744e-05, "loss": 1.0584, "step": 70 }, { "epoch": 0.04319785835969822, "grad_norm": 0.82222980260849, "learning_rate": 9.982528682875851e-05, "loss": 1.0776, "step": 71 }, { "epoch": 0.04380627889997566, "grad_norm": 0.8663562536239624, "learning_rate": 9.981104021955063e-05, "loss": 1.073, "step": 72 }, { "epoch": 0.0444146994402531, "grad_norm": 1.122699499130249, "learning_rate": 9.979623651578881e-05, "loss": 1.0509, "step": 73 }, { "epoch": 0.045023119980530546, "grad_norm": 1.1595953702926636, "learning_rate": 9.978087590151393e-05, "loss": 1.1094, "step": 74 }, { "epoch": 0.045631540520807985, "grad_norm": 1.62815523147583, "learning_rate": 9.976495856769038e-05, "loss": 1.1211, "step": 75 }, { "epoch": 0.045631540520807985, "eval_loss": 1.0243722200393677, "eval_runtime": 0.0649, "eval_samples_per_second": 770.624, "eval_steps_per_second": 46.237, "step": 75 }, { "epoch": 0.046239961061085424, "grad_norm": 2.726969003677368, "learning_rate": 9.97484847122038e-05, "loss": 2.205, "step": 76 }, { "epoch": 0.046848381601362864, "grad_norm": 2.2901804447174072, "learning_rate": 9.973145453985854e-05, "loss": 1.1102, "step": 77 }, { "epoch": 0.0474568021416403, "grad_norm": 2.2033584117889404, "learning_rate": 9.971386826237507e-05, "loss": 1.0353, "step": 78 }, { "epoch": 0.04806522268191774, "grad_norm": 2.2385048866271973, "learning_rate": 9.969572609838744e-05, "loss": 1.0281, "step": 79 }, { "epoch": 0.04867364322219518, "grad_norm": 1.700537919998169, "learning_rate": 9.967702827344057e-05, "loss": 1.0513, "step": 80 }, { "epoch": 0.04928206376247262, "grad_norm": 1.1010593175888062, "learning_rate": 9.965777501998734e-05, "loss": 1.0073, "step": 81 }, { "epoch": 0.04989048430275006, "grad_norm": 0.7716583609580994, "learning_rate": 9.963796657738579e-05, "loss": 1.0205, "step": 82 }, { "epoch": 0.0504989048430275, "grad_norm": 0.7150983810424805, "learning_rate": 9.961760319189612e-05, "loss": 0.924, "step": 83 }, { "epoch": 0.05110732538330494, "grad_norm": 0.9970099925994873, "learning_rate": 9.959668511667762e-05, "loss": 1.0786, "step": 84 }, { "epoch": 0.05171574592358238, "grad_norm": 1.10440194606781, "learning_rate": 9.957521261178554e-05, "loss": 1.0916, "step": 85 }, { "epoch": 0.05232416646385982, "grad_norm": 1.2334967851638794, "learning_rate": 9.95531859441678e-05, "loss": 1.0281, "step": 86 }, { "epoch": 0.05293258700413726, "grad_norm": 1.5450046062469482, "learning_rate": 9.953060538766178e-05, "loss": 0.9459, "step": 87 }, { "epoch": 0.0535410075444147, "grad_norm": 2.800281524658203, "learning_rate": 9.950747122299084e-05, "loss": 2.3543, "step": 88 }, { "epoch": 0.05414942808469214, "grad_norm": 1.7520790100097656, "learning_rate": 9.948378373776079e-05, "loss": 1.0488, "step": 89 }, { "epoch": 0.05475784862496958, "grad_norm": 1.148888111114502, "learning_rate": 9.945954322645642e-05, "loss": 0.8649, "step": 90 }, { "epoch": 0.05536626916524702, "grad_norm": 0.883175253868103, "learning_rate": 9.943474999043775e-05, "loss": 0.9496, "step": 91 }, { "epoch": 0.05597468970552446, "grad_norm": 1.041022539138794, "learning_rate": 9.940940433793637e-05, "loss": 0.9575, "step": 92 }, { "epoch": 0.0565831102458019, "grad_norm": 0.8950981497764587, "learning_rate": 9.938350658405152e-05, "loss": 0.8507, "step": 93 }, { "epoch": 0.05719153078607934, "grad_norm": 0.8415868282318115, "learning_rate": 9.935705705074617e-05, "loss": 0.9419, "step": 94 }, { "epoch": 0.05779995132635678, "grad_norm": 0.8708381652832031, "learning_rate": 9.933005606684317e-05, "loss": 1.0413, "step": 95 }, { "epoch": 0.058408371866634216, "grad_norm": 1.0290625095367432, "learning_rate": 9.930250396802094e-05, "loss": 0.9954, "step": 96 }, { "epoch": 0.059016792406911656, "grad_norm": 0.8220187425613403, "learning_rate": 9.927440109680946e-05, "loss": 0.987, "step": 97 }, { "epoch": 0.059625212947189095, "grad_norm": 0.8632601499557495, "learning_rate": 9.924574780258596e-05, "loss": 0.9248, "step": 98 }, { "epoch": 0.060233633487466534, "grad_norm": 1.202643632888794, "learning_rate": 9.921654444157054e-05, "loss": 0.9264, "step": 99 }, { "epoch": 0.06084205402774397, "grad_norm": 1.6064764261245728, "learning_rate": 9.91867913768218e-05, "loss": 1.032, "step": 100 }, { "epoch": 0.06084205402774397, "eval_loss": 0.901063084602356, "eval_runtime": 0.0644, "eval_samples_per_second": 776.142, "eval_steps_per_second": 46.569, "step": 100 }, { "epoch": 0.06145047456802142, "grad_norm": 2.1135504245758057, "learning_rate": 9.915648897823232e-05, "loss": 1.8061, "step": 101 }, { "epoch": 0.06205889510829886, "grad_norm": 1.5115822553634644, "learning_rate": 9.912563762252399e-05, "loss": 0.9178, "step": 102 }, { "epoch": 0.06266731564857629, "grad_norm": 1.3775817155838013, "learning_rate": 9.909423769324343e-05, "loss": 0.8464, "step": 103 }, { "epoch": 0.06327573618885374, "grad_norm": 1.1930534839630127, "learning_rate": 9.906228958075717e-05, "loss": 0.9817, "step": 104 }, { "epoch": 0.06388415672913117, "grad_norm": 1.158815622329712, "learning_rate": 9.902979368224675e-05, "loss": 0.9052, "step": 105 }, { "epoch": 0.06449257726940862, "grad_norm": 0.8110672831535339, "learning_rate": 9.899675040170388e-05, "loss": 0.8699, "step": 106 }, { "epoch": 0.06510099780968605, "grad_norm": 0.6599923372268677, "learning_rate": 9.896316014992537e-05, "loss": 0.8528, "step": 107 }, { "epoch": 0.0657094183499635, "grad_norm": 0.7083361744880676, "learning_rate": 9.892902334450798e-05, "loss": 0.9154, "step": 108 }, { "epoch": 0.06631783889024094, "grad_norm": 1.0131536722183228, "learning_rate": 9.889434040984332e-05, "loss": 0.9124, "step": 109 }, { "epoch": 0.06692625943051837, "grad_norm": 1.1790049076080322, "learning_rate": 9.885911177711248e-05, "loss": 0.8958, "step": 110 }, { "epoch": 0.06753467997079582, "grad_norm": 1.2120789289474487, "learning_rate": 9.882333788428073e-05, "loss": 0.8547, "step": 111 }, { "epoch": 0.06814310051107325, "grad_norm": 1.194826364517212, "learning_rate": 9.878701917609207e-05, "loss": 0.9218, "step": 112 }, { "epoch": 0.0687515210513507, "grad_norm": 2.966965913772583, "learning_rate": 9.875015610406369e-05, "loss": 2.5956, "step": 113 }, { "epoch": 0.06935994159162813, "grad_norm": 1.3608638048171997, "learning_rate": 9.871274912648033e-05, "loss": 0.9144, "step": 114 }, { "epoch": 0.06996836213190558, "grad_norm": 1.4627468585968018, "learning_rate": 9.867479870838862e-05, "loss": 0.9815, "step": 115 }, { "epoch": 0.07057678267218301, "grad_norm": 1.1806144714355469, "learning_rate": 9.863630532159132e-05, "loss": 0.7937, "step": 116 }, { "epoch": 0.07118520321246045, "grad_norm": 1.0851095914840698, "learning_rate": 9.859726944464137e-05, "loss": 0.8003, "step": 117 }, { "epoch": 0.07179362375273789, "grad_norm": 1.126287579536438, "learning_rate": 9.855769156283603e-05, "loss": 0.9241, "step": 118 }, { "epoch": 0.07240204429301533, "grad_norm": 0.6791806817054749, "learning_rate": 9.851757216821079e-05, "loss": 0.9201, "step": 119 }, { "epoch": 0.07301046483329277, "grad_norm": 0.7268375158309937, "learning_rate": 9.847691175953328e-05, "loss": 0.8126, "step": 120 }, { "epoch": 0.07361888537357021, "grad_norm": 0.8680149912834167, "learning_rate": 9.843571084229707e-05, "loss": 0.8912, "step": 121 }, { "epoch": 0.07422730591384766, "grad_norm": 1.042919397354126, "learning_rate": 9.839396992871535e-05, "loss": 0.9064, "step": 122 }, { "epoch": 0.07483572645412509, "grad_norm": 1.2081691026687622, "learning_rate": 9.835168953771461e-05, "loss": 0.9376, "step": 123 }, { "epoch": 0.07544414699440254, "grad_norm": 1.1437644958496094, "learning_rate": 9.830887019492818e-05, "loss": 0.9418, "step": 124 }, { "epoch": 0.07605256753467997, "grad_norm": 1.2367491722106934, "learning_rate": 9.826551243268966e-05, "loss": 1.015, "step": 125 }, { "epoch": 0.07605256753467997, "eval_loss": 0.8607701063156128, "eval_runtime": 0.0636, "eval_samples_per_second": 785.848, "eval_steps_per_second": 47.151, "step": 125 }, { "epoch": 0.07666098807495741, "grad_norm": 2.4859538078308105, "learning_rate": 9.82216167900263e-05, "loss": 1.8291, "step": 126 }, { "epoch": 0.07726940861523485, "grad_norm": 1.4214491844177246, "learning_rate": 9.81771838126524e-05, "loss": 1.0467, "step": 127 }, { "epoch": 0.0778778291555123, "grad_norm": 1.3560972213745117, "learning_rate": 9.813221405296237e-05, "loss": 1.0159, "step": 128 }, { "epoch": 0.07848624969578973, "grad_norm": 1.501007080078125, "learning_rate": 9.8086708070024e-05, "loss": 0.8479, "step": 129 }, { "epoch": 0.07909467023606717, "grad_norm": 1.5460386276245117, "learning_rate": 9.804066642957143e-05, "loss": 0.9617, "step": 130 }, { "epoch": 0.0797030907763446, "grad_norm": 1.2963954210281372, "learning_rate": 9.799408970399813e-05, "loss": 0.8346, "step": 131 }, { "epoch": 0.08031151131662205, "grad_norm": 0.8653491139411926, "learning_rate": 9.79469784723498e-05, "loss": 0.8009, "step": 132 }, { "epoch": 0.08091993185689948, "grad_norm": 1.0787755250930786, "learning_rate": 9.789933332031717e-05, "loss": 0.9501, "step": 133 }, { "epoch": 0.08152835239717693, "grad_norm": 1.0019638538360596, "learning_rate": 9.78511548402287e-05, "loss": 0.841, "step": 134 }, { "epoch": 0.08213677293745437, "grad_norm": 1.1226329803466797, "learning_rate": 9.780244363104323e-05, "loss": 0.8492, "step": 135 }, { "epoch": 0.08274519347773181, "grad_norm": 1.269909381866455, "learning_rate": 9.775320029834254e-05, "loss": 0.854, "step": 136 }, { "epoch": 0.08335361401800925, "grad_norm": 1.76808762550354, "learning_rate": 9.770342545432383e-05, "loss": 0.8746, "step": 137 }, { "epoch": 0.08396203455828669, "grad_norm": 3.7219080924987793, "learning_rate": 9.765311971779204e-05, "loss": 2.2368, "step": 138 }, { "epoch": 0.08457045509856413, "grad_norm": 1.7717487812042236, "learning_rate": 9.760228371415227e-05, "loss": 1.0042, "step": 139 }, { "epoch": 0.08517887563884156, "grad_norm": 1.4473094940185547, "learning_rate": 9.75509180754019e-05, "loss": 0.8678, "step": 140 }, { "epoch": 0.08578729617911901, "grad_norm": 1.3099236488342285, "learning_rate": 9.749902344012279e-05, "loss": 0.7265, "step": 141 }, { "epoch": 0.08639571671939644, "grad_norm": 1.1434625387191772, "learning_rate": 9.744660045347336e-05, "loss": 0.7999, "step": 142 }, { "epoch": 0.08700413725967389, "grad_norm": 1.1469634771347046, "learning_rate": 9.73936497671805e-05, "loss": 0.8769, "step": 143 }, { "epoch": 0.08761255779995132, "grad_norm": 1.1170365810394287, "learning_rate": 9.734017203953149e-05, "loss": 0.8353, "step": 144 }, { "epoch": 0.08822097834022877, "grad_norm": 0.7087867259979248, "learning_rate": 9.728616793536588e-05, "loss": 0.8558, "step": 145 }, { "epoch": 0.0888293988805062, "grad_norm": 0.7099660634994507, "learning_rate": 9.723163812606716e-05, "loss": 0.7829, "step": 146 }, { "epoch": 0.08943781942078365, "grad_norm": 1.0997332334518433, "learning_rate": 9.71765832895544e-05, "loss": 0.8654, "step": 147 }, { "epoch": 0.09004623996106109, "grad_norm": 1.0360599756240845, "learning_rate": 9.712100411027391e-05, "loss": 0.8817, "step": 148 }, { "epoch": 0.09065466050133852, "grad_norm": 1.460658311843872, "learning_rate": 9.70649012791906e-05, "loss": 0.8495, "step": 149 }, { "epoch": 0.09126308104161597, "grad_norm": 1.6194958686828613, "learning_rate": 9.700827549377955e-05, "loss": 0.929, "step": 150 }, { "epoch": 0.09126308104161597, "eval_loss": 0.8000746369361877, "eval_runtime": 0.0648, "eval_samples_per_second": 771.193, "eval_steps_per_second": 46.272, "step": 150 }, { "epoch": 0.0918715015818934, "grad_norm": 2.0881123542785645, "learning_rate": 9.695112745801716e-05, "loss": 1.903, "step": 151 }, { "epoch": 0.09247992212217085, "grad_norm": 1.1570873260498047, "learning_rate": 9.689345788237256e-05, "loss": 0.9427, "step": 152 }, { "epoch": 0.09308834266244828, "grad_norm": 0.8610666990280151, "learning_rate": 9.683526748379864e-05, "loss": 0.8573, "step": 153 }, { "epoch": 0.09369676320272573, "grad_norm": 0.8190277218818665, "learning_rate": 9.677655698572326e-05, "loss": 0.6991, "step": 154 }, { "epoch": 0.09430518374300316, "grad_norm": 0.8921626210212708, "learning_rate": 9.671732711804022e-05, "loss": 0.7823, "step": 155 }, { "epoch": 0.0949136042832806, "grad_norm": 0.6367238759994507, "learning_rate": 9.665757861710008e-05, "loss": 0.8343, "step": 156 }, { "epoch": 0.09552202482355804, "grad_norm": 0.533492922782898, "learning_rate": 9.659731222570114e-05, "loss": 0.8161, "step": 157 }, { "epoch": 0.09613044536383548, "grad_norm": 0.7728853821754456, "learning_rate": 9.653652869308019e-05, "loss": 0.925, "step": 158 }, { "epoch": 0.09673886590411292, "grad_norm": 0.7567634582519531, "learning_rate": 9.647522877490312e-05, "loss": 0.8247, "step": 159 }, { "epoch": 0.09734728644439036, "grad_norm": 0.6781755685806274, "learning_rate": 9.641341323325554e-05, "loss": 0.8119, "step": 160 }, { "epoch": 0.09795570698466781, "grad_norm": 0.9459850788116455, "learning_rate": 9.635108283663343e-05, "loss": 0.7861, "step": 161 }, { "epoch": 0.09856412752494524, "grad_norm": 0.8214953541755676, "learning_rate": 9.628823835993338e-05, "loss": 0.8647, "step": 162 }, { "epoch": 0.09917254806522269, "grad_norm": 2.5658833980560303, "learning_rate": 9.622488058444314e-05, "loss": 2.3544, "step": 163 }, { "epoch": 0.09978096860550012, "grad_norm": 1.3964160680770874, "learning_rate": 9.61610102978318e-05, "loss": 0.8933, "step": 164 }, { "epoch": 0.10038938914577757, "grad_norm": 1.2962576150894165, "learning_rate": 9.609662829414005e-05, "loss": 0.7469, "step": 165 }, { "epoch": 0.100997809686055, "grad_norm": 1.0267356634140015, "learning_rate": 9.603173537377026e-05, "loss": 0.8354, "step": 166 }, { "epoch": 0.10160623022633244, "grad_norm": 0.7708994150161743, "learning_rate": 9.59663323434766e-05, "loss": 0.9245, "step": 167 }, { "epoch": 0.10221465076660988, "grad_norm": 0.5819109678268433, "learning_rate": 9.590042001635495e-05, "loss": 0.8, "step": 168 }, { "epoch": 0.10282307130688732, "grad_norm": 0.5333255529403687, "learning_rate": 9.583399921183276e-05, "loss": 0.798, "step": 169 }, { "epoch": 0.10343149184716476, "grad_norm": 0.9702023267745972, "learning_rate": 9.5767070755659e-05, "loss": 0.9434, "step": 170 }, { "epoch": 0.1040399123874422, "grad_norm": 0.9722750186920166, "learning_rate": 9.569963547989376e-05, "loss": 0.8818, "step": 171 }, { "epoch": 0.10464833292771963, "grad_norm": 1.0027716159820557, "learning_rate": 9.563169422289797e-05, "loss": 0.7531, "step": 172 }, { "epoch": 0.10525675346799708, "grad_norm": 0.7870794534683228, "learning_rate": 9.556324782932292e-05, "loss": 0.8253, "step": 173 }, { "epoch": 0.10586517400827453, "grad_norm": 0.7732038497924805, "learning_rate": 9.549429715009983e-05, "loss": 0.8425, "step": 174 }, { "epoch": 0.10647359454855196, "grad_norm": 1.3445314168930054, "learning_rate": 9.542484304242927e-05, "loss": 0.8793, "step": 175 }, { "epoch": 0.10647359454855196, "eval_loss": 0.7928541302680969, "eval_runtime": 0.0638, "eval_samples_per_second": 783.396, "eval_steps_per_second": 47.004, "step": 175 }, { "epoch": 0.1070820150888294, "grad_norm": 3.020885705947876, "learning_rate": 9.535488636977042e-05, "loss": 1.6845, "step": 176 }, { "epoch": 0.10769043562910684, "grad_norm": 2.2240676879882812, "learning_rate": 9.528442800183043e-05, "loss": 0.8827, "step": 177 }, { "epoch": 0.10829885616938428, "grad_norm": 1.7883273363113403, "learning_rate": 9.521346881455356e-05, "loss": 0.7705, "step": 178 }, { "epoch": 0.10890727670966172, "grad_norm": 1.2515283823013306, "learning_rate": 9.514200969011025e-05, "loss": 0.7441, "step": 179 }, { "epoch": 0.10951569724993916, "grad_norm": 0.8652520179748535, "learning_rate": 9.507005151688627e-05, "loss": 0.725, "step": 180 }, { "epoch": 0.1101241177902166, "grad_norm": 0.6205310225486755, "learning_rate": 9.499759518947156e-05, "loss": 0.7823, "step": 181 }, { "epoch": 0.11073253833049404, "grad_norm": 0.6425133943557739, "learning_rate": 9.492464160864917e-05, "loss": 0.7767, "step": 182 }, { "epoch": 0.11134095887077147, "grad_norm": 1.1534785032272339, "learning_rate": 9.485119168138405e-05, "loss": 0.8115, "step": 183 }, { "epoch": 0.11194937941104892, "grad_norm": 1.0528301000595093, "learning_rate": 9.477724632081176e-05, "loss": 0.7672, "step": 184 }, { "epoch": 0.11255779995132635, "grad_norm": 1.1875839233398438, "learning_rate": 9.470280644622713e-05, "loss": 0.8329, "step": 185 }, { "epoch": 0.1131662204916038, "grad_norm": 1.0731632709503174, "learning_rate": 9.462787298307285e-05, "loss": 0.8891, "step": 186 }, { "epoch": 0.11377464103188123, "grad_norm": 1.1598612070083618, "learning_rate": 9.455244686292789e-05, "loss": 0.7696, "step": 187 }, { "epoch": 0.11438306157215868, "grad_norm": 3.4111616611480713, "learning_rate": 9.447652902349602e-05, "loss": 1.4939, "step": 188 }, { "epoch": 0.11499148211243612, "grad_norm": 1.8949023485183716, "learning_rate": 9.44001204085941e-05, "loss": 0.8453, "step": 189 }, { "epoch": 0.11559990265271355, "grad_norm": 1.656935214996338, "learning_rate": 9.432322196814033e-05, "loss": 0.7722, "step": 190 }, { "epoch": 0.116208323192991, "grad_norm": 1.8966083526611328, "learning_rate": 9.424583465814249e-05, "loss": 0.7714, "step": 191 }, { "epoch": 0.11681674373326843, "grad_norm": 1.4729326963424683, "learning_rate": 9.4167959440686e-05, "loss": 0.8103, "step": 192 }, { "epoch": 0.11742516427354588, "grad_norm": 1.2855483293533325, "learning_rate": 9.408959728392199e-05, "loss": 0.751, "step": 193 }, { "epoch": 0.11803358481382331, "grad_norm": 0.6022900938987732, "learning_rate": 9.401074916205528e-05, "loss": 0.7532, "step": 194 }, { "epoch": 0.11864200535410076, "grad_norm": 0.6037607789039612, "learning_rate": 9.393141605533224e-05, "loss": 0.683, "step": 195 }, { "epoch": 0.11925042589437819, "grad_norm": 1.0040525197982788, "learning_rate": 9.385159895002859e-05, "loss": 0.7185, "step": 196 }, { "epoch": 0.11985884643465564, "grad_norm": 1.1619277000427246, "learning_rate": 9.37712988384372e-05, "loss": 0.934, "step": 197 }, { "epoch": 0.12046726697493307, "grad_norm": 1.2666449546813965, "learning_rate": 9.369051671885566e-05, "loss": 0.8437, "step": 198 }, { "epoch": 0.12107568751521051, "grad_norm": 1.3021451234817505, "learning_rate": 9.360925359557397e-05, "loss": 0.7939, "step": 199 }, { "epoch": 0.12168410805548795, "grad_norm": 1.3136470317840576, "learning_rate": 9.3527510478862e-05, "loss": 0.8785, "step": 200 }, { "epoch": 0.12168410805548795, "eval_loss": 0.770029604434967, "eval_runtime": 0.0645, "eval_samples_per_second": 775.029, "eval_steps_per_second": 46.502, "step": 200 } ], "logging_steps": 1, "max_steps": 937, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 1, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.3040284166258688e+16, "train_batch_size": 6, "trial_name": null, "trial_params": null }