{ "best_metric": 0.2286633849143982, "best_model_checkpoint": "miner_id_24/checkpoint-100", "epoch": 0.21770682148040638, "eval_steps": 50, "global_step": 150, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001451378809869376, "grad_norm": 0.11486738920211792, "learning_rate": 1.0110000000000001e-05, "loss": 0.2249, "step": 1 }, { "epoch": 0.001451378809869376, "eval_loss": 0.27678582072257996, "eval_runtime": 142.974, "eval_samples_per_second": 2.035, "eval_steps_per_second": 0.511, "step": 1 }, { "epoch": 0.002902757619738752, "grad_norm": 0.11387992650270462, "learning_rate": 2.0220000000000003e-05, "loss": 0.4195, "step": 2 }, { "epoch": 0.0043541364296081275, "grad_norm": 0.1770201325416565, "learning_rate": 3.033e-05, "loss": 0.3746, "step": 3 }, { "epoch": 0.005805515239477504, "grad_norm": 0.15089277923107147, "learning_rate": 4.0440000000000006e-05, "loss": 0.3461, "step": 4 }, { "epoch": 0.00725689404934688, "grad_norm": 0.11748471856117249, "learning_rate": 5.055e-05, "loss": 0.1969, "step": 5 }, { "epoch": 0.008708272859216255, "grad_norm": 0.1325271725654602, "learning_rate": 6.066e-05, "loss": 0.2835, "step": 6 }, { "epoch": 0.010159651669085631, "grad_norm": 0.14989273250102997, "learning_rate": 7.077e-05, "loss": 0.44, "step": 7 }, { "epoch": 0.011611030478955007, "grad_norm": 0.1532517969608307, "learning_rate": 8.088000000000001e-05, "loss": 0.4155, "step": 8 }, { "epoch": 0.013062409288824383, "grad_norm": 0.12538368999958038, "learning_rate": 9.099000000000001e-05, "loss": 0.3613, "step": 9 }, { "epoch": 0.01451378809869376, "grad_norm": 0.13840965926647186, "learning_rate": 0.0001011, "loss": 0.7261, "step": 10 }, { "epoch": 0.015965166908563134, "grad_norm": 0.1670045405626297, "learning_rate": 0.00010056789473684211, "loss": 0.5663, "step": 11 }, { "epoch": 0.01741654571843251, "grad_norm": 0.13095274567604065, "learning_rate": 0.00010003578947368421, "loss": 0.3288, "step": 12 }, { "epoch": 0.018867924528301886, "grad_norm": 0.19772979617118835, "learning_rate": 9.950368421052632e-05, "loss": 0.613, "step": 13 }, { "epoch": 0.020319303338171262, "grad_norm": 0.16937138140201569, "learning_rate": 9.897157894736842e-05, "loss": 0.4953, "step": 14 }, { "epoch": 0.02177068214804064, "grad_norm": 0.1333189755678177, "learning_rate": 9.843947368421053e-05, "loss": 0.2195, "step": 15 }, { "epoch": 0.023222060957910014, "grad_norm": 0.1428212672472, "learning_rate": 9.790736842105264e-05, "loss": 0.2334, "step": 16 }, { "epoch": 0.02467343976777939, "grad_norm": 0.14330728352069855, "learning_rate": 9.737526315789474e-05, "loss": 0.3142, "step": 17 }, { "epoch": 0.026124818577648767, "grad_norm": 0.1493055373430252, "learning_rate": 9.684315789473684e-05, "loss": 0.2551, "step": 18 }, { "epoch": 0.027576197387518143, "grad_norm": 0.16706344485282898, "learning_rate": 9.631105263157895e-05, "loss": 0.5574, "step": 19 }, { "epoch": 0.02902757619738752, "grad_norm": 0.11606133729219437, "learning_rate": 9.577894736842105e-05, "loss": 0.1948, "step": 20 }, { "epoch": 0.030478955007256895, "grad_norm": 0.0992949903011322, "learning_rate": 9.524684210526317e-05, "loss": 0.1042, "step": 21 }, { "epoch": 0.03193033381712627, "grad_norm": 0.15178389847278595, "learning_rate": 9.471473684210526e-05, "loss": 0.4452, "step": 22 }, { "epoch": 0.033381712626995644, "grad_norm": 0.15305830538272858, "learning_rate": 9.418263157894737e-05, "loss": 0.4336, "step": 23 }, { "epoch": 0.03483309143686502, "grad_norm": 0.21439702808856964, "learning_rate": 9.365052631578948e-05, "loss": 0.346, "step": 24 }, { "epoch": 0.036284470246734396, "grad_norm": 0.1342068761587143, "learning_rate": 9.311842105263157e-05, "loss": 0.3454, "step": 25 }, { "epoch": 0.03773584905660377, "grad_norm": 0.1589202582836151, "learning_rate": 9.258631578947368e-05, "loss": 0.4931, "step": 26 }, { "epoch": 0.03918722786647315, "grad_norm": 0.16286569833755493, "learning_rate": 9.20542105263158e-05, "loss": 0.5843, "step": 27 }, { "epoch": 0.040638606676342524, "grad_norm": 0.11265375465154648, "learning_rate": 9.15221052631579e-05, "loss": 0.2569, "step": 28 }, { "epoch": 0.0420899854862119, "grad_norm": 0.18966278433799744, "learning_rate": 9.099000000000001e-05, "loss": 0.4922, "step": 29 }, { "epoch": 0.04354136429608128, "grad_norm": 0.2602640688419342, "learning_rate": 9.045789473684212e-05, "loss": 0.3959, "step": 30 }, { "epoch": 0.04499274310595065, "grad_norm": 0.12830835580825806, "learning_rate": 8.992578947368421e-05, "loss": 0.3903, "step": 31 }, { "epoch": 0.04644412191582003, "grad_norm": 0.10974445939064026, "learning_rate": 8.939368421052632e-05, "loss": 0.2795, "step": 32 }, { "epoch": 0.047895500725689405, "grad_norm": 0.16427350044250488, "learning_rate": 8.886157894736841e-05, "loss": 0.4776, "step": 33 }, { "epoch": 0.04934687953555878, "grad_norm": 0.12569822371006012, "learning_rate": 8.832947368421054e-05, "loss": 0.2615, "step": 34 }, { "epoch": 0.05079825834542816, "grad_norm": 0.10434585809707642, "learning_rate": 8.779736842105264e-05, "loss": 0.2248, "step": 35 }, { "epoch": 0.05224963715529753, "grad_norm": 0.1004101112484932, "learning_rate": 8.726526315789474e-05, "loss": 0.2055, "step": 36 }, { "epoch": 0.05370101596516691, "grad_norm": 0.12491065263748169, "learning_rate": 8.673315789473685e-05, "loss": 0.2482, "step": 37 }, { "epoch": 0.055152394775036286, "grad_norm": 0.11319632083177567, "learning_rate": 8.620105263157896e-05, "loss": 0.2185, "step": 38 }, { "epoch": 0.05660377358490566, "grad_norm": 0.15422259271144867, "learning_rate": 8.566894736842105e-05, "loss": 0.3448, "step": 39 }, { "epoch": 0.05805515239477504, "grad_norm": 0.15866440534591675, "learning_rate": 8.513684210526316e-05, "loss": 0.6739, "step": 40 }, { "epoch": 0.059506531204644414, "grad_norm": 0.1435173898935318, "learning_rate": 8.460473684210527e-05, "loss": 0.2159, "step": 41 }, { "epoch": 0.06095791001451379, "grad_norm": 0.21297410130500793, "learning_rate": 8.407263157894738e-05, "loss": 0.3077, "step": 42 }, { "epoch": 0.062409288824383166, "grad_norm": 0.1251886934041977, "learning_rate": 8.354052631578948e-05, "loss": 0.2113, "step": 43 }, { "epoch": 0.06386066763425254, "grad_norm": 0.1815638542175293, "learning_rate": 8.300842105263158e-05, "loss": 0.6564, "step": 44 }, { "epoch": 0.06531204644412192, "grad_norm": 0.2576989531517029, "learning_rate": 8.247631578947369e-05, "loss": 0.5205, "step": 45 }, { "epoch": 0.06676342525399129, "grad_norm": 0.27275457978248596, "learning_rate": 8.19442105263158e-05, "loss": 0.9175, "step": 46 }, { "epoch": 0.06821480406386067, "grad_norm": 0.26641345024108887, "learning_rate": 8.141210526315789e-05, "loss": 0.812, "step": 47 }, { "epoch": 0.06966618287373004, "grad_norm": 0.2708298861980438, "learning_rate": 8.088000000000001e-05, "loss": 0.6076, "step": 48 }, { "epoch": 0.07111756168359942, "grad_norm": 0.5135082602500916, "learning_rate": 8.03478947368421e-05, "loss": 0.9538, "step": 49 }, { "epoch": 0.07256894049346879, "grad_norm": 0.46205997467041016, "learning_rate": 7.981578947368421e-05, "loss": 1.1091, "step": 50 }, { "epoch": 0.07256894049346879, "eval_loss": 0.24128209054470062, "eval_runtime": 146.2789, "eval_samples_per_second": 1.989, "eval_steps_per_second": 0.499, "step": 50 }, { "epoch": 0.07402031930333818, "grad_norm": 0.15042653679847717, "learning_rate": 7.928368421052632e-05, "loss": 0.5749, "step": 51 }, { "epoch": 0.07547169811320754, "grad_norm": 0.14070290327072144, "learning_rate": 7.875157894736842e-05, "loss": 0.3815, "step": 52 }, { "epoch": 0.07692307692307693, "grad_norm": 0.12616680562496185, "learning_rate": 7.821947368421053e-05, "loss": 0.3485, "step": 53 }, { "epoch": 0.0783744557329463, "grad_norm": 0.1807371824979782, "learning_rate": 7.768736842105263e-05, "loss": 0.6332, "step": 54 }, { "epoch": 0.07982583454281568, "grad_norm": 0.1610403060913086, "learning_rate": 7.715526315789474e-05, "loss": 0.4334, "step": 55 }, { "epoch": 0.08127721335268505, "grad_norm": 0.16740575432777405, "learning_rate": 7.662315789473685e-05, "loss": 0.5169, "step": 56 }, { "epoch": 0.08272859216255443, "grad_norm": 0.13583853840827942, "learning_rate": 7.609105263157895e-05, "loss": 0.3515, "step": 57 }, { "epoch": 0.0841799709724238, "grad_norm": 0.20852766931056976, "learning_rate": 7.555894736842105e-05, "loss": 0.8417, "step": 58 }, { "epoch": 0.08563134978229318, "grad_norm": 0.12840092182159424, "learning_rate": 7.502684210526316e-05, "loss": 0.3527, "step": 59 }, { "epoch": 0.08708272859216255, "grad_norm": 0.20794260501861572, "learning_rate": 7.449473684210526e-05, "loss": 0.8209, "step": 60 }, { "epoch": 0.08853410740203194, "grad_norm": 0.15332041680812836, "learning_rate": 7.396263157894738e-05, "loss": 0.5123, "step": 61 }, { "epoch": 0.0899854862119013, "grad_norm": 0.12152067571878433, "learning_rate": 7.343052631578949e-05, "loss": 0.5442, "step": 62 }, { "epoch": 0.09143686502177069, "grad_norm": 0.15734325349330902, "learning_rate": 7.289842105263158e-05, "loss": 0.4926, "step": 63 }, { "epoch": 0.09288824383164006, "grad_norm": 0.10759459435939789, "learning_rate": 7.236631578947369e-05, "loss": 0.2285, "step": 64 }, { "epoch": 0.09433962264150944, "grad_norm": 0.11365488916635513, "learning_rate": 7.183421052631579e-05, "loss": 0.3032, "step": 65 }, { "epoch": 0.09579100145137881, "grad_norm": 0.13057614862918854, "learning_rate": 7.13021052631579e-05, "loss": 0.4789, "step": 66 }, { "epoch": 0.09724238026124818, "grad_norm": 0.1509314924478531, "learning_rate": 7.077e-05, "loss": 0.5571, "step": 67 }, { "epoch": 0.09869375907111756, "grad_norm": 0.16016532480716705, "learning_rate": 7.023789473684211e-05, "loss": 0.2708, "step": 68 }, { "epoch": 0.10014513788098693, "grad_norm": 0.2185145616531372, "learning_rate": 6.970578947368422e-05, "loss": 0.4046, "step": 69 }, { "epoch": 0.10159651669085631, "grad_norm": 0.14823275804519653, "learning_rate": 6.917368421052633e-05, "loss": 0.5717, "step": 70 }, { "epoch": 0.10304789550072568, "grad_norm": 0.16261740028858185, "learning_rate": 6.864157894736842e-05, "loss": 0.3313, "step": 71 }, { "epoch": 0.10449927431059507, "grad_norm": 0.1639692485332489, "learning_rate": 6.810947368421053e-05, "loss": 0.7474, "step": 72 }, { "epoch": 0.10595065312046444, "grad_norm": 0.11137572675943375, "learning_rate": 6.757736842105264e-05, "loss": 0.3237, "step": 73 }, { "epoch": 0.10740203193033382, "grad_norm": 0.09403427690267563, "learning_rate": 6.704526315789473e-05, "loss": 0.213, "step": 74 }, { "epoch": 0.10885341074020319, "grad_norm": 0.1198626309633255, "learning_rate": 6.651315789473685e-05, "loss": 0.3147, "step": 75 }, { "epoch": 0.11030478955007257, "grad_norm": 0.18822187185287476, "learning_rate": 6.598105263157895e-05, "loss": 0.2479, "step": 76 }, { "epoch": 0.11175616835994194, "grad_norm": 0.08338137716054916, "learning_rate": 6.544894736842106e-05, "loss": 0.1284, "step": 77 }, { "epoch": 0.11320754716981132, "grad_norm": 0.11395663768053055, "learning_rate": 6.491684210526317e-05, "loss": 0.304, "step": 78 }, { "epoch": 0.11465892597968069, "grad_norm": 0.12172822654247284, "learning_rate": 6.438473684210526e-05, "loss": 0.5225, "step": 79 }, { "epoch": 0.11611030478955008, "grad_norm": 0.18436570465564728, "learning_rate": 6.385263157894737e-05, "loss": 1.1369, "step": 80 }, { "epoch": 0.11756168359941944, "grad_norm": 0.08676780015230179, "learning_rate": 6.332052631578948e-05, "loss": 0.2241, "step": 81 }, { "epoch": 0.11901306240928883, "grad_norm": 0.14665937423706055, "learning_rate": 6.278842105263159e-05, "loss": 0.4818, "step": 82 }, { "epoch": 0.1204644412191582, "grad_norm": 0.12407947331666946, "learning_rate": 6.22563157894737e-05, "loss": 0.4061, "step": 83 }, { "epoch": 0.12191582002902758, "grad_norm": 0.12156087160110474, "learning_rate": 6.172421052631579e-05, "loss": 0.4842, "step": 84 }, { "epoch": 0.12336719883889695, "grad_norm": 0.164754256606102, "learning_rate": 6.11921052631579e-05, "loss": 0.6039, "step": 85 }, { "epoch": 0.12481857764876633, "grad_norm": 0.13035471737384796, "learning_rate": 6.066e-05, "loss": 0.5963, "step": 86 }, { "epoch": 0.1262699564586357, "grad_norm": 0.08667127043008804, "learning_rate": 6.012789473684211e-05, "loss": 0.104, "step": 87 }, { "epoch": 0.12772133526850507, "grad_norm": 0.12500996887683868, "learning_rate": 5.959578947368421e-05, "loss": 0.1519, "step": 88 }, { "epoch": 0.12917271407837447, "grad_norm": 0.1052846610546112, "learning_rate": 5.9063684210526324e-05, "loss": 0.168, "step": 89 }, { "epoch": 0.13062409288824384, "grad_norm": 0.16326619684696198, "learning_rate": 5.8531578947368425e-05, "loss": 1.0778, "step": 90 }, { "epoch": 0.1320754716981132, "grad_norm": 0.18031665682792664, "learning_rate": 5.7999473684210527e-05, "loss": 0.8578, "step": 91 }, { "epoch": 0.13352685050798258, "grad_norm": 0.14327001571655273, "learning_rate": 5.7467368421052635e-05, "loss": 0.8367, "step": 92 }, { "epoch": 0.13497822931785197, "grad_norm": 0.12158166617155075, "learning_rate": 5.6935263157894736e-05, "loss": 0.2978, "step": 93 }, { "epoch": 0.13642960812772134, "grad_norm": 0.17898587882518768, "learning_rate": 5.640315789473684e-05, "loss": 0.7248, "step": 94 }, { "epoch": 0.1378809869375907, "grad_norm": 0.1620013415813446, "learning_rate": 5.587105263157895e-05, "loss": 0.5909, "step": 95 }, { "epoch": 0.13933236574746008, "grad_norm": 0.24596630036830902, "learning_rate": 5.533894736842106e-05, "loss": 1.3, "step": 96 }, { "epoch": 0.14078374455732948, "grad_norm": 0.28974536061286926, "learning_rate": 5.480684210526316e-05, "loss": 1.0699, "step": 97 }, { "epoch": 0.14223512336719885, "grad_norm": 0.2533659338951111, "learning_rate": 5.4274736842105264e-05, "loss": 0.6266, "step": 98 }, { "epoch": 0.14368650217706821, "grad_norm": 0.36984333395957947, "learning_rate": 5.374263157894737e-05, "loss": 1.3249, "step": 99 }, { "epoch": 0.14513788098693758, "grad_norm": 0.4221994876861572, "learning_rate": 5.3210526315789474e-05, "loss": 0.9726, "step": 100 }, { "epoch": 0.14513788098693758, "eval_loss": 0.2286633849143982, "eval_runtime": 145.162, "eval_samples_per_second": 2.005, "eval_steps_per_second": 0.503, "step": 100 }, { "epoch": 0.14658925979680695, "grad_norm": 0.13860496878623962, "learning_rate": 5.2678421052631576e-05, "loss": 0.7931, "step": 101 }, { "epoch": 0.14804063860667635, "grad_norm": 0.1530522257089615, "learning_rate": 5.214631578947369e-05, "loss": 0.4883, "step": 102 }, { "epoch": 0.14949201741654572, "grad_norm": 0.09421262890100479, "learning_rate": 5.161421052631579e-05, "loss": 0.1757, "step": 103 }, { "epoch": 0.1509433962264151, "grad_norm": 0.11253266036510468, "learning_rate": 5.10821052631579e-05, "loss": 0.4355, "step": 104 }, { "epoch": 0.15239477503628446, "grad_norm": 0.11226090788841248, "learning_rate": 5.055e-05, "loss": 0.3406, "step": 105 }, { "epoch": 0.15384615384615385, "grad_norm": 0.09969495981931686, "learning_rate": 5.0017894736842104e-05, "loss": 0.33, "step": 106 }, { "epoch": 0.15529753265602322, "grad_norm": 0.14994554221630096, "learning_rate": 4.948578947368421e-05, "loss": 0.7712, "step": 107 }, { "epoch": 0.1567489114658926, "grad_norm": 0.16747573018074036, "learning_rate": 4.895368421052632e-05, "loss": 0.4805, "step": 108 }, { "epoch": 0.15820029027576196, "grad_norm": 0.15101346373558044, "learning_rate": 4.842157894736842e-05, "loss": 0.6745, "step": 109 }, { "epoch": 0.15965166908563136, "grad_norm": 0.14325077831745148, "learning_rate": 4.7889473684210523e-05, "loss": 0.2481, "step": 110 }, { "epoch": 0.16110304789550073, "grad_norm": 0.6589614152908325, "learning_rate": 4.735736842105263e-05, "loss": 0.385, "step": 111 }, { "epoch": 0.1625544267053701, "grad_norm": 0.14817237854003906, "learning_rate": 4.682526315789474e-05, "loss": 0.4519, "step": 112 }, { "epoch": 0.16400580551523947, "grad_norm": 0.12075196951627731, "learning_rate": 4.629315789473684e-05, "loss": 0.32, "step": 113 }, { "epoch": 0.16545718432510886, "grad_norm": 0.10896284133195877, "learning_rate": 4.576105263157895e-05, "loss": 0.252, "step": 114 }, { "epoch": 0.16690856313497823, "grad_norm": 0.15797024965286255, "learning_rate": 4.522894736842106e-05, "loss": 0.7193, "step": 115 }, { "epoch": 0.1683599419448476, "grad_norm": 0.12811946868896484, "learning_rate": 4.469684210526316e-05, "loss": 0.5991, "step": 116 }, { "epoch": 0.16981132075471697, "grad_norm": 0.17125706374645233, "learning_rate": 4.416473684210527e-05, "loss": 0.6612, "step": 117 }, { "epoch": 0.17126269956458637, "grad_norm": 0.21033529937267303, "learning_rate": 4.363263157894737e-05, "loss": 0.637, "step": 118 }, { "epoch": 0.17271407837445574, "grad_norm": 0.14343442022800446, "learning_rate": 4.310052631578948e-05, "loss": 0.3919, "step": 119 }, { "epoch": 0.1741654571843251, "grad_norm": 0.09223167598247528, "learning_rate": 4.256842105263158e-05, "loss": 0.2206, "step": 120 }, { "epoch": 0.17561683599419448, "grad_norm": 0.10201866924762726, "learning_rate": 4.203631578947369e-05, "loss": 0.2657, "step": 121 }, { "epoch": 0.17706821480406387, "grad_norm": 0.09830533713102341, "learning_rate": 4.150421052631579e-05, "loss": 0.1767, "step": 122 }, { "epoch": 0.17851959361393324, "grad_norm": 0.11305135488510132, "learning_rate": 4.09721052631579e-05, "loss": 0.3528, "step": 123 }, { "epoch": 0.1799709724238026, "grad_norm": 0.11795450747013092, "learning_rate": 4.0440000000000006e-05, "loss": 0.2826, "step": 124 }, { "epoch": 0.18142235123367198, "grad_norm": 0.10810278356075287, "learning_rate": 3.990789473684211e-05, "loss": 0.3728, "step": 125 }, { "epoch": 0.18287373004354138, "grad_norm": 0.07889194041490555, "learning_rate": 3.937578947368421e-05, "loss": 0.1485, "step": 126 }, { "epoch": 0.18432510885341075, "grad_norm": 0.12167462706565857, "learning_rate": 3.884368421052632e-05, "loss": 0.3191, "step": 127 }, { "epoch": 0.18577648766328012, "grad_norm": 0.13652844727039337, "learning_rate": 3.8311578947368426e-05, "loss": 0.3272, "step": 128 }, { "epoch": 0.18722786647314948, "grad_norm": 0.11355964839458466, "learning_rate": 3.777947368421053e-05, "loss": 0.2216, "step": 129 }, { "epoch": 0.18867924528301888, "grad_norm": 0.12072530388832092, "learning_rate": 3.724736842105263e-05, "loss": 0.2511, "step": 130 }, { "epoch": 0.19013062409288825, "grad_norm": 0.11075890809297562, "learning_rate": 3.6715263157894744e-05, "loss": 0.1769, "step": 131 }, { "epoch": 0.19158200290275762, "grad_norm": 0.09322332590818405, "learning_rate": 3.6183157894736845e-05, "loss": 0.2052, "step": 132 }, { "epoch": 0.193033381712627, "grad_norm": 0.12875616550445557, "learning_rate": 3.565105263157895e-05, "loss": 0.1961, "step": 133 }, { "epoch": 0.19448476052249636, "grad_norm": 0.11314037442207336, "learning_rate": 3.5118947368421055e-05, "loss": 0.2227, "step": 134 }, { "epoch": 0.19593613933236576, "grad_norm": 0.10332886129617691, "learning_rate": 3.458684210526316e-05, "loss": 0.339, "step": 135 }, { "epoch": 0.19738751814223512, "grad_norm": 0.17729254066944122, "learning_rate": 3.4054736842105265e-05, "loss": 0.2802, "step": 136 }, { "epoch": 0.1988388969521045, "grad_norm": 0.08777690678834915, "learning_rate": 3.3522631578947366e-05, "loss": 0.1391, "step": 137 }, { "epoch": 0.20029027576197386, "grad_norm": 0.1536087989807129, "learning_rate": 3.2990526315789475e-05, "loss": 0.3643, "step": 138 }, { "epoch": 0.20174165457184326, "grad_norm": 0.11051812022924423, "learning_rate": 3.245842105263158e-05, "loss": 0.1991, "step": 139 }, { "epoch": 0.20319303338171263, "grad_norm": 0.09295262396335602, "learning_rate": 3.1926315789473685e-05, "loss": 0.2913, "step": 140 }, { "epoch": 0.204644412191582, "grad_norm": 0.13505689799785614, "learning_rate": 3.139421052631579e-05, "loss": 0.8371, "step": 141 }, { "epoch": 0.20609579100145137, "grad_norm": 0.1267169862985611, "learning_rate": 3.0862105263157894e-05, "loss": 0.6024, "step": 142 }, { "epoch": 0.20754716981132076, "grad_norm": 0.13407814502716064, "learning_rate": 3.033e-05, "loss": 0.4159, "step": 143 }, { "epoch": 0.20899854862119013, "grad_norm": 0.11469519138336182, "learning_rate": 2.9797894736842104e-05, "loss": 0.2402, "step": 144 }, { "epoch": 0.2104499274310595, "grad_norm": 0.169059157371521, "learning_rate": 2.9265789473684213e-05, "loss": 0.3534, "step": 145 }, { "epoch": 0.21190130624092887, "grad_norm": 0.19773763418197632, "learning_rate": 2.8733684210526317e-05, "loss": 0.5001, "step": 146 }, { "epoch": 0.21335268505079827, "grad_norm": 0.4277418553829193, "learning_rate": 2.820157894736842e-05, "loss": 0.7581, "step": 147 }, { "epoch": 0.21480406386066764, "grad_norm": 0.2791711688041687, "learning_rate": 2.766947368421053e-05, "loss": 0.8721, "step": 148 }, { "epoch": 0.216255442670537, "grad_norm": 0.40901464223861694, "learning_rate": 2.7137368421052632e-05, "loss": 0.8766, "step": 149 }, { "epoch": 0.21770682148040638, "grad_norm": 0.39813101291656494, "learning_rate": 2.6605263157894737e-05, "loss": 0.5074, "step": 150 }, { "epoch": 0.21770682148040638, "eval_loss": 0.23037858307361603, "eval_runtime": 143.626, "eval_samples_per_second": 2.026, "eval_steps_per_second": 0.508, "step": 150 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 1 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.8064531192283136e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }