{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0564542753383954, "eval_steps": 200, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002641135688345989, "grad_norm": 38990.80078125, "learning_rate": 1.0000000000000002e-06, "loss": 39.1249, "step": 1 }, { "epoch": 0.002641135688345989, "eval_loss": 10.096823692321777, "eval_runtime": 2.1873, "eval_samples_per_second": 226.308, "eval_steps_per_second": 28.346, "step": 1 }, { "epoch": 0.005282271376691978, "grad_norm": 22047.08203125, "learning_rate": 2.0000000000000003e-06, "loss": 36.867, "step": 2 }, { "epoch": 0.007923407065037967, "grad_norm": 81682.6796875, "learning_rate": 3e-06, "loss": 39.9853, "step": 3 }, { "epoch": 0.010564542753383956, "grad_norm": 26123.87109375, "learning_rate": 4.000000000000001e-06, "loss": 38.4879, "step": 4 }, { "epoch": 0.013205678441729944, "grad_norm": 29102.25390625, "learning_rate": 5e-06, "loss": 37.9034, "step": 5 }, { "epoch": 0.015846814130075933, "grad_norm": 24646.03125, "learning_rate": 6e-06, "loss": 37.7639, "step": 6 }, { "epoch": 0.01848794981842192, "grad_norm": 56427.59375, "learning_rate": 7.000000000000001e-06, "loss": 37.4074, "step": 7 }, { "epoch": 0.02112908550676791, "grad_norm": 28639.47265625, "learning_rate": 8.000000000000001e-06, "loss": 37.1164, "step": 8 }, { "epoch": 0.0237702211951139, "grad_norm": 18426.78515625, "learning_rate": 9e-06, "loss": 37.2912, "step": 9 }, { "epoch": 0.02641135688345989, "grad_norm": 29681.62109375, "learning_rate": 1e-05, "loss": 37.1851, "step": 10 }, { "epoch": 0.029052492571805876, "grad_norm": 61535.3671875, "learning_rate": 1.1000000000000001e-05, "loss": 38.4262, "step": 11 }, { "epoch": 0.03169362826015187, "grad_norm": 28930.455078125, "learning_rate": 1.2e-05, "loss": 37.3699, "step": 12 }, { "epoch": 0.034334763948497854, "grad_norm": 31548.685546875, "learning_rate": 1.3000000000000001e-05, "loss": 36.4952, "step": 13 }, { "epoch": 0.03697589963684384, "grad_norm": 18435.33203125, "learning_rate": 1.4000000000000001e-05, "loss": 36.2176, "step": 14 }, { "epoch": 0.03961703532518983, "grad_norm": 28708.28515625, "learning_rate": 1.5e-05, "loss": 36.6194, "step": 15 }, { "epoch": 0.04225817101353582, "grad_norm": 20564.423828125, "learning_rate": 1.6000000000000003e-05, "loss": 36.3482, "step": 16 }, { "epoch": 0.04489930670188181, "grad_norm": 17695.943359375, "learning_rate": 1.7000000000000003e-05, "loss": 35.715, "step": 17 }, { "epoch": 0.0475404423902278, "grad_norm": 15335.0712890625, "learning_rate": 1.8e-05, "loss": 34.6268, "step": 18 }, { "epoch": 0.050181578078573784, "grad_norm": 13583.33203125, "learning_rate": 1.9e-05, "loss": 35.5606, "step": 19 }, { "epoch": 0.05282271376691978, "grad_norm": 26019.890625, "learning_rate": 2e-05, "loss": 35.5296, "step": 20 }, { "epoch": 0.055463849455265765, "grad_norm": 14379.9052734375, "learning_rate": 2.1e-05, "loss": 34.4806, "step": 21 }, { "epoch": 0.05810498514361175, "grad_norm": 27736.314453125, "learning_rate": 2.2000000000000003e-05, "loss": 36.2501, "step": 22 }, { "epoch": 0.06074612083195774, "grad_norm": 260855.0, "learning_rate": 2.3000000000000003e-05, "loss": 150.3672, "step": 23 }, { "epoch": 0.06338725652030373, "grad_norm": 422515.90625, "learning_rate": 2.4e-05, "loss": 292.9531, "step": 24 }, { "epoch": 0.06602839220864971, "grad_norm": 1018015.1875, "learning_rate": 2.5e-05, "loss": 355.2773, "step": 25 }, { "epoch": 0.06866952789699571, "grad_norm": 817252.75, "learning_rate": 2.6000000000000002e-05, "loss": 268.3635, "step": 26 }, { "epoch": 0.0713106635853417, "grad_norm": 650477.3125, "learning_rate": 2.7000000000000002e-05, "loss": 304.9252, "step": 27 }, { "epoch": 0.07395179927368768, "grad_norm": 405537.28125, "learning_rate": 2.8000000000000003e-05, "loss": 298.1852, "step": 28 }, { "epoch": 0.07659293496203368, "grad_norm": 1309428.0, "learning_rate": 2.9e-05, "loss": 405.8535, "step": 29 }, { "epoch": 0.07923407065037966, "grad_norm": 574544.0625, "learning_rate": 3e-05, "loss": 341.7129, "step": 30 }, { "epoch": 0.08187520633872565, "grad_norm": 425538.625, "learning_rate": 3.1e-05, "loss": 260.0619, "step": 31 }, { "epoch": 0.08451634202707164, "grad_norm": 566855.6875, "learning_rate": 3.2000000000000005e-05, "loss": 272.2906, "step": 32 }, { "epoch": 0.08715747771541763, "grad_norm": 300651.28125, "learning_rate": 3.3e-05, "loss": 79.3429, "step": 33 }, { "epoch": 0.08979861340376362, "grad_norm": 14083.6494140625, "learning_rate": 3.4000000000000007e-05, "loss": 35.4547, "step": 34 }, { "epoch": 0.09243974909210961, "grad_norm": 22081.201171875, "learning_rate": 3.5e-05, "loss": 34.5699, "step": 35 }, { "epoch": 0.0950808847804556, "grad_norm": 11022.6884765625, "learning_rate": 3.6e-05, "loss": 34.3495, "step": 36 }, { "epoch": 0.09772202046880159, "grad_norm": 11900.990234375, "learning_rate": 3.7e-05, "loss": 35.6864, "step": 37 }, { "epoch": 0.10036315615714757, "grad_norm": 13156.771484375, "learning_rate": 3.8e-05, "loss": 34.4444, "step": 38 }, { "epoch": 0.10300429184549356, "grad_norm": 11813.6083984375, "learning_rate": 3.9000000000000006e-05, "loss": 34.9737, "step": 39 }, { "epoch": 0.10564542753383956, "grad_norm": 15030.2021484375, "learning_rate": 4e-05, "loss": 34.0348, "step": 40 }, { "epoch": 0.10828656322218554, "grad_norm": 11196.2529296875, "learning_rate": 4.1e-05, "loss": 34.3456, "step": 41 }, { "epoch": 0.11092769891053153, "grad_norm": 11016.130859375, "learning_rate": 4.2e-05, "loss": 34.3275, "step": 42 }, { "epoch": 0.11356883459887751, "grad_norm": 14342.5283203125, "learning_rate": 4.3e-05, "loss": 33.6461, "step": 43 }, { "epoch": 0.1162099702872235, "grad_norm": 13592.828125, "learning_rate": 4.4000000000000006e-05, "loss": 35.0608, "step": 44 }, { "epoch": 0.1188511059755695, "grad_norm": 14278.0205078125, "learning_rate": 4.5e-05, "loss": 33.6034, "step": 45 }, { "epoch": 0.12149224166391548, "grad_norm": 15676.8076171875, "learning_rate": 4.600000000000001e-05, "loss": 34.7689, "step": 46 }, { "epoch": 0.12413337735226147, "grad_norm": 16533.037109375, "learning_rate": 4.7e-05, "loss": 34.1397, "step": 47 }, { "epoch": 0.12677451304060747, "grad_norm": 17516.21484375, "learning_rate": 4.8e-05, "loss": 35.4853, "step": 48 }, { "epoch": 0.12941564872895345, "grad_norm": 22093.806640625, "learning_rate": 4.9e-05, "loss": 36.8646, "step": 49 }, { "epoch": 0.13205678441729943, "grad_norm": 34389.921875, "learning_rate": 5e-05, "loss": 39.5266, "step": 50 }, { "epoch": 0.13469792010564544, "grad_norm": 7149.9775390625, "learning_rate": 5.1000000000000006e-05, "loss": 35.8088, "step": 51 }, { "epoch": 0.13733905579399142, "grad_norm": 6511.89306640625, "learning_rate": 5.2000000000000004e-05, "loss": 35.0745, "step": 52 }, { "epoch": 0.1399801914823374, "grad_norm": 11293.515625, "learning_rate": 5.300000000000001e-05, "loss": 33.7064, "step": 53 }, { "epoch": 0.1426213271706834, "grad_norm": 7394.4853515625, "learning_rate": 5.4000000000000005e-05, "loss": 34.612, "step": 54 }, { "epoch": 0.14526246285902938, "grad_norm": 7513.56982421875, "learning_rate": 5.500000000000001e-05, "loss": 34.2795, "step": 55 }, { "epoch": 0.14790359854737536, "grad_norm": 12561.0849609375, "learning_rate": 5.6000000000000006e-05, "loss": 34.1079, "step": 56 }, { "epoch": 0.15054473423572137, "grad_norm": 7255.42724609375, "learning_rate": 5.6999999999999996e-05, "loss": 33.7773, "step": 57 }, { "epoch": 0.15318586992406735, "grad_norm": 8305.197265625, "learning_rate": 5.8e-05, "loss": 33.5425, "step": 58 }, { "epoch": 0.15582700561241333, "grad_norm": 7724.32666015625, "learning_rate": 5.9e-05, "loss": 34.3069, "step": 59 }, { "epoch": 0.1584681413007593, "grad_norm": 6973.86669921875, "learning_rate": 6e-05, "loss": 31.8323, "step": 60 }, { "epoch": 0.16110927698910532, "grad_norm": 8178.408203125, "learning_rate": 6.1e-05, "loss": 33.4728, "step": 61 }, { "epoch": 0.1637504126774513, "grad_norm": 7446.3310546875, "learning_rate": 6.2e-05, "loss": 32.2049, "step": 62 }, { "epoch": 0.16639154836579728, "grad_norm": 7538.81494140625, "learning_rate": 6.3e-05, "loss": 31.9451, "step": 63 }, { "epoch": 0.1690326840541433, "grad_norm": 7067.33154296875, "learning_rate": 6.400000000000001e-05, "loss": 32.0696, "step": 64 }, { "epoch": 0.17167381974248927, "grad_norm": 7199.02294921875, "learning_rate": 6.500000000000001e-05, "loss": 31.7234, "step": 65 }, { "epoch": 0.17431495543083525, "grad_norm": 6351.2900390625, "learning_rate": 6.6e-05, "loss": 31.4103, "step": 66 }, { "epoch": 0.17695609111918126, "grad_norm": 9954.1572265625, "learning_rate": 6.7e-05, "loss": 31.2581, "step": 67 }, { "epoch": 0.17959722680752724, "grad_norm": 6812.11083984375, "learning_rate": 6.800000000000001e-05, "loss": 31.0586, "step": 68 }, { "epoch": 0.18223836249587322, "grad_norm": 6788.81787109375, "learning_rate": 6.9e-05, "loss": 31.2012, "step": 69 }, { "epoch": 0.18487949818421923, "grad_norm": 6330.77880859375, "learning_rate": 7e-05, "loss": 31.758, "step": 70 }, { "epoch": 0.1875206338725652, "grad_norm": 6925.2958984375, "learning_rate": 7.1e-05, "loss": 31.7811, "step": 71 }, { "epoch": 0.1901617695609112, "grad_norm": 15530.548828125, "learning_rate": 7.2e-05, "loss": 38.1656, "step": 72 }, { "epoch": 0.19280290524925717, "grad_norm": 248175.0, "learning_rate": 7.3e-05, "loss": 314.6807, "step": 73 }, { "epoch": 0.19544404093760318, "grad_norm": 348192.3125, "learning_rate": 7.4e-05, "loss": 319.9785, "step": 74 }, { "epoch": 0.19808517662594916, "grad_norm": 399153.90625, "learning_rate": 7.500000000000001e-05, "loss": 245.481, "step": 75 }, { "epoch": 0.20072631231429514, "grad_norm": 293389.5, "learning_rate": 7.6e-05, "loss": 269.3301, "step": 76 }, { "epoch": 0.20336744800264114, "grad_norm": 536027.375, "learning_rate": 7.7e-05, "loss": 240.3848, "step": 77 }, { "epoch": 0.20600858369098712, "grad_norm": 340636.96875, "learning_rate": 7.800000000000001e-05, "loss": 257.0401, "step": 78 }, { "epoch": 0.2086497193793331, "grad_norm": 343861.0625, "learning_rate": 7.900000000000001e-05, "loss": 246.5806, "step": 79 }, { "epoch": 0.2112908550676791, "grad_norm": 377362.75, "learning_rate": 8e-05, "loss": 204.2622, "step": 80 }, { "epoch": 0.2139319907560251, "grad_norm": 495172.15625, "learning_rate": 8.1e-05, "loss": 152.9565, "step": 81 }, { "epoch": 0.21657312644437107, "grad_norm": 410514.21875, "learning_rate": 8.2e-05, "loss": 120.1336, "step": 82 }, { "epoch": 0.21921426213271708, "grad_norm": 37318.89453125, "learning_rate": 8.3e-05, "loss": 43.507, "step": 83 }, { "epoch": 0.22185539782106306, "grad_norm": 46563.8515625, "learning_rate": 8.4e-05, "loss": 42.6661, "step": 84 }, { "epoch": 0.22449653350940904, "grad_norm": 25882.45703125, "learning_rate": 8.5e-05, "loss": 41.1904, "step": 85 }, { "epoch": 0.22713766919775502, "grad_norm": 21462.017578125, "learning_rate": 8.6e-05, "loss": 35.6957, "step": 86 }, { "epoch": 0.22977880488610103, "grad_norm": 11826.3798828125, "learning_rate": 8.7e-05, "loss": 33.1654, "step": 87 }, { "epoch": 0.232419940574447, "grad_norm": 10408.4365234375, "learning_rate": 8.800000000000001e-05, "loss": 32.1405, "step": 88 }, { "epoch": 0.235061076262793, "grad_norm": 9028.2587890625, "learning_rate": 8.900000000000001e-05, "loss": 30.806, "step": 89 }, { "epoch": 0.237702211951139, "grad_norm": 14064.7021484375, "learning_rate": 9e-05, "loss": 32.0673, "step": 90 }, { "epoch": 0.24034334763948498, "grad_norm": 10274.6611328125, "learning_rate": 9.1e-05, "loss": 30.8923, "step": 91 }, { "epoch": 0.24298448332783096, "grad_norm": 13376.0947265625, "learning_rate": 9.200000000000001e-05, "loss": 32.0376, "step": 92 }, { "epoch": 0.24562561901617697, "grad_norm": 13412.4970703125, "learning_rate": 9.300000000000001e-05, "loss": 32.5937, "step": 93 }, { "epoch": 0.24826675470452295, "grad_norm": 17289.099609375, "learning_rate": 9.4e-05, "loss": 32.0219, "step": 94 }, { "epoch": 0.2509078903928689, "grad_norm": 10165.4990234375, "learning_rate": 9.5e-05, "loss": 32.7753, "step": 95 }, { "epoch": 0.25354902608121493, "grad_norm": 16371.439453125, "learning_rate": 9.6e-05, "loss": 31.7399, "step": 96 }, { "epoch": 0.2561901617695609, "grad_norm": 28360.642578125, "learning_rate": 9.7e-05, "loss": 32.4525, "step": 97 }, { "epoch": 0.2588312974579069, "grad_norm": 19952.9296875, "learning_rate": 9.8e-05, "loss": 33.4285, "step": 98 }, { "epoch": 0.2614724331462529, "grad_norm": 20724.11328125, "learning_rate": 9.900000000000001e-05, "loss": 34.1331, "step": 99 }, { "epoch": 0.26411356883459886, "grad_norm": 38431.6328125, "learning_rate": 0.0001, "loss": 40.0183, "step": 100 }, { "epoch": 0.26675470452294486, "grad_norm": 9295.7626953125, "learning_rate": 9.99999993018897e-05, "loss": 32.908, "step": 101 }, { "epoch": 0.26939584021129087, "grad_norm": 9068.3134765625, "learning_rate": 9.999999720755877e-05, "loss": 32.7796, "step": 102 }, { "epoch": 0.2720369758996368, "grad_norm": 9507.033203125, "learning_rate": 9.99999937170073e-05, "loss": 33.0735, "step": 103 }, { "epoch": 0.27467811158798283, "grad_norm": 9898.73046875, "learning_rate": 9.999998883023537e-05, "loss": 34.4524, "step": 104 }, { "epoch": 0.27731924727632884, "grad_norm": 8197.7294921875, "learning_rate": 9.999998254724313e-05, "loss": 34.3429, "step": 105 }, { "epoch": 0.2799603829646748, "grad_norm": 7723.392578125, "learning_rate": 9.999997486803075e-05, "loss": 33.3132, "step": 106 }, { "epoch": 0.2826015186530208, "grad_norm": 9174.4091796875, "learning_rate": 9.999996579259843e-05, "loss": 32.9465, "step": 107 }, { "epoch": 0.2852426543413668, "grad_norm": 10098.0283203125, "learning_rate": 9.999995532094644e-05, "loss": 34.1124, "step": 108 }, { "epoch": 0.28788379002971276, "grad_norm": 7904.126953125, "learning_rate": 9.999994345307508e-05, "loss": 32.68, "step": 109 }, { "epoch": 0.29052492571805877, "grad_norm": 7395.32177734375, "learning_rate": 9.999993018898466e-05, "loss": 32.1147, "step": 110 }, { "epoch": 0.2931660614064048, "grad_norm": 15490.7314453125, "learning_rate": 9.999991552867558e-05, "loss": 32.7157, "step": 111 }, { "epoch": 0.29580719709475073, "grad_norm": 6962.9326171875, "learning_rate": 9.99998994721482e-05, "loss": 32.5103, "step": 112 }, { "epoch": 0.29844833278309674, "grad_norm": 6746.60546875, "learning_rate": 9.999988201940302e-05, "loss": 31.9245, "step": 113 }, { "epoch": 0.30108946847144274, "grad_norm": 6943.94140625, "learning_rate": 9.999986317044051e-05, "loss": 30.399, "step": 114 }, { "epoch": 0.3037306041597887, "grad_norm": 6095.4384765625, "learning_rate": 9.999984292526118e-05, "loss": 29.941, "step": 115 }, { "epoch": 0.3063717398481347, "grad_norm": 6518.970703125, "learning_rate": 9.999982128386562e-05, "loss": 30.5093, "step": 116 }, { "epoch": 0.3090128755364807, "grad_norm": 5806.0927734375, "learning_rate": 9.99997982462544e-05, "loss": 29.6937, "step": 117 }, { "epoch": 0.31165401122482667, "grad_norm": 6216.46435546875, "learning_rate": 9.999977381242821e-05, "loss": 29.7115, "step": 118 }, { "epoch": 0.3142951469131727, "grad_norm": 5445.48828125, "learning_rate": 9.999974798238769e-05, "loss": 28.9644, "step": 119 }, { "epoch": 0.3169362826015186, "grad_norm": 4930.64453125, "learning_rate": 9.99997207561336e-05, "loss": 29.803, "step": 120 }, { "epoch": 0.31957741828986463, "grad_norm": 5866.5478515625, "learning_rate": 9.999969213366667e-05, "loss": 29.2732, "step": 121 }, { "epoch": 0.32221855397821064, "grad_norm": 13160.4111328125, "learning_rate": 9.99996621149877e-05, "loss": 31.3179, "step": 122 }, { "epoch": 0.3248596896665566, "grad_norm": 577362.4375, "learning_rate": 9.999963070009755e-05, "loss": 192.1116, "step": 123 }, { "epoch": 0.3275008253549026, "grad_norm": 447577.625, "learning_rate": 9.999959788899706e-05, "loss": 353.353, "step": 124 }, { "epoch": 0.3301419610432486, "grad_norm": 422884.03125, "learning_rate": 9.999956368168719e-05, "loss": 328.2871, "step": 125 }, { "epoch": 0.33278309673159456, "grad_norm": 217372.875, "learning_rate": 9.999952807816888e-05, "loss": 311.041, "step": 126 }, { "epoch": 0.33542423241994057, "grad_norm": 197269.15625, "learning_rate": 9.99994910784431e-05, "loss": 331.5454, "step": 127 }, { "epoch": 0.3380653681082866, "grad_norm": 447190.15625, "learning_rate": 9.999945268251092e-05, "loss": 292.5098, "step": 128 }, { "epoch": 0.34070650379663253, "grad_norm": 156708.53125, "learning_rate": 9.999941289037338e-05, "loss": 329.5899, "step": 129 }, { "epoch": 0.34334763948497854, "grad_norm": 214527.265625, "learning_rate": 9.999937170203162e-05, "loss": 295.6437, "step": 130 }, { "epoch": 0.34598877517332455, "grad_norm": 144792.09375, "learning_rate": 9.999932911748678e-05, "loss": 321.7724, "step": 131 }, { "epoch": 0.3486299108616705, "grad_norm": 183092.328125, "learning_rate": 9.999928513674004e-05, "loss": 138.0811, "step": 132 }, { "epoch": 0.3512710465500165, "grad_norm": 6041.10107421875, "learning_rate": 9.999923975979262e-05, "loss": 30.1601, "step": 133 }, { "epoch": 0.3539121822383625, "grad_norm": 5054.18798828125, "learning_rate": 9.999919298664582e-05, "loss": 29.4563, "step": 134 }, { "epoch": 0.35655331792670847, "grad_norm": 9742.12890625, "learning_rate": 9.999914481730092e-05, "loss": 29.5483, "step": 135 }, { "epoch": 0.3591944536150545, "grad_norm": 13321.4970703125, "learning_rate": 9.999909525175927e-05, "loss": 29.7589, "step": 136 }, { "epoch": 0.3618355893034005, "grad_norm": 9211.091796875, "learning_rate": 9.999904429002225e-05, "loss": 30.0795, "step": 137 }, { "epoch": 0.36447672499174644, "grad_norm": 10673.2529296875, "learning_rate": 9.99989919320913e-05, "loss": 29.8073, "step": 138 }, { "epoch": 0.36711786068009244, "grad_norm": 9673.37109375, "learning_rate": 9.999893817796786e-05, "loss": 30.8933, "step": 139 }, { "epoch": 0.36975899636843845, "grad_norm": 10085.38671875, "learning_rate": 9.999888302765345e-05, "loss": 29.8822, "step": 140 }, { "epoch": 0.3724001320567844, "grad_norm": 10791.521484375, "learning_rate": 9.99988264811496e-05, "loss": 30.1218, "step": 141 }, { "epoch": 0.3750412677451304, "grad_norm": 11358.93359375, "learning_rate": 9.99987685384579e-05, "loss": 30.735, "step": 142 }, { "epoch": 0.3776824034334764, "grad_norm": 7013.380859375, "learning_rate": 9.999870919957996e-05, "loss": 29.9077, "step": 143 }, { "epoch": 0.3803235391218224, "grad_norm": 7458.63525390625, "learning_rate": 9.999864846451744e-05, "loss": 30.7425, "step": 144 }, { "epoch": 0.3829646748101684, "grad_norm": 8038.50732421875, "learning_rate": 9.999858633327201e-05, "loss": 31.053, "step": 145 }, { "epoch": 0.38560581049851433, "grad_norm": 7841.15283203125, "learning_rate": 9.999852280584544e-05, "loss": 30.7345, "step": 146 }, { "epoch": 0.38824694618686034, "grad_norm": 7719.5048828125, "learning_rate": 9.999845788223949e-05, "loss": 30.9241, "step": 147 }, { "epoch": 0.39088808187520635, "grad_norm": 13179.359375, "learning_rate": 9.999839156245598e-05, "loss": 31.945, "step": 148 }, { "epoch": 0.3935292175635523, "grad_norm": 11153.3046875, "learning_rate": 9.999832384649674e-05, "loss": 34.644, "step": 149 }, { "epoch": 0.3961703532518983, "grad_norm": 47252.56640625, "learning_rate": 9.999825473436369e-05, "loss": 39.1459, "step": 150 }, { "epoch": 0.3988114889402443, "grad_norm": 5307.408203125, "learning_rate": 9.999818422605875e-05, "loss": 32.3124, "step": 151 }, { "epoch": 0.40145262462859027, "grad_norm": 8414.1484375, "learning_rate": 9.999811232158389e-05, "loss": 31.5456, "step": 152 }, { "epoch": 0.4040937603169363, "grad_norm": 5779.16943359375, "learning_rate": 9.999803902094109e-05, "loss": 32.0291, "step": 153 }, { "epoch": 0.4067348960052823, "grad_norm": 6989.2958984375, "learning_rate": 9.999796432413244e-05, "loss": 32.4468, "step": 154 }, { "epoch": 0.40937603169362824, "grad_norm": 10169.005859375, "learning_rate": 9.999788823116001e-05, "loss": 33.1476, "step": 155 }, { "epoch": 0.41201716738197425, "grad_norm": 6967.77197265625, "learning_rate": 9.999781074202592e-05, "loss": 32.4884, "step": 156 }, { "epoch": 0.41465830307032026, "grad_norm": 8052.6611328125, "learning_rate": 9.999773185673232e-05, "loss": 33.8162, "step": 157 }, { "epoch": 0.4172994387586662, "grad_norm": 13675.26953125, "learning_rate": 9.999765157528145e-05, "loss": 33.4981, "step": 158 }, { "epoch": 0.4199405744470122, "grad_norm": 23900.8515625, "learning_rate": 9.99975698976755e-05, "loss": 33.6758, "step": 159 }, { "epoch": 0.4225817101353582, "grad_norm": 8697.0146484375, "learning_rate": 9.99974868239168e-05, "loss": 33.4007, "step": 160 }, { "epoch": 0.4252228458237042, "grad_norm": 7423.0234375, "learning_rate": 9.999740235400765e-05, "loss": 32.8796, "step": 161 }, { "epoch": 0.4278639815120502, "grad_norm": 8968.0107421875, "learning_rate": 9.999731648795041e-05, "loss": 35.1091, "step": 162 }, { "epoch": 0.4305051172003962, "grad_norm": 8960.2470703125, "learning_rate": 9.999722922574749e-05, "loss": 34.6028, "step": 163 }, { "epoch": 0.43314625288874214, "grad_norm": 9324.4716796875, "learning_rate": 9.999714056740129e-05, "loss": 35.0468, "step": 164 }, { "epoch": 0.43578738857708815, "grad_norm": 15031.443359375, "learning_rate": 9.999705051291432e-05, "loss": 33.8078, "step": 165 }, { "epoch": 0.43842852426543416, "grad_norm": 10380.2470703125, "learning_rate": 9.999695906228908e-05, "loss": 34.8672, "step": 166 }, { "epoch": 0.4410696599537801, "grad_norm": 18920.16796875, "learning_rate": 9.999686621552813e-05, "loss": 34.697, "step": 167 }, { "epoch": 0.4437107956421261, "grad_norm": 17273.609375, "learning_rate": 9.999677197263406e-05, "loss": 35.5471, "step": 168 }, { "epoch": 0.44635193133047213, "grad_norm": 10327.810546875, "learning_rate": 9.999667633360952e-05, "loss": 33.3773, "step": 169 }, { "epoch": 0.4489930670188181, "grad_norm": 15529.2939453125, "learning_rate": 9.999657929845714e-05, "loss": 35.3255, "step": 170 }, { "epoch": 0.4516342027071641, "grad_norm": 15885.65625, "learning_rate": 9.999648086717966e-05, "loss": 35.0333, "step": 171 }, { "epoch": 0.45427533839551004, "grad_norm": 16440.353515625, "learning_rate": 9.999638103977982e-05, "loss": 36.1782, "step": 172 }, { "epoch": 0.45691647408385605, "grad_norm": 623966.3125, "learning_rate": 9.999627981626041e-05, "loss": 117.4766, "step": 173 }, { "epoch": 0.45955760977220206, "grad_norm": 443642.1875, "learning_rate": 9.999617719662426e-05, "loss": 256.9298, "step": 174 }, { "epoch": 0.462198745460548, "grad_norm": 426303.78125, "learning_rate": 9.999607318087423e-05, "loss": 213.1021, "step": 175 }, { "epoch": 0.464839881148894, "grad_norm": 753837.9375, "learning_rate": 9.999596776901322e-05, "loss": 234.3458, "step": 176 }, { "epoch": 0.46748101683724, "grad_norm": 500841.875, "learning_rate": 9.999586096104419e-05, "loss": 242.1502, "step": 177 }, { "epoch": 0.470122152525586, "grad_norm": 488348.28125, "learning_rate": 9.99957527569701e-05, "loss": 259.3533, "step": 178 }, { "epoch": 0.472763288213932, "grad_norm": 599034.6875, "learning_rate": 9.999564315679398e-05, "loss": 254.9457, "step": 179 }, { "epoch": 0.475404423902278, "grad_norm": 740236.3125, "learning_rate": 9.99955321605189e-05, "loss": 200.2197, "step": 180 }, { "epoch": 0.47804555959062395, "grad_norm": 279145.40625, "learning_rate": 9.999541976814796e-05, "loss": 211.6974, "step": 181 }, { "epoch": 0.48068669527896996, "grad_norm": 565175.5625, "learning_rate": 9.999530597968428e-05, "loss": 152.1028, "step": 182 }, { "epoch": 0.48332783096731596, "grad_norm": 29268.025390625, "learning_rate": 9.999519079513107e-05, "loss": 37.5746, "step": 183 }, { "epoch": 0.4859689666556619, "grad_norm": 19322.490234375, "learning_rate": 9.999507421449151e-05, "loss": 38.4138, "step": 184 }, { "epoch": 0.4886101023440079, "grad_norm": 27010.8203125, "learning_rate": 9.999495623776886e-05, "loss": 35.2608, "step": 185 }, { "epoch": 0.49125123803235393, "grad_norm": 25924.7890625, "learning_rate": 9.999483686496645e-05, "loss": 38.389, "step": 186 }, { "epoch": 0.4938923737206999, "grad_norm": 33607.66015625, "learning_rate": 9.999471609608757e-05, "loss": 35.7422, "step": 187 }, { "epoch": 0.4965335094090459, "grad_norm": 19824.349609375, "learning_rate": 9.999459393113561e-05, "loss": 37.8325, "step": 188 }, { "epoch": 0.4991746450973919, "grad_norm": 16384.638671875, "learning_rate": 9.9994470370114e-05, "loss": 37.3911, "step": 189 }, { "epoch": 0.5018157807857379, "grad_norm": 15732.8330078125, "learning_rate": 9.999434541302616e-05, "loss": 35.9949, "step": 190 }, { "epoch": 0.5044569164740839, "grad_norm": 23623.61328125, "learning_rate": 9.99942190598756e-05, "loss": 36.3237, "step": 191 }, { "epoch": 0.5070980521624299, "grad_norm": 32387.189453125, "learning_rate": 9.999409131066583e-05, "loss": 36.7266, "step": 192 }, { "epoch": 0.5097391878507759, "grad_norm": 19656.185546875, "learning_rate": 9.999396216540044e-05, "loss": 36.9575, "step": 193 }, { "epoch": 0.5123803235391218, "grad_norm": 20705.455078125, "learning_rate": 9.999383162408304e-05, "loss": 37.2048, "step": 194 }, { "epoch": 0.5150214592274678, "grad_norm": 21470.52734375, "learning_rate": 9.999369968671723e-05, "loss": 36.1668, "step": 195 }, { "epoch": 0.5176625949158138, "grad_norm": 19358.25, "learning_rate": 9.999356635330674e-05, "loss": 33.7397, "step": 196 }, { "epoch": 0.5203037306041598, "grad_norm": 19253.916015625, "learning_rate": 9.999343162385529e-05, "loss": 36.7927, "step": 197 }, { "epoch": 0.5229448662925058, "grad_norm": 41119.46875, "learning_rate": 9.99932954983666e-05, "loss": 36.7557, "step": 198 }, { "epoch": 0.5255860019808518, "grad_norm": 23741.87109375, "learning_rate": 9.999315797684451e-05, "loss": 38.4819, "step": 199 }, { "epoch": 0.5282271376691977, "grad_norm": 33874.09765625, "learning_rate": 9.999301905929286e-05, "loss": 42.3858, "step": 200 }, { "epoch": 0.5282271376691977, "eval_loss": 7.334134578704834, "eval_runtime": 2.2174, "eval_samples_per_second": 223.237, "eval_steps_per_second": 27.961, "step": 200 }, { "epoch": 0.5308682733575437, "grad_norm": 15801.6083984375, "learning_rate": 9.999287874571552e-05, "loss": 38.8128, "step": 201 }, { "epoch": 0.5335094090458897, "grad_norm": 12974.27734375, "learning_rate": 9.99927370361164e-05, "loss": 38.6081, "step": 202 }, { "epoch": 0.5361505447342357, "grad_norm": 12007.9013671875, "learning_rate": 9.999259393049947e-05, "loss": 37.1496, "step": 203 }, { "epoch": 0.5387916804225817, "grad_norm": 13070.220703125, "learning_rate": 9.999244942886871e-05, "loss": 38.7187, "step": 204 }, { "epoch": 0.5414328161109278, "grad_norm": 16807.220703125, "learning_rate": 9.999230353122819e-05, "loss": 41.07, "step": 205 }, { "epoch": 0.5440739517992736, "grad_norm": 14268.9052734375, "learning_rate": 9.999215623758194e-05, "loss": 40.1817, "step": 206 }, { "epoch": 0.5467150874876197, "grad_norm": 13336.4287109375, "learning_rate": 9.99920075479341e-05, "loss": 37.3859, "step": 207 }, { "epoch": 0.5493562231759657, "grad_norm": 15000.0390625, "learning_rate": 9.999185746228882e-05, "loss": 37.9181, "step": 208 }, { "epoch": 0.5519973588643117, "grad_norm": 11059.775390625, "learning_rate": 9.999170598065028e-05, "loss": 37.7867, "step": 209 }, { "epoch": 0.5546384945526577, "grad_norm": 12954.494140625, "learning_rate": 9.999155310302273e-05, "loss": 38.3371, "step": 210 }, { "epoch": 0.5572796302410036, "grad_norm": 10920.3037109375, "learning_rate": 9.999139882941043e-05, "loss": 35.1785, "step": 211 }, { "epoch": 0.5599207659293496, "grad_norm": 15022.30078125, "learning_rate": 9.999124315981766e-05, "loss": 35.528, "step": 212 }, { "epoch": 0.5625619016176956, "grad_norm": 10339.8525390625, "learning_rate": 9.999108609424881e-05, "loss": 34.5773, "step": 213 }, { "epoch": 0.5652030373060416, "grad_norm": 9615.1484375, "learning_rate": 9.999092763270823e-05, "loss": 34.6027, "step": 214 }, { "epoch": 0.5678441729943876, "grad_norm": 13707.630859375, "learning_rate": 9.999076777520037e-05, "loss": 34.8469, "step": 215 }, { "epoch": 0.5704853086827336, "grad_norm": 13718.404296875, "learning_rate": 9.99906065217297e-05, "loss": 34.0409, "step": 216 }, { "epoch": 0.5731264443710795, "grad_norm": 12160.12109375, "learning_rate": 9.99904438723007e-05, "loss": 32.9267, "step": 217 }, { "epoch": 0.5757675800594255, "grad_norm": 9693.056640625, "learning_rate": 9.999027982691793e-05, "loss": 33.0474, "step": 218 }, { "epoch": 0.5784087157477715, "grad_norm": 14817.9755859375, "learning_rate": 9.999011438558595e-05, "loss": 33.6275, "step": 219 }, { "epoch": 0.5810498514361175, "grad_norm": 12656.400390625, "learning_rate": 9.99899475483094e-05, "loss": 33.9675, "step": 220 }, { "epoch": 0.5836909871244635, "grad_norm": 17197.283203125, "learning_rate": 9.998977931509291e-05, "loss": 35.6857, "step": 221 }, { "epoch": 0.5863321228128096, "grad_norm": 215147.109375, "learning_rate": 9.998960968594121e-05, "loss": 88.1464, "step": 222 }, { "epoch": 0.5889732585011554, "grad_norm": 625456.3125, "learning_rate": 9.998943866085903e-05, "loss": 186.8345, "step": 223 }, { "epoch": 0.5916143941895015, "grad_norm": 491068.96875, "learning_rate": 9.998926623985114e-05, "loss": 158.0338, "step": 224 }, { "epoch": 0.5942555298778475, "grad_norm": 626101.125, "learning_rate": 9.998909242292235e-05, "loss": 218.7658, "step": 225 }, { "epoch": 0.5968966655661935, "grad_norm": 303837.34375, "learning_rate": 9.998891721007752e-05, "loss": 186.0703, "step": 226 }, { "epoch": 0.5995378012545395, "grad_norm": 354231.84375, "learning_rate": 9.998874060132155e-05, "loss": 162.2602, "step": 227 }, { "epoch": 0.6021789369428855, "grad_norm": 570096.0625, "learning_rate": 9.998856259665936e-05, "loss": 165.2661, "step": 228 }, { "epoch": 0.6048200726312314, "grad_norm": 405688.65625, "learning_rate": 9.998838319609591e-05, "loss": 159.5345, "step": 229 }, { "epoch": 0.6074612083195774, "grad_norm": 592211.125, "learning_rate": 9.998820239963624e-05, "loss": 141.6046, "step": 230 }, { "epoch": 0.6101023440079234, "grad_norm": 678225.0625, "learning_rate": 9.998802020728537e-05, "loss": 84.9725, "step": 231 }, { "epoch": 0.6127434796962694, "grad_norm": 22088.375, "learning_rate": 9.998783661904843e-05, "loss": 38.1227, "step": 232 }, { "epoch": 0.6153846153846154, "grad_norm": 19927.962890625, "learning_rate": 9.99876516349305e-05, "loss": 37.8816, "step": 233 }, { "epoch": 0.6180257510729614, "grad_norm": 33203.27734375, "learning_rate": 9.998746525493674e-05, "loss": 34.0087, "step": 234 }, { "epoch": 0.6206668867613073, "grad_norm": 10135.03515625, "learning_rate": 9.99872774790724e-05, "loss": 34.0175, "step": 235 }, { "epoch": 0.6233080224496533, "grad_norm": 11513.166015625, "learning_rate": 9.99870883073427e-05, "loss": 32.6651, "step": 236 }, { "epoch": 0.6259491581379993, "grad_norm": 7397.00732421875, "learning_rate": 9.998689773975291e-05, "loss": 32.2064, "step": 237 }, { "epoch": 0.6285902938263453, "grad_norm": 10573.4638671875, "learning_rate": 9.998670577630838e-05, "loss": 32.1057, "step": 238 }, { "epoch": 0.6312314295146914, "grad_norm": 10578.8310546875, "learning_rate": 9.998651241701445e-05, "loss": 32.1381, "step": 239 }, { "epoch": 0.6338725652030373, "grad_norm": 9302.189453125, "learning_rate": 9.998631766187651e-05, "loss": 32.8179, "step": 240 }, { "epoch": 0.6365137008913833, "grad_norm": 8694.892578125, "learning_rate": 9.998612151090003e-05, "loss": 32.7711, "step": 241 }, { "epoch": 0.6391548365797293, "grad_norm": 10467.7099609375, "learning_rate": 9.998592396409047e-05, "loss": 33.1121, "step": 242 }, { "epoch": 0.6417959722680753, "grad_norm": 11832.251953125, "learning_rate": 9.998572502145334e-05, "loss": 32.8568, "step": 243 }, { "epoch": 0.6444371079564213, "grad_norm": 14376.9228515625, "learning_rate": 9.998552468299421e-05, "loss": 32.5907, "step": 244 }, { "epoch": 0.6470782436447673, "grad_norm": 13190.787109375, "learning_rate": 9.998532294871866e-05, "loss": 32.6583, "step": 245 }, { "epoch": 0.6497193793331132, "grad_norm": 10301.1328125, "learning_rate": 9.998511981863232e-05, "loss": 31.7794, "step": 246 }, { "epoch": 0.6523605150214592, "grad_norm": 18970.587890625, "learning_rate": 9.998491529274089e-05, "loss": 32.5321, "step": 247 }, { "epoch": 0.6550016507098052, "grad_norm": 10323.8408203125, "learning_rate": 9.998470937105006e-05, "loss": 32.6962, "step": 248 }, { "epoch": 0.6576427863981512, "grad_norm": 13553.1123046875, "learning_rate": 9.998450205356557e-05, "loss": 34.1782, "step": 249 }, { "epoch": 0.6602839220864972, "grad_norm": 34080.28125, "learning_rate": 9.998429334029323e-05, "loss": 37.3095, "step": 250 }, { "epoch": 0.6629250577748432, "grad_norm": 12205.15234375, "learning_rate": 9.998408323123887e-05, "loss": 33.7182, "step": 251 }, { "epoch": 0.6655661934631891, "grad_norm": 11019.15234375, "learning_rate": 9.998387172640834e-05, "loss": 34.2941, "step": 252 }, { "epoch": 0.6682073291515351, "grad_norm": 10185.3310546875, "learning_rate": 9.998365882580756e-05, "loss": 34.5573, "step": 253 }, { "epoch": 0.6708484648398811, "grad_norm": 8710.2685546875, "learning_rate": 9.998344452944247e-05, "loss": 33.6592, "step": 254 }, { "epoch": 0.6734896005282272, "grad_norm": 8050.28759765625, "learning_rate": 9.998322883731903e-05, "loss": 33.1733, "step": 255 }, { "epoch": 0.6761307362165732, "grad_norm": 6891.90673828125, "learning_rate": 9.998301174944332e-05, "loss": 32.2699, "step": 256 }, { "epoch": 0.6787718719049192, "grad_norm": 6904.37060546875, "learning_rate": 9.998279326582134e-05, "loss": 33.2969, "step": 257 }, { "epoch": 0.6814130075932651, "grad_norm": 6681.41162109375, "learning_rate": 9.998257338645924e-05, "loss": 32.5617, "step": 258 }, { "epoch": 0.6840541432816111, "grad_norm": 7499.51025390625, "learning_rate": 9.998235211136312e-05, "loss": 31.2502, "step": 259 }, { "epoch": 0.6866952789699571, "grad_norm": 5850.79931640625, "learning_rate": 9.99821294405392e-05, "loss": 31.384, "step": 260 }, { "epoch": 0.6893364146583031, "grad_norm": 5846.03271484375, "learning_rate": 9.998190537399366e-05, "loss": 31.2545, "step": 261 }, { "epoch": 0.6919775503466491, "grad_norm": 7224.54833984375, "learning_rate": 9.998167991173277e-05, "loss": 31.2568, "step": 262 }, { "epoch": 0.6946186860349951, "grad_norm": 6079.56982421875, "learning_rate": 9.998145305376286e-05, "loss": 31.7204, "step": 263 }, { "epoch": 0.697259821723341, "grad_norm": 7802.859375, "learning_rate": 9.99812248000902e-05, "loss": 30.3375, "step": 264 }, { "epoch": 0.699900957411687, "grad_norm": 7014.5146484375, "learning_rate": 9.998099515072122e-05, "loss": 30.6416, "step": 265 }, { "epoch": 0.702542093100033, "grad_norm": 6766.64208984375, "learning_rate": 9.998076410566229e-05, "loss": 30.4145, "step": 266 }, { "epoch": 0.705183228788379, "grad_norm": 6723.0986328125, "learning_rate": 9.99805316649199e-05, "loss": 29.3229, "step": 267 }, { "epoch": 0.707824364476725, "grad_norm": 8847.9677734375, "learning_rate": 9.998029782850051e-05, "loss": 29.2886, "step": 268 }, { "epoch": 0.7104655001650709, "grad_norm": 5896.45458984375, "learning_rate": 9.998006259641068e-05, "loss": 29.5852, "step": 269 }, { "epoch": 0.7131066358534169, "grad_norm": 7112.9150390625, "learning_rate": 9.997982596865695e-05, "loss": 29.5084, "step": 270 }, { "epoch": 0.715747771541763, "grad_norm": 8039.98876953125, "learning_rate": 9.997958794524594e-05, "loss": 31.9893, "step": 271 }, { "epoch": 0.718388907230109, "grad_norm": 179267.265625, "learning_rate": 9.99793485261843e-05, "loss": 140.9562, "step": 272 }, { "epoch": 0.721030042918455, "grad_norm": 578681.125, "learning_rate": 9.997910771147872e-05, "loss": 262.198, "step": 273 }, { "epoch": 0.723671178606801, "grad_norm": 322541.34375, "learning_rate": 9.99788655011359e-05, "loss": 237.3132, "step": 274 }, { "epoch": 0.7263123142951469, "grad_norm": 235946.640625, "learning_rate": 9.997862189516263e-05, "loss": 300.6354, "step": 275 }, { "epoch": 0.7289534499834929, "grad_norm": 262057.515625, "learning_rate": 9.99783768935657e-05, "loss": 209.6862, "step": 276 }, { "epoch": 0.7315945856718389, "grad_norm": 221274.765625, "learning_rate": 9.997813049635195e-05, "loss": 208.7495, "step": 277 }, { "epoch": 0.7342357213601849, "grad_norm": 363778.46875, "learning_rate": 9.997788270352827e-05, "loss": 234.0036, "step": 278 }, { "epoch": 0.7368768570485309, "grad_norm": 198016.546875, "learning_rate": 9.997763351510157e-05, "loss": 221.2396, "step": 279 }, { "epoch": 0.7395179927368769, "grad_norm": 383717.4375, "learning_rate": 9.997738293107881e-05, "loss": 166.7505, "step": 280 }, { "epoch": 0.7421591284252228, "grad_norm": 471310.09375, "learning_rate": 9.9977130951467e-05, "loss": 155.5116, "step": 281 }, { "epoch": 0.7448002641135688, "grad_norm": 135402.15625, "learning_rate": 9.997687757627316e-05, "loss": 71.9904, "step": 282 }, { "epoch": 0.7474413998019148, "grad_norm": 6735.1005859375, "learning_rate": 9.997662280550437e-05, "loss": 30.8698, "step": 283 }, { "epoch": 0.7500825354902608, "grad_norm": 11189.4736328125, "learning_rate": 9.997636663916776e-05, "loss": 30.6788, "step": 284 }, { "epoch": 0.7527236711786068, "grad_norm": 9472.00390625, "learning_rate": 9.997610907727046e-05, "loss": 32.5548, "step": 285 }, { "epoch": 0.7553648068669528, "grad_norm": 10074.7333984375, "learning_rate": 9.997585011981966e-05, "loss": 30.9945, "step": 286 }, { "epoch": 0.7580059425552987, "grad_norm": 11928.4619140625, "learning_rate": 9.997558976682262e-05, "loss": 30.6684, "step": 287 }, { "epoch": 0.7606470782436447, "grad_norm": 13231.986328125, "learning_rate": 9.997532801828658e-05, "loss": 30.9457, "step": 288 }, { "epoch": 0.7632882139319908, "grad_norm": 8904.8466796875, "learning_rate": 9.997506487421888e-05, "loss": 31.3361, "step": 289 }, { "epoch": 0.7659293496203368, "grad_norm": 9125.240234375, "learning_rate": 9.997480033462683e-05, "loss": 30.7196, "step": 290 }, { "epoch": 0.7685704853086828, "grad_norm": 9812.6181640625, "learning_rate": 9.997453439951784e-05, "loss": 30.7277, "step": 291 }, { "epoch": 0.7712116209970287, "grad_norm": 7082.22607421875, "learning_rate": 9.997426706889935e-05, "loss": 31.2053, "step": 292 }, { "epoch": 0.7738527566853747, "grad_norm": 9316.9384765625, "learning_rate": 9.997399834277878e-05, "loss": 31.5169, "step": 293 }, { "epoch": 0.7764938923737207, "grad_norm": 19302.771484375, "learning_rate": 9.997372822116368e-05, "loss": 31.651, "step": 294 }, { "epoch": 0.7791350280620667, "grad_norm": 10954.8271484375, "learning_rate": 9.99734567040616e-05, "loss": 30.4, "step": 295 }, { "epoch": 0.7817761637504127, "grad_norm": 9081.9521484375, "learning_rate": 9.997318379148007e-05, "loss": 30.8718, "step": 296 }, { "epoch": 0.7844172994387587, "grad_norm": 6827.958984375, "learning_rate": 9.997290948342673e-05, "loss": 31.0843, "step": 297 }, { "epoch": 0.7870584351271046, "grad_norm": 10805.7939453125, "learning_rate": 9.997263377990926e-05, "loss": 31.6845, "step": 298 }, { "epoch": 0.7896995708154506, "grad_norm": 11347.0078125, "learning_rate": 9.997235668093535e-05, "loss": 33.4166, "step": 299 }, { "epoch": 0.7923407065037966, "grad_norm": 16983.841796875, "learning_rate": 9.997207818651274e-05, "loss": 35.7603, "step": 300 }, { "epoch": 0.7949818421921426, "grad_norm": 3815.614990234375, "learning_rate": 9.997179829664918e-05, "loss": 33.1237, "step": 301 }, { "epoch": 0.7976229778804886, "grad_norm": 4439.759765625, "learning_rate": 9.997151701135253e-05, "loss": 32.6201, "step": 302 }, { "epoch": 0.8002641135688346, "grad_norm": 6584.0, "learning_rate": 9.997123433063062e-05, "loss": 31.9738, "step": 303 }, { "epoch": 0.8029052492571805, "grad_norm": 8394.333984375, "learning_rate": 9.997095025449134e-05, "loss": 34.1952, "step": 304 }, { "epoch": 0.8055463849455266, "grad_norm": 8264.888671875, "learning_rate": 9.997066478294262e-05, "loss": 34.1646, "step": 305 }, { "epoch": 0.8081875206338726, "grad_norm": 6815.27587890625, "learning_rate": 9.997037791599245e-05, "loss": 32.8399, "step": 306 }, { "epoch": 0.8108286563222186, "grad_norm": 6638.54296875, "learning_rate": 9.997008965364884e-05, "loss": 32.737, "step": 307 }, { "epoch": 0.8134697920105646, "grad_norm": 6356.19287109375, "learning_rate": 9.996979999591983e-05, "loss": 33.2864, "step": 308 }, { "epoch": 0.8161109276989106, "grad_norm": 10876.560546875, "learning_rate": 9.996950894281349e-05, "loss": 32.8353, "step": 309 }, { "epoch": 0.8187520633872565, "grad_norm": 18334.380859375, "learning_rate": 9.996921649433796e-05, "loss": 33.1125, "step": 310 }, { "epoch": 0.8213931990756025, "grad_norm": 5925.57080078125, "learning_rate": 9.996892265050144e-05, "loss": 33.4775, "step": 311 }, { "epoch": 0.8240343347639485, "grad_norm": 5512.29541015625, "learning_rate": 9.99686274113121e-05, "loss": 32.4073, "step": 312 }, { "epoch": 0.8266754704522945, "grad_norm": 6770.63232421875, "learning_rate": 9.996833077677819e-05, "loss": 33.0255, "step": 313 }, { "epoch": 0.8293166061406405, "grad_norm": 9025.830078125, "learning_rate": 9.9968032746908e-05, "loss": 31.7732, "step": 314 }, { "epoch": 0.8319577418289865, "grad_norm": 5815.4296875, "learning_rate": 9.996773332170983e-05, "loss": 31.5946, "step": 315 }, { "epoch": 0.8345988775173324, "grad_norm": 7221.68603515625, "learning_rate": 9.996743250119209e-05, "loss": 31.5973, "step": 316 }, { "epoch": 0.8372400132056784, "grad_norm": 7172.86962890625, "learning_rate": 9.996713028536313e-05, "loss": 31.4948, "step": 317 }, { "epoch": 0.8398811488940244, "grad_norm": 11000.0458984375, "learning_rate": 9.99668266742314e-05, "loss": 31.3127, "step": 318 }, { "epoch": 0.8425222845823704, "grad_norm": 8431.4716796875, "learning_rate": 9.99665216678054e-05, "loss": 30.8608, "step": 319 }, { "epoch": 0.8451634202707164, "grad_norm": 7308.78466796875, "learning_rate": 9.996621526609364e-05, "loss": 30.8716, "step": 320 }, { "epoch": 0.8478045559590623, "grad_norm": 8358.787109375, "learning_rate": 9.996590746910467e-05, "loss": 31.0737, "step": 321 }, { "epoch": 0.8504456916474084, "grad_norm": 29319.46484375, "learning_rate": 9.996559827684709e-05, "loss": 46.2402, "step": 322 }, { "epoch": 0.8530868273357544, "grad_norm": 903961.25, "learning_rate": 9.996528768932951e-05, "loss": 161.367, "step": 323 }, { "epoch": 0.8557279630241004, "grad_norm": 436229.9375, "learning_rate": 9.996497570656062e-05, "loss": 215.2534, "step": 324 }, { "epoch": 0.8583690987124464, "grad_norm": 310716.5, "learning_rate": 9.996466232854915e-05, "loss": 218.9532, "step": 325 }, { "epoch": 0.8610102344007924, "grad_norm": 935038.75, "learning_rate": 9.996434755530384e-05, "loss": 204.1668, "step": 326 }, { "epoch": 0.8636513700891383, "grad_norm": 577125.0625, "learning_rate": 9.996403138683347e-05, "loss": 225.0228, "step": 327 }, { "epoch": 0.8662925057774843, "grad_norm": 429562.09375, "learning_rate": 9.996371382314686e-05, "loss": 221.4529, "step": 328 }, { "epoch": 0.8689336414658303, "grad_norm": 469087.0625, "learning_rate": 9.996339486425291e-05, "loss": 161.132, "step": 329 }, { "epoch": 0.8715747771541763, "grad_norm": 946113.1875, "learning_rate": 9.99630745101605e-05, "loss": 169.4336, "step": 330 }, { "epoch": 0.8742159128425223, "grad_norm": 537740.1875, "learning_rate": 9.996275276087859e-05, "loss": 166.9042, "step": 331 }, { "epoch": 0.8768570485308683, "grad_norm": 377986.5, "learning_rate": 9.996242961641615e-05, "loss": 139.4483, "step": 332 }, { "epoch": 0.8794981842192142, "grad_norm": 6949.21044921875, "learning_rate": 9.996210507678223e-05, "loss": 32.8323, "step": 333 }, { "epoch": 0.8821393199075602, "grad_norm": 6551.869140625, "learning_rate": 9.996177914198586e-05, "loss": 31.1956, "step": 334 }, { "epoch": 0.8847804555959062, "grad_norm": 8210.8974609375, "learning_rate": 9.996145181203615e-05, "loss": 30.2494, "step": 335 }, { "epoch": 0.8874215912842522, "grad_norm": 12632.7666015625, "learning_rate": 9.996112308694225e-05, "loss": 30.7789, "step": 336 }, { "epoch": 0.8900627269725983, "grad_norm": 11905.80078125, "learning_rate": 9.996079296671334e-05, "loss": 30.9992, "step": 337 }, { "epoch": 0.8927038626609443, "grad_norm": 11776.396484375, "learning_rate": 9.996046145135865e-05, "loss": 30.6118, "step": 338 }, { "epoch": 0.8953449983492902, "grad_norm": 10494.625, "learning_rate": 9.99601285408874e-05, "loss": 30.6983, "step": 339 }, { "epoch": 0.8979861340376362, "grad_norm": 8309.9296875, "learning_rate": 9.995979423530892e-05, "loss": 30.6617, "step": 340 }, { "epoch": 0.9006272697259822, "grad_norm": 11482.9853515625, "learning_rate": 9.995945853463253e-05, "loss": 30.5696, "step": 341 }, { "epoch": 0.9032684054143282, "grad_norm": 8950.994140625, "learning_rate": 9.995912143886763e-05, "loss": 29.6077, "step": 342 }, { "epoch": 0.9059095411026742, "grad_norm": 8950.931640625, "learning_rate": 9.995878294802357e-05, "loss": 30.4176, "step": 343 }, { "epoch": 0.9085506767910201, "grad_norm": 6688.57470703125, "learning_rate": 9.995844306210988e-05, "loss": 29.8723, "step": 344 }, { "epoch": 0.9111918124793661, "grad_norm": 7882.67431640625, "learning_rate": 9.995810178113599e-05, "loss": 30.049, "step": 345 }, { "epoch": 0.9138329481677121, "grad_norm": 9309.5625, "learning_rate": 9.995775910511147e-05, "loss": 30.2998, "step": 346 }, { "epoch": 0.9164740838560581, "grad_norm": 9403.8974609375, "learning_rate": 9.995741503404587e-05, "loss": 30.4171, "step": 347 }, { "epoch": 0.9191152195444041, "grad_norm": 10254.1376953125, "learning_rate": 9.995706956794879e-05, "loss": 32.398, "step": 348 }, { "epoch": 0.9217563552327501, "grad_norm": 11519.509765625, "learning_rate": 9.99567227068299e-05, "loss": 33.4377, "step": 349 }, { "epoch": 0.924397490921096, "grad_norm": 17227.236328125, "learning_rate": 9.995637445069887e-05, "loss": 36.9788, "step": 350 }, { "epoch": 0.927038626609442, "grad_norm": 8033.53369140625, "learning_rate": 9.995602479956545e-05, "loss": 32.1, "step": 351 }, { "epoch": 0.929679762297788, "grad_norm": 10333.927734375, "learning_rate": 9.995567375343937e-05, "loss": 32.4024, "step": 352 }, { "epoch": 0.932320897986134, "grad_norm": 5577.73486328125, "learning_rate": 9.995532131233044e-05, "loss": 33.2651, "step": 353 }, { "epoch": 0.93496203367448, "grad_norm": 5001.80615234375, "learning_rate": 9.99549674762485e-05, "loss": 33.2199, "step": 354 }, { "epoch": 0.9376031693628261, "grad_norm": 6995.62255859375, "learning_rate": 9.995461224520345e-05, "loss": 33.0332, "step": 355 }, { "epoch": 0.940244305051172, "grad_norm": 5345.10888671875, "learning_rate": 9.995425561920519e-05, "loss": 32.4465, "step": 356 }, { "epoch": 0.942885440739518, "grad_norm": 5311.36376953125, "learning_rate": 9.99538975982637e-05, "loss": 33.3183, "step": 357 }, { "epoch": 0.945526576427864, "grad_norm": 4239.72021484375, "learning_rate": 9.995353818238895e-05, "loss": 30.5123, "step": 358 }, { "epoch": 0.94816771211621, "grad_norm": 6135.8544921875, "learning_rate": 9.9953177371591e-05, "loss": 30.1126, "step": 359 }, { "epoch": 0.950808847804556, "grad_norm": 3885.701904296875, "learning_rate": 9.995281516587991e-05, "loss": 30.1448, "step": 360 }, { "epoch": 0.953449983492902, "grad_norm": 17259.177734375, "learning_rate": 9.99524515652658e-05, "loss": 30.9694, "step": 361 }, { "epoch": 0.9560911191812479, "grad_norm": 5949.1728515625, "learning_rate": 9.995208656975884e-05, "loss": 30.8493, "step": 362 }, { "epoch": 0.9587322548695939, "grad_norm": 231986.453125, "learning_rate": 9.995172017936919e-05, "loss": 141.9035, "step": 363 }, { "epoch": 0.9613733905579399, "grad_norm": 103330.5546875, "learning_rate": 9.99513523941071e-05, "loss": 188.4911, "step": 364 }, { "epoch": 0.9640145262462859, "grad_norm": 307991.03125, "learning_rate": 9.995098321398284e-05, "loss": 160.2285, "step": 365 }, { "epoch": 0.9666556619346319, "grad_norm": 190517.765625, "learning_rate": 9.995061263900671e-05, "loss": 152.4148, "step": 366 }, { "epoch": 0.9692967976229779, "grad_norm": 134986.078125, "learning_rate": 9.995024066918908e-05, "loss": 119.8174, "step": 367 }, { "epoch": 0.9719379333113238, "grad_norm": 215117.609375, "learning_rate": 9.994986730454031e-05, "loss": 125.8479, "step": 368 }, { "epoch": 0.9745790689996698, "grad_norm": 22283.35546875, "learning_rate": 9.994949254507084e-05, "loss": 34.5446, "step": 369 }, { "epoch": 0.9772202046880158, "grad_norm": 12405.2236328125, "learning_rate": 9.994911639079112e-05, "loss": 35.1761, "step": 370 }, { "epoch": 0.9798613403763619, "grad_norm": 13683.02734375, "learning_rate": 9.994873884171167e-05, "loss": 32.3272, "step": 371 }, { "epoch": 0.9825024760647079, "grad_norm": 8963.904296875, "learning_rate": 9.994835989784305e-05, "loss": 31.0019, "step": 372 }, { "epoch": 0.9851436117530538, "grad_norm": 19926.2734375, "learning_rate": 9.994797955919581e-05, "loss": 30.4514, "step": 373 }, { "epoch": 0.9877847474413998, "grad_norm": 37221.25, "learning_rate": 9.994759782578058e-05, "loss": 32.0492, "step": 374 }, { "epoch": 0.9904258831297458, "grad_norm": 10019.828125, "learning_rate": 9.994721469760801e-05, "loss": 31.7783, "step": 375 }, { "epoch": 0.9930670188180918, "grad_norm": 8898.4228515625, "learning_rate": 9.994683017468883e-05, "loss": 30.9381, "step": 376 }, { "epoch": 0.9957081545064378, "grad_norm": 13350.8203125, "learning_rate": 9.994644425703374e-05, "loss": 32.4939, "step": 377 }, { "epoch": 0.9983492901947838, "grad_norm": 29945.037109375, "learning_rate": 9.994605694465355e-05, "loss": 34.0366, "step": 378 }, { "epoch": 1.0009904258831297, "grad_norm": 19480.009765625, "learning_rate": 9.994566823755907e-05, "loss": 37.1069, "step": 379 }, { "epoch": 1.0036315615714757, "grad_norm": 4824.83544921875, "learning_rate": 9.99452781357611e-05, "loss": 35.9486, "step": 380 }, { "epoch": 1.0062726972598217, "grad_norm": 4898.34423828125, "learning_rate": 9.994488663927062e-05, "loss": 34.3521, "step": 381 }, { "epoch": 1.0089138329481677, "grad_norm": 7551.79736328125, "learning_rate": 9.994449374809851e-05, "loss": 36.7028, "step": 382 }, { "epoch": 1.0115549686365137, "grad_norm": 8357.705078125, "learning_rate": 9.994409946225574e-05, "loss": 36.5134, "step": 383 }, { "epoch": 1.0141961043248597, "grad_norm": 5780.6787109375, "learning_rate": 9.994370378175332e-05, "loss": 37.3621, "step": 384 }, { "epoch": 1.0168372400132057, "grad_norm": 5624.93896484375, "learning_rate": 9.994330670660235e-05, "loss": 37.6676, "step": 385 }, { "epoch": 1.0194783757015518, "grad_norm": 6545.541015625, "learning_rate": 9.994290823681385e-05, "loss": 37.2885, "step": 386 }, { "epoch": 1.0221195113898978, "grad_norm": 9896.431640625, "learning_rate": 9.994250837239897e-05, "loss": 37.8031, "step": 387 }, { "epoch": 1.0247606470782435, "grad_norm": 6628.89453125, "learning_rate": 9.994210711336891e-05, "loss": 39.4998, "step": 388 }, { "epoch": 1.0274017827665896, "grad_norm": 7230.349609375, "learning_rate": 9.994170445973483e-05, "loss": 37.6952, "step": 389 }, { "epoch": 1.0300429184549356, "grad_norm": 5001.923828125, "learning_rate": 9.994130041150798e-05, "loss": 37.2387, "step": 390 }, { "epoch": 1.0326840541432816, "grad_norm": 8473.236328125, "learning_rate": 9.994089496869968e-05, "loss": 37.7243, "step": 391 }, { "epoch": 1.0353251898316276, "grad_norm": 12679.2109375, "learning_rate": 9.994048813132119e-05, "loss": 35.9025, "step": 392 }, { "epoch": 1.0379663255199736, "grad_norm": 7488.9248046875, "learning_rate": 9.994007989938392e-05, "loss": 36.2572, "step": 393 }, { "epoch": 1.0406074612083196, "grad_norm": 8192.458984375, "learning_rate": 9.993967027289927e-05, "loss": 38.7854, "step": 394 }, { "epoch": 1.0432485968966656, "grad_norm": 6160.6787109375, "learning_rate": 9.993925925187865e-05, "loss": 35.9352, "step": 395 }, { "epoch": 1.0458897325850116, "grad_norm": 6419.31103515625, "learning_rate": 9.993884683633354e-05, "loss": 37.7825, "step": 396 }, { "epoch": 1.0485308682733576, "grad_norm": 8226.6005859375, "learning_rate": 9.993843302627549e-05, "loss": 35.5052, "step": 397 }, { "epoch": 1.0511720039617036, "grad_norm": 8380.81640625, "learning_rate": 9.993801782171603e-05, "loss": 36.5649, "step": 398 }, { "epoch": 1.0538131396500496, "grad_norm": 10895.78515625, "learning_rate": 9.993760122266676e-05, "loss": 37.1919, "step": 399 }, { "epoch": 1.0564542753383954, "grad_norm": 14454.5390625, "learning_rate": 9.99371832291393e-05, "loss": 38.3564, "step": 400 }, { "epoch": 1.0564542753383954, "eval_loss": 8.69857406616211, "eval_runtime": 2.1301, "eval_samples_per_second": 232.388, "eval_steps_per_second": 29.107, "step": 400 } ], "logging_steps": 1, "max_steps": 18900, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1043682507620352.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }