{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.26490066225165565, "eval_steps": 25, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0026490066225165563, "grad_norm": 10.047337532043457, "learning_rate": 2e-05, "loss": 8.8879, "step": 1 }, { "epoch": 0.0026490066225165563, "eval_loss": 2.289355754852295, "eval_runtime": 6.4257, "eval_samples_per_second": 24.744, "eval_steps_per_second": 12.45, "step": 1 }, { "epoch": 0.005298013245033113, "grad_norm": 8.795169830322266, "learning_rate": 4e-05, "loss": 10.083, "step": 2 }, { "epoch": 0.007947019867549669, "grad_norm": 9.06167221069336, "learning_rate": 6e-05, "loss": 9.4119, "step": 3 }, { "epoch": 0.010596026490066225, "grad_norm": 8.713191032409668, "learning_rate": 8e-05, "loss": 9.5243, "step": 4 }, { "epoch": 0.013245033112582781, "grad_norm": 10.839781761169434, "learning_rate": 0.0001, "loss": 9.3628, "step": 5 }, { "epoch": 0.015894039735099338, "grad_norm": 9.976048469543457, "learning_rate": 0.00012, "loss": 8.335, "step": 6 }, { "epoch": 0.018543046357615896, "grad_norm": 9.373313903808594, "learning_rate": 0.00014, "loss": 8.1384, "step": 7 }, { "epoch": 0.02119205298013245, "grad_norm": 14.212782859802246, "learning_rate": 0.00016, "loss": 8.1568, "step": 8 }, { "epoch": 0.02384105960264901, "grad_norm": 11.925253868103027, "learning_rate": 0.00018, "loss": 8.3016, "step": 9 }, { "epoch": 0.026490066225165563, "grad_norm": 11.655914306640625, "learning_rate": 0.0002, "loss": 7.6722, "step": 10 }, { "epoch": 0.02913907284768212, "grad_norm": 10.125999450683594, "learning_rate": 0.0001999390827019096, "loss": 7.6015, "step": 11 }, { "epoch": 0.031788079470198675, "grad_norm": 10.365988731384277, "learning_rate": 0.00019975640502598244, "loss": 6.6575, "step": 12 }, { "epoch": 0.03443708609271523, "grad_norm": 9.90079116821289, "learning_rate": 0.00019945218953682734, "loss": 7.5634, "step": 13 }, { "epoch": 0.03708609271523179, "grad_norm": 8.313066482543945, "learning_rate": 0.00019902680687415705, "loss": 7.8489, "step": 14 }, { "epoch": 0.039735099337748346, "grad_norm": 9.736310005187988, "learning_rate": 0.00019848077530122083, "loss": 8.3458, "step": 15 }, { "epoch": 0.0423841059602649, "grad_norm": 9.37343978881836, "learning_rate": 0.00019781476007338058, "loss": 7.0014, "step": 16 }, { "epoch": 0.045033112582781455, "grad_norm": 8.530896186828613, "learning_rate": 0.00019702957262759965, "loss": 6.7211, "step": 17 }, { "epoch": 0.04768211920529802, "grad_norm": 8.414729118347168, "learning_rate": 0.0001961261695938319, "loss": 6.8012, "step": 18 }, { "epoch": 0.05033112582781457, "grad_norm": 25.417264938354492, "learning_rate": 0.00019510565162951537, "loss": 7.5455, "step": 19 }, { "epoch": 0.052980132450331126, "grad_norm": 6.27047872543335, "learning_rate": 0.00019396926207859084, "loss": 6.9449, "step": 20 }, { "epoch": 0.05562913907284768, "grad_norm": 6.81449556350708, "learning_rate": 0.00019271838545667876, "loss": 6.7678, "step": 21 }, { "epoch": 0.05827814569536424, "grad_norm": 7.924989700317383, "learning_rate": 0.0001913545457642601, "loss": 7.2337, "step": 22 }, { "epoch": 0.060927152317880796, "grad_norm": 6.0346293449401855, "learning_rate": 0.0001898794046299167, "loss": 7.1267, "step": 23 }, { "epoch": 0.06357615894039735, "grad_norm": 7.905987739562988, "learning_rate": 0.00018829475928589271, "loss": 7.5959, "step": 24 }, { "epoch": 0.06622516556291391, "grad_norm": 6.775877475738525, "learning_rate": 0.00018660254037844388, "loss": 7.3511, "step": 25 }, { "epoch": 0.06622516556291391, "eval_loss": 1.7976596355438232, "eval_runtime": 6.4278, "eval_samples_per_second": 24.736, "eval_steps_per_second": 12.446, "step": 25 }, { "epoch": 0.06887417218543046, "grad_norm": 9.095159530639648, "learning_rate": 0.0001848048096156426, "loss": 7.3748, "step": 26 }, { "epoch": 0.07152317880794702, "grad_norm": 8.567487716674805, "learning_rate": 0.00018290375725550417, "loss": 6.9435, "step": 27 }, { "epoch": 0.07417218543046358, "grad_norm": 8.406890869140625, "learning_rate": 0.00018090169943749476, "loss": 7.2, "step": 28 }, { "epoch": 0.07682119205298013, "grad_norm": 7.356227397918701, "learning_rate": 0.00017880107536067218, "loss": 8.3478, "step": 29 }, { "epoch": 0.07947019867549669, "grad_norm": 7.5263776779174805, "learning_rate": 0.0001766044443118978, "loss": 7.1226, "step": 30 }, { "epoch": 0.08211920529801324, "grad_norm": 7.3591766357421875, "learning_rate": 0.00017431448254773944, "loss": 6.8925, "step": 31 }, { "epoch": 0.0847682119205298, "grad_norm": 6.77475643157959, "learning_rate": 0.0001719339800338651, "loss": 6.9482, "step": 32 }, { "epoch": 0.08741721854304636, "grad_norm": 7.577714443206787, "learning_rate": 0.00016946583704589973, "loss": 7.6245, "step": 33 }, { "epoch": 0.09006622516556291, "grad_norm": 8.364958763122559, "learning_rate": 0.00016691306063588583, "loss": 6.0576, "step": 34 }, { "epoch": 0.09271523178807947, "grad_norm": 6.96998929977417, "learning_rate": 0.00016427876096865394, "loss": 7.0662, "step": 35 }, { "epoch": 0.09536423841059603, "grad_norm": 7.5627593994140625, "learning_rate": 0.0001615661475325658, "loss": 8.0993, "step": 36 }, { "epoch": 0.09801324503311258, "grad_norm": 7.6041107177734375, "learning_rate": 0.00015877852522924732, "loss": 7.5794, "step": 37 }, { "epoch": 0.10066225165562914, "grad_norm": 6.256470680236816, "learning_rate": 0.0001559192903470747, "loss": 7.9102, "step": 38 }, { "epoch": 0.10331125827814569, "grad_norm": 6.799134731292725, "learning_rate": 0.0001529919264233205, "loss": 6.9581, "step": 39 }, { "epoch": 0.10596026490066225, "grad_norm": 6.711977481842041, "learning_rate": 0.00015000000000000001, "loss": 7.503, "step": 40 }, { "epoch": 0.10860927152317881, "grad_norm": 6.807905197143555, "learning_rate": 0.00014694715627858908, "loss": 8.0716, "step": 41 }, { "epoch": 0.11125827814569536, "grad_norm": 9.389381408691406, "learning_rate": 0.00014383711467890774, "loss": 6.842, "step": 42 }, { "epoch": 0.11390728476821192, "grad_norm": 6.579544544219971, "learning_rate": 0.00014067366430758004, "loss": 6.9505, "step": 43 }, { "epoch": 0.11655629139072848, "grad_norm": 6.2934184074401855, "learning_rate": 0.00013746065934159123, "loss": 6.7775, "step": 44 }, { "epoch": 0.11920529801324503, "grad_norm": 6.887556076049805, "learning_rate": 0.00013420201433256689, "loss": 7.4255, "step": 45 }, { "epoch": 0.12185430463576159, "grad_norm": 6.750312805175781, "learning_rate": 0.00013090169943749476, "loss": 6.7287, "step": 46 }, { "epoch": 0.12450331125827814, "grad_norm": 5.679995059967041, "learning_rate": 0.0001275637355816999, "loss": 6.5704, "step": 47 }, { "epoch": 0.1271523178807947, "grad_norm": 8.375008583068848, "learning_rate": 0.00012419218955996676, "loss": 6.5227, "step": 48 }, { "epoch": 0.12980132450331125, "grad_norm": 6.405532360076904, "learning_rate": 0.00012079116908177593, "loss": 7.0324, "step": 49 }, { "epoch": 0.13245033112582782, "grad_norm": 8.176584243774414, "learning_rate": 0.00011736481776669306, "loss": 6.9371, "step": 50 }, { "epoch": 0.13245033112582782, "eval_loss": 1.7544549703598022, "eval_runtime": 6.4152, "eval_samples_per_second": 24.785, "eval_steps_per_second": 12.47, "step": 50 }, { "epoch": 0.13509933774834437, "grad_norm": 8.039058685302734, "learning_rate": 0.00011391731009600654, "loss": 6.7803, "step": 51 }, { "epoch": 0.13774834437086092, "grad_norm": 5.868615627288818, "learning_rate": 0.00011045284632676536, "loss": 7.0988, "step": 52 }, { "epoch": 0.1403973509933775, "grad_norm": 6.132235527038574, "learning_rate": 0.00010697564737441252, "loss": 6.5921, "step": 53 }, { "epoch": 0.14304635761589404, "grad_norm": 6.561174392700195, "learning_rate": 0.00010348994967025012, "loss": 6.8622, "step": 54 }, { "epoch": 0.1456953642384106, "grad_norm": 6.470300674438477, "learning_rate": 0.0001, "loss": 7.5364, "step": 55 }, { "epoch": 0.14834437086092717, "grad_norm": 7.035355567932129, "learning_rate": 9.651005032974994e-05, "loss": 6.9205, "step": 56 }, { "epoch": 0.1509933774834437, "grad_norm": 6.995288848876953, "learning_rate": 9.302435262558747e-05, "loss": 6.7877, "step": 57 }, { "epoch": 0.15364238410596026, "grad_norm": 7.111328601837158, "learning_rate": 8.954715367323468e-05, "loss": 6.6399, "step": 58 }, { "epoch": 0.1562913907284768, "grad_norm": 6.084195137023926, "learning_rate": 8.608268990399349e-05, "loss": 7.2281, "step": 59 }, { "epoch": 0.15894039735099338, "grad_norm": 7.517582893371582, "learning_rate": 8.263518223330697e-05, "loss": 6.7618, "step": 60 }, { "epoch": 0.16158940397350993, "grad_norm": 7.0816874504089355, "learning_rate": 7.920883091822408e-05, "loss": 7.3287, "step": 61 }, { "epoch": 0.16423841059602648, "grad_norm": 6.235696792602539, "learning_rate": 7.580781044003324e-05, "loss": 6.5588, "step": 62 }, { "epoch": 0.16688741721854305, "grad_norm": 9.056792259216309, "learning_rate": 7.243626441830009e-05, "loss": 7.5109, "step": 63 }, { "epoch": 0.1695364238410596, "grad_norm": 7.157319068908691, "learning_rate": 6.909830056250527e-05, "loss": 6.6009, "step": 64 }, { "epoch": 0.17218543046357615, "grad_norm": 6.171271324157715, "learning_rate": 6.579798566743314e-05, "loss": 7.2154, "step": 65 }, { "epoch": 0.17483443708609273, "grad_norm": 6.188497543334961, "learning_rate": 6.25393406584088e-05, "loss": 7.397, "step": 66 }, { "epoch": 0.17748344370860927, "grad_norm": 5.943967819213867, "learning_rate": 5.9326335692419995e-05, "loss": 7.0184, "step": 67 }, { "epoch": 0.18013245033112582, "grad_norm": 7.671113967895508, "learning_rate": 5.616288532109225e-05, "loss": 6.6367, "step": 68 }, { "epoch": 0.1827814569536424, "grad_norm": 6.130111217498779, "learning_rate": 5.305284372141095e-05, "loss": 6.6626, "step": 69 }, { "epoch": 0.18543046357615894, "grad_norm": 9.644905090332031, "learning_rate": 5.000000000000002e-05, "loss": 7.2466, "step": 70 }, { "epoch": 0.1880794701986755, "grad_norm": 6.909838676452637, "learning_rate": 4.700807357667952e-05, "loss": 7.1044, "step": 71 }, { "epoch": 0.19072847682119207, "grad_norm": 6.281887531280518, "learning_rate": 4.4080709652925336e-05, "loss": 6.2072, "step": 72 }, { "epoch": 0.19337748344370861, "grad_norm": 6.749195575714111, "learning_rate": 4.12214747707527e-05, "loss": 6.725, "step": 73 }, { "epoch": 0.19602649006622516, "grad_norm": 5.841734886169434, "learning_rate": 3.843385246743417e-05, "loss": 6.9915, "step": 74 }, { "epoch": 0.1986754966887417, "grad_norm": 5.667914390563965, "learning_rate": 3.5721239031346066e-05, "loss": 5.9325, "step": 75 }, { "epoch": 0.1986754966887417, "eval_loss": 1.7390562295913696, "eval_runtime": 6.4222, "eval_samples_per_second": 24.758, "eval_steps_per_second": 12.457, "step": 75 }, { "epoch": 0.20132450331125828, "grad_norm": 6.836699962615967, "learning_rate": 3.308693936411421e-05, "loss": 6.6397, "step": 76 }, { "epoch": 0.20397350993377483, "grad_norm": 7.016422748565674, "learning_rate": 3.053416295410026e-05, "loss": 6.9548, "step": 77 }, { "epoch": 0.20662251655629138, "grad_norm": 5.971282482147217, "learning_rate": 2.8066019966134904e-05, "loss": 6.9251, "step": 78 }, { "epoch": 0.20927152317880796, "grad_norm": 5.713533878326416, "learning_rate": 2.5685517452260567e-05, "loss": 6.7799, "step": 79 }, { "epoch": 0.2119205298013245, "grad_norm": 6.973901748657227, "learning_rate": 2.339555568810221e-05, "loss": 6.9072, "step": 80 }, { "epoch": 0.21456953642384105, "grad_norm": 6.7183685302734375, "learning_rate": 2.119892463932781e-05, "loss": 6.9971, "step": 81 }, { "epoch": 0.21721854304635763, "grad_norm": 6.681792259216309, "learning_rate": 1.9098300562505266e-05, "loss": 7.233, "step": 82 }, { "epoch": 0.21986754966887417, "grad_norm": 8.250860214233398, "learning_rate": 1.7096242744495837e-05, "loss": 7.0648, "step": 83 }, { "epoch": 0.22251655629139072, "grad_norm": 6.514826774597168, "learning_rate": 1.5195190384357404e-05, "loss": 6.2877, "step": 84 }, { "epoch": 0.2251655629139073, "grad_norm": 6.5112786293029785, "learning_rate": 1.339745962155613e-05, "loss": 6.7192, "step": 85 }, { "epoch": 0.22781456953642384, "grad_norm": 6.814388751983643, "learning_rate": 1.1705240714107302e-05, "loss": 7.2217, "step": 86 }, { "epoch": 0.2304635761589404, "grad_norm": 5.921008586883545, "learning_rate": 1.0120595370083318e-05, "loss": 7.1328, "step": 87 }, { "epoch": 0.23311258278145697, "grad_norm": 5.6203932762146, "learning_rate": 8.645454235739903e-06, "loss": 5.9535, "step": 88 }, { "epoch": 0.23576158940397351, "grad_norm": 6.021057605743408, "learning_rate": 7.281614543321269e-06, "loss": 6.8638, "step": 89 }, { "epoch": 0.23841059602649006, "grad_norm": 7.241350173950195, "learning_rate": 6.030737921409169e-06, "loss": 7.3687, "step": 90 }, { "epoch": 0.2410596026490066, "grad_norm": 6.832858562469482, "learning_rate": 4.8943483704846475e-06, "loss": 6.4625, "step": 91 }, { "epoch": 0.24370860927152319, "grad_norm": 6.067267894744873, "learning_rate": 3.873830406168111e-06, "loss": 6.7384, "step": 92 }, { "epoch": 0.24635761589403973, "grad_norm": 7.329842567443848, "learning_rate": 2.970427372400353e-06, "loss": 6.5468, "step": 93 }, { "epoch": 0.24900662251655628, "grad_norm": 5.869202136993408, "learning_rate": 2.1852399266194314e-06, "loss": 7.0172, "step": 94 }, { "epoch": 0.25165562913907286, "grad_norm": 5.81821870803833, "learning_rate": 1.5192246987791981e-06, "loss": 6.985, "step": 95 }, { "epoch": 0.2543046357615894, "grad_norm": 7.167514801025391, "learning_rate": 9.731931258429638e-07, "loss": 7.0374, "step": 96 }, { "epoch": 0.25695364238410595, "grad_norm": 6.343252182006836, "learning_rate": 5.478104631726711e-07, "loss": 5.98, "step": 97 }, { "epoch": 0.2596026490066225, "grad_norm": 6.643914222717285, "learning_rate": 2.4359497401758024e-07, "loss": 6.9213, "step": 98 }, { "epoch": 0.26225165562913905, "grad_norm": 5.431641101837158, "learning_rate": 6.09172980904238e-08, "loss": 7.2081, "step": 99 }, { "epoch": 0.26490066225165565, "grad_norm": 6.084027290344238, "learning_rate": 0.0, "loss": 6.7836, "step": 100 }, { "epoch": 0.26490066225165565, "eval_loss": 1.736009955406189, "eval_runtime": 6.4349, "eval_samples_per_second": 24.709, "eval_steps_per_second": 12.432, "step": 100 } ], "logging_steps": 1, "max_steps": 100, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.002425230445773e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }