{ "best_metric": 0.875745415687561, "best_model_checkpoint": "miner_id_24/checkpoint-500", "epoch": 0.03259133722256624, "eval_steps": 50, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 6.518267444513249e-05, "eval_loss": 2.850663900375366, "eval_runtime": 443.5942, "eval_samples_per_second": 14.563, "eval_steps_per_second": 3.641, "step": 1 }, { "epoch": 0.0006518267444513248, "grad_norm": 3.971789598464966, "learning_rate": 4.16e-05, "loss": 2.0815, "step": 10 }, { "epoch": 0.0013036534889026496, "grad_norm": 2.8578591346740723, "learning_rate": 8.32e-05, "loss": 1.8557, "step": 20 }, { "epoch": 0.0019554802333539745, "grad_norm": 3.0689802169799805, "learning_rate": 0.0001248, "loss": 1.5229, "step": 30 }, { "epoch": 0.002607306977805299, "grad_norm": 4.3151397705078125, "learning_rate": 0.0001664, "loss": 1.3099, "step": 40 }, { "epoch": 0.0032591337222566243, "grad_norm": 4.729147434234619, "learning_rate": 0.000208, "loss": 1.6873, "step": 50 }, { "epoch": 0.0032591337222566243, "eval_loss": 1.273123025894165, "eval_runtime": 445.0677, "eval_samples_per_second": 14.515, "eval_steps_per_second": 3.629, "step": 50 }, { "epoch": 0.003910960466707949, "grad_norm": 2.3496737480163574, "learning_rate": 0.0002077466612270217, "loss": 1.0568, "step": 60 }, { "epoch": 0.004562787211159274, "grad_norm": 2.499488353729248, "learning_rate": 0.0002069878791491233, "loss": 1.1359, "step": 70 }, { "epoch": 0.005214613955610598, "grad_norm": 3.1299209594726562, "learning_rate": 0.00020572735047631578, "loss": 1.2264, "step": 80 }, { "epoch": 0.005866440700061924, "grad_norm": 3.3837058544158936, "learning_rate": 0.00020397121637758515, "loss": 1.2004, "step": 90 }, { "epoch": 0.006518267444513249, "grad_norm": 2.973175287246704, "learning_rate": 0.00020172803256173445, "loss": 1.545, "step": 100 }, { "epoch": 0.006518267444513249, "eval_loss": 1.275679111480713, "eval_runtime": 446.3022, "eval_samples_per_second": 14.474, "eval_steps_per_second": 3.619, "step": 100 }, { "epoch": 0.007170094188964573, "grad_norm": 2.3844895362854004, "learning_rate": 0.00019900872759483047, "loss": 1.0244, "step": 110 }, { "epoch": 0.007821920933415898, "grad_norm": 2.4352855682373047, "learning_rate": 0.0001958265496573284, "loss": 1.1891, "step": 120 }, { "epoch": 0.008473747677867223, "grad_norm": 3.231959819793701, "learning_rate": 0.00019219700200026827, "loss": 1.2308, "step": 130 }, { "epoch": 0.009125574422318547, "grad_norm": 2.5154852867126465, "learning_rate": 0.0001881377674149945, "loss": 1.2077, "step": 140 }, { "epoch": 0.009777401166769872, "grad_norm": 3.0093817710876465, "learning_rate": 0.00018366862208437368, "loss": 1.3988, "step": 150 }, { "epoch": 0.009777401166769872, "eval_loss": 1.1874332427978516, "eval_runtime": 447.657, "eval_samples_per_second": 14.431, "eval_steps_per_second": 3.608, "step": 150 }, { "epoch": 0.010429227911221197, "grad_norm": 2.5995352268218994, "learning_rate": 0.00017881133923521971, "loss": 1.0265, "step": 160 }, { "epoch": 0.011081054655672521, "grad_norm": 2.5440680980682373, "learning_rate": 0.00017358958306132124, "loss": 0.9651, "step": 170 }, { "epoch": 0.011732881400123848, "grad_norm": 2.6204676628112793, "learning_rate": 0.00016802879343386844, "loss": 1.0322, "step": 180 }, { "epoch": 0.012384708144575172, "grad_norm": 2.2899601459503174, "learning_rate": 0.00016215606196095766, "loss": 1.1977, "step": 190 }, { "epoch": 0.013036534889026497, "grad_norm": 3.2127535343170166, "learning_rate": 0.000156, "loss": 1.6102, "step": 200 }, { "epoch": 0.013036534889026497, "eval_loss": 1.1086701154708862, "eval_runtime": 446.4333, "eval_samples_per_second": 14.47, "eval_steps_per_second": 3.618, "step": 200 }, { "epoch": 0.013688361633477822, "grad_norm": 1.3613508939743042, "learning_rate": 0.00014959059926606403, "loss": 0.838, "step": 210 }, { "epoch": 0.014340188377929147, "grad_norm": 1.6068763732910156, "learning_rate": 0.00014295908571525487, "loss": 0.9718, "step": 220 }, { "epoch": 0.014992015122380471, "grad_norm": 2.977311849594116, "learning_rate": 0.00013613776741499452, "loss": 1.0673, "step": 230 }, { "epoch": 0.015643841866831796, "grad_norm": 2.3362767696380615, "learning_rate": 0.00012915987714236542, "loss": 1.1417, "step": 240 }, { "epoch": 0.01629566861128312, "grad_norm": 5.354211807250977, "learning_rate": 0.00012205941047736077, "loss": 1.3597, "step": 250 }, { "epoch": 0.01629566861128312, "eval_loss": 1.0507456064224243, "eval_runtime": 446.5699, "eval_samples_per_second": 14.466, "eval_steps_per_second": 3.616, "step": 250 }, { "epoch": 0.016947495355734445, "grad_norm": 2.0789878368377686, "learning_rate": 0.00011487096017983597, "loss": 0.8857, "step": 260 }, { "epoch": 0.01759932210018577, "grad_norm": 1.4197766780853271, "learning_rate": 0.00010762954765706012, "loss": 0.903, "step": 270 }, { "epoch": 0.018251148844637095, "grad_norm": 2.561424732208252, "learning_rate": 0.00010037045234293992, "loss": 0.9548, "step": 280 }, { "epoch": 0.01890297558908842, "grad_norm": 1.98018217086792, "learning_rate": 9.312903982016405e-05, "loss": 1.1046, "step": 290 }, { "epoch": 0.019554802333539744, "grad_norm": 5.7000837326049805, "learning_rate": 8.594058952263925e-05, "loss": 1.3285, "step": 300 }, { "epoch": 0.019554802333539744, "eval_loss": 0.9808639287948608, "eval_runtime": 446.8663, "eval_samples_per_second": 14.456, "eval_steps_per_second": 3.614, "step": 300 }, { "epoch": 0.02020662907799107, "grad_norm": 1.3622760772705078, "learning_rate": 7.884012285763457e-05, "loss": 0.7478, "step": 310 }, { "epoch": 0.020858455822442393, "grad_norm": 1.826834797859192, "learning_rate": 7.186223258500548e-05, "loss": 0.8517, "step": 320 }, { "epoch": 0.021510282566893718, "grad_norm": 1.2498234510421753, "learning_rate": 6.504091428474514e-05, "loss": 1.0577, "step": 330 }, { "epoch": 0.022162109311345043, "grad_norm": 2.1019372940063477, "learning_rate": 5.840940073393593e-05, "loss": 0.8884, "step": 340 }, { "epoch": 0.02281393605579637, "grad_norm": 2.712697744369507, "learning_rate": 5.200000000000002e-05, "loss": 1.0915, "step": 350 }, { "epoch": 0.02281393605579637, "eval_loss": 0.9416553974151611, "eval_runtime": 449.3829, "eval_samples_per_second": 14.375, "eval_steps_per_second": 3.594, "step": 350 }, { "epoch": 0.023465762800247696, "grad_norm": 1.5010788440704346, "learning_rate": 4.5843938039042344e-05, "loss": 0.829, "step": 360 }, { "epoch": 0.02411758954469902, "grad_norm": 3.3226091861724854, "learning_rate": 3.997120656613154e-05, "loss": 0.9773, "step": 370 }, { "epoch": 0.024769416289150345, "grad_norm": 1.6384238004684448, "learning_rate": 3.441041693867878e-05, "loss": 0.8725, "step": 380 }, { "epoch": 0.02542124303360167, "grad_norm": 2.352992534637451, "learning_rate": 2.9188660764780296e-05, "loss": 0.9907, "step": 390 }, { "epoch": 0.026073069778052994, "grad_norm": 3.1694495677948, "learning_rate": 2.4331377915626298e-05, "loss": 1.2449, "step": 400 }, { "epoch": 0.026073069778052994, "eval_loss": 0.8955165147781372, "eval_runtime": 446.7108, "eval_samples_per_second": 14.461, "eval_steps_per_second": 3.615, "step": 400 }, { "epoch": 0.02672489652250432, "grad_norm": 2.0528738498687744, "learning_rate": 1.9862232585005475e-05, "loss": 0.7577, "step": 410 }, { "epoch": 0.027376723266955644, "grad_norm": 1.3070178031921387, "learning_rate": 1.58029979997317e-05, "loss": 0.7595, "step": 420 }, { "epoch": 0.02802855001140697, "grad_norm": 2.6197099685668945, "learning_rate": 1.2173450342671593e-05, "loss": 0.7412, "step": 430 }, { "epoch": 0.028680376755858293, "grad_norm": 2.77477765083313, "learning_rate": 8.991272405169498e-06, "loss": 0.9327, "step": 440 }, { "epoch": 0.029332203500309618, "grad_norm": 4.5697340965271, "learning_rate": 6.271967438265535e-06, "loss": 1.0734, "step": 450 }, { "epoch": 0.029332203500309618, "eval_loss": 0.876220166683197, "eval_runtime": 443.5926, "eval_samples_per_second": 14.563, "eval_steps_per_second": 3.641, "step": 450 }, { "epoch": 0.029984030244760942, "grad_norm": 2.2378225326538086, "learning_rate": 4.028783622414835e-06, "loss": 0.6365, "step": 460 }, { "epoch": 0.030635856989212267, "grad_norm": 2.819683074951172, "learning_rate": 2.272649523684208e-06, "loss": 0.8311, "step": 470 }, { "epoch": 0.03128768373366359, "grad_norm": 1.4101600646972656, "learning_rate": 1.0121208508766823e-06, "loss": 0.9517, "step": 480 }, { "epoch": 0.03193951047811492, "grad_norm": 1.9987350702285767, "learning_rate": 2.533387729782834e-07, "loss": 0.8854, "step": 490 }, { "epoch": 0.03259133722256624, "grad_norm": 5.414979934692383, "learning_rate": 0.0, "loss": 1.0317, "step": 500 }, { "epoch": 0.03259133722256624, "eval_loss": 0.875745415687561, "eval_runtime": 444.3276, "eval_samples_per_second": 14.539, "eval_steps_per_second": 3.635, "step": 500 } ], "logging_steps": 10, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0400815009234944e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }