|
{ |
|
"best_metric": 0.7650073170661926, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-25", |
|
"epoch": 1.8790697674418606, |
|
"eval_steps": 25, |
|
"global_step": 25, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07441860465116279, |
|
"grad_norm": 26.82486915588379, |
|
"learning_rate": 5e-05, |
|
"loss": 15.5404, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.07441860465116279, |
|
"eval_loss": 1.1030958890914917, |
|
"eval_runtime": 4.0595, |
|
"eval_samples_per_second": 12.317, |
|
"eval_steps_per_second": 3.202, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.14883720930232558, |
|
"grad_norm": 32.55805969238281, |
|
"learning_rate": 0.0001, |
|
"loss": 17.9188, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.22325581395348837, |
|
"grad_norm": 23.376680374145508, |
|
"learning_rate": 9.985407886603945e-05, |
|
"loss": 17.5531, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.29767441860465116, |
|
"grad_norm": 15.263604164123535, |
|
"learning_rate": 9.941726181870608e-05, |
|
"loss": 14.8715, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.37209302325581395, |
|
"grad_norm": 13.026528358459473, |
|
"learning_rate": 9.869238178417235e-05, |
|
"loss": 14.1927, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.44651162790697674, |
|
"grad_norm": 11.496667861938477, |
|
"learning_rate": 9.768413988762156e-05, |
|
"loss": 14.7444, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.5209302325581395, |
|
"grad_norm": 8.687058448791504, |
|
"learning_rate": 9.639907496464709e-05, |
|
"loss": 13.8784, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.5953488372093023, |
|
"grad_norm": 7.8401970863342285, |
|
"learning_rate": 9.484552115439445e-05, |
|
"loss": 13.6697, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.6697674418604651, |
|
"grad_norm": 7.2838263511657715, |
|
"learning_rate": 9.303355384947076e-05, |
|
"loss": 13.5577, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.7441860465116279, |
|
"grad_norm": 7.735969066619873, |
|
"learning_rate": 9.097492435315756e-05, |
|
"loss": 12.9802, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.8186046511627907, |
|
"grad_norm": 6.482202053070068, |
|
"learning_rate": 8.868298366769954e-05, |
|
"loss": 13.1186, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.8930232558139535, |
|
"grad_norm": 6.737044811248779, |
|
"learning_rate": 8.617259590793198e-05, |
|
"loss": 13.2019, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.9674418604651163, |
|
"grad_norm": 8.719466209411621, |
|
"learning_rate": 8.346004190179198e-05, |
|
"loss": 13.6937, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 1.0604651162790697, |
|
"grad_norm": 5.885293483734131, |
|
"learning_rate": 8.056291360290201e-05, |
|
"loss": 11.8573, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 1.1348837209302325, |
|
"grad_norm": 7.148478031158447, |
|
"learning_rate": 7.75e-05, |
|
"loss": 12.4117, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 1.2093023255813953, |
|
"grad_norm": 7.730895519256592, |
|
"learning_rate": 7.429116526313744e-05, |
|
"loss": 12.0587, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 1.283720930232558, |
|
"grad_norm": 7.085984230041504, |
|
"learning_rate": 7.095721991691411e-05, |
|
"loss": 11.7461, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 1.3581395348837209, |
|
"grad_norm": 6.469642639160156, |
|
"learning_rate": 6.751978587624037e-05, |
|
"loss": 11.8886, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.4325581395348836, |
|
"grad_norm": 6.971822738647461, |
|
"learning_rate": 6.400115621992201e-05, |
|
"loss": 11.9513, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 1.5069767441860464, |
|
"grad_norm": 6.892801284790039, |
|
"learning_rate": 6.042415061148954e-05, |
|
"loss": 11.5344, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 1.5813953488372094, |
|
"grad_norm": 5.958037853240967, |
|
"learning_rate": 5.681196730492368e-05, |
|
"loss": 11.6245, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 1.655813953488372, |
|
"grad_norm": 6.668700218200684, |
|
"learning_rate": 5.318803269507634e-05, |
|
"loss": 11.5543, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 1.730232558139535, |
|
"grad_norm": 7.723108291625977, |
|
"learning_rate": 4.9575849388510473e-05, |
|
"loss": 11.489, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 1.8046511627906976, |
|
"grad_norm": 5.082211494445801, |
|
"learning_rate": 4.599884378007802e-05, |
|
"loss": 11.2579, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 1.8790697674418606, |
|
"grad_norm": 6.550709247589111, |
|
"learning_rate": 4.248021412375963e-05, |
|
"loss": 11.8115, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 1.8790697674418606, |
|
"eval_loss": 0.7650073170661926, |
|
"eval_runtime": 3.3463, |
|
"eval_samples_per_second": 14.942, |
|
"eval_steps_per_second": 3.885, |
|
"step": 25 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 41, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 1, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.829013890367488e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|