cite-text-analysis's picture
Upload folder using huggingface_hub
a3e89c7 verified
raw
history blame
1.39 kB
{
"best_metric": 0.5389063030773956,
"best_model_checkpoint": "case-analysis-roberta-base/checkpoint-224",
"epoch": 1.0,
"eval_steps": 500,
"global_step": 224,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.0,
"eval_accuracy": 0.7527839643652561,
"eval_f1_macro": 0.5389063030773956,
"eval_f1_micro": 0.7527839643652561,
"eval_f1_weighted": 0.7300520074128541,
"eval_loss": 0.8771109580993652,
"eval_macro_fpr": 0.13144039552419665,
"eval_macro_sensitivity": 0.5401895401895401,
"eval_macro_specificity": 0.8833499134100214,
"eval_precision": 0.7119932520850578,
"eval_precision_macro": 0.5403637213518124,
"eval_recall": 0.7527839643652561,
"eval_recall_macro": 0.5401895401895401,
"eval_runtime": 6.9264,
"eval_samples_per_second": 64.825,
"eval_steps_per_second": 8.229,
"eval_weighted_fpr": 0.09866666666666667,
"eval_weighted_sensitivity": 0.7527839643652561,
"eval_weighted_specificity": 0.7806156892748298,
"step": 224
}
],
"logging_steps": 500,
"max_steps": 6720,
"num_input_tokens_seen": 0,
"num_train_epochs": 30,
"save_steps": 500,
"total_flos": 471503477932032.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}