eddysang commited on
Commit
c9fcc1f
·
verified ·
1 Parent(s): f4a2c3d

Training in progress, step 90, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8338d0371a035aa1a6e38b77dff221fcf9ba83d3e29f4789534a13eabc6e82f8
3
  size 319876032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c582e34b740c2ab84777b72c509a0e97c269aac69112da8e8698713a54feeae8
3
  size 319876032
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6264c706641e81281ccc8cdc92c1bd7506e0ccf8ae3a5ca2071df4683ad54811
3
  size 640009682
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5ccefb6bcb213e033c6d906dfab3f18aaf2958fcc017bb013b20f4d401e0fe3
3
  size 640009682
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ad7d53f2bd686d5b08e71c45b51b2fc986926916fcd4884c9b343639657183c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5134bfc96a8a009684ec48127ff4fc03b2f2e61a0fbe126a34bef4ab7025231
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09ab64d5ace66796aa9c2fa2fc4e0206d69a1eb4ef03f574f85ee8eb16a64b71
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:429c9f0a4ff1909f8cc1cafc3bf6eb6ff29e4592139b5344f1aee9b75eae7379
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.74310692669805,
5
  "eval_steps": 50,
6
- "global_step": 81,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -590,6 +590,69 @@
590
  "learning_rate": 1.99258117923236e-05,
591
  "loss": 15.7491,
592
  "step": 81
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
593
  }
594
  ],
595
  "logging_steps": 1,
@@ -609,7 +672,7 @@
609
  "attributes": {}
610
  }
611
  },
612
- "total_flos": 4.259916615505674e+17,
613
  "train_batch_size": 2,
614
  "trial_name": null,
615
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.936785474108944,
5
  "eval_steps": 50,
6
+ "global_step": 90,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
590
  "learning_rate": 1.99258117923236e-05,
591
  "loss": 15.7491,
592
  "step": 81
593
+ },
594
+ {
595
+ "epoch": 1.7646267652992602,
596
+ "grad_norm": 18.65260887145996,
597
+ "learning_rate": 1.796955257999768e-05,
598
+ "loss": 18.4079,
599
+ "step": 82
600
+ },
601
+ {
602
+ "epoch": 1.7861466039004708,
603
+ "grad_norm": 20.124731063842773,
604
+ "learning_rate": 1.6101230183944144e-05,
605
+ "loss": 21.8745,
606
+ "step": 83
607
+ },
608
+ {
609
+ "epoch": 1.8076664425016813,
610
+ "grad_norm": 15.53337287902832,
611
+ "learning_rate": 1.4323725421878949e-05,
612
+ "loss": 14.7667,
613
+ "step": 84
614
+ },
615
+ {
616
+ "epoch": 1.8291862811028916,
617
+ "grad_norm": 24.60688018798828,
618
+ "learning_rate": 1.2639779077309098e-05,
619
+ "loss": 23.2765,
620
+ "step": 85
621
+ },
622
+ {
623
+ "epoch": 1.8507061197041024,
624
+ "grad_norm": 18.01706886291504,
625
+ "learning_rate": 1.1051987673443085e-05,
626
+ "loss": 19.2231,
627
+ "step": 86
628
+ },
629
+ {
630
+ "epoch": 1.8722259583053127,
631
+ "grad_norm": 17.668659210205078,
632
+ "learning_rate": 9.56279946954021e-06,
633
+ "loss": 23.5257,
634
+ "step": 87
635
+ },
636
+ {
637
+ "epoch": 1.8937457969065232,
638
+ "grad_norm": 20.853605270385742,
639
+ "learning_rate": 8.174510685872415e-06,
640
+ "loss": 22.4145,
641
+ "step": 88
642
+ },
643
+ {
644
+ "epoch": 1.9152656355077338,
645
+ "grad_norm": 20.111770629882812,
646
+ "learning_rate": 6.889261963118898e-06,
647
+ "loss": 18.1747,
648
+ "step": 89
649
+ },
650
+ {
651
+ "epoch": 1.936785474108944,
652
+ "grad_norm": 19.906658172607422,
653
+ "learning_rate": 5.709035061653494e-06,
654
+ "loss": 23.4895,
655
+ "step": 90
656
  }
657
  ],
658
  "logging_steps": 1,
 
672
  "attributes": {}
673
  }
674
  },
675
+ "total_flos": 4.7332406838951936e+17,
676
  "train_batch_size": 2,
677
  "trial_name": null,
678
  "trial_params": null