rak-r05 commited on
Commit
298667d
·
verified ·
1 Parent(s): f0e5db3

Training in progress, step 114, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5184a9f855392b369617a571c4fe8c407c7dbd8eba55d46eea4846faa6eef1da
3
  size 598799664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe3df2f7dc7783032ccc2f13267df9276dc009f2004bc181b5608456e4475d0
3
  size 598799664
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:123b917ce2349c1a74d8844eac29f8c6cf2dc99b9a864c096be4ff65462b6328
3
  size 42898516
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf68138b10f1cb89c7a858ed47f9e3aaaeccfa47ef6b36311059258c91f72bdc
3
  size 42898516
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51a053afc2b69b6892ee05d33f3fb3716eac75fd7fbab8c36a6324da423adb8d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b72207d94b98ed69df38c6423929558151487763a83e0e8622b137fe48d72e4
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc8e5f98da37dd5d6e76ad1dcd9560295d5ddc12e8c320cbe967800911c3d6c0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d35a71cd96ab153cf58aa68810dc4405b5e1a7fbfef5501d233ae84e34aa51c4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0708955223880597,
5
  "eval_steps": 38,
6
- "global_step": 76,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -563,6 +563,280 @@
563
  "eval_samples_per_second": 7.506,
564
  "eval_steps_per_second": 3.753,
565
  "step": 76
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
566
  }
567
  ],
568
  "logging_steps": 1,
@@ -582,7 +856,7 @@
582
  "attributes": {}
583
  }
584
  },
585
- "total_flos": 9963999823134720.0,
586
  "train_batch_size": 2,
587
  "trial_name": null,
588
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.10634328358208955,
5
  "eval_steps": 38,
6
+ "global_step": 114,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
563
  "eval_samples_per_second": 7.506,
564
  "eval_steps_per_second": 3.753,
565
  "step": 76
566
+ },
567
+ {
568
+ "epoch": 0.07182835820895522,
569
+ "grad_norm": 1.2228177785873413,
570
+ "learning_rate": 0.00021345380112240797,
571
+ "loss": 2.1149,
572
+ "step": 77
573
+ },
574
+ {
575
+ "epoch": 0.07276119402985075,
576
+ "grad_norm": 1.0541841983795166,
577
+ "learning_rate": 0.00020897296607010301,
578
+ "loss": 1.6306,
579
+ "step": 78
580
+ },
581
+ {
582
+ "epoch": 0.07369402985074627,
583
+ "grad_norm": 1.1272865533828735,
584
+ "learning_rate": 0.00020448761285916104,
585
+ "loss": 1.9262,
586
+ "step": 79
587
+ },
588
+ {
589
+ "epoch": 0.07462686567164178,
590
+ "grad_norm": 1.1912484169006348,
591
+ "learning_rate": 0.0002,
592
+ "loss": 2.0926,
593
+ "step": 80
594
+ },
595
+ {
596
+ "epoch": 0.07555970149253731,
597
+ "grad_norm": 1.2342007160186768,
598
+ "learning_rate": 0.00019551238714083903,
599
+ "loss": 2.3557,
600
+ "step": 81
601
+ },
602
+ {
603
+ "epoch": 0.07649253731343283,
604
+ "grad_norm": 1.068264365196228,
605
+ "learning_rate": 0.00019102703392989709,
606
+ "loss": 1.9768,
607
+ "step": 82
608
+ },
609
+ {
610
+ "epoch": 0.07742537313432836,
611
+ "grad_norm": 1.035170555114746,
612
+ "learning_rate": 0.00018654619887759207,
613
+ "loss": 1.6881,
614
+ "step": 83
615
+ },
616
+ {
617
+ "epoch": 0.07835820895522388,
618
+ "grad_norm": 1.2881474494934082,
619
+ "learning_rate": 0.00018207213821931333,
620
+ "loss": 2.1289,
621
+ "step": 84
622
+ },
623
+ {
624
+ "epoch": 0.07929104477611941,
625
+ "grad_norm": 1.1358542442321777,
626
+ "learning_rate": 0.00017760710477933845,
627
+ "loss": 2.2829,
628
+ "step": 85
629
+ },
630
+ {
631
+ "epoch": 0.08022388059701492,
632
+ "grad_norm": 1.2520966529846191,
633
+ "learning_rate": 0.00017315334683646897,
634
+ "loss": 1.552,
635
+ "step": 86
636
+ },
637
+ {
638
+ "epoch": 0.08115671641791045,
639
+ "grad_norm": 0.9834119081497192,
640
+ "learning_rate": 0.00016871310699195379,
641
+ "loss": 1.5413,
642
+ "step": 87
643
+ },
644
+ {
645
+ "epoch": 0.08208955223880597,
646
+ "grad_norm": 1.043672800064087,
647
+ "learning_rate": 0.00016428862104027268,
648
+ "loss": 1.7337,
649
+ "step": 88
650
+ },
651
+ {
652
+ "epoch": 0.0830223880597015,
653
+ "grad_norm": 1.0788074731826782,
654
+ "learning_rate": 0.00015988211684334546,
655
+ "loss": 1.3079,
656
+ "step": 89
657
+ },
658
+ {
659
+ "epoch": 0.08395522388059702,
660
+ "grad_norm": 1.0924961566925049,
661
+ "learning_rate": 0.00015549581320873715,
662
+ "loss": 1.7402,
663
+ "step": 90
664
+ },
665
+ {
666
+ "epoch": 0.08488805970149253,
667
+ "grad_norm": 1.1585627794265747,
668
+ "learning_rate": 0.00015113191877242117,
669
+ "loss": 1.7314,
670
+ "step": 91
671
+ },
672
+ {
673
+ "epoch": 0.08582089552238806,
674
+ "grad_norm": 1.0373111963272095,
675
+ "learning_rate": 0.00014679263088666499,
676
+ "loss": 1.7103,
677
+ "step": 92
678
+ },
679
+ {
680
+ "epoch": 0.08675373134328358,
681
+ "grad_norm": 1.178009271621704,
682
+ "learning_rate": 0.00014248013451359656,
683
+ "loss": 1.8889,
684
+ "step": 93
685
+ },
686
+ {
687
+ "epoch": 0.08768656716417911,
688
+ "grad_norm": 1.110259771347046,
689
+ "learning_rate": 0.00013819660112501054,
690
+ "loss": 1.685,
691
+ "step": 94
692
+ },
693
+ {
694
+ "epoch": 0.08861940298507463,
695
+ "grad_norm": 1.0290873050689697,
696
+ "learning_rate": 0.00013394418760896666,
697
+ "loss": 1.6533,
698
+ "step": 95
699
+ },
700
+ {
701
+ "epoch": 0.08955223880597014,
702
+ "grad_norm": 1.0740941762924194,
703
+ "learning_rate": 0.00012972503518373144,
704
+ "loss": 1.8436,
705
+ "step": 96
706
+ },
707
+ {
708
+ "epoch": 0.09048507462686567,
709
+ "grad_norm": 1.1851738691329956,
710
+ "learning_rate": 0.00012554126831961098,
711
+ "loss": 1.8081,
712
+ "step": 97
713
+ },
714
+ {
715
+ "epoch": 0.0914179104477612,
716
+ "grad_norm": 1.0965279340744019,
717
+ "learning_rate": 0.0001213949936692153,
718
+ "loss": 1.8133,
719
+ "step": 98
720
+ },
721
+ {
722
+ "epoch": 0.09235074626865672,
723
+ "grad_norm": 1.0990545749664307,
724
+ "learning_rate": 0.00011728829900669591,
725
+ "loss": 1.7193,
726
+ "step": 99
727
+ },
728
+ {
729
+ "epoch": 0.09328358208955224,
730
+ "grad_norm": 1.0958396196365356,
731
+ "learning_rate": 0.00011322325217648839,
732
+ "loss": 1.3802,
733
+ "step": 100
734
+ },
735
+ {
736
+ "epoch": 0.09421641791044776,
737
+ "grad_norm": 1.0536209344863892,
738
+ "learning_rate": 0.00010920190005209065,
739
+ "loss": 1.8329,
740
+ "step": 101
741
+ },
742
+ {
743
+ "epoch": 0.09514925373134328,
744
+ "grad_norm": 1.0651001930236816,
745
+ "learning_rate": 0.00010522626750540028,
746
+ "loss": 1.6494,
747
+ "step": 102
748
+ },
749
+ {
750
+ "epoch": 0.0960820895522388,
751
+ "grad_norm": 1.1998820304870605,
752
+ "learning_rate": 0.00010129835638713063,
753
+ "loss": 2.0276,
754
+ "step": 103
755
+ },
756
+ {
757
+ "epoch": 0.09701492537313433,
758
+ "grad_norm": 0.9894289374351501,
759
+ "learning_rate": 9.74201445188188e-05,
760
+ "loss": 1.5171,
761
+ "step": 104
762
+ },
763
+ {
764
+ "epoch": 0.09794776119402986,
765
+ "grad_norm": 1.2795695066452026,
766
+ "learning_rate": 9.359358469693271e-05,
767
+ "loss": 1.7987,
768
+ "step": 105
769
+ },
770
+ {
771
+ "epoch": 0.09888059701492537,
772
+ "grad_norm": 1.0911823511123657,
773
+ "learning_rate": 8.982060370957952e-05,
774
+ "loss": 1.6949,
775
+ "step": 106
776
+ },
777
+ {
778
+ "epoch": 0.09981343283582089,
779
+ "grad_norm": 0.9669733047485352,
780
+ "learning_rate": 8.610310136630962e-05,
781
+ "loss": 1.405,
782
+ "step": 107
783
+ },
784
+ {
785
+ "epoch": 0.10074626865671642,
786
+ "grad_norm": 1.1521021127700806,
787
+ "learning_rate": 8.24429495415054e-05,
788
+ "loss": 1.8475,
789
+ "step": 108
790
+ },
791
+ {
792
+ "epoch": 0.10167910447761194,
793
+ "grad_norm": 1.085671305656433,
794
+ "learning_rate": 7.884199123183605e-05,
795
+ "loss": 1.5308,
796
+ "step": 109
797
+ },
798
+ {
799
+ "epoch": 0.10261194029850747,
800
+ "grad_norm": 1.141276240348816,
801
+ "learning_rate": 7.530203962825331e-05,
802
+ "loss": 1.8675,
803
+ "step": 110
804
+ },
805
+ {
806
+ "epoch": 0.10354477611940298,
807
+ "grad_norm": 1.0789172649383545,
808
+ "learning_rate": 7.182487720299517e-05,
809
+ "loss": 1.7724,
810
+ "step": 111
811
+ },
812
+ {
813
+ "epoch": 0.1044776119402985,
814
+ "grad_norm": 1.0939713716506958,
815
+ "learning_rate": 6.841225481205749e-05,
816
+ "loss": 1.5574,
817
+ "step": 112
818
+ },
819
+ {
820
+ "epoch": 0.10541044776119403,
821
+ "grad_norm": 1.2875022888183594,
822
+ "learning_rate": 6.506589081358514e-05,
823
+ "loss": 1.7821,
824
+ "step": 113
825
+ },
826
+ {
827
+ "epoch": 0.10634328358208955,
828
+ "grad_norm": 1.1060545444488525,
829
+ "learning_rate": 6.178747020262707e-05,
830
+ "loss": 1.543,
831
+ "step": 114
832
+ },
833
+ {
834
+ "epoch": 0.10634328358208955,
835
+ "eval_loss": 0.44875675439834595,
836
+ "eval_runtime": 61.3854,
837
+ "eval_samples_per_second": 7.363,
838
+ "eval_steps_per_second": 3.682,
839
+ "step": 114
840
  }
841
  ],
842
  "logging_steps": 1,
 
856
  "attributes": {}
857
  }
858
  },
859
+ "total_flos": 1.494599973470208e+16,
860
  "train_batch_size": 2,
861
  "trial_name": null,
862
  "trial_params": null