eddysang commited on
Commit
39a1f64
·
verified ·
1 Parent(s): 69b7430

Training in progress, step 85, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f124e3c2fc8ec891656fbfca5ea9b6718202104b53c2f90f6f237b8efab7e7d6
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:376b17e9e9631961076e3bbeb3993d19795f20b76c2f2005817821e13465e6e1
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34aed8fc1414c28cb971615850dd2195d1ae883d038da558c27b755343438bb1
3
  size 671466706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3727032b69e6649af9f27beaabc3fe9ad69040d535a04f1eb340b53219c80bb
3
  size 671466706
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4c0dfbada95d07c631cece0f7fdca52ecc156135c62bc34b4e8b1a466156bc4
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:427ad7b17c78789ef97429efbe9d6211a7d09f2d08a147aa825d611691ab1bf8
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f4e77fd2a3bb3f08929494d77da2f57f8781f91a45852bcf8f71a5777dd088c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b383ad1d61ff4e9bbd86bd276c043e414782d2bb7de68ada3e289a786eb79681
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.10011502185415229,
5
  "eval_steps": 50,
6
- "global_step": 68,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -499,6 +499,125 @@
499
  "learning_rate": 0.00012518479547691435,
500
  "loss": 0.0903,
501
  "step": 68
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
502
  }
503
  ],
504
  "logging_steps": 1,
@@ -518,7 +637,7 @@
518
  "attributes": {}
519
  }
520
  },
521
- "total_flos": 7.176620549823529e+17,
522
  "train_batch_size": 2,
523
  "trial_name": null,
524
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.12514377731769036,
5
  "eval_steps": 50,
6
+ "global_step": 85,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
499
  "learning_rate": 0.00012518479547691435,
500
  "loss": 0.0903,
501
  "step": 68
502
+ },
503
+ {
504
+ "epoch": 0.10158730158730159,
505
+ "grad_norm": 0.1023801937699318,
506
+ "learning_rate": 0.00012420442717428804,
507
+ "loss": 0.0845,
508
+ "step": 69
509
+ },
510
+ {
511
+ "epoch": 0.10305958132045089,
512
+ "grad_norm": 0.10249310731887817,
513
+ "learning_rate": 0.00012320907072649044,
514
+ "loss": 0.0539,
515
+ "step": 70
516
+ },
517
+ {
518
+ "epoch": 0.10453186105360018,
519
+ "grad_norm": 0.10111914575099945,
520
+ "learning_rate": 0.0001221990293287378,
521
+ "loss": 0.0424,
522
+ "step": 71
523
+ },
524
+ {
525
+ "epoch": 0.10600414078674948,
526
+ "grad_norm": 0.16136892139911652,
527
+ "learning_rate": 0.00012117461064942435,
528
+ "loss": 0.1277,
529
+ "step": 72
530
+ },
531
+ {
532
+ "epoch": 0.10747642051989878,
533
+ "grad_norm": 0.13022761046886444,
534
+ "learning_rate": 0.00012013612673640363,
535
+ "loss": 0.118,
536
+ "step": 73
537
+ },
538
+ {
539
+ "epoch": 0.10894870025304808,
540
+ "grad_norm": 0.10115568339824677,
541
+ "learning_rate": 0.00011908389392193547,
542
+ "loss": 0.0554,
543
+ "step": 74
544
+ },
545
+ {
546
+ "epoch": 0.11042097998619738,
547
+ "grad_norm": 0.1352306455373764,
548
+ "learning_rate": 0.00011801823272632844,
549
+ "loss": 0.0683,
550
+ "step": 75
551
+ },
552
+ {
553
+ "epoch": 0.11189325971934667,
554
+ "grad_norm": 0.11654029786586761,
555
+ "learning_rate": 0.00011693946776030599,
556
+ "loss": 0.0656,
557
+ "step": 76
558
+ },
559
+ {
560
+ "epoch": 0.11336553945249597,
561
+ "grad_norm": 0.1405310332775116,
562
+ "learning_rate": 0.00011584792762612703,
563
+ "loss": 0.0681,
564
+ "step": 77
565
+ },
566
+ {
567
+ "epoch": 0.11483781918564527,
568
+ "grad_norm": 0.19620081782341003,
569
+ "learning_rate": 0.00011474394481749035,
570
+ "loss": 0.1183,
571
+ "step": 78
572
+ },
573
+ {
574
+ "epoch": 0.11631009891879457,
575
+ "grad_norm": 0.09413562715053558,
576
+ "learning_rate": 0.00011362785561825406,
577
+ "loss": 0.0377,
578
+ "step": 79
579
+ },
580
+ {
581
+ "epoch": 0.11778237865194387,
582
+ "grad_norm": 0.10567747801542282,
583
+ "learning_rate": 0.0001125,
584
+ "loss": 0.0776,
585
+ "step": 80
586
+ },
587
+ {
588
+ "epoch": 0.11925465838509317,
589
+ "grad_norm": 0.15690375864505768,
590
+ "learning_rate": 0.00011136072151847529,
591
+ "loss": 0.0366,
592
+ "step": 81
593
+ },
594
+ {
595
+ "epoch": 0.12072693811824246,
596
+ "grad_norm": 0.09006724506616592,
597
+ "learning_rate": 0.00011021036720894179,
598
+ "loss": 0.0319,
599
+ "step": 82
600
+ },
601
+ {
602
+ "epoch": 0.12219921785139176,
603
+ "grad_norm": 0.1135464459657669,
604
+ "learning_rate": 0.00010904928748046599,
605
+ "loss": 0.0482,
606
+ "step": 83
607
+ },
608
+ {
609
+ "epoch": 0.12367149758454106,
610
+ "grad_norm": 0.09596288949251175,
611
+ "learning_rate": 0.0001078778360091808,
612
+ "loss": 0.0454,
613
+ "step": 84
614
+ },
615
+ {
616
+ "epoch": 0.12514377731769036,
617
+ "grad_norm": 0.17637494206428528,
618
+ "learning_rate": 0.00010669636963055245,
619
+ "loss": 0.1116,
620
+ "step": 85
621
  }
622
  ],
623
  "logging_steps": 1,
 
637
  "attributes": {}
638
  }
639
  },
640
+ "total_flos": 8.995021026974761e+17,
641
  "train_batch_size": 2,
642
  "trial_name": null,
643
  "trial_params": null