eddysang commited on
Commit
5db24a7
·
verified ·
1 Parent(s): ad303a1

Training in progress, step 85, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e86ee6e154e49069091d893455757e17295ee85019f64a663e52903b19cfc1cf
3
  size 97728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c93154ec6a8dec35705ce9d1c2f4a5c55f118f8418e41f2d63409d33becad085
3
  size 97728
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e1a75388d3fb482b018a16ca66ce2bf8d2b482576fc3daf3f841c59f4c236cb
3
  size 212298
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f4de1512f949d496d512471d0c604d3dad5cd6125348a1d0f27f1b0c5a351ea
3
  size 212298
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:051417fc01d33b78aa641f032182ab64da1fcc15fe5a876899f334dcea2dc357
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdd3ea04947084511a30ef65c56c4e673088b8f8e3ac5e88f4ddcd31448dda70
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f4e77fd2a3bb3f08929494d77da2f57f8781f91a45852bcf8f71a5777dd088c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b383ad1d61ff4e9bbd86bd276c043e414782d2bb7de68ada3e289a786eb79681
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.23038644785600848,
5
  "eval_steps": 50,
6
- "global_step": 68,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -499,6 +499,125 @@
499
  "learning_rate": 0.00012518479547691435,
500
  "loss": 10.3485,
501
  "step": 68
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
502
  }
503
  ],
504
  "logging_steps": 1,
@@ -518,7 +637,7 @@
518
  "attributes": {}
519
  }
520
  },
521
- "total_flos": 29122375581696.0,
522
  "train_batch_size": 2,
523
  "trial_name": null,
524
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.2879830598200106,
5
  "eval_steps": 50,
6
+ "global_step": 85,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
499
  "learning_rate": 0.00012518479547691435,
500
  "loss": 10.3485,
501
  "step": 68
502
+ },
503
+ {
504
+ "epoch": 0.23377448385389094,
505
+ "grad_norm": 0.028004931285977364,
506
+ "learning_rate": 0.00012420442717428804,
507
+ "loss": 10.3488,
508
+ "step": 69
509
+ },
510
+ {
511
+ "epoch": 0.23716251985177342,
512
+ "grad_norm": 0.030328616499900818,
513
+ "learning_rate": 0.00012320907072649044,
514
+ "loss": 10.3479,
515
+ "step": 70
516
+ },
517
+ {
518
+ "epoch": 0.2405505558496559,
519
+ "grad_norm": 0.027720727026462555,
520
+ "learning_rate": 0.0001221990293287378,
521
+ "loss": 10.3477,
522
+ "step": 71
523
+ },
524
+ {
525
+ "epoch": 0.24393859184753838,
526
+ "grad_norm": 0.02903708443045616,
527
+ "learning_rate": 0.00012117461064942435,
528
+ "loss": 10.3479,
529
+ "step": 72
530
+ },
531
+ {
532
+ "epoch": 0.24732662784542087,
533
+ "grad_norm": 0.026270205155014992,
534
+ "learning_rate": 0.00012013612673640363,
535
+ "loss": 10.3473,
536
+ "step": 73
537
+ },
538
+ {
539
+ "epoch": 0.2507146638433033,
540
+ "grad_norm": 0.022023234516382217,
541
+ "learning_rate": 0.00011908389392193547,
542
+ "loss": 10.3462,
543
+ "step": 74
544
+ },
545
+ {
546
+ "epoch": 0.25410269984118583,
547
+ "grad_norm": 0.018086234107613564,
548
+ "learning_rate": 0.00011801823272632844,
549
+ "loss": 10.3471,
550
+ "step": 75
551
+ },
552
+ {
553
+ "epoch": 0.2574907358390683,
554
+ "grad_norm": 0.017115121707320213,
555
+ "learning_rate": 0.00011693946776030599,
556
+ "loss": 10.3481,
557
+ "step": 76
558
+ },
559
+ {
560
+ "epoch": 0.26087877183695074,
561
+ "grad_norm": 0.022446399554610252,
562
+ "learning_rate": 0.00011584792762612703,
563
+ "loss": 10.3469,
564
+ "step": 77
565
+ },
566
+ {
567
+ "epoch": 0.26426680783483325,
568
+ "grad_norm": 0.023726455867290497,
569
+ "learning_rate": 0.00011474394481749035,
570
+ "loss": 10.3475,
571
+ "step": 78
572
+ },
573
+ {
574
+ "epoch": 0.2676548438327157,
575
+ "grad_norm": 0.02203121967613697,
576
+ "learning_rate": 0.00011362785561825406,
577
+ "loss": 10.3484,
578
+ "step": 79
579
+ },
580
+ {
581
+ "epoch": 0.2710428798305982,
582
+ "grad_norm": 0.016930948942899704,
583
+ "learning_rate": 0.0001125,
584
+ "loss": 10.3476,
585
+ "step": 80
586
+ },
587
+ {
588
+ "epoch": 0.27443091582848067,
589
+ "grad_norm": 0.017117898911237717,
590
+ "learning_rate": 0.00011136072151847529,
591
+ "loss": 10.3474,
592
+ "step": 81
593
+ },
594
+ {
595
+ "epoch": 0.2778189518263632,
596
+ "grad_norm": 0.02053241617977619,
597
+ "learning_rate": 0.00011021036720894179,
598
+ "loss": 10.3467,
599
+ "step": 82
600
+ },
601
+ {
602
+ "epoch": 0.28120698782424564,
603
+ "grad_norm": 0.022768596187233925,
604
+ "learning_rate": 0.00010904928748046599,
605
+ "loss": 10.3479,
606
+ "step": 83
607
+ },
608
+ {
609
+ "epoch": 0.2845950238221281,
610
+ "grad_norm": 0.016698352992534637,
611
+ "learning_rate": 0.0001078778360091808,
612
+ "loss": 10.3468,
613
+ "step": 84
614
+ },
615
+ {
616
+ "epoch": 0.2879830598200106,
617
+ "grad_norm": 0.020199725404381752,
618
+ "learning_rate": 0.00010669636963055245,
619
+ "loss": 10.3473,
620
+ "step": 85
621
  }
622
  ],
623
  "logging_steps": 1,
 
637
  "attributes": {}
638
  }
639
  },
640
+ "total_flos": 36406310731776.0,
641
  "train_batch_size": 2,
642
  "trial_name": null,
643
  "trial_params": null