eddysang commited on
Commit
2de65e0
·
verified ·
1 Parent(s): f636840

Training in progress, step 68, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2690da81d83e89e9ee7b7dc7113de883b592b46176c59f118aaff5ad1b1c42cc
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f124e3c2fc8ec891656fbfca5ea9b6718202104b53c2f90f6f237b8efab7e7d6
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:580ef17b80875f7190859d2da30842b5c7fe663cfb653f5c46ee771354997856
3
  size 671466706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34aed8fc1414c28cb971615850dd2195d1ae883d038da558c27b755343438bb1
3
  size 671466706
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02f54eb84b382db409dccc97cebf46240cf5b6285939222273bf59b8c0558286
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4c0dfbada95d07c631cece0f7fdca52ecc156135c62bc34b4e8b1a466156bc4
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c86702c0a3caad6c51746e54805a7289de03dff9cc5abc148a58966cf1f4d339
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f4e77fd2a3bb3f08929494d77da2f57f8781f91a45852bcf8f71a5777dd088c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.07508626639061422,
5
  "eval_steps": 50,
6
- "global_step": 51,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -380,6 +380,125 @@
380
  "learning_rate": 0.00013928754755265842,
381
  "loss": 0.1427,
382
  "step": 51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
383
  }
384
  ],
385
  "logging_steps": 1,
@@ -399,7 +518,7 @@
399
  "attributes": {}
400
  }
401
  },
402
- "total_flos": 5.401115673671762e+17,
403
  "train_batch_size": 2,
404
  "trial_name": null,
405
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.10011502185415229,
5
  "eval_steps": 50,
6
+ "global_step": 68,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
380
  "learning_rate": 0.00013928754755265842,
381
  "loss": 0.1427,
382
  "step": 51
383
+ },
384
+ {
385
+ "epoch": 0.07655854612376352,
386
+ "grad_norm": 0.12584620714187622,
387
+ "learning_rate": 0.00013860360721173193,
388
+ "loss": 0.0863,
389
+ "step": 52
390
+ },
391
+ {
392
+ "epoch": 0.07803082585691282,
393
+ "grad_norm": 0.11394777148962021,
394
+ "learning_rate": 0.0001379002925959068,
395
+ "loss": 0.0691,
396
+ "step": 53
397
+ },
398
+ {
399
+ "epoch": 0.07950310559006211,
400
+ "grad_norm": 0.10938312113285065,
401
+ "learning_rate": 0.0001371778179416281,
402
+ "loss": 0.0826,
403
+ "step": 54
404
+ },
405
+ {
406
+ "epoch": 0.08097538532321141,
407
+ "grad_norm": 0.12558647990226746,
408
+ "learning_rate": 0.00013643640332167438,
409
+ "loss": 0.0766,
410
+ "step": 55
411
+ },
412
+ {
413
+ "epoch": 0.08244766505636071,
414
+ "grad_norm": 0.10897130519151688,
415
+ "learning_rate": 0.00013567627457812106,
416
+ "loss": 0.062,
417
+ "step": 56
418
+ },
419
+ {
420
+ "epoch": 0.08391994478951001,
421
+ "grad_norm": 0.09030541032552719,
422
+ "learning_rate": 0.00013489766325354695,
423
+ "loss": 0.0393,
424
+ "step": 57
425
+ },
426
+ {
427
+ "epoch": 0.08539222452265931,
428
+ "grad_norm": 0.1143849790096283,
429
+ "learning_rate": 0.00013410080652050412,
430
+ "loss": 0.1219,
431
+ "step": 58
432
+ },
433
+ {
434
+ "epoch": 0.0868645042558086,
435
+ "grad_norm": 0.1283547431230545,
436
+ "learning_rate": 0.0001332859471092728,
437
+ "loss": 0.1195,
438
+ "step": 59
439
+ },
440
+ {
441
+ "epoch": 0.0883367839889579,
442
+ "grad_norm": 0.11828132718801498,
443
+ "learning_rate": 0.00013245333323392333,
444
+ "loss": 0.0925,
445
+ "step": 60
446
+ },
447
+ {
448
+ "epoch": 0.0898090637221072,
449
+ "grad_norm": 0.10878538340330124,
450
+ "learning_rate": 0.0001316032185167079,
451
+ "loss": 0.0542,
452
+ "step": 61
453
+ },
454
+ {
455
+ "epoch": 0.0912813434552565,
456
+ "grad_norm": 0.1320784091949463,
457
+ "learning_rate": 0.00013073586191080457,
458
+ "loss": 0.1005,
459
+ "step": 62
460
+ },
461
+ {
462
+ "epoch": 0.0927536231884058,
463
+ "grad_norm": 0.10936389863491058,
464
+ "learning_rate": 0.00012985152762143778,
465
+ "loss": 0.0526,
466
+ "step": 63
467
+ },
468
+ {
469
+ "epoch": 0.0942259029215551,
470
+ "grad_norm": 0.13538584113121033,
471
+ "learning_rate": 0.00012895048502539882,
472
+ "loss": 0.0956,
473
+ "step": 64
474
+ },
475
+ {
476
+ "epoch": 0.0956981826547044,
477
+ "grad_norm": 0.13087764382362366,
478
+ "learning_rate": 0.00012803300858899104,
479
+ "loss": 0.0795,
480
+ "step": 65
481
+ },
482
+ {
483
+ "epoch": 0.09717046238785369,
484
+ "grad_norm": 0.1554757058620453,
485
+ "learning_rate": 0.0001270993777844248,
486
+ "loss": 0.1455,
487
+ "step": 66
488
+ },
489
+ {
490
+ "epoch": 0.09864274212100299,
491
+ "grad_norm": 0.12131679803133011,
492
+ "learning_rate": 0.0001261498770046874,
493
+ "loss": 0.0625,
494
+ "step": 67
495
+ },
496
+ {
497
+ "epoch": 0.10011502185415229,
498
+ "grad_norm": 0.11470583826303482,
499
+ "learning_rate": 0.00012518479547691435,
500
+ "loss": 0.0903,
501
+ "step": 68
502
  }
503
  ],
504
  "logging_steps": 1,
 
518
  "attributes": {}
519
  }
520
  },
521
+ "total_flos": 7.176620549823529e+17,
522
  "train_batch_size": 2,
523
  "trial_name": null,
524
  "trial_params": null