CooperW commited on
Commit
57dd742
1 Parent(s): 3c8dcee

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df5e8f19285789390f518cd5c2ee6d18c6b4079c83217593f13cb6312b900641
3
  size 956362232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b6cd1a6cad6794e059cf43e749927da0aee5a92493c3b7658ede9969f510853
3
  size 956362232
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe3c1155940699321c9319f31ff481eb14d2a068276d35c6ad4bd1d8ece41f38
3
  size 486148756
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0e28ae24814d7cb3e928f6f1b556bc26177ed0db6e0d35231ce41e0886afa5a
3
  size 486148756
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8eccac390378b22aeb148c260b00bec01d948946d8363d5282899af673e0e86
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a60c7d771c1fd156acee762fba03c724cb41829a3f71df370ecd1d20b134982
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.03790570954750059,
5
  "eval_steps": 500,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -357,6 +357,356 @@
357
  "learning_rate": 0.00011736481776669306,
358
  "loss": 0.116,
359
  "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
360
  }
361
  ],
362
  "logging_steps": 1,
@@ -371,12 +721,12 @@
371
  "should_evaluate": false,
372
  "should_log": false,
373
  "should_save": true,
374
- "should_training_stop": false
375
  },
376
  "attributes": {}
377
  }
378
  },
379
- "total_flos": 5.812811709692314e+16,
380
  "train_batch_size": 8,
381
  "trial_name": null,
382
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.07581141909500118,
5
  "eval_steps": 500,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
357
  "learning_rate": 0.00011736481776669306,
358
  "loss": 0.116,
359
  "step": 50
360
+ },
361
+ {
362
+ "epoch": 0.038663823738450606,
363
+ "grad_norm": 0.0628269761800766,
364
+ "learning_rate": 0.00011391731009600654,
365
+ "loss": 0.1068,
366
+ "step": 51
367
+ },
368
+ {
369
+ "epoch": 0.039421937929400616,
370
+ "grad_norm": 0.07782096415758133,
371
+ "learning_rate": 0.00011045284632676536,
372
+ "loss": 0.1124,
373
+ "step": 52
374
+ },
375
+ {
376
+ "epoch": 0.040180052120350626,
377
+ "grad_norm": 0.07580537348985672,
378
+ "learning_rate": 0.00010697564737441252,
379
+ "loss": 0.1038,
380
+ "step": 53
381
+ },
382
+ {
383
+ "epoch": 0.04093816631130064,
384
+ "grad_norm": 0.05523601919412613,
385
+ "learning_rate": 0.00010348994967025012,
386
+ "loss": 0.1,
387
+ "step": 54
388
+ },
389
+ {
390
+ "epoch": 0.04169628050225065,
391
+ "grad_norm": 0.056455422192811966,
392
+ "learning_rate": 0.0001,
393
+ "loss": 0.1046,
394
+ "step": 55
395
+ },
396
+ {
397
+ "epoch": 0.04245439469320066,
398
+ "grad_norm": 0.052510254085063934,
399
+ "learning_rate": 9.651005032974994e-05,
400
+ "loss": 0.1025,
401
+ "step": 56
402
+ },
403
+ {
404
+ "epoch": 0.04321250888415067,
405
+ "grad_norm": 0.053201328963041306,
406
+ "learning_rate": 9.302435262558747e-05,
407
+ "loss": 0.0956,
408
+ "step": 57
409
+ },
410
+ {
411
+ "epoch": 0.04397062307510069,
412
+ "grad_norm": 0.05429815128445625,
413
+ "learning_rate": 8.954715367323468e-05,
414
+ "loss": 0.0915,
415
+ "step": 58
416
+ },
417
+ {
418
+ "epoch": 0.0447287372660507,
419
+ "grad_norm": 0.05565072223544121,
420
+ "learning_rate": 8.608268990399349e-05,
421
+ "loss": 0.1037,
422
+ "step": 59
423
+ },
424
+ {
425
+ "epoch": 0.04548685145700071,
426
+ "grad_norm": 0.06064201518893242,
427
+ "learning_rate": 8.263518223330697e-05,
428
+ "loss": 0.0996,
429
+ "step": 60
430
+ },
431
+ {
432
+ "epoch": 0.04624496564795072,
433
+ "grad_norm": 0.053189653903245926,
434
+ "learning_rate": 7.920883091822408e-05,
435
+ "loss": 0.0972,
436
+ "step": 61
437
+ },
438
+ {
439
+ "epoch": 0.04700307983890074,
440
+ "grad_norm": 0.0630035251379013,
441
+ "learning_rate": 7.580781044003324e-05,
442
+ "loss": 0.0957,
443
+ "step": 62
444
+ },
445
+ {
446
+ "epoch": 0.04776119402985075,
447
+ "grad_norm": 0.05781060457229614,
448
+ "learning_rate": 7.243626441830009e-05,
449
+ "loss": 0.1116,
450
+ "step": 63
451
+ },
452
+ {
453
+ "epoch": 0.04851930822080076,
454
+ "grad_norm": 0.05752211809158325,
455
+ "learning_rate": 6.909830056250527e-05,
456
+ "loss": 0.1015,
457
+ "step": 64
458
+ },
459
+ {
460
+ "epoch": 0.04927742241175077,
461
+ "grad_norm": 0.05063620209693909,
462
+ "learning_rate": 6.579798566743314e-05,
463
+ "loss": 0.1083,
464
+ "step": 65
465
+ },
466
+ {
467
+ "epoch": 0.050035536602700784,
468
+ "grad_norm": 0.0517071969807148,
469
+ "learning_rate": 6.25393406584088e-05,
470
+ "loss": 0.0946,
471
+ "step": 66
472
+ },
473
+ {
474
+ "epoch": 0.050793650793650794,
475
+ "grad_norm": 0.04985162615776062,
476
+ "learning_rate": 5.9326335692419995e-05,
477
+ "loss": 0.1058,
478
+ "step": 67
479
+ },
480
+ {
481
+ "epoch": 0.051551764984600804,
482
+ "grad_norm": 0.04708476364612579,
483
+ "learning_rate": 5.616288532109225e-05,
484
+ "loss": 0.0921,
485
+ "step": 68
486
+ },
487
+ {
488
+ "epoch": 0.05230987917555082,
489
+ "grad_norm": 0.05587763711810112,
490
+ "learning_rate": 5.305284372141095e-05,
491
+ "loss": 0.0927,
492
+ "step": 69
493
+ },
494
+ {
495
+ "epoch": 0.05306799336650083,
496
+ "grad_norm": 0.05187711864709854,
497
+ "learning_rate": 5.000000000000002e-05,
498
+ "loss": 0.1022,
499
+ "step": 70
500
+ },
501
+ {
502
+ "epoch": 0.05382610755745084,
503
+ "grad_norm": 0.05270811542868614,
504
+ "learning_rate": 4.700807357667952e-05,
505
+ "loss": 0.1022,
506
+ "step": 71
507
+ },
508
+ {
509
+ "epoch": 0.05458422174840085,
510
+ "grad_norm": 0.050687965005636215,
511
+ "learning_rate": 4.4080709652925336e-05,
512
+ "loss": 0.0999,
513
+ "step": 72
514
+ },
515
+ {
516
+ "epoch": 0.05534233593935087,
517
+ "grad_norm": 0.04412781447172165,
518
+ "learning_rate": 4.12214747707527e-05,
519
+ "loss": 0.0989,
520
+ "step": 73
521
+ },
522
+ {
523
+ "epoch": 0.05610045013030088,
524
+ "grad_norm": 0.04820292443037033,
525
+ "learning_rate": 3.843385246743417e-05,
526
+ "loss": 0.0882,
527
+ "step": 74
528
+ },
529
+ {
530
+ "epoch": 0.05685856432125089,
531
+ "grad_norm": 0.05101883411407471,
532
+ "learning_rate": 3.5721239031346066e-05,
533
+ "loss": 0.0896,
534
+ "step": 75
535
+ },
536
+ {
537
+ "epoch": 0.0576166785122009,
538
+ "grad_norm": 0.056556086987257004,
539
+ "learning_rate": 3.308693936411421e-05,
540
+ "loss": 0.0951,
541
+ "step": 76
542
+ },
543
+ {
544
+ "epoch": 0.058374792703150914,
545
+ "grad_norm": 0.05035807937383652,
546
+ "learning_rate": 3.053416295410026e-05,
547
+ "loss": 0.0911,
548
+ "step": 77
549
+ },
550
+ {
551
+ "epoch": 0.059132906894100924,
552
+ "grad_norm": 0.043333761394023895,
553
+ "learning_rate": 2.8066019966134904e-05,
554
+ "loss": 0.089,
555
+ "step": 78
556
+ },
557
+ {
558
+ "epoch": 0.059891021085050934,
559
+ "grad_norm": 0.047586455941200256,
560
+ "learning_rate": 2.5685517452260567e-05,
561
+ "loss": 0.0874,
562
+ "step": 79
563
+ },
564
+ {
565
+ "epoch": 0.06064913527600095,
566
+ "grad_norm": 0.0488692931830883,
567
+ "learning_rate": 2.339555568810221e-05,
568
+ "loss": 0.0925,
569
+ "step": 80
570
+ },
571
+ {
572
+ "epoch": 0.06140724946695096,
573
+ "grad_norm": 0.05427223816514015,
574
+ "learning_rate": 2.119892463932781e-05,
575
+ "loss": 0.0927,
576
+ "step": 81
577
+ },
578
+ {
579
+ "epoch": 0.06216536365790097,
580
+ "grad_norm": 0.057675670832395554,
581
+ "learning_rate": 1.9098300562505266e-05,
582
+ "loss": 0.0978,
583
+ "step": 82
584
+ },
585
+ {
586
+ "epoch": 0.06292347784885098,
587
+ "grad_norm": 0.05663346126675606,
588
+ "learning_rate": 1.7096242744495837e-05,
589
+ "loss": 0.1041,
590
+ "step": 83
591
+ },
592
+ {
593
+ "epoch": 0.06368159203980099,
594
+ "grad_norm": 0.054945193231105804,
595
+ "learning_rate": 1.5195190384357404e-05,
596
+ "loss": 0.0847,
597
+ "step": 84
598
+ },
599
+ {
600
+ "epoch": 0.064439706230751,
601
+ "grad_norm": 0.05341991409659386,
602
+ "learning_rate": 1.339745962155613e-05,
603
+ "loss": 0.0877,
604
+ "step": 85
605
+ },
606
+ {
607
+ "epoch": 0.06519782042170102,
608
+ "grad_norm": 0.0533662885427475,
609
+ "learning_rate": 1.1705240714107302e-05,
610
+ "loss": 0.1052,
611
+ "step": 86
612
+ },
613
+ {
614
+ "epoch": 0.06595593461265103,
615
+ "grad_norm": 0.045474544167518616,
616
+ "learning_rate": 1.0120595370083318e-05,
617
+ "loss": 0.0863,
618
+ "step": 87
619
+ },
620
+ {
621
+ "epoch": 0.06671404880360104,
622
+ "grad_norm": 0.043375492095947266,
623
+ "learning_rate": 8.645454235739903e-06,
624
+ "loss": 0.0904,
625
+ "step": 88
626
+ },
627
+ {
628
+ "epoch": 0.06747216299455105,
629
+ "grad_norm": 0.05601764842867851,
630
+ "learning_rate": 7.281614543321269e-06,
631
+ "loss": 0.094,
632
+ "step": 89
633
+ },
634
+ {
635
+ "epoch": 0.06823027718550106,
636
+ "grad_norm": 0.0548299178481102,
637
+ "learning_rate": 6.030737921409169e-06,
638
+ "loss": 0.0857,
639
+ "step": 90
640
+ },
641
+ {
642
+ "epoch": 0.06898839137645107,
643
+ "grad_norm": 0.04874192178249359,
644
+ "learning_rate": 4.8943483704846475e-06,
645
+ "loss": 0.0885,
646
+ "step": 91
647
+ },
648
+ {
649
+ "epoch": 0.06974650556740108,
650
+ "grad_norm": 0.05287107825279236,
651
+ "learning_rate": 3.873830406168111e-06,
652
+ "loss": 0.1027,
653
+ "step": 92
654
+ },
655
+ {
656
+ "epoch": 0.0705046197583511,
657
+ "grad_norm": 0.04792382940649986,
658
+ "learning_rate": 2.970427372400353e-06,
659
+ "loss": 0.0849,
660
+ "step": 93
661
+ },
662
+ {
663
+ "epoch": 0.07126273394930112,
664
+ "grad_norm": 0.04913964122533798,
665
+ "learning_rate": 2.1852399266194314e-06,
666
+ "loss": 0.0848,
667
+ "step": 94
668
+ },
669
+ {
670
+ "epoch": 0.07202084814025113,
671
+ "grad_norm": 0.04982119798660278,
672
+ "learning_rate": 1.5192246987791981e-06,
673
+ "loss": 0.0959,
674
+ "step": 95
675
+ },
676
+ {
677
+ "epoch": 0.07277896233120114,
678
+ "grad_norm": 0.04819797724485397,
679
+ "learning_rate": 9.731931258429638e-07,
680
+ "loss": 0.0881,
681
+ "step": 96
682
+ },
683
+ {
684
+ "epoch": 0.07353707652215115,
685
+ "grad_norm": 0.050941213965415955,
686
+ "learning_rate": 5.478104631726711e-07,
687
+ "loss": 0.0957,
688
+ "step": 97
689
+ },
690
+ {
691
+ "epoch": 0.07429519071310116,
692
+ "grad_norm": 0.046391792595386505,
693
+ "learning_rate": 2.4359497401758024e-07,
694
+ "loss": 0.0859,
695
+ "step": 98
696
+ },
697
+ {
698
+ "epoch": 0.07505330490405117,
699
+ "grad_norm": 0.0439014732837677,
700
+ "learning_rate": 6.09172980904238e-08,
701
+ "loss": 0.0963,
702
+ "step": 99
703
+ },
704
+ {
705
+ "epoch": 0.07581141909500118,
706
+ "grad_norm": 0.04594559594988823,
707
+ "learning_rate": 0.0,
708
+ "loss": 0.091,
709
+ "step": 100
710
  }
711
  ],
712
  "logging_steps": 1,
 
721
  "should_evaluate": false,
722
  "should_log": false,
723
  "should_save": true,
724
+ "should_training_stop": true
725
  },
726
  "attributes": {}
727
  }
728
  },
729
+ "total_flos": 1.162003273287598e+17,
730
  "train_batch_size": 8,
731
  "trial_name": null,
732
  "trial_params": null