{ "best_metric": null, "best_model_checkpoint": null, "epoch": 15.0, "eval_steps": 500, "global_step": 4785, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003134796238244514, "grad_norm": 1200.0, "learning_rate": 4.175365344467641e-07, "loss": 56.0196, "step": 1 }, { "epoch": 0.01567398119122257, "grad_norm": 1104.0, "learning_rate": 2.0876826722338207e-06, "loss": 57.1065, "step": 5 }, { "epoch": 0.03134796238244514, "grad_norm": 780.0, "learning_rate": 4.175365344467641e-06, "loss": 53.982, "step": 10 }, { "epoch": 0.047021943573667714, "grad_norm": 378.0, "learning_rate": 6.2630480167014616e-06, "loss": 39.7621, "step": 15 }, { "epoch": 0.06269592476489028, "grad_norm": 119.0, "learning_rate": 8.350730688935283e-06, "loss": 31.2881, "step": 20 }, { "epoch": 0.07836990595611286, "grad_norm": 59.0, "learning_rate": 1.0438413361169103e-05, "loss": 29.3676, "step": 25 }, { "epoch": 0.09404388714733543, "grad_norm": 27.75, "learning_rate": 1.2526096033402923e-05, "loss": 27.221, "step": 30 }, { "epoch": 0.109717868338558, "grad_norm": 17.0, "learning_rate": 1.4613778705636743e-05, "loss": 25.3228, "step": 35 }, { "epoch": 0.12539184952978055, "grad_norm": 11.375, "learning_rate": 1.6701461377870565e-05, "loss": 24.7964, "step": 40 }, { "epoch": 0.14106583072100312, "grad_norm": 14.6875, "learning_rate": 1.8789144050104384e-05, "loss": 23.6672, "step": 45 }, { "epoch": 0.15673981191222572, "grad_norm": 23.5, "learning_rate": 2.0876826722338206e-05, "loss": 22.938, "step": 50 }, { "epoch": 0.1724137931034483, "grad_norm": 46.75, "learning_rate": 2.2964509394572024e-05, "loss": 20.5447, "step": 55 }, { "epoch": 0.18808777429467086, "grad_norm": 102.5, "learning_rate": 2.5052192066805846e-05, "loss": 15.6089, "step": 60 }, { "epoch": 0.20376175548589343, "grad_norm": 21.875, "learning_rate": 2.7139874739039668e-05, "loss": 6.57, "step": 65 }, { "epoch": 0.219435736677116, "grad_norm": 7.15625, "learning_rate": 2.9227557411273487e-05, "loss": 2.9537, "step": 70 }, { "epoch": 0.23510971786833856, "grad_norm": 3.953125, "learning_rate": 3.131524008350731e-05, "loss": 2.4063, "step": 75 }, { "epoch": 0.2507836990595611, "grad_norm": 3.234375, "learning_rate": 3.340292275574113e-05, "loss": 2.1568, "step": 80 }, { "epoch": 0.2664576802507837, "grad_norm": 4.75, "learning_rate": 3.5490605427974946e-05, "loss": 1.9549, "step": 85 }, { "epoch": 0.28213166144200624, "grad_norm": 5.90625, "learning_rate": 3.757828810020877e-05, "loss": 1.8178, "step": 90 }, { "epoch": 0.29780564263322884, "grad_norm": 16.625, "learning_rate": 3.966597077244259e-05, "loss": 1.7357, "step": 95 }, { "epoch": 0.31347962382445144, "grad_norm": 10.625, "learning_rate": 4.175365344467641e-05, "loss": 1.6526, "step": 100 }, { "epoch": 0.329153605015674, "grad_norm": 16.875, "learning_rate": 4.3841336116910233e-05, "loss": 1.6119, "step": 105 }, { "epoch": 0.3448275862068966, "grad_norm": 15.9375, "learning_rate": 4.592901878914405e-05, "loss": 1.5619, "step": 110 }, { "epoch": 0.3605015673981191, "grad_norm": 6.25, "learning_rate": 4.801670146137787e-05, "loss": 1.553, "step": 115 }, { "epoch": 0.3761755485893417, "grad_norm": 16.875, "learning_rate": 5.010438413361169e-05, "loss": 1.5499, "step": 120 }, { "epoch": 0.39184952978056425, "grad_norm": 18.875, "learning_rate": 5.219206680584552e-05, "loss": 1.5358, "step": 125 }, { "epoch": 0.40752351097178685, "grad_norm": 10.5, "learning_rate": 5.4279749478079336e-05, "loss": 1.5022, "step": 130 }, { "epoch": 0.4231974921630094, "grad_norm": 8.6875, "learning_rate": 5.636743215031316e-05, "loss": 1.4674, "step": 135 }, { "epoch": 0.438871473354232, "grad_norm": 3.25, "learning_rate": 5.8455114822546973e-05, "loss": 1.4238, "step": 140 }, { "epoch": 0.45454545454545453, "grad_norm": 9.0, "learning_rate": 6.05427974947808e-05, "loss": 1.3893, "step": 145 }, { "epoch": 0.4702194357366771, "grad_norm": 25.875, "learning_rate": 6.263048016701462e-05, "loss": 1.3631, "step": 150 }, { "epoch": 0.48589341692789967, "grad_norm": 5.21875, "learning_rate": 6.471816283924845e-05, "loss": 1.3578, "step": 155 }, { "epoch": 0.5015673981191222, "grad_norm": 4.96875, "learning_rate": 6.680584551148226e-05, "loss": 1.2742, "step": 160 }, { "epoch": 0.5172413793103449, "grad_norm": 7.21875, "learning_rate": 6.889352818371608e-05, "loss": 1.2685, "step": 165 }, { "epoch": 0.5329153605015674, "grad_norm": 8.5, "learning_rate": 7.098121085594989e-05, "loss": 1.2819, "step": 170 }, { "epoch": 0.54858934169279, "grad_norm": 9.1875, "learning_rate": 7.306889352818372e-05, "loss": 1.284, "step": 175 }, { "epoch": 0.5642633228840125, "grad_norm": 6.40625, "learning_rate": 7.515657620041754e-05, "loss": 1.2541, "step": 180 }, { "epoch": 0.5799373040752351, "grad_norm": 11.5625, "learning_rate": 7.724425887265136e-05, "loss": 1.25, "step": 185 }, { "epoch": 0.5956112852664577, "grad_norm": 11.875, "learning_rate": 7.933194154488518e-05, "loss": 1.2324, "step": 190 }, { "epoch": 0.6112852664576802, "grad_norm": 4.6875, "learning_rate": 8.141962421711901e-05, "loss": 1.2093, "step": 195 }, { "epoch": 0.6269592476489029, "grad_norm": 10.0, "learning_rate": 8.350730688935282e-05, "loss": 1.2225, "step": 200 }, { "epoch": 0.6426332288401254, "grad_norm": 2.03125, "learning_rate": 8.559498956158665e-05, "loss": 1.1948, "step": 205 }, { "epoch": 0.658307210031348, "grad_norm": 6.125, "learning_rate": 8.768267223382047e-05, "loss": 1.1823, "step": 210 }, { "epoch": 0.6739811912225705, "grad_norm": 4.34375, "learning_rate": 8.977035490605428e-05, "loss": 1.1812, "step": 215 }, { "epoch": 0.6896551724137931, "grad_norm": 5.3125, "learning_rate": 9.18580375782881e-05, "loss": 1.1833, "step": 220 }, { "epoch": 0.7053291536050157, "grad_norm": 9.875, "learning_rate": 9.394572025052193e-05, "loss": 1.1606, "step": 225 }, { "epoch": 0.7210031347962382, "grad_norm": 2.609375, "learning_rate": 9.603340292275574e-05, "loss": 1.1662, "step": 230 }, { "epoch": 0.7366771159874608, "grad_norm": 2.96875, "learning_rate": 9.812108559498957e-05, "loss": 1.1573, "step": 235 }, { "epoch": 0.7523510971786834, "grad_norm": 2.0, "learning_rate": 0.00010020876826722338, "loss": 1.1386, "step": 240 }, { "epoch": 0.768025078369906, "grad_norm": 5.3125, "learning_rate": 0.00010229645093945721, "loss": 1.1822, "step": 245 }, { "epoch": 0.7836990595611285, "grad_norm": 8.125, "learning_rate": 0.00010438413361169104, "loss": 1.1415, "step": 250 }, { "epoch": 0.799373040752351, "grad_norm": 4.65625, "learning_rate": 0.00010647181628392484, "loss": 1.1534, "step": 255 }, { "epoch": 0.8150470219435737, "grad_norm": 1.0703125, "learning_rate": 0.00010855949895615867, "loss": 1.1338, "step": 260 }, { "epoch": 0.8307210031347962, "grad_norm": 3.03125, "learning_rate": 0.00011064718162839249, "loss": 1.1435, "step": 265 }, { "epoch": 0.8463949843260188, "grad_norm": 2.90625, "learning_rate": 0.00011273486430062632, "loss": 1.1213, "step": 270 }, { "epoch": 0.8620689655172413, "grad_norm": 3.875, "learning_rate": 0.00011482254697286012, "loss": 1.123, "step": 275 }, { "epoch": 0.877742946708464, "grad_norm": 9.8125, "learning_rate": 0.00011691022964509395, "loss": 1.154, "step": 280 }, { "epoch": 0.8934169278996865, "grad_norm": 3.15625, "learning_rate": 0.00011899791231732778, "loss": 1.1346, "step": 285 }, { "epoch": 0.9090909090909091, "grad_norm": 15.75, "learning_rate": 0.0001210855949895616, "loss": 1.0998, "step": 290 }, { "epoch": 0.9247648902821317, "grad_norm": 4.375, "learning_rate": 0.0001231732776617954, "loss": 1.0742, "step": 295 }, { "epoch": 0.9404388714733543, "grad_norm": 3.015625, "learning_rate": 0.00012526096033402923, "loss": 1.0956, "step": 300 }, { "epoch": 0.9561128526645768, "grad_norm": 3.796875, "learning_rate": 0.00012734864300626306, "loss": 1.0698, "step": 305 }, { "epoch": 0.9717868338557993, "grad_norm": 2.75, "learning_rate": 0.0001294363256784969, "loss": 1.0526, "step": 310 }, { "epoch": 0.987460815047022, "grad_norm": 2.078125, "learning_rate": 0.0001315240083507307, "loss": 1.0589, "step": 315 }, { "epoch": 1.0, "eval_loss": 1.7604742050170898, "eval_runtime": 0.8071, "eval_samples_per_second": 2.478, "eval_steps_per_second": 1.239, "step": 319 }, { "epoch": 1.0031347962382444, "grad_norm": 2.359375, "learning_rate": 0.00013361169102296452, "loss": 1.06, "step": 320 }, { "epoch": 1.0188087774294672, "grad_norm": 34.0, "learning_rate": 0.00013569937369519835, "loss": 1.0232, "step": 325 }, { "epoch": 1.0344827586206897, "grad_norm": 8.6875, "learning_rate": 0.00013778705636743215, "loss": 1.0624, "step": 330 }, { "epoch": 1.0501567398119123, "grad_norm": 10.8125, "learning_rate": 0.00013987473903966598, "loss": 1.0581, "step": 335 }, { "epoch": 1.0658307210031348, "grad_norm": 3.109375, "learning_rate": 0.00014196242171189978, "loss": 1.0274, "step": 340 }, { "epoch": 1.0815047021943573, "grad_norm": 3.828125, "learning_rate": 0.0001440501043841336, "loss": 1.0301, "step": 345 }, { "epoch": 1.09717868338558, "grad_norm": 10.25, "learning_rate": 0.00014613778705636744, "loss": 1.0221, "step": 350 }, { "epoch": 1.1128526645768024, "grad_norm": 4.0625, "learning_rate": 0.00014822546972860124, "loss": 1.0185, "step": 355 }, { "epoch": 1.1285266457680252, "grad_norm": 6.65625, "learning_rate": 0.00015031315240083507, "loss": 1.0146, "step": 360 }, { "epoch": 1.1442006269592477, "grad_norm": 4.21875, "learning_rate": 0.0001524008350730689, "loss": 1.0134, "step": 365 }, { "epoch": 1.1598746081504703, "grad_norm": 5.375, "learning_rate": 0.00015448851774530273, "loss": 1.0372, "step": 370 }, { "epoch": 1.1755485893416928, "grad_norm": 1.046875, "learning_rate": 0.00015657620041753653, "loss": 1.0096, "step": 375 }, { "epoch": 1.1912225705329154, "grad_norm": 3.734375, "learning_rate": 0.00015866388308977036, "loss": 1.0143, "step": 380 }, { "epoch": 1.206896551724138, "grad_norm": 3.265625, "learning_rate": 0.0001607515657620042, "loss": 1.0159, "step": 385 }, { "epoch": 1.2225705329153604, "grad_norm": 1.875, "learning_rate": 0.00016283924843423802, "loss": 1.0069, "step": 390 }, { "epoch": 1.238244514106583, "grad_norm": 1.9609375, "learning_rate": 0.00016492693110647182, "loss": 1.0019, "step": 395 }, { "epoch": 1.2539184952978055, "grad_norm": 5.0, "learning_rate": 0.00016701461377870565, "loss": 0.9979, "step": 400 }, { "epoch": 1.2695924764890283, "grad_norm": 0.85546875, "learning_rate": 0.00016910229645093947, "loss": 1.0066, "step": 405 }, { "epoch": 1.2852664576802508, "grad_norm": 3.25, "learning_rate": 0.0001711899791231733, "loss": 1.0002, "step": 410 }, { "epoch": 1.3009404388714734, "grad_norm": 0.92578125, "learning_rate": 0.0001732776617954071, "loss": 1.0036, "step": 415 }, { "epoch": 1.316614420062696, "grad_norm": 3.125, "learning_rate": 0.00017536534446764093, "loss": 1.003, "step": 420 }, { "epoch": 1.3322884012539185, "grad_norm": 3.390625, "learning_rate": 0.00017745302713987476, "loss": 1.0207, "step": 425 }, { "epoch": 1.347962382445141, "grad_norm": 9.3125, "learning_rate": 0.00017954070981210856, "loss": 1.0097, "step": 430 }, { "epoch": 1.3636363636363638, "grad_norm": 2.140625, "learning_rate": 0.0001816283924843424, "loss": 0.9984, "step": 435 }, { "epoch": 1.3793103448275863, "grad_norm": 0.93359375, "learning_rate": 0.0001837160751565762, "loss": 1.019, "step": 440 }, { "epoch": 1.3949843260188088, "grad_norm": 1.4140625, "learning_rate": 0.00018580375782881002, "loss": 1.0093, "step": 445 }, { "epoch": 1.4106583072100314, "grad_norm": 2.90625, "learning_rate": 0.00018789144050104385, "loss": 0.9906, "step": 450 }, { "epoch": 1.426332288401254, "grad_norm": 1.8203125, "learning_rate": 0.00018997912317327765, "loss": 0.9956, "step": 455 }, { "epoch": 1.4420062695924765, "grad_norm": 4.21875, "learning_rate": 0.00019206680584551148, "loss": 1.0026, "step": 460 }, { "epoch": 1.457680250783699, "grad_norm": 2.46875, "learning_rate": 0.0001941544885177453, "loss": 0.9779, "step": 465 }, { "epoch": 1.4733542319749215, "grad_norm": 1.0859375, "learning_rate": 0.00019624217118997914, "loss": 0.9981, "step": 470 }, { "epoch": 1.489028213166144, "grad_norm": 2.34375, "learning_rate": 0.00019832985386221294, "loss": 1.0087, "step": 475 }, { "epoch": 1.5047021943573666, "grad_norm": 5.71875, "learning_rate": 0.0001999999733852936, "loss": 1.0154, "step": 480 }, { "epoch": 1.5203761755485894, "grad_norm": 2.125, "learning_rate": 0.00019999904187205744, "loss": 1.0196, "step": 485 }, { "epoch": 1.536050156739812, "grad_norm": 5.1875, "learning_rate": 0.00019999677963766844, "loss": 1.0196, "step": 490 }, { "epoch": 1.5517241379310345, "grad_norm": 3.015625, "learning_rate": 0.00019999318671223102, "loss": 1.003, "step": 495 }, { "epoch": 1.567398119122257, "grad_norm": 2.453125, "learning_rate": 0.0001999882631435574, "loss": 1.0022, "step": 500 }, { "epoch": 1.5830721003134798, "grad_norm": 5.78125, "learning_rate": 0.00019998200899716724, "loss": 0.9932, "step": 505 }, { "epoch": 1.5987460815047023, "grad_norm": 3.625, "learning_rate": 0.00019997442435628653, "loss": 1.0083, "step": 510 }, { "epoch": 1.6144200626959249, "grad_norm": 6.375, "learning_rate": 0.00019996550932184666, "loss": 1.0153, "step": 515 }, { "epoch": 1.6300940438871474, "grad_norm": 1.59375, "learning_rate": 0.00019995526401248302, "loss": 1.0158, "step": 520 }, { "epoch": 1.64576802507837, "grad_norm": 4.84375, "learning_rate": 0.00019994368856453341, "loss": 1.0308, "step": 525 }, { "epoch": 1.6614420062695925, "grad_norm": 2.796875, "learning_rate": 0.00019993078313203632, "loss": 0.9834, "step": 530 }, { "epoch": 1.677115987460815, "grad_norm": 1.28125, "learning_rate": 0.0001999165478867286, "loss": 0.9937, "step": 535 }, { "epoch": 1.6927899686520376, "grad_norm": 1.9921875, "learning_rate": 0.00019990098301804357, "loss": 0.9884, "step": 540 }, { "epoch": 1.70846394984326, "grad_norm": 1.671875, "learning_rate": 0.00019988408873310815, "loss": 0.9846, "step": 545 }, { "epoch": 1.7241379310344827, "grad_norm": 0.9921875, "learning_rate": 0.00019986586525674036, "loss": 0.9711, "step": 550 }, { "epoch": 1.7398119122257052, "grad_norm": 1.921875, "learning_rate": 0.00019984631283144616, "loss": 0.9789, "step": 555 }, { "epoch": 1.7554858934169277, "grad_norm": 1.953125, "learning_rate": 0.0001998254317174163, "loss": 0.984, "step": 560 }, { "epoch": 1.7711598746081505, "grad_norm": 3.046875, "learning_rate": 0.00019980322219252284, "loss": 0.9609, "step": 565 }, { "epoch": 1.786833855799373, "grad_norm": 1.7578125, "learning_rate": 0.0001997796845523155, "loss": 0.9997, "step": 570 }, { "epoch": 1.8025078369905956, "grad_norm": 2.9375, "learning_rate": 0.00019975481911001762, "loss": 0.9772, "step": 575 }, { "epoch": 1.8181818181818183, "grad_norm": 1.5234375, "learning_rate": 0.00019972862619652203, "loss": 0.9939, "step": 580 }, { "epoch": 1.8338557993730409, "grad_norm": 1.6953125, "learning_rate": 0.00019970110616038673, "loss": 0.9794, "step": 585 }, { "epoch": 1.8495297805642634, "grad_norm": 3.109375, "learning_rate": 0.0001996722593678302, "loss": 0.9702, "step": 590 }, { "epoch": 1.865203761755486, "grad_norm": 1.2578125, "learning_rate": 0.00019964208620272647, "loss": 0.9667, "step": 595 }, { "epoch": 1.8808777429467085, "grad_norm": 0.8203125, "learning_rate": 0.00019961058706660005, "loss": 0.9632, "step": 600 }, { "epoch": 1.896551724137931, "grad_norm": 12.5, "learning_rate": 0.00019957776237862067, "loss": 0.9554, "step": 605 }, { "epoch": 1.9122257053291536, "grad_norm": 1.6328125, "learning_rate": 0.00019954361257559756, "loss": 0.9902, "step": 610 }, { "epoch": 1.9278996865203761, "grad_norm": 1.328125, "learning_rate": 0.0001995081381119737, "loss": 0.998, "step": 615 }, { "epoch": 1.9435736677115987, "grad_norm": 2.28125, "learning_rate": 0.00019947133945981987, "loss": 0.9781, "step": 620 }, { "epoch": 1.9592476489028212, "grad_norm": 2.796875, "learning_rate": 0.00019943321710882815, "loss": 0.9956, "step": 625 }, { "epoch": 1.9749216300940438, "grad_norm": 3.046875, "learning_rate": 0.0001993937715663056, "loss": 0.9776, "step": 630 }, { "epoch": 1.9905956112852663, "grad_norm": 1.9140625, "learning_rate": 0.00019935300335716748, "loss": 1.0, "step": 635 }, { "epoch": 2.0, "eval_loss": 1.7305909395217896, "eval_runtime": 0.8036, "eval_samples_per_second": 2.489, "eval_steps_per_second": 1.244, "step": 638 }, { "epoch": 2.006269592476489, "grad_norm": 3.921875, "learning_rate": 0.00019931091302393008, "loss": 0.9559, "step": 640 }, { "epoch": 2.0219435736677114, "grad_norm": 1.6015625, "learning_rate": 0.00019926750112670382, "loss": 0.8412, "step": 645 }, { "epoch": 2.0376175548589344, "grad_norm": 1.8203125, "learning_rate": 0.00019922276824318547, "loss": 0.8475, "step": 650 }, { "epoch": 2.053291536050157, "grad_norm": 1.3359375, "learning_rate": 0.0001991767149686507, "loss": 0.8512, "step": 655 }, { "epoch": 2.0689655172413794, "grad_norm": 1.0, "learning_rate": 0.0001991293419159461, "loss": 0.8301, "step": 660 }, { "epoch": 2.084639498432602, "grad_norm": 2.390625, "learning_rate": 0.00019908064971548085, "loss": 0.8622, "step": 665 }, { "epoch": 2.1003134796238245, "grad_norm": 1.1953125, "learning_rate": 0.0001990306390152186, "loss": 0.8375, "step": 670 }, { "epoch": 2.115987460815047, "grad_norm": 3.390625, "learning_rate": 0.00019897931048066877, "loss": 0.8571, "step": 675 }, { "epoch": 2.1316614420062696, "grad_norm": 1.8203125, "learning_rate": 0.00019892666479487744, "loss": 0.904, "step": 680 }, { "epoch": 2.147335423197492, "grad_norm": 2.859375, "learning_rate": 0.00019887270265841868, "loss": 0.8602, "step": 685 }, { "epoch": 2.1630094043887147, "grad_norm": 2.125, "learning_rate": 0.00019881742478938496, "loss": 0.8618, "step": 690 }, { "epoch": 2.1786833855799372, "grad_norm": 3.09375, "learning_rate": 0.00019876083192337757, "loss": 0.8737, "step": 695 }, { "epoch": 2.19435736677116, "grad_norm": 3.9375, "learning_rate": 0.00019870292481349698, "loss": 0.8688, "step": 700 }, { "epoch": 2.2100313479623823, "grad_norm": 1.7578125, "learning_rate": 0.00019864370423033274, "loss": 0.8821, "step": 705 }, { "epoch": 2.225705329153605, "grad_norm": 1.3203125, "learning_rate": 0.00019858317096195323, "loss": 0.8746, "step": 710 }, { "epoch": 2.2413793103448274, "grad_norm": 2.796875, "learning_rate": 0.00019852132581389513, "loss": 0.8742, "step": 715 }, { "epoch": 2.2570532915360504, "grad_norm": 2.421875, "learning_rate": 0.00019845816960915286, "loss": 0.8747, "step": 720 }, { "epoch": 2.2727272727272725, "grad_norm": 2.046875, "learning_rate": 0.0001983937031881674, "loss": 0.843, "step": 725 }, { "epoch": 2.2884012539184955, "grad_norm": 3.140625, "learning_rate": 0.0001983279274088153, "loss": 0.8682, "step": 730 }, { "epoch": 2.304075235109718, "grad_norm": 2.9375, "learning_rate": 0.00019826084314639714, "loss": 0.856, "step": 735 }, { "epoch": 2.3197492163009406, "grad_norm": 1.21875, "learning_rate": 0.00019819245129362595, "loss": 0.8663, "step": 740 }, { "epoch": 2.335423197492163, "grad_norm": 2.21875, "learning_rate": 0.00019812275276061533, "loss": 0.8483, "step": 745 }, { "epoch": 2.3510971786833856, "grad_norm": 9.875, "learning_rate": 0.00019805174847486721, "loss": 0.8416, "step": 750 }, { "epoch": 2.366771159874608, "grad_norm": 0.78515625, "learning_rate": 0.00019797943938125977, "loss": 0.8743, "step": 755 }, { "epoch": 2.3824451410658307, "grad_norm": 1.21875, "learning_rate": 0.00019790582644203458, "loss": 0.8529, "step": 760 }, { "epoch": 2.3981191222570533, "grad_norm": 6.15625, "learning_rate": 0.00019783091063678402, "loss": 0.8628, "step": 765 }, { "epoch": 2.413793103448276, "grad_norm": 2.65625, "learning_rate": 0.00019775469296243807, "loss": 0.8689, "step": 770 }, { "epoch": 2.4294670846394983, "grad_norm": 1.59375, "learning_rate": 0.0001976771744332512, "loss": 0.8671, "step": 775 }, { "epoch": 2.445141065830721, "grad_norm": 2.0, "learning_rate": 0.00019759835608078877, "loss": 0.8832, "step": 780 }, { "epoch": 2.4608150470219434, "grad_norm": 1.1796875, "learning_rate": 0.00019751823895391323, "loss": 0.878, "step": 785 }, { "epoch": 2.476489028213166, "grad_norm": 1.5, "learning_rate": 0.00019743682411877046, "loss": 0.8882, "step": 790 }, { "epoch": 2.492163009404389, "grad_norm": 1.7265625, "learning_rate": 0.00019735411265877522, "loss": 0.8934, "step": 795 }, { "epoch": 2.507836990595611, "grad_norm": 1.4140625, "learning_rate": 0.00019727010567459696, "loss": 0.8815, "step": 800 }, { "epoch": 2.523510971786834, "grad_norm": 1.1171875, "learning_rate": 0.00019718480428414505, "loss": 0.8925, "step": 805 }, { "epoch": 2.5391849529780566, "grad_norm": 1.4765625, "learning_rate": 0.00019709820962255409, "loss": 0.8956, "step": 810 }, { "epoch": 2.554858934169279, "grad_norm": 3.546875, "learning_rate": 0.00019701032284216857, "loss": 0.8828, "step": 815 }, { "epoch": 2.5705329153605017, "grad_norm": 1.125, "learning_rate": 0.00019692114511252767, "loss": 0.9194, "step": 820 }, { "epoch": 2.586206896551724, "grad_norm": 1.734375, "learning_rate": 0.00019683067762034967, "loss": 0.8825, "step": 825 }, { "epoch": 2.6018808777429467, "grad_norm": 4.1875, "learning_rate": 0.00019673892156951613, "loss": 0.8725, "step": 830 }, { "epoch": 2.6175548589341693, "grad_norm": 1.296875, "learning_rate": 0.00019664587818105596, "loss": 0.877, "step": 835 }, { "epoch": 2.633228840125392, "grad_norm": 0.89453125, "learning_rate": 0.0001965515486931291, "loss": 0.8764, "step": 840 }, { "epoch": 2.6489028213166144, "grad_norm": 0.8515625, "learning_rate": 0.00019645593436101, "loss": 0.8571, "step": 845 }, { "epoch": 2.664576802507837, "grad_norm": 2.40625, "learning_rate": 0.00019635903645707096, "loss": 0.887, "step": 850 }, { "epoch": 2.6802507836990594, "grad_norm": 0.8203125, "learning_rate": 0.00019626085627076528, "loss": 0.8755, "step": 855 }, { "epoch": 2.695924764890282, "grad_norm": 1.4140625, "learning_rate": 0.00019616139510861, "loss": 0.8664, "step": 860 }, { "epoch": 2.7115987460815045, "grad_norm": 0.9296875, "learning_rate": 0.00019606065429416848, "loss": 0.8888, "step": 865 }, { "epoch": 2.7272727272727275, "grad_norm": 1.4296875, "learning_rate": 0.00019595863516803293, "loss": 0.8772, "step": 870 }, { "epoch": 2.7429467084639496, "grad_norm": 1.5859375, "learning_rate": 0.0001958553390878064, "loss": 0.8919, "step": 875 }, { "epoch": 2.7586206896551726, "grad_norm": 1.609375, "learning_rate": 0.00019575076742808488, "loss": 0.8806, "step": 880 }, { "epoch": 2.774294670846395, "grad_norm": 0.81640625, "learning_rate": 0.00019564492158043891, "loss": 0.8722, "step": 885 }, { "epoch": 2.7899686520376177, "grad_norm": 1.359375, "learning_rate": 0.0001955378029533951, "loss": 0.8887, "step": 890 }, { "epoch": 2.80564263322884, "grad_norm": 1.1015625, "learning_rate": 0.00019542941297241722, "loss": 0.9079, "step": 895 }, { "epoch": 2.8213166144200628, "grad_norm": 2.0, "learning_rate": 0.00019531975307988763, "loss": 0.877, "step": 900 }, { "epoch": 2.8369905956112853, "grad_norm": 1.4140625, "learning_rate": 0.00019520882473508762, "loss": 0.8953, "step": 905 }, { "epoch": 2.852664576802508, "grad_norm": 5.03125, "learning_rate": 0.00019509662941417826, "loss": 0.8886, "step": 910 }, { "epoch": 2.8683385579937304, "grad_norm": 1.15625, "learning_rate": 0.00019498316861018086, "loss": 0.9104, "step": 915 }, { "epoch": 2.884012539184953, "grad_norm": 1.6015625, "learning_rate": 0.0001948684438329566, "loss": 0.8815, "step": 920 }, { "epoch": 2.8996865203761755, "grad_norm": 1.8125, "learning_rate": 0.00019475245660918717, "loss": 0.8718, "step": 925 }, { "epoch": 2.915360501567398, "grad_norm": 6.46875, "learning_rate": 0.00019463520848235377, "loss": 0.8774, "step": 930 }, { "epoch": 2.9310344827586206, "grad_norm": 1.015625, "learning_rate": 0.000194516701012717, "loss": 0.8853, "step": 935 }, { "epoch": 2.946708463949843, "grad_norm": 1.5546875, "learning_rate": 0.00019439693577729593, "loss": 0.8833, "step": 940 }, { "epoch": 2.962382445141066, "grad_norm": 1.796875, "learning_rate": 0.0001942759143698472, "loss": 0.8867, "step": 945 }, { "epoch": 2.978056426332288, "grad_norm": 1.0703125, "learning_rate": 0.0001941536384008437, "loss": 0.8923, "step": 950 }, { "epoch": 2.993730407523511, "grad_norm": 1.75, "learning_rate": 0.0001940301094974531, "loss": 0.8741, "step": 955 }, { "epoch": 3.0, "eval_loss": 1.8115239143371582, "eval_runtime": 0.806, "eval_samples_per_second": 2.481, "eval_steps_per_second": 1.241, "step": 957 }, { "epoch": 3.0094043887147337, "grad_norm": 0.7578125, "learning_rate": 0.00019390532930351652, "loss": 0.7936, "step": 960 }, { "epoch": 3.0250783699059562, "grad_norm": 1.7734375, "learning_rate": 0.00019377929947952626, "loss": 0.7324, "step": 965 }, { "epoch": 3.040752351097179, "grad_norm": 1.7109375, "learning_rate": 0.00019365202170260393, "loss": 0.726, "step": 970 }, { "epoch": 3.0564263322884013, "grad_norm": 1.0859375, "learning_rate": 0.000193523497666478, "loss": 0.7355, "step": 975 }, { "epoch": 3.072100313479624, "grad_norm": 2.578125, "learning_rate": 0.00019339372908146147, "loss": 0.7393, "step": 980 }, { "epoch": 3.0877742946708464, "grad_norm": 1.0, "learning_rate": 0.00019326271767442884, "loss": 0.736, "step": 985 }, { "epoch": 3.103448275862069, "grad_norm": 1.3203125, "learning_rate": 0.00019313046518879337, "loss": 0.7157, "step": 990 }, { "epoch": 3.1191222570532915, "grad_norm": 1.015625, "learning_rate": 0.00019299697338448369, "loss": 0.7231, "step": 995 }, { "epoch": 3.134796238244514, "grad_norm": 1.2734375, "learning_rate": 0.0001928622440379205, "loss": 0.7221, "step": 1000 }, { "epoch": 3.1504702194357366, "grad_norm": 0.9140625, "learning_rate": 0.0001927262789419929, "loss": 0.7352, "step": 1005 }, { "epoch": 3.166144200626959, "grad_norm": 0.9375, "learning_rate": 0.0001925890799060345, "loss": 0.7196, "step": 1010 }, { "epoch": 3.1818181818181817, "grad_norm": 0.90625, "learning_rate": 0.00019245064875579942, "loss": 0.7269, "step": 1015 }, { "epoch": 3.197492163009404, "grad_norm": 0.98828125, "learning_rate": 0.00019231098733343783, "loss": 0.7225, "step": 1020 }, { "epoch": 3.2131661442006267, "grad_norm": 1.0625, "learning_rate": 0.00019217009749747174, "loss": 0.734, "step": 1025 }, { "epoch": 3.2288401253918497, "grad_norm": 0.828125, "learning_rate": 0.0001920279811227699, "loss": 0.7387, "step": 1030 }, { "epoch": 3.2445141065830723, "grad_norm": 1.5703125, "learning_rate": 0.00019188464010052312, "loss": 0.7303, "step": 1035 }, { "epoch": 3.260188087774295, "grad_norm": 1.53125, "learning_rate": 0.00019174007633821893, "loss": 0.7565, "step": 1040 }, { "epoch": 3.2758620689655173, "grad_norm": 2.921875, "learning_rate": 0.00019159429175961634, "loss": 0.7588, "step": 1045 }, { "epoch": 3.29153605015674, "grad_norm": 1.0, "learning_rate": 0.0001914472883047202, "loss": 0.7452, "step": 1050 }, { "epoch": 3.3072100313479624, "grad_norm": 1.1953125, "learning_rate": 0.00019129906792975527, "loss": 0.7395, "step": 1055 }, { "epoch": 3.322884012539185, "grad_norm": 1.203125, "learning_rate": 0.0001911496326071404, "loss": 0.7429, "step": 1060 }, { "epoch": 3.3385579937304075, "grad_norm": 0.92578125, "learning_rate": 0.00019099898432546202, "loss": 0.7643, "step": 1065 }, { "epoch": 3.35423197492163, "grad_norm": 1.3203125, "learning_rate": 0.00019084712508944793, "loss": 0.755, "step": 1070 }, { "epoch": 3.3699059561128526, "grad_norm": 1.8671875, "learning_rate": 0.00019069405691994045, "loss": 0.7381, "step": 1075 }, { "epoch": 3.385579937304075, "grad_norm": 1.46875, "learning_rate": 0.00019053978185386964, "loss": 0.7546, "step": 1080 }, { "epoch": 3.4012539184952977, "grad_norm": 2.609375, "learning_rate": 0.00019038430194422606, "loss": 0.7624, "step": 1085 }, { "epoch": 3.41692789968652, "grad_norm": 1.3203125, "learning_rate": 0.00019022761926003359, "loss": 0.7657, "step": 1090 }, { "epoch": 3.4326018808777428, "grad_norm": 1.15625, "learning_rate": 0.00019006973588632184, "loss": 0.7433, "step": 1095 }, { "epoch": 3.4482758620689653, "grad_norm": 1.6015625, "learning_rate": 0.0001899106539240984, "loss": 0.7767, "step": 1100 }, { "epoch": 3.4639498432601883, "grad_norm": 33.75, "learning_rate": 0.00018975037549032086, "loss": 0.755, "step": 1105 }, { "epoch": 3.479623824451411, "grad_norm": 0.8515625, "learning_rate": 0.0001895889027178687, "loss": 0.7631, "step": 1110 }, { "epoch": 3.4952978056426334, "grad_norm": 1.1015625, "learning_rate": 0.0001894262377555148, "loss": 0.7545, "step": 1115 }, { "epoch": 3.510971786833856, "grad_norm": 0.9765625, "learning_rate": 0.00018926238276789704, "loss": 0.7491, "step": 1120 }, { "epoch": 3.5266457680250785, "grad_norm": 1.0234375, "learning_rate": 0.0001890973399354892, "loss": 0.7663, "step": 1125 }, { "epoch": 3.542319749216301, "grad_norm": 1.0703125, "learning_rate": 0.00018893111145457225, "loss": 0.755, "step": 1130 }, { "epoch": 3.5579937304075235, "grad_norm": 1.8046875, "learning_rate": 0.00018876369953720496, "loss": 0.7681, "step": 1135 }, { "epoch": 3.573667711598746, "grad_norm": 5.125, "learning_rate": 0.00018859510641119448, "loss": 0.766, "step": 1140 }, { "epoch": 3.5893416927899686, "grad_norm": 1.3828125, "learning_rate": 0.00018842533432006662, "loss": 0.7801, "step": 1145 }, { "epoch": 3.605015673981191, "grad_norm": 1.0625, "learning_rate": 0.00018825438552303621, "loss": 0.7647, "step": 1150 }, { "epoch": 3.6206896551724137, "grad_norm": 1.109375, "learning_rate": 0.00018808226229497684, "loss": 0.7768, "step": 1155 }, { "epoch": 3.6363636363636362, "grad_norm": 1.4921875, "learning_rate": 0.00018790896692639068, "loss": 0.7786, "step": 1160 }, { "epoch": 3.652037617554859, "grad_norm": 1.2890625, "learning_rate": 0.00018773450172337793, "loss": 0.762, "step": 1165 }, { "epoch": 3.6677115987460818, "grad_norm": 1.0078125, "learning_rate": 0.00018755886900760619, "loss": 0.7612, "step": 1170 }, { "epoch": 3.683385579937304, "grad_norm": 0.81640625, "learning_rate": 0.00018738207111627958, "loss": 0.7718, "step": 1175 }, { "epoch": 3.699059561128527, "grad_norm": 0.89453125, "learning_rate": 0.00018720411040210752, "loss": 0.7577, "step": 1180 }, { "epoch": 3.714733542319749, "grad_norm": 0.8828125, "learning_rate": 0.00018702498923327366, "loss": 0.7429, "step": 1185 }, { "epoch": 3.730407523510972, "grad_norm": 0.828125, "learning_rate": 0.00018684470999340405, "loss": 0.7552, "step": 1190 }, { "epoch": 3.7460815047021945, "grad_norm": 1.46875, "learning_rate": 0.00018666327508153567, "loss": 0.7606, "step": 1195 }, { "epoch": 3.761755485893417, "grad_norm": 0.89453125, "learning_rate": 0.0001864806869120844, "loss": 0.7678, "step": 1200 }, { "epoch": 3.7774294670846396, "grad_norm": 1.078125, "learning_rate": 0.00018629694791481296, "loss": 0.7985, "step": 1205 }, { "epoch": 3.793103448275862, "grad_norm": 5.125, "learning_rate": 0.00018611206053479842, "loss": 0.7712, "step": 1210 }, { "epoch": 3.8087774294670846, "grad_norm": 1.78125, "learning_rate": 0.00018592602723239984, "loss": 0.7745, "step": 1215 }, { "epoch": 3.824451410658307, "grad_norm": 0.89453125, "learning_rate": 0.00018573885048322547, "loss": 0.7684, "step": 1220 }, { "epoch": 3.8401253918495297, "grad_norm": 0.83984375, "learning_rate": 0.00018555053277809975, "loss": 0.7811, "step": 1225 }, { "epoch": 3.8557993730407523, "grad_norm": 1.1484375, "learning_rate": 0.00018536107662303026, "loss": 0.7732, "step": 1230 }, { "epoch": 3.871473354231975, "grad_norm": 1.015625, "learning_rate": 0.00018517048453917424, "loss": 0.7668, "step": 1235 }, { "epoch": 3.8871473354231973, "grad_norm": 0.921875, "learning_rate": 0.00018497875906280515, "loss": 0.759, "step": 1240 }, { "epoch": 3.9028213166144203, "grad_norm": 1.46875, "learning_rate": 0.00018478590274527898, "loss": 0.7763, "step": 1245 }, { "epoch": 3.9184952978056424, "grad_norm": 1.6171875, "learning_rate": 0.0001845919181530001, "loss": 0.7633, "step": 1250 }, { "epoch": 3.9341692789968654, "grad_norm": 0.7265625, "learning_rate": 0.00018439680786738722, "loss": 0.7853, "step": 1255 }, { "epoch": 3.9498432601880875, "grad_norm": 1.46875, "learning_rate": 0.00018420057448483905, "loss": 0.7856, "step": 1260 }, { "epoch": 3.9655172413793105, "grad_norm": 1.359375, "learning_rate": 0.00018400322061669982, "loss": 0.7831, "step": 1265 }, { "epoch": 3.981191222570533, "grad_norm": 0.75390625, "learning_rate": 0.00018380474888922426, "loss": 0.7952, "step": 1270 }, { "epoch": 3.9968652037617556, "grad_norm": 1.0625, "learning_rate": 0.000183605161943543, "loss": 0.7735, "step": 1275 }, { "epoch": 4.0, "eval_loss": 1.923946738243103, "eval_runtime": 0.8, "eval_samples_per_second": 2.5, "eval_steps_per_second": 1.25, "step": 1276 }, { "epoch": 4.012539184952978, "grad_norm": 0.84375, "learning_rate": 0.0001834044624356272, "loss": 0.6611, "step": 1280 }, { "epoch": 4.028213166144201, "grad_norm": 0.8984375, "learning_rate": 0.0001832026530362532, "loss": 0.5993, "step": 1285 }, { "epoch": 4.043887147335423, "grad_norm": 1.3125, "learning_rate": 0.00018299973643096714, "loss": 0.6197, "step": 1290 }, { "epoch": 4.059561128526646, "grad_norm": 1.1640625, "learning_rate": 0.00018279571532004907, "loss": 0.6147, "step": 1295 }, { "epoch": 4.075235109717869, "grad_norm": 1.1171875, "learning_rate": 0.00018259059241847707, "loss": 0.6295, "step": 1300 }, { "epoch": 4.090909090909091, "grad_norm": 0.97265625, "learning_rate": 0.00018238437045589115, "loss": 0.6219, "step": 1305 }, { "epoch": 4.106583072100314, "grad_norm": 0.98828125, "learning_rate": 0.00018217705217655689, "loss": 0.6033, "step": 1310 }, { "epoch": 4.122257053291536, "grad_norm": 0.875, "learning_rate": 0.0001819686403393289, "loss": 0.622, "step": 1315 }, { "epoch": 4.137931034482759, "grad_norm": 0.890625, "learning_rate": 0.00018175913771761417, "loss": 0.6166, "step": 1320 }, { "epoch": 4.153605015673981, "grad_norm": 1.234375, "learning_rate": 0.0001815485470993351, "loss": 0.6335, "step": 1325 }, { "epoch": 4.169278996865204, "grad_norm": 0.88671875, "learning_rate": 0.00018133687128689242, "loss": 0.6204, "step": 1330 }, { "epoch": 4.184952978056426, "grad_norm": 1.09375, "learning_rate": 0.000181124113097128, "loss": 0.635, "step": 1335 }, { "epoch": 4.200626959247649, "grad_norm": 0.83203125, "learning_rate": 0.00018091027536128716, "loss": 0.6224, "step": 1340 }, { "epoch": 4.216300940438871, "grad_norm": 0.890625, "learning_rate": 0.00018069536092498112, "loss": 0.6314, "step": 1345 }, { "epoch": 4.231974921630094, "grad_norm": 0.80859375, "learning_rate": 0.00018047937264814917, "loss": 0.6421, "step": 1350 }, { "epoch": 4.247648902821316, "grad_norm": 2.25, "learning_rate": 0.00018026231340502057, "loss": 0.6335, "step": 1355 }, { "epoch": 4.263322884012539, "grad_norm": 1.0, "learning_rate": 0.00018004418608407626, "loss": 0.6365, "step": 1360 }, { "epoch": 4.278996865203762, "grad_norm": 1.2421875, "learning_rate": 0.00017982499358801037, "loss": 0.6289, "step": 1365 }, { "epoch": 4.294670846394984, "grad_norm": 0.95703125, "learning_rate": 0.00017960473883369186, "loss": 0.6297, "step": 1370 }, { "epoch": 4.310344827586207, "grad_norm": 1.484375, "learning_rate": 0.00017938342475212532, "loss": 0.6496, "step": 1375 }, { "epoch": 4.326018808777429, "grad_norm": 0.9140625, "learning_rate": 0.00017916105428841234, "loss": 0.6454, "step": 1380 }, { "epoch": 4.341692789968652, "grad_norm": 1.21875, "learning_rate": 0.00017893763040171203, "loss": 0.6452, "step": 1385 }, { "epoch": 4.3573667711598745, "grad_norm": 2.25, "learning_rate": 0.00017871315606520183, "loss": 0.6542, "step": 1390 }, { "epoch": 4.3730407523510975, "grad_norm": 1.21875, "learning_rate": 0.0001784876342660378, "loss": 0.6448, "step": 1395 }, { "epoch": 4.38871473354232, "grad_norm": 0.95703125, "learning_rate": 0.00017826106800531498, "loss": 0.63, "step": 1400 }, { "epoch": 4.4043887147335425, "grad_norm": 1.828125, "learning_rate": 0.0001780334602980275, "loss": 0.6317, "step": 1405 }, { "epoch": 4.420062695924765, "grad_norm": 1.75, "learning_rate": 0.0001778048141730282, "loss": 0.6375, "step": 1410 }, { "epoch": 4.435736677115988, "grad_norm": 1.046875, "learning_rate": 0.00017757513267298856, "loss": 0.645, "step": 1415 }, { "epoch": 4.45141065830721, "grad_norm": 1.6328125, "learning_rate": 0.00017734441885435828, "loss": 0.6431, "step": 1420 }, { "epoch": 4.467084639498433, "grad_norm": 0.9296875, "learning_rate": 0.00017711267578732423, "loss": 0.6494, "step": 1425 }, { "epoch": 4.482758620689655, "grad_norm": 1.2421875, "learning_rate": 0.00017687990655577008, "loss": 0.6464, "step": 1430 }, { "epoch": 4.498432601880878, "grad_norm": 33.75, "learning_rate": 0.00017664611425723486, "loss": 0.6598, "step": 1435 }, { "epoch": 4.514106583072101, "grad_norm": 1.5234375, "learning_rate": 0.00017641130200287197, "loss": 0.6602, "step": 1440 }, { "epoch": 4.529780564263323, "grad_norm": 1.34375, "learning_rate": 0.00017617547291740767, "loss": 0.6443, "step": 1445 }, { "epoch": 4.545454545454545, "grad_norm": 1.2890625, "learning_rate": 0.00017593863013909956, "loss": 0.6441, "step": 1450 }, { "epoch": 4.561128526645768, "grad_norm": 1.375, "learning_rate": 0.00017570077681969474, "loss": 0.6405, "step": 1455 }, { "epoch": 4.576802507836991, "grad_norm": 2.15625, "learning_rate": 0.00017546191612438804, "loss": 0.6605, "step": 1460 }, { "epoch": 4.592476489028213, "grad_norm": 1.1796875, "learning_rate": 0.0001752220512317797, "loss": 0.6572, "step": 1465 }, { "epoch": 4.608150470219436, "grad_norm": 1.96875, "learning_rate": 0.00017498118533383316, "loss": 0.6396, "step": 1470 }, { "epoch": 4.623824451410658, "grad_norm": 1.2265625, "learning_rate": 0.0001747393216358326, "loss": 0.656, "step": 1475 }, { "epoch": 4.639498432601881, "grad_norm": 1.0390625, "learning_rate": 0.00017449646335634017, "loss": 0.6602, "step": 1480 }, { "epoch": 4.655172413793103, "grad_norm": 0.8203125, "learning_rate": 0.00017425261372715345, "loss": 0.6519, "step": 1485 }, { "epoch": 4.670846394984326, "grad_norm": 0.875, "learning_rate": 0.00017400777599326203, "loss": 0.6475, "step": 1490 }, { "epoch": 4.686520376175548, "grad_norm": 0.84765625, "learning_rate": 0.00017376195341280468, "loss": 0.6656, "step": 1495 }, { "epoch": 4.702194357366771, "grad_norm": 0.88671875, "learning_rate": 0.00017351514925702583, "loss": 0.6655, "step": 1500 }, { "epoch": 4.717868338557993, "grad_norm": 1.4296875, "learning_rate": 0.00017326736681023204, "loss": 0.672, "step": 1505 }, { "epoch": 4.733542319749216, "grad_norm": 0.98828125, "learning_rate": 0.00017301860936974834, "loss": 0.6635, "step": 1510 }, { "epoch": 4.749216300940439, "grad_norm": 1.15625, "learning_rate": 0.00017276888024587433, "loss": 0.6666, "step": 1515 }, { "epoch": 4.764890282131661, "grad_norm": 1.265625, "learning_rate": 0.00017251818276184012, "loss": 0.6692, "step": 1520 }, { "epoch": 4.7805642633228835, "grad_norm": 0.80859375, "learning_rate": 0.0001722665202537621, "loss": 0.6578, "step": 1525 }, { "epoch": 4.7962382445141065, "grad_norm": 0.92578125, "learning_rate": 0.00017201389607059863, "loss": 0.6607, "step": 1530 }, { "epoch": 4.8119122257053295, "grad_norm": 2.140625, "learning_rate": 0.00017176031357410537, "loss": 0.6538, "step": 1535 }, { "epoch": 4.827586206896552, "grad_norm": 1.828125, "learning_rate": 0.0001715057761387905, "loss": 0.6703, "step": 1540 }, { "epoch": 4.843260188087775, "grad_norm": 0.87109375, "learning_rate": 0.00017125028715187, "loss": 0.6761, "step": 1545 }, { "epoch": 4.858934169278997, "grad_norm": 0.83984375, "learning_rate": 0.0001709938500132225, "loss": 0.6616, "step": 1550 }, { "epoch": 4.87460815047022, "grad_norm": 1.015625, "learning_rate": 0.00017073646813534388, "loss": 0.6597, "step": 1555 }, { "epoch": 4.890282131661442, "grad_norm": 0.8984375, "learning_rate": 0.00017047814494330207, "loss": 0.6733, "step": 1560 }, { "epoch": 4.905956112852665, "grad_norm": 1.859375, "learning_rate": 0.00017021888387469135, "loss": 0.6737, "step": 1565 }, { "epoch": 4.921630094043887, "grad_norm": 0.890625, "learning_rate": 0.00016995868837958665, "loss": 0.6736, "step": 1570 }, { "epoch": 4.93730407523511, "grad_norm": 0.921875, "learning_rate": 0.0001696975619204977, "loss": 0.68, "step": 1575 }, { "epoch": 4.952978056426332, "grad_norm": 2.0, "learning_rate": 0.0001694355079723227, "loss": 0.6755, "step": 1580 }, { "epoch": 4.968652037617555, "grad_norm": 0.78125, "learning_rate": 0.0001691725300223025, "loss": 0.6827, "step": 1585 }, { "epoch": 4.984326018808778, "grad_norm": 0.859375, "learning_rate": 0.0001689086315699738, "loss": 0.6681, "step": 1590 }, { "epoch": 5.0, "grad_norm": 0.91015625, "learning_rate": 0.00016864381612712276, "loss": 0.6599, "step": 1595 }, { "epoch": 5.0, "eval_loss": 2.0616867542266846, "eval_runtime": 0.7937, "eval_samples_per_second": 2.52, "eval_steps_per_second": 1.26, "step": 1595 }, { "epoch": 5.015673981191223, "grad_norm": 1.9296875, "learning_rate": 0.00016837808721773827, "loss": 0.5239, "step": 1600 }, { "epoch": 5.031347962382445, "grad_norm": 1.125, "learning_rate": 0.000168111448377965, "loss": 0.509, "step": 1605 }, { "epoch": 5.047021943573668, "grad_norm": 2.5625, "learning_rate": 0.0001678439031560564, "loss": 0.5041, "step": 1610 }, { "epoch": 5.06269592476489, "grad_norm": 0.8203125, "learning_rate": 0.00016757545511232746, "loss": 0.5195, "step": 1615 }, { "epoch": 5.078369905956113, "grad_norm": 0.8984375, "learning_rate": 0.00016730610781910728, "loss": 0.4963, "step": 1620 }, { "epoch": 5.094043887147335, "grad_norm": 0.88671875, "learning_rate": 0.00016703586486069164, "loss": 0.5159, "step": 1625 }, { "epoch": 5.109717868338558, "grad_norm": 1.0859375, "learning_rate": 0.0001667647298332952, "loss": 0.5111, "step": 1630 }, { "epoch": 5.12539184952978, "grad_norm": 0.9765625, "learning_rate": 0.00016649270634500366, "loss": 0.5309, "step": 1635 }, { "epoch": 5.141065830721003, "grad_norm": 0.9296875, "learning_rate": 0.00016621979801572585, "loss": 0.5254, "step": 1640 }, { "epoch": 5.156739811912225, "grad_norm": 0.8359375, "learning_rate": 0.00016594600847714538, "loss": 0.5274, "step": 1645 }, { "epoch": 5.172413793103448, "grad_norm": 0.8984375, "learning_rate": 0.0001656713413726725, "loss": 0.5239, "step": 1650 }, { "epoch": 5.1880877742946705, "grad_norm": 0.984375, "learning_rate": 0.00016539580035739547, "loss": 0.5246, "step": 1655 }, { "epoch": 5.2037617554858935, "grad_norm": 1.6953125, "learning_rate": 0.00016511938909803204, "loss": 0.5367, "step": 1660 }, { "epoch": 5.219435736677116, "grad_norm": 0.9453125, "learning_rate": 0.00016484211127288048, "loss": 0.5356, "step": 1665 }, { "epoch": 5.235109717868339, "grad_norm": 0.953125, "learning_rate": 0.00016456397057177085, "loss": 0.5367, "step": 1670 }, { "epoch": 5.250783699059561, "grad_norm": 1.25, "learning_rate": 0.00016428497069601578, "loss": 0.534, "step": 1675 }, { "epoch": 5.266457680250784, "grad_norm": 0.9296875, "learning_rate": 0.00016400511535836118, "loss": 0.5476, "step": 1680 }, { "epoch": 5.282131661442007, "grad_norm": 1.1015625, "learning_rate": 0.0001637244082829369, "loss": 0.544, "step": 1685 }, { "epoch": 5.297805642633229, "grad_norm": 0.84375, "learning_rate": 0.00016344285320520717, "loss": 0.5414, "step": 1690 }, { "epoch": 5.313479623824452, "grad_norm": 1.0, "learning_rate": 0.00016316045387192087, "loss": 0.5435, "step": 1695 }, { "epoch": 5.329153605015674, "grad_norm": 0.85546875, "learning_rate": 0.00016287721404106167, "loss": 0.5412, "step": 1700 }, { "epoch": 5.344827586206897, "grad_norm": 1.0, "learning_rate": 0.00016259313748179802, "loss": 0.5448, "step": 1705 }, { "epoch": 5.360501567398119, "grad_norm": 0.76171875, "learning_rate": 0.000162308227974433, "loss": 0.5523, "step": 1710 }, { "epoch": 5.376175548589342, "grad_norm": 1.015625, "learning_rate": 0.00016202248931035404, "loss": 0.5382, "step": 1715 }, { "epoch": 5.391849529780564, "grad_norm": 1.0546875, "learning_rate": 0.0001617359252919824, "loss": 0.5427, "step": 1720 }, { "epoch": 5.407523510971787, "grad_norm": 1.203125, "learning_rate": 0.00016144853973272262, "loss": 0.5426, "step": 1725 }, { "epoch": 5.423197492163009, "grad_norm": 1.09375, "learning_rate": 0.00016116033645691174, "loss": 0.5514, "step": 1730 }, { "epoch": 5.438871473354232, "grad_norm": 1.1953125, "learning_rate": 0.00016087131929976852, "loss": 0.5471, "step": 1735 }, { "epoch": 5.454545454545454, "grad_norm": 0.8046875, "learning_rate": 0.00016058149210734223, "loss": 0.5617, "step": 1740 }, { "epoch": 5.470219435736677, "grad_norm": 0.90234375, "learning_rate": 0.0001602908587364616, "loss": 0.5598, "step": 1745 }, { "epoch": 5.485893416927899, "grad_norm": 0.875, "learning_rate": 0.00015999942305468338, "loss": 0.5457, "step": 1750 }, { "epoch": 5.501567398119122, "grad_norm": 0.99609375, "learning_rate": 0.000159707188940241, "loss": 0.5667, "step": 1755 }, { "epoch": 5.517241379310345, "grad_norm": 2.90625, "learning_rate": 0.00015941416028199298, "loss": 0.552, "step": 1760 }, { "epoch": 5.532915360501567, "grad_norm": 0.8046875, "learning_rate": 0.00015912034097937094, "loss": 0.5408, "step": 1765 }, { "epoch": 5.54858934169279, "grad_norm": 0.953125, "learning_rate": 0.00015882573494232797, "loss": 0.5516, "step": 1770 }, { "epoch": 5.564263322884012, "grad_norm": 0.94921875, "learning_rate": 0.00015853034609128648, "loss": 0.5649, "step": 1775 }, { "epoch": 5.579937304075235, "grad_norm": 1.25, "learning_rate": 0.00015823417835708606, "loss": 0.5666, "step": 1780 }, { "epoch": 5.5956112852664575, "grad_norm": 1.515625, "learning_rate": 0.00015793723568093118, "loss": 0.5653, "step": 1785 }, { "epoch": 5.61128526645768, "grad_norm": 1.4296875, "learning_rate": 0.00015763952201433866, "loss": 0.5663, "step": 1790 }, { "epoch": 5.6269592476489025, "grad_norm": 0.84765625, "learning_rate": 0.00015734104131908522, "loss": 0.5582, "step": 1795 }, { "epoch": 5.6426332288401255, "grad_norm": 0.94140625, "learning_rate": 0.00015704179756715467, "loss": 0.5569, "step": 1800 }, { "epoch": 5.658307210031348, "grad_norm": 0.83203125, "learning_rate": 0.00015674179474068508, "loss": 0.5618, "step": 1805 }, { "epoch": 5.673981191222571, "grad_norm": 0.890625, "learning_rate": 0.00015644103683191575, "loss": 0.5636, "step": 1810 }, { "epoch": 5.689655172413794, "grad_norm": 0.80859375, "learning_rate": 0.00015613952784313418, "loss": 0.5562, "step": 1815 }, { "epoch": 5.705329153605016, "grad_norm": 1.7421875, "learning_rate": 0.00015583727178662262, "loss": 0.5566, "step": 1820 }, { "epoch": 5.721003134796238, "grad_norm": 1.4375, "learning_rate": 0.00015553427268460496, "loss": 0.5591, "step": 1825 }, { "epoch": 5.736677115987461, "grad_norm": 0.98046875, "learning_rate": 0.00015523053456919294, "loss": 0.5666, "step": 1830 }, { "epoch": 5.752351097178684, "grad_norm": 0.98828125, "learning_rate": 0.00015492606148233265, "loss": 0.5664, "step": 1835 }, { "epoch": 5.768025078369906, "grad_norm": 0.84375, "learning_rate": 0.00015462085747575068, "loss": 0.5624, "step": 1840 }, { "epoch": 5.783699059561129, "grad_norm": 0.85546875, "learning_rate": 0.00015431492661090022, "loss": 0.5587, "step": 1845 }, { "epoch": 5.799373040752351, "grad_norm": 0.99609375, "learning_rate": 0.00015400827295890702, "loss": 0.5645, "step": 1850 }, { "epoch": 5.815047021943574, "grad_norm": 1.03125, "learning_rate": 0.0001537009006005152, "loss": 0.5631, "step": 1855 }, { "epoch": 5.830721003134796, "grad_norm": 1.046875, "learning_rate": 0.0001533928136260329, "loss": 0.5712, "step": 1860 }, { "epoch": 5.846394984326019, "grad_norm": 1.140625, "learning_rate": 0.00015308401613527796, "loss": 0.5676, "step": 1865 }, { "epoch": 5.862068965517241, "grad_norm": 0.9375, "learning_rate": 0.00015277451223752326, "loss": 0.5761, "step": 1870 }, { "epoch": 5.877742946708464, "grad_norm": 0.98046875, "learning_rate": 0.00015246430605144216, "loss": 0.5685, "step": 1875 }, { "epoch": 5.893416927899686, "grad_norm": 0.83203125, "learning_rate": 0.00015215340170505348, "loss": 0.5678, "step": 1880 }, { "epoch": 5.909090909090909, "grad_norm": 1.6484375, "learning_rate": 0.0001518418033356668, "loss": 0.569, "step": 1885 }, { "epoch": 5.924764890282132, "grad_norm": 0.9765625, "learning_rate": 0.00015152951508982726, "loss": 0.5669, "step": 1890 }, { "epoch": 5.940438871473354, "grad_norm": 0.91015625, "learning_rate": 0.0001512165411232604, "loss": 0.5759, "step": 1895 }, { "epoch": 5.956112852664576, "grad_norm": 0.87890625, "learning_rate": 0.00015090288560081692, "loss": 0.5641, "step": 1900 }, { "epoch": 5.971786833855799, "grad_norm": 1.1953125, "learning_rate": 0.0001505885526964172, "loss": 0.574, "step": 1905 }, { "epoch": 5.987460815047022, "grad_norm": 1.078125, "learning_rate": 0.00015027354659299578, "loss": 0.5764, "step": 1910 }, { "epoch": 6.0, "eval_loss": 2.323450803756714, "eval_runtime": 0.8121, "eval_samples_per_second": 2.463, "eval_steps_per_second": 1.231, "step": 1914 }, { "epoch": 6.003134796238244, "grad_norm": 0.96484375, "learning_rate": 0.00014995787148244563, "loss": 0.5514, "step": 1915 }, { "epoch": 6.018808777429467, "grad_norm": 1.1484375, "learning_rate": 0.00014964153156556245, "loss": 0.4392, "step": 1920 }, { "epoch": 6.0344827586206895, "grad_norm": 1.015625, "learning_rate": 0.00014932453105198884, "loss": 0.4499, "step": 1925 }, { "epoch": 6.0501567398119125, "grad_norm": 1.0625, "learning_rate": 0.0001490068741601581, "loss": 0.4379, "step": 1930 }, { "epoch": 6.065830721003135, "grad_norm": 0.84765625, "learning_rate": 0.00014868856511723814, "loss": 0.431, "step": 1935 }, { "epoch": 6.081504702194358, "grad_norm": 0.98046875, "learning_rate": 0.00014836960815907532, "loss": 0.4428, "step": 1940 }, { "epoch": 6.09717868338558, "grad_norm": 0.89453125, "learning_rate": 0.000148050007530138, "loss": 0.4482, "step": 1945 }, { "epoch": 6.112852664576803, "grad_norm": 1.453125, "learning_rate": 0.00014772976748346015, "loss": 0.4478, "step": 1950 }, { "epoch": 6.128526645768025, "grad_norm": 2.453125, "learning_rate": 0.00014740889228058462, "loss": 0.4414, "step": 1955 }, { "epoch": 6.144200626959248, "grad_norm": 1.1328125, "learning_rate": 0.0001470873861915065, "loss": 0.4466, "step": 1960 }, { "epoch": 6.15987460815047, "grad_norm": 1.2109375, "learning_rate": 0.00014676525349461637, "loss": 0.4356, "step": 1965 }, { "epoch": 6.175548589341693, "grad_norm": 0.8984375, "learning_rate": 0.00014644249847664317, "loss": 0.4526, "step": 1970 }, { "epoch": 6.191222570532915, "grad_norm": 1.1796875, "learning_rate": 0.00014611912543259742, "loss": 0.442, "step": 1975 }, { "epoch": 6.206896551724138, "grad_norm": 1.21875, "learning_rate": 0.00014579513866571378, "loss": 0.4463, "step": 1980 }, { "epoch": 6.222570532915361, "grad_norm": 0.875, "learning_rate": 0.00014547054248739404, "loss": 0.444, "step": 1985 }, { "epoch": 6.238244514106583, "grad_norm": 0.96484375, "learning_rate": 0.0001451453412171496, "loss": 0.4513, "step": 1990 }, { "epoch": 6.253918495297806, "grad_norm": 0.9140625, "learning_rate": 0.000144819539182544, "loss": 0.4494, "step": 1995 }, { "epoch": 6.269592476489028, "grad_norm": 0.92578125, "learning_rate": 0.00014449314071913533, "loss": 0.4493, "step": 2000 }, { "epoch": 6.285266457680251, "grad_norm": 0.95703125, "learning_rate": 0.00014416615017041868, "loss": 0.4605, "step": 2005 }, { "epoch": 6.300940438871473, "grad_norm": 0.8515625, "learning_rate": 0.00014383857188776807, "loss": 0.4642, "step": 2010 }, { "epoch": 6.316614420062696, "grad_norm": 0.91015625, "learning_rate": 0.00014351041023037884, "loss": 0.4526, "step": 2015 }, { "epoch": 6.332288401253918, "grad_norm": 0.87109375, "learning_rate": 0.00014318166956520936, "loss": 0.4634, "step": 2020 }, { "epoch": 6.347962382445141, "grad_norm": 0.89453125, "learning_rate": 0.00014285235426692315, "loss": 0.4608, "step": 2025 }, { "epoch": 6.363636363636363, "grad_norm": 0.9296875, "learning_rate": 0.00014252246871783051, "loss": 0.4588, "step": 2030 }, { "epoch": 6.379310344827586, "grad_norm": 0.83203125, "learning_rate": 0.00014219201730783024, "loss": 0.467, "step": 2035 }, { "epoch": 6.394984326018808, "grad_norm": 0.97265625, "learning_rate": 0.0001418610044343514, "loss": 0.4641, "step": 2040 }, { "epoch": 6.410658307210031, "grad_norm": 0.9609375, "learning_rate": 0.00014152943450229443, "loss": 0.4647, "step": 2045 }, { "epoch": 6.4263322884012535, "grad_norm": 0.875, "learning_rate": 0.00014119731192397284, "loss": 0.4627, "step": 2050 }, { "epoch": 6.4420062695924765, "grad_norm": 1.1796875, "learning_rate": 0.0001408646411190544, "loss": 0.4664, "step": 2055 }, { "epoch": 6.4576802507836994, "grad_norm": 1.0625, "learning_rate": 0.0001405314265145023, "loss": 0.4697, "step": 2060 }, { "epoch": 6.4733542319749215, "grad_norm": 0.95703125, "learning_rate": 0.0001401976725445162, "loss": 0.4664, "step": 2065 }, { "epoch": 6.4890282131661445, "grad_norm": 0.9375, "learning_rate": 0.0001398633836504734, "loss": 0.4692, "step": 2070 }, { "epoch": 6.504702194357367, "grad_norm": 0.8828125, "learning_rate": 0.00013952856428086952, "loss": 0.4658, "step": 2075 }, { "epoch": 6.52037617554859, "grad_norm": 0.87109375, "learning_rate": 0.00013919321889125941, "loss": 0.4711, "step": 2080 }, { "epoch": 6.536050156739812, "grad_norm": 0.859375, "learning_rate": 0.0001388573519441979, "loss": 0.4675, "step": 2085 }, { "epoch": 6.551724137931035, "grad_norm": 0.94140625, "learning_rate": 0.00013852096790918026, "loss": 0.4677, "step": 2090 }, { "epoch": 6.567398119122257, "grad_norm": 0.9140625, "learning_rate": 0.00013818407126258293, "loss": 0.4728, "step": 2095 }, { "epoch": 6.58307210031348, "grad_norm": 2.75, "learning_rate": 0.0001378466664876038, "loss": 0.4769, "step": 2100 }, { "epoch": 6.598746081504702, "grad_norm": 2.203125, "learning_rate": 0.00013750875807420259, "loss": 0.463, "step": 2105 }, { "epoch": 6.614420062695925, "grad_norm": 1.1015625, "learning_rate": 0.00013717035051904114, "loss": 0.4663, "step": 2110 }, { "epoch": 6.630094043887147, "grad_norm": 0.953125, "learning_rate": 0.00013683144832542352, "loss": 0.4699, "step": 2115 }, { "epoch": 6.64576802507837, "grad_norm": 0.96875, "learning_rate": 0.00013649205600323609, "loss": 0.4703, "step": 2120 }, { "epoch": 6.661442006269592, "grad_norm": 0.91015625, "learning_rate": 0.00013615217806888755, "loss": 0.4643, "step": 2125 }, { "epoch": 6.677115987460815, "grad_norm": 1.078125, "learning_rate": 0.0001358118190452488, "loss": 0.4669, "step": 2130 }, { "epoch": 6.692789968652038, "grad_norm": 0.9375, "learning_rate": 0.0001354709834615928, "loss": 0.4703, "step": 2135 }, { "epoch": 6.70846394984326, "grad_norm": 0.875, "learning_rate": 0.00013512967585353413, "loss": 0.4714, "step": 2140 }, { "epoch": 6.724137931034483, "grad_norm": 0.95703125, "learning_rate": 0.00013478790076296892, "loss": 0.4658, "step": 2145 }, { "epoch": 6.739811912225705, "grad_norm": 0.97265625, "learning_rate": 0.00013444566273801414, "loss": 0.4649, "step": 2150 }, { "epoch": 6.755485893416928, "grad_norm": 0.83984375, "learning_rate": 0.00013410296633294727, "loss": 0.4783, "step": 2155 }, { "epoch": 6.77115987460815, "grad_norm": 1.171875, "learning_rate": 0.00013375981610814545, "loss": 0.4741, "step": 2160 }, { "epoch": 6.786833855799373, "grad_norm": 0.89453125, "learning_rate": 0.00013341621663002514, "loss": 0.4651, "step": 2165 }, { "epoch": 6.802507836990595, "grad_norm": 0.9765625, "learning_rate": 0.0001330721724709811, "loss": 0.4682, "step": 2170 }, { "epoch": 6.818181818181818, "grad_norm": 0.8671875, "learning_rate": 0.00013272768820932554, "loss": 0.4761, "step": 2175 }, { "epoch": 6.83385579937304, "grad_norm": 0.953125, "learning_rate": 0.0001323827684292273, "loss": 0.467, "step": 2180 }, { "epoch": 6.849529780564263, "grad_norm": 0.9453125, "learning_rate": 0.0001320374177206509, "loss": 0.4719, "step": 2185 }, { "epoch": 6.8652037617554855, "grad_norm": 0.87890625, "learning_rate": 0.00013169164067929526, "loss": 0.4829, "step": 2190 }, { "epoch": 6.8808777429467085, "grad_norm": 0.9140625, "learning_rate": 0.00013134544190653274, "loss": 0.4743, "step": 2195 }, { "epoch": 6.896551724137931, "grad_norm": 0.78515625, "learning_rate": 0.00013099882600934773, "loss": 0.4701, "step": 2200 }, { "epoch": 6.912225705329154, "grad_norm": 0.94140625, "learning_rate": 0.00013065179760027556, "loss": 0.4703, "step": 2205 }, { "epoch": 6.927899686520377, "grad_norm": 0.89453125, "learning_rate": 0.00013030436129734082, "loss": 0.4802, "step": 2210 }, { "epoch": 6.943573667711599, "grad_norm": 1.2265625, "learning_rate": 0.00012995652172399623, "loss": 0.4781, "step": 2215 }, { "epoch": 6.959247648902822, "grad_norm": 1.265625, "learning_rate": 0.00012960828350906095, "loss": 0.4838, "step": 2220 }, { "epoch": 6.974921630094044, "grad_norm": 1.546875, "learning_rate": 0.00012925965128665897, "loss": 0.4751, "step": 2225 }, { "epoch": 6.990595611285267, "grad_norm": 0.921875, "learning_rate": 0.0001289106296961574, "loss": 0.4817, "step": 2230 }, { "epoch": 7.0, "eval_loss": 2.6759016513824463, "eval_runtime": 0.7898, "eval_samples_per_second": 2.532, "eval_steps_per_second": 1.266, "step": 2233 }, { "epoch": 7.006269592476489, "grad_norm": 0.78125, "learning_rate": 0.00012856122338210493, "loss": 0.4256, "step": 2235 }, { "epoch": 7.021943573667712, "grad_norm": 0.8828125, "learning_rate": 0.00012821143699416984, "loss": 0.3775, "step": 2240 }, { "epoch": 7.037617554858934, "grad_norm": 0.73828125, "learning_rate": 0.00012786127518707818, "loss": 0.3705, "step": 2245 }, { "epoch": 7.053291536050157, "grad_norm": 0.859375, "learning_rate": 0.00012751074262055178, "loss": 0.3732, "step": 2250 }, { "epoch": 7.068965517241379, "grad_norm": 0.86328125, "learning_rate": 0.00012715984395924643, "loss": 0.3737, "step": 2255 }, { "epoch": 7.084639498432602, "grad_norm": 0.9296875, "learning_rate": 0.00012680858387268952, "loss": 0.3788, "step": 2260 }, { "epoch": 7.100313479623824, "grad_norm": 0.890625, "learning_rate": 0.00012645696703521818, "loss": 0.3711, "step": 2265 }, { "epoch": 7.115987460815047, "grad_norm": 1.015625, "learning_rate": 0.00012610499812591673, "loss": 0.3725, "step": 2270 }, { "epoch": 7.131661442006269, "grad_norm": 0.90234375, "learning_rate": 0.0001257526818285549, "loss": 0.371, "step": 2275 }, { "epoch": 7.147335423197492, "grad_norm": 0.9296875, "learning_rate": 0.0001254000228315251, "loss": 0.3751, "step": 2280 }, { "epoch": 7.163009404388715, "grad_norm": 0.921875, "learning_rate": 0.00012504702582778008, "loss": 0.3798, "step": 2285 }, { "epoch": 7.178683385579937, "grad_norm": 0.765625, "learning_rate": 0.00012469369551477074, "loss": 0.3826, "step": 2290 }, { "epoch": 7.19435736677116, "grad_norm": 0.8671875, "learning_rate": 0.0001243400365943833, "loss": 0.3846, "step": 2295 }, { "epoch": 7.210031347962382, "grad_norm": 0.87890625, "learning_rate": 0.00012398605377287694, "loss": 0.383, "step": 2300 }, { "epoch": 7.225705329153605, "grad_norm": 0.78125, "learning_rate": 0.000123631751760821, "loss": 0.387, "step": 2305 }, { "epoch": 7.241379310344827, "grad_norm": 0.796875, "learning_rate": 0.00012327713527303255, "loss": 0.3752, "step": 2310 }, { "epoch": 7.25705329153605, "grad_norm": 0.9453125, "learning_rate": 0.0001229222090285134, "loss": 0.3832, "step": 2315 }, { "epoch": 7.2727272727272725, "grad_norm": 0.84765625, "learning_rate": 0.00012256697775038741, "loss": 0.39, "step": 2320 }, { "epoch": 7.2884012539184955, "grad_norm": 0.828125, "learning_rate": 0.00012221144616583765, "loss": 0.3902, "step": 2325 }, { "epoch": 7.304075235109718, "grad_norm": 0.92578125, "learning_rate": 0.00012185561900604341, "loss": 0.376, "step": 2330 }, { "epoch": 7.3197492163009406, "grad_norm": 0.84765625, "learning_rate": 0.00012149950100611738, "loss": 0.3873, "step": 2335 }, { "epoch": 7.335423197492163, "grad_norm": 0.80078125, "learning_rate": 0.00012114309690504249, "loss": 0.388, "step": 2340 }, { "epoch": 7.351097178683386, "grad_norm": 0.88671875, "learning_rate": 0.00012078641144560898, "loss": 0.3889, "step": 2345 }, { "epoch": 7.366771159874608, "grad_norm": 0.796875, "learning_rate": 0.00012042944937435116, "loss": 0.395, "step": 2350 }, { "epoch": 7.382445141065831, "grad_norm": 0.84765625, "learning_rate": 0.00012007221544148435, "loss": 0.3957, "step": 2355 }, { "epoch": 7.398119122257054, "grad_norm": 0.80078125, "learning_rate": 0.00011971471440084157, "loss": 0.393, "step": 2360 }, { "epoch": 7.413793103448276, "grad_norm": 0.91015625, "learning_rate": 0.00011935695100981041, "loss": 0.3884, "step": 2365 }, { "epoch": 7.429467084639499, "grad_norm": 0.84375, "learning_rate": 0.00011899893002926958, "loss": 0.3907, "step": 2370 }, { "epoch": 7.445141065830721, "grad_norm": 0.84375, "learning_rate": 0.00011864065622352568, "loss": 0.3865, "step": 2375 }, { "epoch": 7.460815047021944, "grad_norm": 0.94140625, "learning_rate": 0.00011828213436024968, "loss": 0.3866, "step": 2380 }, { "epoch": 7.476489028213166, "grad_norm": 0.83984375, "learning_rate": 0.00011792336921041359, "loss": 0.3878, "step": 2385 }, { "epoch": 7.492163009404389, "grad_norm": 0.9140625, "learning_rate": 0.00011756436554822685, "loss": 0.3933, "step": 2390 }, { "epoch": 7.507836990595611, "grad_norm": 0.89453125, "learning_rate": 0.00011720512815107292, "loss": 0.3825, "step": 2395 }, { "epoch": 7.523510971786834, "grad_norm": 0.82421875, "learning_rate": 0.00011684566179944567, "loss": 0.3892, "step": 2400 }, { "epoch": 7.539184952978056, "grad_norm": 0.83203125, "learning_rate": 0.00011648597127688567, "loss": 0.3918, "step": 2405 }, { "epoch": 7.554858934169279, "grad_norm": 0.8515625, "learning_rate": 0.00011612606136991665, "loss": 0.3952, "step": 2410 }, { "epoch": 7.570532915360501, "grad_norm": 0.84375, "learning_rate": 0.00011576593686798181, "loss": 0.3941, "step": 2415 }, { "epoch": 7.586206896551724, "grad_norm": 1.1015625, "learning_rate": 0.00011540560256337995, "loss": 0.3996, "step": 2420 }, { "epoch": 7.601880877742946, "grad_norm": 0.921875, "learning_rate": 0.00011504506325120184, "loss": 0.3973, "step": 2425 }, { "epoch": 7.617554858934169, "grad_norm": 0.84765625, "learning_rate": 0.0001146843237292663, "loss": 0.3944, "step": 2430 }, { "epoch": 7.633228840125392, "grad_norm": 0.921875, "learning_rate": 0.0001143233887980565, "loss": 0.4008, "step": 2435 }, { "epoch": 7.648902821316614, "grad_norm": 0.83203125, "learning_rate": 0.00011396226326065593, "loss": 0.3967, "step": 2440 }, { "epoch": 7.664576802507837, "grad_norm": 0.87890625, "learning_rate": 0.00011360095192268454, "loss": 0.3979, "step": 2445 }, { "epoch": 7.6802507836990594, "grad_norm": 0.84765625, "learning_rate": 0.00011323945959223477, "loss": 0.3957, "step": 2450 }, { "epoch": 7.695924764890282, "grad_norm": 0.8984375, "learning_rate": 0.00011287779107980766, "loss": 0.398, "step": 2455 }, { "epoch": 7.7115987460815045, "grad_norm": 0.88671875, "learning_rate": 0.0001125159511982487, "loss": 0.3976, "step": 2460 }, { "epoch": 7.7272727272727275, "grad_norm": 0.8671875, "learning_rate": 0.00011215394476268387, "loss": 0.3941, "step": 2465 }, { "epoch": 7.74294670846395, "grad_norm": 1.0390625, "learning_rate": 0.00011179177659045554, "loss": 0.4061, "step": 2470 }, { "epoch": 7.758620689655173, "grad_norm": 1.03125, "learning_rate": 0.00011142945150105839, "loss": 0.3991, "step": 2475 }, { "epoch": 7.774294670846395, "grad_norm": 0.86328125, "learning_rate": 0.00011106697431607518, "loss": 0.3863, "step": 2480 }, { "epoch": 7.789968652037618, "grad_norm": 0.90625, "learning_rate": 0.00011070434985911271, "loss": 0.3923, "step": 2485 }, { "epoch": 7.80564263322884, "grad_norm": 0.84375, "learning_rate": 0.0001103415829557376, "loss": 0.3988, "step": 2490 }, { "epoch": 7.821316614420063, "grad_norm": 0.8515625, "learning_rate": 0.00010997867843341198, "loss": 0.3868, "step": 2495 }, { "epoch": 7.836990595611285, "grad_norm": 0.8984375, "learning_rate": 0.0001096156411214294, "loss": 0.3949, "step": 2500 }, { "epoch": 7.852664576802508, "grad_norm": 0.765625, "learning_rate": 0.00010925247585085044, "loss": 0.4005, "step": 2505 }, { "epoch": 7.868338557993731, "grad_norm": 0.84375, "learning_rate": 0.00010888918745443845, "loss": 0.3974, "step": 2510 }, { "epoch": 7.884012539184953, "grad_norm": 0.921875, "learning_rate": 0.00010852578076659535, "loss": 0.4004, "step": 2515 }, { "epoch": 7.899686520376176, "grad_norm": 0.80859375, "learning_rate": 0.00010816226062329706, "loss": 0.3997, "step": 2520 }, { "epoch": 7.915360501567398, "grad_norm": 0.91796875, "learning_rate": 0.00010779863186202943, "loss": 0.4002, "step": 2525 }, { "epoch": 7.931034482758621, "grad_norm": 0.92578125, "learning_rate": 0.00010743489932172366, "loss": 0.3973, "step": 2530 }, { "epoch": 7.946708463949843, "grad_norm": 0.9453125, "learning_rate": 0.00010707106784269196, "loss": 0.3968, "step": 2535 }, { "epoch": 7.962382445141066, "grad_norm": 0.953125, "learning_rate": 0.00010670714226656315, "loss": 0.397, "step": 2540 }, { "epoch": 7.978056426332288, "grad_norm": 0.90625, "learning_rate": 0.00010634312743621832, "loss": 0.4065, "step": 2545 }, { "epoch": 7.993730407523511, "grad_norm": 0.84375, "learning_rate": 0.00010597902819572619, "loss": 0.4016, "step": 2550 }, { "epoch": 8.0, "eval_loss": 3.0132577419281006, "eval_runtime": 0.7738, "eval_samples_per_second": 2.585, "eval_steps_per_second": 1.292, "step": 2552 }, { "epoch": 8.009404388714733, "grad_norm": 0.76953125, "learning_rate": 0.00010561484939027877, "loss": 0.361, "step": 2555 }, { "epoch": 8.025078369905955, "grad_norm": 0.88671875, "learning_rate": 0.00010525059586612693, "loss": 0.3253, "step": 2560 }, { "epoch": 8.04075235109718, "grad_norm": 0.765625, "learning_rate": 0.0001048862724705158, "loss": 0.3205, "step": 2565 }, { "epoch": 8.056426332288401, "grad_norm": 0.72265625, "learning_rate": 0.00010452188405162033, "loss": 0.3241, "step": 2570 }, { "epoch": 8.072100313479623, "grad_norm": 0.828125, "learning_rate": 0.00010415743545848072, "loss": 0.3193, "step": 2575 }, { "epoch": 8.087774294670846, "grad_norm": 0.79296875, "learning_rate": 0.00010379293154093796, "loss": 0.3212, "step": 2580 }, { "epoch": 8.10344827586207, "grad_norm": 0.80859375, "learning_rate": 0.00010342837714956928, "loss": 0.3161, "step": 2585 }, { "epoch": 8.119122257053291, "grad_norm": 0.765625, "learning_rate": 0.00010306377713562354, "loss": 0.321, "step": 2590 }, { "epoch": 8.134796238244514, "grad_norm": 0.83203125, "learning_rate": 0.00010269913635095676, "loss": 0.3245, "step": 2595 }, { "epoch": 8.150470219435737, "grad_norm": 0.83984375, "learning_rate": 0.00010233445964796749, "loss": 0.3279, "step": 2600 }, { "epoch": 8.16614420062696, "grad_norm": 0.765625, "learning_rate": 0.00010196975187953221, "loss": 0.3233, "step": 2605 }, { "epoch": 8.181818181818182, "grad_norm": 0.78515625, "learning_rate": 0.00010160501789894086, "loss": 0.3207, "step": 2610 }, { "epoch": 8.197492163009404, "grad_norm": 0.7890625, "learning_rate": 0.0001012402625598322, "loss": 0.3217, "step": 2615 }, { "epoch": 8.213166144200628, "grad_norm": 0.76953125, "learning_rate": 0.00010087549071612919, "loss": 0.3255, "step": 2620 }, { "epoch": 8.22884012539185, "grad_norm": 0.82421875, "learning_rate": 0.00010051070722197438, "loss": 0.3256, "step": 2625 }, { "epoch": 8.244514106583072, "grad_norm": 0.74609375, "learning_rate": 0.0001001459169316654, "loss": 0.3233, "step": 2630 }, { "epoch": 8.260188087774294, "grad_norm": 0.8046875, "learning_rate": 9.978112469959033e-05, "loss": 0.3232, "step": 2635 }, { "epoch": 8.275862068965518, "grad_norm": 0.80859375, "learning_rate": 9.941633538016315e-05, "loss": 0.3237, "step": 2640 }, { "epoch": 8.29153605015674, "grad_norm": 0.79296875, "learning_rate": 9.90515538277589e-05, "loss": 0.3307, "step": 2645 }, { "epoch": 8.307210031347962, "grad_norm": 0.765625, "learning_rate": 9.868678489664945e-05, "loss": 0.3228, "step": 2650 }, { "epoch": 8.322884012539184, "grad_norm": 0.75390625, "learning_rate": 9.832203344093855e-05, "loss": 0.3238, "step": 2655 }, { "epoch": 8.338557993730408, "grad_norm": 0.78125, "learning_rate": 9.795730431449759e-05, "loss": 0.3302, "step": 2660 }, { "epoch": 8.35423197492163, "grad_norm": 0.734375, "learning_rate": 9.759260237090058e-05, "loss": 0.3243, "step": 2665 }, { "epoch": 8.369905956112852, "grad_norm": 0.84765625, "learning_rate": 9.722793246336006e-05, "loss": 0.3255, "step": 2670 }, { "epoch": 8.385579937304076, "grad_norm": 0.7734375, "learning_rate": 9.686329944466203e-05, "loss": 0.323, "step": 2675 }, { "epoch": 8.401253918495298, "grad_norm": 0.8125, "learning_rate": 9.649870816710172e-05, "loss": 0.3278, "step": 2680 }, { "epoch": 8.41692789968652, "grad_norm": 0.875, "learning_rate": 9.613416348241887e-05, "loss": 0.3282, "step": 2685 }, { "epoch": 8.432601880877742, "grad_norm": 0.90234375, "learning_rate": 9.576967024173323e-05, "loss": 0.3226, "step": 2690 }, { "epoch": 8.448275862068966, "grad_norm": 0.80078125, "learning_rate": 9.540523329547984e-05, "loss": 0.3361, "step": 2695 }, { "epoch": 8.463949843260188, "grad_norm": 0.8671875, "learning_rate": 9.504085749334479e-05, "loss": 0.3309, "step": 2700 }, { "epoch": 8.47962382445141, "grad_norm": 0.74609375, "learning_rate": 9.467654768420032e-05, "loss": 0.325, "step": 2705 }, { "epoch": 8.495297805642632, "grad_norm": 0.78515625, "learning_rate": 9.431230871604067e-05, "loss": 0.3265, "step": 2710 }, { "epoch": 8.510971786833856, "grad_norm": 0.8203125, "learning_rate": 9.394814543591719e-05, "loss": 0.3302, "step": 2715 }, { "epoch": 8.526645768025078, "grad_norm": 0.77734375, "learning_rate": 9.358406268987417e-05, "loss": 0.3299, "step": 2720 }, { "epoch": 8.5423197492163, "grad_norm": 0.80859375, "learning_rate": 9.322006532288411e-05, "loss": 0.3303, "step": 2725 }, { "epoch": 8.557993730407524, "grad_norm": 0.87109375, "learning_rate": 9.285615817878342e-05, "loss": 0.3246, "step": 2730 }, { "epoch": 8.573667711598747, "grad_norm": 0.7578125, "learning_rate": 9.249234610020779e-05, "loss": 0.3256, "step": 2735 }, { "epoch": 8.589341692789969, "grad_norm": 0.80859375, "learning_rate": 9.212863392852793e-05, "loss": 0.3286, "step": 2740 }, { "epoch": 8.60501567398119, "grad_norm": 0.8359375, "learning_rate": 9.176502650378499e-05, "loss": 0.3301, "step": 2745 }, { "epoch": 8.620689655172415, "grad_norm": 0.73828125, "learning_rate": 9.140152866462629e-05, "loss": 0.3345, "step": 2750 }, { "epoch": 8.636363636363637, "grad_norm": 0.76171875, "learning_rate": 9.103814524824073e-05, "loss": 0.3335, "step": 2755 }, { "epoch": 8.652037617554859, "grad_norm": 0.73828125, "learning_rate": 9.067488109029474e-05, "loss": 0.3287, "step": 2760 }, { "epoch": 8.66771159874608, "grad_norm": 0.77734375, "learning_rate": 9.031174102486752e-05, "loss": 0.3286, "step": 2765 }, { "epoch": 8.683385579937305, "grad_norm": 0.79296875, "learning_rate": 8.994872988438711e-05, "loss": 0.3283, "step": 2770 }, { "epoch": 8.699059561128527, "grad_norm": 0.80078125, "learning_rate": 8.958585249956578e-05, "loss": 0.3308, "step": 2775 }, { "epoch": 8.714733542319749, "grad_norm": 0.859375, "learning_rate": 8.922311369933598e-05, "loss": 0.3308, "step": 2780 }, { "epoch": 8.730407523510971, "grad_norm": 0.859375, "learning_rate": 8.886051831078582e-05, "loss": 0.3314, "step": 2785 }, { "epoch": 8.746081504702195, "grad_norm": 0.8046875, "learning_rate": 8.849807115909513e-05, "loss": 0.33, "step": 2790 }, { "epoch": 8.761755485893417, "grad_norm": 0.86328125, "learning_rate": 8.8135777067471e-05, "loss": 0.329, "step": 2795 }, { "epoch": 8.77742946708464, "grad_norm": 0.84375, "learning_rate": 8.777364085708378e-05, "loss": 0.3356, "step": 2800 }, { "epoch": 8.793103448275861, "grad_norm": 0.83984375, "learning_rate": 8.741166734700273e-05, "loss": 0.3384, "step": 2805 }, { "epoch": 8.808777429467085, "grad_norm": 0.80859375, "learning_rate": 8.704986135413212e-05, "loss": 0.3324, "step": 2810 }, { "epoch": 8.824451410658307, "grad_norm": 0.78125, "learning_rate": 8.668822769314691e-05, "loss": 0.3291, "step": 2815 }, { "epoch": 8.84012539184953, "grad_norm": 0.78515625, "learning_rate": 8.632677117642892e-05, "loss": 0.3358, "step": 2820 }, { "epoch": 8.855799373040753, "grad_norm": 0.90625, "learning_rate": 8.596549661400248e-05, "loss": 0.3331, "step": 2825 }, { "epoch": 8.871473354231975, "grad_norm": 0.7890625, "learning_rate": 8.560440881347071e-05, "loss": 0.3306, "step": 2830 }, { "epoch": 8.887147335423197, "grad_norm": 0.796875, "learning_rate": 8.524351257995135e-05, "loss": 0.3322, "step": 2835 }, { "epoch": 8.90282131661442, "grad_norm": 0.78515625, "learning_rate": 8.488281271601302e-05, "loss": 0.3246, "step": 2840 }, { "epoch": 8.918495297805643, "grad_norm": 0.81640625, "learning_rate": 8.452231402161099e-05, "loss": 0.3339, "step": 2845 }, { "epoch": 8.934169278996865, "grad_norm": 0.74609375, "learning_rate": 8.416202129402371e-05, "loss": 0.3291, "step": 2850 }, { "epoch": 8.949843260188088, "grad_norm": 0.765625, "learning_rate": 8.380193932778857e-05, "loss": 0.3268, "step": 2855 }, { "epoch": 8.96551724137931, "grad_norm": 0.84375, "learning_rate": 8.344207291463843e-05, "loss": 0.3258, "step": 2860 }, { "epoch": 8.981191222570533, "grad_norm": 0.77734375, "learning_rate": 8.30824268434376e-05, "loss": 0.3345, "step": 2865 }, { "epoch": 8.996865203761756, "grad_norm": 0.796875, "learning_rate": 8.27230059001184e-05, "loss": 0.327, "step": 2870 }, { "epoch": 9.0, "eval_loss": 3.4537088871002197, "eval_runtime": 0.7983, "eval_samples_per_second": 2.505, "eval_steps_per_second": 1.253, "step": 2871 }, { "epoch": 9.012539184952978, "grad_norm": 0.5625, "learning_rate": 8.2363814867617e-05, "loss": 0.2932, "step": 2875 }, { "epoch": 9.0282131661442, "grad_norm": 0.7734375, "learning_rate": 8.200485852581036e-05, "loss": 0.2798, "step": 2880 }, { "epoch": 9.043887147335424, "grad_norm": 0.78125, "learning_rate": 8.16461416514522e-05, "loss": 0.2751, "step": 2885 }, { "epoch": 9.059561128526646, "grad_norm": 0.71484375, "learning_rate": 8.12876690181096e-05, "loss": 0.2771, "step": 2890 }, { "epoch": 9.075235109717868, "grad_norm": 0.66796875, "learning_rate": 8.092944539609937e-05, "loss": 0.281, "step": 2895 }, { "epoch": 9.090909090909092, "grad_norm": 0.7109375, "learning_rate": 8.057147555242473e-05, "loss": 0.2709, "step": 2900 }, { "epoch": 9.106583072100314, "grad_norm": 0.6875, "learning_rate": 8.021376425071175e-05, "loss": 0.2787, "step": 2905 }, { "epoch": 9.122257053291536, "grad_norm": 0.64453125, "learning_rate": 7.985631625114603e-05, "loss": 0.2817, "step": 2910 }, { "epoch": 9.137931034482758, "grad_norm": 0.68359375, "learning_rate": 7.94991363104092e-05, "loss": 0.2798, "step": 2915 }, { "epoch": 9.153605015673982, "grad_norm": 0.72265625, "learning_rate": 7.914222918161589e-05, "loss": 0.2804, "step": 2920 }, { "epoch": 9.169278996865204, "grad_norm": 0.71484375, "learning_rate": 7.878559961425025e-05, "loss": 0.2778, "step": 2925 }, { "epoch": 9.184952978056426, "grad_norm": 0.67578125, "learning_rate": 7.842925235410288e-05, "loss": 0.2832, "step": 2930 }, { "epoch": 9.200626959247648, "grad_norm": 0.703125, "learning_rate": 7.807319214320747e-05, "loss": 0.2781, "step": 2935 }, { "epoch": 9.216300940438872, "grad_norm": 0.67578125, "learning_rate": 7.771742371977811e-05, "loss": 0.277, "step": 2940 }, { "epoch": 9.231974921630094, "grad_norm": 0.69921875, "learning_rate": 7.73619518181457e-05, "loss": 0.2701, "step": 2945 }, { "epoch": 9.247648902821316, "grad_norm": 0.73046875, "learning_rate": 7.700678116869543e-05, "loss": 0.2771, "step": 2950 }, { "epoch": 9.263322884012538, "grad_norm": 0.7265625, "learning_rate": 7.66519164978035e-05, "loss": 0.2738, "step": 2955 }, { "epoch": 9.278996865203762, "grad_norm": 0.77734375, "learning_rate": 7.629736252777445e-05, "loss": 0.2735, "step": 2960 }, { "epoch": 9.294670846394984, "grad_norm": 0.6875, "learning_rate": 7.594312397677809e-05, "loss": 0.2773, "step": 2965 }, { "epoch": 9.310344827586206, "grad_norm": 0.68359375, "learning_rate": 7.558920555878696e-05, "loss": 0.2764, "step": 2970 }, { "epoch": 9.32601880877743, "grad_norm": 0.83984375, "learning_rate": 7.523561198351342e-05, "loss": 0.2828, "step": 2975 }, { "epoch": 9.341692789968652, "grad_norm": 0.74609375, "learning_rate": 7.48823479563471e-05, "loss": 0.2834, "step": 2980 }, { "epoch": 9.357366771159874, "grad_norm": 0.87109375, "learning_rate": 7.452941817829212e-05, "loss": 0.2848, "step": 2985 }, { "epoch": 9.373040752351097, "grad_norm": 0.66015625, "learning_rate": 7.417682734590469e-05, "loss": 0.2801, "step": 2990 }, { "epoch": 9.38871473354232, "grad_norm": 0.71875, "learning_rate": 7.382458015123057e-05, "loss": 0.2822, "step": 2995 }, { "epoch": 9.404388714733543, "grad_norm": 0.69921875, "learning_rate": 7.347268128174265e-05, "loss": 0.2736, "step": 3000 }, { "epoch": 9.420062695924765, "grad_norm": 0.703125, "learning_rate": 7.31211354202784e-05, "loss": 0.2801, "step": 3005 }, { "epoch": 9.435736677115987, "grad_norm": 0.703125, "learning_rate": 7.276994724497787e-05, "loss": 0.2799, "step": 3010 }, { "epoch": 9.45141065830721, "grad_norm": 0.6953125, "learning_rate": 7.241912142922109e-05, "loss": 0.2768, "step": 3015 }, { "epoch": 9.467084639498433, "grad_norm": 0.72265625, "learning_rate": 7.206866264156623e-05, "loss": 0.2851, "step": 3020 }, { "epoch": 9.482758620689655, "grad_norm": 0.78125, "learning_rate": 7.171857554568706e-05, "loss": 0.284, "step": 3025 }, { "epoch": 9.498432601880877, "grad_norm": 0.73046875, "learning_rate": 7.136886480031138e-05, "loss": 0.2826, "step": 3030 }, { "epoch": 9.5141065830721, "grad_norm": 0.7265625, "learning_rate": 7.101953505915857e-05, "loss": 0.2823, "step": 3035 }, { "epoch": 9.529780564263323, "grad_norm": 0.7109375, "learning_rate": 7.067059097087796e-05, "loss": 0.2767, "step": 3040 }, { "epoch": 9.545454545454545, "grad_norm": 0.71484375, "learning_rate": 7.03220371789868e-05, "loss": 0.2808, "step": 3045 }, { "epoch": 9.561128526645769, "grad_norm": 0.6953125, "learning_rate": 6.997387832180864e-05, "loss": 0.2847, "step": 3050 }, { "epoch": 9.576802507836991, "grad_norm": 0.7109375, "learning_rate": 6.962611903241132e-05, "loss": 0.2829, "step": 3055 }, { "epoch": 9.592476489028213, "grad_norm": 0.6875, "learning_rate": 6.92787639385457e-05, "loss": 0.2815, "step": 3060 }, { "epoch": 9.608150470219435, "grad_norm": 0.94140625, "learning_rate": 6.893181766258373e-05, "loss": 0.281, "step": 3065 }, { "epoch": 9.623824451410659, "grad_norm": 0.69140625, "learning_rate": 6.858528482145716e-05, "loss": 0.2807, "step": 3070 }, { "epoch": 9.639498432601881, "grad_norm": 0.6953125, "learning_rate": 6.823917002659596e-05, "loss": 0.2884, "step": 3075 }, { "epoch": 9.655172413793103, "grad_norm": 0.6484375, "learning_rate": 6.789347788386706e-05, "loss": 0.2789, "step": 3080 }, { "epoch": 9.670846394984325, "grad_norm": 0.73828125, "learning_rate": 6.754821299351299e-05, "loss": 0.2733, "step": 3085 }, { "epoch": 9.68652037617555, "grad_norm": 0.6953125, "learning_rate": 6.720337995009076e-05, "loss": 0.2876, "step": 3090 }, { "epoch": 9.702194357366771, "grad_norm": 0.6875, "learning_rate": 6.68589833424105e-05, "loss": 0.2856, "step": 3095 }, { "epoch": 9.717868338557993, "grad_norm": 0.77734375, "learning_rate": 6.651502775347469e-05, "loss": 0.2829, "step": 3100 }, { "epoch": 9.733542319749215, "grad_norm": 0.71484375, "learning_rate": 6.617151776041692e-05, "loss": 0.2875, "step": 3105 }, { "epoch": 9.74921630094044, "grad_norm": 0.69140625, "learning_rate": 6.582845793444119e-05, "loss": 0.2746, "step": 3110 }, { "epoch": 9.764890282131661, "grad_norm": 0.703125, "learning_rate": 6.548585284076084e-05, "loss": 0.288, "step": 3115 }, { "epoch": 9.780564263322884, "grad_norm": 0.7421875, "learning_rate": 6.514370703853806e-05, "loss": 0.2807, "step": 3120 }, { "epoch": 9.796238244514107, "grad_norm": 0.7265625, "learning_rate": 6.480202508082298e-05, "loss": 0.2858, "step": 3125 }, { "epoch": 9.81191222570533, "grad_norm": 0.69921875, "learning_rate": 6.44608115144933e-05, "loss": 0.2793, "step": 3130 }, { "epoch": 9.827586206896552, "grad_norm": 0.7109375, "learning_rate": 6.412007088019364e-05, "loss": 0.2818, "step": 3135 }, { "epoch": 9.843260188087774, "grad_norm": 0.75390625, "learning_rate": 6.377980771227509e-05, "loss": 0.2864, "step": 3140 }, { "epoch": 9.858934169278998, "grad_norm": 0.73046875, "learning_rate": 6.344002653873504e-05, "loss": 0.2819, "step": 3145 }, { "epoch": 9.87460815047022, "grad_norm": 0.75, "learning_rate": 6.31007318811567e-05, "loss": 0.2833, "step": 3150 }, { "epoch": 9.890282131661442, "grad_norm": 0.7265625, "learning_rate": 6.276192825464918e-05, "loss": 0.2888, "step": 3155 }, { "epoch": 9.905956112852664, "grad_norm": 0.66796875, "learning_rate": 6.242362016778713e-05, "loss": 0.2819, "step": 3160 }, { "epoch": 9.921630094043888, "grad_norm": 0.703125, "learning_rate": 6.208581212255104e-05, "loss": 0.2858, "step": 3165 }, { "epoch": 9.93730407523511, "grad_norm": 0.6875, "learning_rate": 6.1748508614267e-05, "loss": 0.2793, "step": 3170 }, { "epoch": 9.952978056426332, "grad_norm": 0.73046875, "learning_rate": 6.141171413154722e-05, "loss": 0.2795, "step": 3175 }, { "epoch": 9.968652037617554, "grad_norm": 0.6640625, "learning_rate": 6.107543315623001e-05, "loss": 0.2794, "step": 3180 }, { "epoch": 9.984326018808778, "grad_norm": 0.7578125, "learning_rate": 6.073967016332041e-05, "loss": 0.2833, "step": 3185 }, { "epoch": 10.0, "grad_norm": 0.6484375, "learning_rate": 6.040442962093029e-05, "loss": 0.2814, "step": 3190 }, { "epoch": 10.0, "eval_loss": 3.8273370265960693, "eval_runtime": 0.8006, "eval_samples_per_second": 2.498, "eval_steps_per_second": 1.249, "step": 3190 }, { "epoch": 10.015673981191222, "grad_norm": 0.53515625, "learning_rate": 6.006971599021928e-05, "loss": 0.2526, "step": 3195 }, { "epoch": 10.031347962382446, "grad_norm": 0.59375, "learning_rate": 5.973553372533509e-05, "loss": 0.2478, "step": 3200 }, { "epoch": 10.047021943573668, "grad_norm": 0.6484375, "learning_rate": 5.9401887273354475e-05, "loss": 0.2487, "step": 3205 }, { "epoch": 10.06269592476489, "grad_norm": 0.6640625, "learning_rate": 5.9068781074223824e-05, "loss": 0.2522, "step": 3210 }, { "epoch": 10.078369905956112, "grad_norm": 0.6015625, "learning_rate": 5.8736219560700324e-05, "loss": 0.2503, "step": 3215 }, { "epoch": 10.094043887147336, "grad_norm": 0.65625, "learning_rate": 5.840420715829272e-05, "loss": 0.247, "step": 3220 }, { "epoch": 10.109717868338558, "grad_norm": 0.67578125, "learning_rate": 5.807274828520266e-05, "loss": 0.2521, "step": 3225 }, { "epoch": 10.12539184952978, "grad_norm": 0.6640625, "learning_rate": 5.774184735226571e-05, "loss": 0.2484, "step": 3230 }, { "epoch": 10.141065830721002, "grad_norm": 0.59375, "learning_rate": 5.741150876289283e-05, "loss": 0.2474, "step": 3235 }, { "epoch": 10.156739811912226, "grad_norm": 0.58984375, "learning_rate": 5.708173691301153e-05, "loss": 0.2506, "step": 3240 }, { "epoch": 10.172413793103448, "grad_norm": 0.61328125, "learning_rate": 5.675253619100772e-05, "loss": 0.2544, "step": 3245 }, { "epoch": 10.18808777429467, "grad_norm": 0.66015625, "learning_rate": 5.642391097766693e-05, "loss": 0.249, "step": 3250 }, { "epoch": 10.203761755485893, "grad_norm": 1.2578125, "learning_rate": 5.609586564611631e-05, "loss": 0.2535, "step": 3255 }, { "epoch": 10.219435736677116, "grad_norm": 0.57421875, "learning_rate": 5.576840456176631e-05, "loss": 0.2504, "step": 3260 }, { "epoch": 10.235109717868339, "grad_norm": 0.66015625, "learning_rate": 5.544153208225265e-05, "loss": 0.2524, "step": 3265 }, { "epoch": 10.25078369905956, "grad_norm": 0.65234375, "learning_rate": 5.511525255737815e-05, "loss": 0.2549, "step": 3270 }, { "epoch": 10.266457680250785, "grad_norm": 0.625, "learning_rate": 5.478957032905514e-05, "loss": 0.2521, "step": 3275 }, { "epoch": 10.282131661442007, "grad_norm": 0.6171875, "learning_rate": 5.446448973124736e-05, "loss": 0.2531, "step": 3280 }, { "epoch": 10.297805642633229, "grad_norm": 0.59375, "learning_rate": 5.414001508991264e-05, "loss": 0.2533, "step": 3285 }, { "epoch": 10.31347962382445, "grad_norm": 0.63671875, "learning_rate": 5.3816150722944916e-05, "loss": 0.2544, "step": 3290 }, { "epoch": 10.329153605015675, "grad_norm": 0.6328125, "learning_rate": 5.3492900940117264e-05, "loss": 0.2491, "step": 3295 }, { "epoch": 10.344827586206897, "grad_norm": 0.63671875, "learning_rate": 5.3170270043024015e-05, "loss": 0.2522, "step": 3300 }, { "epoch": 10.360501567398119, "grad_norm": 0.62890625, "learning_rate": 5.284826232502399e-05, "loss": 0.249, "step": 3305 }, { "epoch": 10.376175548589341, "grad_norm": 0.77734375, "learning_rate": 5.252688207118297e-05, "loss": 0.2459, "step": 3310 }, { "epoch": 10.391849529780565, "grad_norm": 0.625, "learning_rate": 5.220613355821704e-05, "loss": 0.2503, "step": 3315 }, { "epoch": 10.407523510971787, "grad_norm": 0.6328125, "learning_rate": 5.188602105443533e-05, "loss": 0.2543, "step": 3320 }, { "epoch": 10.423197492163009, "grad_norm": 0.65234375, "learning_rate": 5.156654881968348e-05, "loss": 0.2509, "step": 3325 }, { "epoch": 10.438871473354231, "grad_norm": 0.65234375, "learning_rate": 5.124772110528684e-05, "loss": 0.2524, "step": 3330 }, { "epoch": 10.454545454545455, "grad_norm": 0.6796875, "learning_rate": 5.0929542153993926e-05, "loss": 0.2525, "step": 3335 }, { "epoch": 10.470219435736677, "grad_norm": 0.62890625, "learning_rate": 5.061201619991984e-05, "loss": 0.248, "step": 3340 }, { "epoch": 10.4858934169279, "grad_norm": 0.72265625, "learning_rate": 5.029514746849018e-05, "loss": 0.2516, "step": 3345 }, { "epoch": 10.501567398119121, "grad_norm": 0.75, "learning_rate": 4.9978940176384514e-05, "loss": 0.2509, "step": 3350 }, { "epoch": 10.517241379310345, "grad_norm": 0.625, "learning_rate": 4.966339853148055e-05, "loss": 0.2506, "step": 3355 }, { "epoch": 10.532915360501567, "grad_norm": 0.6171875, "learning_rate": 4.934852673279787e-05, "loss": 0.254, "step": 3360 }, { "epoch": 10.54858934169279, "grad_norm": 0.63671875, "learning_rate": 4.9034328970442275e-05, "loss": 0.2519, "step": 3365 }, { "epoch": 10.564263322884013, "grad_norm": 0.625, "learning_rate": 4.8720809425549916e-05, "loss": 0.2541, "step": 3370 }, { "epoch": 10.579937304075235, "grad_norm": 0.6484375, "learning_rate": 4.8407972270231704e-05, "loss": 0.2563, "step": 3375 }, { "epoch": 10.595611285266457, "grad_norm": 0.60546875, "learning_rate": 4.809582166751765e-05, "loss": 0.2529, "step": 3380 }, { "epoch": 10.61128526645768, "grad_norm": 0.7265625, "learning_rate": 4.778436177130173e-05, "loss": 0.2546, "step": 3385 }, { "epoch": 10.626959247648903, "grad_norm": 0.6328125, "learning_rate": 4.747359672628631e-05, "loss": 0.2545, "step": 3390 }, { "epoch": 10.642633228840126, "grad_norm": 0.62890625, "learning_rate": 4.7163530667927226e-05, "loss": 0.2564, "step": 3395 }, { "epoch": 10.658307210031348, "grad_norm": 0.66015625, "learning_rate": 4.685416772237864e-05, "loss": 0.2506, "step": 3400 }, { "epoch": 10.67398119122257, "grad_norm": 0.6640625, "learning_rate": 4.65455120064382e-05, "loss": 0.2505, "step": 3405 }, { "epoch": 10.689655172413794, "grad_norm": 0.68359375, "learning_rate": 4.623756762749207e-05, "loss": 0.2499, "step": 3410 }, { "epoch": 10.705329153605016, "grad_norm": 0.6484375, "learning_rate": 4.593033868346059e-05, "loss": 0.2529, "step": 3415 }, { "epoch": 10.721003134796238, "grad_norm": 0.63671875, "learning_rate": 4.5623829262743414e-05, "loss": 0.2549, "step": 3420 }, { "epoch": 10.736677115987462, "grad_norm": 0.59375, "learning_rate": 4.531804344416536e-05, "loss": 0.2524, "step": 3425 }, { "epoch": 10.752351097178684, "grad_norm": 0.6015625, "learning_rate": 4.501298529692194e-05, "loss": 0.2556, "step": 3430 }, { "epoch": 10.768025078369906, "grad_norm": 0.64453125, "learning_rate": 4.470865888052537e-05, "loss": 0.2518, "step": 3435 }, { "epoch": 10.783699059561128, "grad_norm": 0.640625, "learning_rate": 4.4405068244750446e-05, "loss": 0.2544, "step": 3440 }, { "epoch": 10.799373040752352, "grad_norm": 0.61328125, "learning_rate": 4.410221742958073e-05, "loss": 0.2493, "step": 3445 }, { "epoch": 10.815047021943574, "grad_norm": 0.76953125, "learning_rate": 4.380011046515461e-05, "loss": 0.2556, "step": 3450 }, { "epoch": 10.830721003134796, "grad_norm": 0.6640625, "learning_rate": 4.349875137171196e-05, "loss": 0.2525, "step": 3455 }, { "epoch": 10.846394984326018, "grad_norm": 0.63671875, "learning_rate": 4.3198144159540346e-05, "loss": 0.2532, "step": 3460 }, { "epoch": 10.862068965517242, "grad_norm": 0.625, "learning_rate": 4.289829282892188e-05, "loss": 0.2494, "step": 3465 }, { "epoch": 10.877742946708464, "grad_norm": 0.61328125, "learning_rate": 4.2599201370079875e-05, "loss": 0.2495, "step": 3470 }, { "epoch": 10.893416927899686, "grad_norm": 0.6484375, "learning_rate": 4.230087376312582e-05, "loss": 0.2483, "step": 3475 }, { "epoch": 10.909090909090908, "grad_norm": 0.71875, "learning_rate": 4.2003313978006244e-05, "loss": 0.2505, "step": 3480 }, { "epoch": 10.924764890282132, "grad_norm": 0.609375, "learning_rate": 4.170652597445016e-05, "loss": 0.2499, "step": 3485 }, { "epoch": 10.940438871473354, "grad_norm": 0.69921875, "learning_rate": 4.1410513701916086e-05, "loss": 0.2501, "step": 3490 }, { "epoch": 10.956112852664576, "grad_norm": 0.66796875, "learning_rate": 4.111528109953975e-05, "loss": 0.2547, "step": 3495 }, { "epoch": 10.971786833855798, "grad_norm": 0.60546875, "learning_rate": 4.0820832096081415e-05, "loss": 0.252, "step": 3500 }, { "epoch": 10.987460815047022, "grad_norm": 0.625, "learning_rate": 4.052717060987386e-05, "loss": 0.2539, "step": 3505 }, { "epoch": 11.0, "eval_loss": 4.135468006134033, "eval_runtime": 0.805, "eval_samples_per_second": 2.485, "eval_steps_per_second": 1.242, "step": 3509 }, { "epoch": 11.003134796238244, "grad_norm": 0.5078125, "learning_rate": 4.023430054876999e-05, "loss": 0.2514, "step": 3510 }, { "epoch": 11.018808777429467, "grad_norm": 0.52734375, "learning_rate": 3.994222581009107e-05, "loss": 0.2418, "step": 3515 }, { "epoch": 11.03448275862069, "grad_norm": 0.5390625, "learning_rate": 3.965095028057461e-05, "loss": 0.2369, "step": 3520 }, { "epoch": 11.050156739811912, "grad_norm": 0.57421875, "learning_rate": 3.936047783632286e-05, "loss": 0.2349, "step": 3525 }, { "epoch": 11.065830721003135, "grad_norm": 0.6015625, "learning_rate": 3.907081234275109e-05, "loss": 0.2392, "step": 3530 }, { "epoch": 11.081504702194357, "grad_norm": 0.5859375, "learning_rate": 3.878195765453626e-05, "loss": 0.2316, "step": 3535 }, { "epoch": 11.09717868338558, "grad_norm": 0.53515625, "learning_rate": 3.849391761556559e-05, "loss": 0.2321, "step": 3540 }, { "epoch": 11.112852664576803, "grad_norm": 0.52734375, "learning_rate": 3.820669605888556e-05, "loss": 0.235, "step": 3545 }, { "epoch": 11.128526645768025, "grad_norm": 0.55078125, "learning_rate": 3.79202968066508e-05, "loss": 0.2333, "step": 3550 }, { "epoch": 11.144200626959247, "grad_norm": 0.55078125, "learning_rate": 3.7634723670073294e-05, "loss": 0.2376, "step": 3555 }, { "epoch": 11.15987460815047, "grad_norm": 0.6171875, "learning_rate": 3.7349980449371516e-05, "loss": 0.2379, "step": 3560 }, { "epoch": 11.175548589341693, "grad_norm": 0.5625, "learning_rate": 3.706607093372012e-05, "loss": 0.2344, "step": 3565 }, { "epoch": 11.191222570532915, "grad_norm": 0.55078125, "learning_rate": 3.67829989011992e-05, "loss": 0.2355, "step": 3570 }, { "epoch": 11.206896551724139, "grad_norm": 0.5703125, "learning_rate": 3.65007681187443e-05, "loss": 0.2332, "step": 3575 }, { "epoch": 11.22257053291536, "grad_norm": 0.6015625, "learning_rate": 3.621938234209613e-05, "loss": 0.2346, "step": 3580 }, { "epoch": 11.238244514106583, "grad_norm": 0.6015625, "learning_rate": 3.5938845315750666e-05, "loss": 0.2385, "step": 3585 }, { "epoch": 11.253918495297805, "grad_norm": 0.5546875, "learning_rate": 3.565916077290914e-05, "loss": 0.2367, "step": 3590 }, { "epoch": 11.269592476489029, "grad_norm": 0.66796875, "learning_rate": 3.5380332435428655e-05, "loss": 0.2383, "step": 3595 }, { "epoch": 11.285266457680251, "grad_norm": 0.62890625, "learning_rate": 3.510236401377236e-05, "loss": 0.2373, "step": 3600 }, { "epoch": 11.300940438871473, "grad_norm": 0.6015625, "learning_rate": 3.482525920696036e-05, "loss": 0.2352, "step": 3605 }, { "epoch": 11.316614420062695, "grad_norm": 0.5234375, "learning_rate": 3.454902170252019e-05, "loss": 0.2343, "step": 3610 }, { "epoch": 11.33228840125392, "grad_norm": 0.64453125, "learning_rate": 3.4273655176438014e-05, "loss": 0.2317, "step": 3615 }, { "epoch": 11.347962382445141, "grad_norm": 0.59765625, "learning_rate": 3.3999163293109534e-05, "loss": 0.2375, "step": 3620 }, { "epoch": 11.363636363636363, "grad_norm": 0.625, "learning_rate": 3.372554970529137e-05, "loss": 0.2384, "step": 3625 }, { "epoch": 11.379310344827585, "grad_norm": 0.63671875, "learning_rate": 3.345281805405219e-05, "loss": 0.2385, "step": 3630 }, { "epoch": 11.39498432601881, "grad_norm": 0.63671875, "learning_rate": 3.318097196872464e-05, "loss": 0.2429, "step": 3635 }, { "epoch": 11.410658307210031, "grad_norm": 0.58984375, "learning_rate": 3.291001506685666e-05, "loss": 0.2351, "step": 3640 }, { "epoch": 11.426332288401253, "grad_norm": 0.5546875, "learning_rate": 3.2639950954163644e-05, "loss": 0.2377, "step": 3645 }, { "epoch": 11.442006269592476, "grad_norm": 0.72265625, "learning_rate": 3.23707832244803e-05, "loss": 0.2353, "step": 3650 }, { "epoch": 11.4576802507837, "grad_norm": 0.5390625, "learning_rate": 3.2102515459712876e-05, "loss": 0.2361, "step": 3655 }, { "epoch": 11.473354231974922, "grad_norm": 0.61328125, "learning_rate": 3.1835151229791435e-05, "loss": 0.2357, "step": 3660 }, { "epoch": 11.489028213166144, "grad_norm": 0.54296875, "learning_rate": 3.1568694092622475e-05, "loss": 0.2359, "step": 3665 }, { "epoch": 11.504702194357368, "grad_norm": 0.5390625, "learning_rate": 3.1303147594041394e-05, "loss": 0.2333, "step": 3670 }, { "epoch": 11.52037617554859, "grad_norm": 0.5859375, "learning_rate": 3.1038515267765545e-05, "loss": 0.236, "step": 3675 }, { "epoch": 11.536050156739812, "grad_norm": 0.64453125, "learning_rate": 3.0774800635346934e-05, "loss": 0.2323, "step": 3680 }, { "epoch": 11.551724137931034, "grad_norm": 0.578125, "learning_rate": 3.0512007206125638e-05, "loss": 0.2358, "step": 3685 }, { "epoch": 11.567398119122258, "grad_norm": 0.6328125, "learning_rate": 3.0250138477182886e-05, "loss": 0.2379, "step": 3690 }, { "epoch": 11.58307210031348, "grad_norm": 0.640625, "learning_rate": 2.9989197933294687e-05, "loss": 0.2353, "step": 3695 }, { "epoch": 11.598746081504702, "grad_norm": 0.56640625, "learning_rate": 2.9729189046885266e-05, "loss": 0.2359, "step": 3700 }, { "epoch": 11.614420062695924, "grad_norm": 0.5546875, "learning_rate": 2.947011527798107e-05, "loss": 0.2375, "step": 3705 }, { "epoch": 11.630094043887148, "grad_norm": 0.609375, "learning_rate": 2.9211980074164514e-05, "loss": 0.2364, "step": 3710 }, { "epoch": 11.64576802507837, "grad_norm": 0.62109375, "learning_rate": 2.895478687052835e-05, "loss": 0.2393, "step": 3715 }, { "epoch": 11.661442006269592, "grad_norm": 0.59375, "learning_rate": 2.8698539089629662e-05, "loss": 0.235, "step": 3720 }, { "epoch": 11.677115987460816, "grad_norm": 0.56640625, "learning_rate": 2.844324014144457e-05, "loss": 0.2341, "step": 3725 }, { "epoch": 11.692789968652038, "grad_norm": 0.56640625, "learning_rate": 2.818889342332275e-05, "loss": 0.2299, "step": 3730 }, { "epoch": 11.70846394984326, "grad_norm": 0.61328125, "learning_rate": 2.793550231994222e-05, "loss": 0.2388, "step": 3735 }, { "epoch": 11.724137931034482, "grad_norm": 0.57421875, "learning_rate": 2.768307020326425e-05, "loss": 0.2341, "step": 3740 }, { "epoch": 11.739811912225706, "grad_norm": 0.578125, "learning_rate": 2.7431600432488657e-05, "loss": 0.234, "step": 3745 }, { "epoch": 11.755485893416928, "grad_norm": 0.60546875, "learning_rate": 2.7181096354008884e-05, "loss": 0.2364, "step": 3750 }, { "epoch": 11.77115987460815, "grad_norm": 0.62890625, "learning_rate": 2.6931561301367646e-05, "loss": 0.2331, "step": 3755 }, { "epoch": 11.786833855799372, "grad_norm": 0.60546875, "learning_rate": 2.6682998595212505e-05, "loss": 0.2357, "step": 3760 }, { "epoch": 11.802507836990596, "grad_norm": 0.65625, "learning_rate": 2.6435411543251677e-05, "loss": 0.2389, "step": 3765 }, { "epoch": 11.818181818181818, "grad_norm": 0.60546875, "learning_rate": 2.6188803440209942e-05, "loss": 0.241, "step": 3770 }, { "epoch": 11.83385579937304, "grad_norm": 0.62109375, "learning_rate": 2.5943177567785015e-05, "loss": 0.2361, "step": 3775 }, { "epoch": 11.849529780564263, "grad_norm": 0.578125, "learning_rate": 2.5698537194603566e-05, "loss": 0.2352, "step": 3780 }, { "epoch": 11.865203761755486, "grad_norm": 0.58203125, "learning_rate": 2.5454885576178067e-05, "loss": 0.2389, "step": 3785 }, { "epoch": 11.880877742946709, "grad_norm": 0.71875, "learning_rate": 2.5212225954863132e-05, "loss": 0.2367, "step": 3790 }, { "epoch": 11.89655172413793, "grad_norm": 0.5703125, "learning_rate": 2.4970561559812645e-05, "loss": 0.2383, "step": 3795 }, { "epoch": 11.912225705329153, "grad_norm": 0.61328125, "learning_rate": 2.472989560693665e-05, "loss": 0.2314, "step": 3800 }, { "epoch": 11.927899686520377, "grad_norm": 0.66796875, "learning_rate": 2.449023129885859e-05, "loss": 0.2388, "step": 3805 }, { "epoch": 11.943573667711599, "grad_norm": 0.56640625, "learning_rate": 2.425157182487262e-05, "loss": 0.2383, "step": 3810 }, { "epoch": 11.95924764890282, "grad_norm": 0.67578125, "learning_rate": 2.401392036090132e-05, "loss": 0.2384, "step": 3815 }, { "epoch": 11.974921630094045, "grad_norm": 0.62890625, "learning_rate": 2.3777280069453245e-05, "loss": 0.2358, "step": 3820 }, { "epoch": 11.990595611285267, "grad_norm": 0.55859375, "learning_rate": 2.3541654099581e-05, "loss": 0.233, "step": 3825 }, { "epoch": 12.0, "eval_loss": 4.354940891265869, "eval_runtime": 0.795, "eval_samples_per_second": 2.516, "eval_steps_per_second": 1.258, "step": 3828 }, { "epoch": 12.006269592476489, "grad_norm": 0.5546875, "learning_rate": 2.330704558683926e-05, "loss": 0.2349, "step": 3830 }, { "epoch": 12.021943573667711, "grad_norm": 0.51171875, "learning_rate": 2.307345765324306e-05, "loss": 0.2297, "step": 3835 }, { "epoch": 12.037617554858935, "grad_norm": 0.5390625, "learning_rate": 2.284089340722618e-05, "loss": 0.2337, "step": 3840 }, { "epoch": 12.053291536050157, "grad_norm": 0.51953125, "learning_rate": 2.2609355943599942e-05, "loss": 0.2295, "step": 3845 }, { "epoch": 12.068965517241379, "grad_norm": 0.578125, "learning_rate": 2.2378848343511804e-05, "loss": 0.2287, "step": 3850 }, { "epoch": 12.084639498432601, "grad_norm": 0.5859375, "learning_rate": 2.214937367440463e-05, "loss": 0.2281, "step": 3855 }, { "epoch": 12.100313479623825, "grad_norm": 0.56640625, "learning_rate": 2.192093498997555e-05, "loss": 0.2299, "step": 3860 }, { "epoch": 12.115987460815047, "grad_norm": 0.58984375, "learning_rate": 2.169353533013565e-05, "loss": 0.2287, "step": 3865 }, { "epoch": 12.13166144200627, "grad_norm": 0.5390625, "learning_rate": 2.1467177720969268e-05, "loss": 0.2281, "step": 3870 }, { "epoch": 12.147335423197493, "grad_norm": 0.490234375, "learning_rate": 2.12418651746939e-05, "loss": 0.2238, "step": 3875 }, { "epoch": 12.163009404388715, "grad_norm": 0.56640625, "learning_rate": 2.101760068961992e-05, "loss": 0.2331, "step": 3880 }, { "epoch": 12.178683385579937, "grad_norm": 0.5234375, "learning_rate": 2.0794387250110913e-05, "loss": 0.2286, "step": 3885 }, { "epoch": 12.19435736677116, "grad_norm": 0.53125, "learning_rate": 2.0572227826543755e-05, "loss": 0.2323, "step": 3890 }, { "epoch": 12.210031347962383, "grad_norm": 0.5390625, "learning_rate": 2.0351125375269264e-05, "loss": 0.2269, "step": 3895 }, { "epoch": 12.225705329153605, "grad_norm": 0.51171875, "learning_rate": 2.0131082838572655e-05, "loss": 0.2288, "step": 3900 }, { "epoch": 12.241379310344827, "grad_norm": 0.546875, "learning_rate": 1.991210314463461e-05, "loss": 0.236, "step": 3905 }, { "epoch": 12.25705329153605, "grad_norm": 0.54296875, "learning_rate": 1.969418920749214e-05, "loss": 0.229, "step": 3910 }, { "epoch": 12.272727272727273, "grad_norm": 0.58203125, "learning_rate": 1.9477343926999913e-05, "loss": 0.2305, "step": 3915 }, { "epoch": 12.288401253918495, "grad_norm": 0.52734375, "learning_rate": 1.9261570188791555e-05, "loss": 0.2287, "step": 3920 }, { "epoch": 12.304075235109718, "grad_norm": 0.55859375, "learning_rate": 1.9046870864241384e-05, "loss": 0.2274, "step": 3925 }, { "epoch": 12.31974921630094, "grad_norm": 0.52734375, "learning_rate": 1.8833248810426073e-05, "loss": 0.2282, "step": 3930 }, { "epoch": 12.335423197492164, "grad_norm": 0.54296875, "learning_rate": 1.8620706870086723e-05, "loss": 0.2282, "step": 3935 }, { "epoch": 12.351097178683386, "grad_norm": 0.578125, "learning_rate": 1.8409247871591006e-05, "loss": 0.2323, "step": 3940 }, { "epoch": 12.366771159874608, "grad_norm": 0.57421875, "learning_rate": 1.8198874628895524e-05, "loss": 0.228, "step": 3945 }, { "epoch": 12.38244514106583, "grad_norm": 0.55078125, "learning_rate": 1.798958994150829e-05, "loss": 0.2227, "step": 3950 }, { "epoch": 12.398119122257054, "grad_norm": 0.50390625, "learning_rate": 1.7781396594451637e-05, "loss": 0.235, "step": 3955 }, { "epoch": 12.413793103448276, "grad_norm": 0.55078125, "learning_rate": 1.757429735822499e-05, "loss": 0.2303, "step": 3960 }, { "epoch": 12.429467084639498, "grad_norm": 0.6875, "learning_rate": 1.7368294988768097e-05, "loss": 0.2318, "step": 3965 }, { "epoch": 12.445141065830722, "grad_norm": 0.59765625, "learning_rate": 1.716339222742436e-05, "loss": 0.2305, "step": 3970 }, { "epoch": 12.460815047021944, "grad_norm": 0.54296875, "learning_rate": 1.695959180090425e-05, "loss": 0.2268, "step": 3975 }, { "epoch": 12.476489028213166, "grad_norm": 0.65234375, "learning_rate": 1.6756896421249168e-05, "loss": 0.2279, "step": 3980 }, { "epoch": 12.492163009404388, "grad_norm": 0.57421875, "learning_rate": 1.6555308785795232e-05, "loss": 0.2267, "step": 3985 }, { "epoch": 12.507836990595612, "grad_norm": 0.59765625, "learning_rate": 1.6354831577137485e-05, "loss": 0.2287, "step": 3990 }, { "epoch": 12.523510971786834, "grad_norm": 0.58984375, "learning_rate": 1.6155467463094066e-05, "loss": 0.2248, "step": 3995 }, { "epoch": 12.539184952978056, "grad_norm": 0.5703125, "learning_rate": 1.5957219096670883e-05, "loss": 0.2299, "step": 4000 }, { "epoch": 12.554858934169278, "grad_norm": 0.59765625, "learning_rate": 1.576008911602609e-05, "loss": 0.2288, "step": 4005 }, { "epoch": 12.570532915360502, "grad_norm": 0.56640625, "learning_rate": 1.5564080144435212e-05, "loss": 0.2318, "step": 4010 }, { "epoch": 12.586206896551724, "grad_norm": 0.5625, "learning_rate": 1.536919479025609e-05, "loss": 0.233, "step": 4015 }, { "epoch": 12.601880877742946, "grad_norm": 0.55078125, "learning_rate": 1.517543564689422e-05, "loss": 0.2253, "step": 4020 }, { "epoch": 12.61755485893417, "grad_norm": 0.60546875, "learning_rate": 1.4982805292768165e-05, "loss": 0.2266, "step": 4025 }, { "epoch": 12.633228840125392, "grad_norm": 0.5546875, "learning_rate": 1.4791306291275398e-05, "loss": 0.2272, "step": 4030 }, { "epoch": 12.648902821316614, "grad_norm": 0.55078125, "learning_rate": 1.4600941190758022e-05, "loss": 0.2304, "step": 4035 }, { "epoch": 12.664576802507836, "grad_norm": 0.56640625, "learning_rate": 1.4411712524469012e-05, "loss": 0.2314, "step": 4040 }, { "epoch": 12.68025078369906, "grad_norm": 0.59375, "learning_rate": 1.4223622810538328e-05, "loss": 0.2303, "step": 4045 }, { "epoch": 12.695924764890282, "grad_norm": 0.5546875, "learning_rate": 1.4036674551939599e-05, "loss": 0.2323, "step": 4050 }, { "epoch": 12.711598746081505, "grad_norm": 0.6015625, "learning_rate": 1.385087023645667e-05, "loss": 0.2307, "step": 4055 }, { "epoch": 12.727272727272727, "grad_norm": 0.515625, "learning_rate": 1.3666212336650586e-05, "loss": 0.2235, "step": 4060 }, { "epoch": 12.74294670846395, "grad_norm": 0.51953125, "learning_rate": 1.3482703309826584e-05, "loss": 0.2285, "step": 4065 }, { "epoch": 12.758620689655173, "grad_norm": 0.578125, "learning_rate": 1.330034559800154e-05, "loss": 0.2311, "step": 4070 }, { "epoch": 12.774294670846395, "grad_norm": 0.578125, "learning_rate": 1.31191416278713e-05, "loss": 0.2286, "step": 4075 }, { "epoch": 12.789968652037617, "grad_norm": 0.55859375, "learning_rate": 1.293909381077858e-05, "loss": 0.2294, "step": 4080 }, { "epoch": 12.80564263322884, "grad_norm": 0.58203125, "learning_rate": 1.2760204542680654e-05, "loss": 0.2309, "step": 4085 }, { "epoch": 12.821316614420063, "grad_norm": 0.57421875, "learning_rate": 1.2582476204117755e-05, "loss": 0.2294, "step": 4090 }, { "epoch": 12.836990595611285, "grad_norm": 0.5234375, "learning_rate": 1.2405911160181072e-05, "loss": 0.2241, "step": 4095 }, { "epoch": 12.852664576802507, "grad_norm": 0.52734375, "learning_rate": 1.2230511760481533e-05, "loss": 0.2253, "step": 4100 }, { "epoch": 12.86833855799373, "grad_norm": 0.5703125, "learning_rate": 1.2056280339118397e-05, "loss": 0.2358, "step": 4105 }, { "epoch": 12.884012539184953, "grad_norm": 0.5625, "learning_rate": 1.188321921464829e-05, "loss": 0.229, "step": 4110 }, { "epoch": 12.899686520376175, "grad_norm": 0.578125, "learning_rate": 1.1711330690054211e-05, "loss": 0.2299, "step": 4115 }, { "epoch": 12.915360501567399, "grad_norm": 0.54296875, "learning_rate": 1.1540617052715074e-05, "loss": 0.2283, "step": 4120 }, { "epoch": 12.931034482758621, "grad_norm": 0.55859375, "learning_rate": 1.1371080574375114e-05, "loss": 0.2297, "step": 4125 }, { "epoch": 12.946708463949843, "grad_norm": 0.5859375, "learning_rate": 1.1202723511113766e-05, "loss": 0.2338, "step": 4130 }, { "epoch": 12.962382445141065, "grad_norm": 0.6171875, "learning_rate": 1.1035548103315484e-05, "loss": 0.2337, "step": 4135 }, { "epoch": 12.978056426332289, "grad_norm": 0.5703125, "learning_rate": 1.086955657564015e-05, "loss": 0.2294, "step": 4140 }, { "epoch": 12.993730407523511, "grad_norm": 0.6328125, "learning_rate": 1.0704751136993251e-05, "loss": 0.2281, "step": 4145 }, { "epoch": 13.0, "eval_loss": 4.455935478210449, "eval_runtime": 0.8034, "eval_samples_per_second": 2.489, "eval_steps_per_second": 1.245, "step": 4147 }, { "epoch": 13.009404388714733, "grad_norm": 0.50390625, "learning_rate": 1.0541133980496686e-05, "loss": 0.2257, "step": 4150 }, { "epoch": 13.025078369905955, "grad_norm": 0.57421875, "learning_rate": 1.0378707283459376e-05, "loss": 0.2262, "step": 4155 }, { "epoch": 13.04075235109718, "grad_norm": 0.49609375, "learning_rate": 1.0217473207348483e-05, "loss": 0.2247, "step": 4160 }, { "epoch": 13.056426332288401, "grad_norm": 0.5859375, "learning_rate": 1.0057433897760493e-05, "loss": 0.2277, "step": 4165 }, { "epoch": 13.072100313479623, "grad_norm": 0.53515625, "learning_rate": 9.898591484392793e-06, "loss": 0.2256, "step": 4170 }, { "epoch": 13.087774294670846, "grad_norm": 0.5625, "learning_rate": 9.74094808101519e-06, "loss": 0.23, "step": 4175 }, { "epoch": 13.10344827586207, "grad_norm": 0.60546875, "learning_rate": 9.584505785441932e-06, "loss": 0.2266, "step": 4180 }, { "epoch": 13.119122257053291, "grad_norm": 0.55078125, "learning_rate": 9.429266679503657e-06, "loss": 0.2283, "step": 4185 }, { "epoch": 13.134796238244514, "grad_norm": 0.515625, "learning_rate": 9.275232829019787e-06, "loss": 0.2257, "step": 4190 }, { "epoch": 13.150470219435737, "grad_norm": 0.51171875, "learning_rate": 9.122406283771002e-06, "loss": 0.2307, "step": 4195 }, { "epoch": 13.16614420062696, "grad_norm": 0.4921875, "learning_rate": 8.970789077471953e-06, "loss": 0.2259, "step": 4200 }, { "epoch": 13.181818181818182, "grad_norm": 0.54296875, "learning_rate": 8.82038322774419e-06, "loss": 0.2286, "step": 4205 }, { "epoch": 13.197492163009404, "grad_norm": 0.5625, "learning_rate": 8.671190736089373e-06, "loss": 0.2277, "step": 4210 }, { "epoch": 13.213166144200628, "grad_norm": 0.73046875, "learning_rate": 8.523213587862533e-06, "loss": 0.2287, "step": 4215 }, { "epoch": 13.22884012539185, "grad_norm": 0.55859375, "learning_rate": 8.376453752245795e-06, "loss": 0.2266, "step": 4220 }, { "epoch": 13.244514106583072, "grad_norm": 0.54296875, "learning_rate": 8.230913182222e-06, "loss": 0.2264, "step": 4225 }, { "epoch": 13.260188087774294, "grad_norm": 0.5625, "learning_rate": 8.086593814548882e-06, "loss": 0.2258, "step": 4230 }, { "epoch": 13.275862068965518, "grad_norm": 0.5703125, "learning_rate": 7.943497569733183e-06, "loss": 0.2295, "step": 4235 }, { "epoch": 13.29153605015674, "grad_norm": 0.578125, "learning_rate": 7.801626352005186e-06, "loss": 0.2254, "step": 4240 }, { "epoch": 13.307210031347962, "grad_norm": 0.51171875, "learning_rate": 7.66098204929323e-06, "loss": 0.2284, "step": 4245 }, { "epoch": 13.322884012539184, "grad_norm": 0.578125, "learning_rate": 7.521566533198765e-06, "loss": 0.2263, "step": 4250 }, { "epoch": 13.338557993730408, "grad_norm": 0.53125, "learning_rate": 7.383381658971311e-06, "loss": 0.2279, "step": 4255 }, { "epoch": 13.35423197492163, "grad_norm": 0.4921875, "learning_rate": 7.246429265483856e-06, "loss": 0.2238, "step": 4260 }, { "epoch": 13.369905956112852, "grad_norm": 0.5234375, "learning_rate": 7.1107111752083175e-06, "loss": 0.2273, "step": 4265 }, { "epoch": 13.385579937304076, "grad_norm": 0.53125, "learning_rate": 6.976229194191352e-06, "loss": 0.2244, "step": 4270 }, { "epoch": 13.401253918495298, "grad_norm": 0.53515625, "learning_rate": 6.842985112030253e-06, "loss": 0.2256, "step": 4275 }, { "epoch": 13.41692789968652, "grad_norm": 0.53125, "learning_rate": 6.710980701849223e-06, "loss": 0.2246, "step": 4280 }, { "epoch": 13.432601880877742, "grad_norm": 0.498046875, "learning_rate": 6.580217720275661e-06, "loss": 0.2275, "step": 4285 }, { "epoch": 13.448275862068966, "grad_norm": 0.5234375, "learning_rate": 6.450697907416936e-06, "loss": 0.2269, "step": 4290 }, { "epoch": 13.463949843260188, "grad_norm": 0.5234375, "learning_rate": 6.3224229868370845e-06, "loss": 0.2311, "step": 4295 }, { "epoch": 13.47962382445141, "grad_norm": 0.5859375, "learning_rate": 6.19539466553396e-06, "loss": 0.2254, "step": 4300 }, { "epoch": 13.495297805642632, "grad_norm": 0.58984375, "learning_rate": 6.0696146339165095e-06, "loss": 0.2289, "step": 4305 }, { "epoch": 13.510971786833856, "grad_norm": 0.55078125, "learning_rate": 5.945084565782277e-06, "loss": 0.2319, "step": 4310 }, { "epoch": 13.526645768025078, "grad_norm": 0.498046875, "learning_rate": 5.82180611829507e-06, "loss": 0.224, "step": 4315 }, { "epoch": 13.5423197492163, "grad_norm": 0.5703125, "learning_rate": 5.699780931963006e-06, "loss": 0.2297, "step": 4320 }, { "epoch": 13.557993730407524, "grad_norm": 0.54296875, "learning_rate": 5.5790106306165766e-06, "loss": 0.228, "step": 4325 }, { "epoch": 13.573667711598747, "grad_norm": 0.6171875, "learning_rate": 5.459496821387166e-06, "loss": 0.2315, "step": 4330 }, { "epoch": 13.589341692789969, "grad_norm": 0.53515625, "learning_rate": 5.341241094685523e-06, "loss": 0.2283, "step": 4335 }, { "epoch": 13.60501567398119, "grad_norm": 0.6953125, "learning_rate": 5.2242450241806964e-06, "loss": 0.2277, "step": 4340 }, { "epoch": 13.620689655172415, "grad_norm": 0.55859375, "learning_rate": 5.108510166779068e-06, "loss": 0.2264, "step": 4345 }, { "epoch": 13.636363636363637, "grad_norm": 0.51953125, "learning_rate": 4.994038062603645e-06, "loss": 0.232, "step": 4350 }, { "epoch": 13.652037617554859, "grad_norm": 0.546875, "learning_rate": 4.880830234973499e-06, "loss": 0.2254, "step": 4355 }, { "epoch": 13.66771159874608, "grad_norm": 0.58984375, "learning_rate": 4.7688881903835915e-06, "loss": 0.2277, "step": 4360 }, { "epoch": 13.683385579937305, "grad_norm": 0.515625, "learning_rate": 4.658213418484636e-06, "loss": 0.2292, "step": 4365 }, { "epoch": 13.699059561128527, "grad_norm": 0.5, "learning_rate": 4.548807392063359e-06, "loss": 0.2246, "step": 4370 }, { "epoch": 13.714733542319749, "grad_norm": 0.51171875, "learning_rate": 4.4406715670228474e-06, "loss": 0.226, "step": 4375 }, { "epoch": 13.730407523510971, "grad_norm": 0.5546875, "learning_rate": 4.333807382363197e-06, "loss": 0.2292, "step": 4380 }, { "epoch": 13.746081504702195, "grad_norm": 0.55078125, "learning_rate": 4.22821626016231e-06, "loss": 0.2289, "step": 4385 }, { "epoch": 13.761755485893417, "grad_norm": 0.5234375, "learning_rate": 4.123899605557091e-06, "loss": 0.2247, "step": 4390 }, { "epoch": 13.77742946708464, "grad_norm": 0.55078125, "learning_rate": 4.020858806724592e-06, "loss": 0.2221, "step": 4395 }, { "epoch": 13.793103448275861, "grad_norm": 0.55078125, "learning_rate": 3.91909523486369e-06, "loss": 0.2277, "step": 4400 }, { "epoch": 13.808777429467085, "grad_norm": 0.55859375, "learning_rate": 3.818610244176702e-06, "loss": 0.2298, "step": 4405 }, { "epoch": 13.824451410658307, "grad_norm": 0.625, "learning_rate": 3.719405171851487e-06, "loss": 0.2223, "step": 4410 }, { "epoch": 13.84012539184953, "grad_norm": 0.546875, "learning_rate": 3.621481338043564e-06, "loss": 0.2269, "step": 4415 }, { "epoch": 13.855799373040753, "grad_norm": 0.57421875, "learning_rate": 3.5248400458586127e-06, "loss": 0.2266, "step": 4420 }, { "epoch": 13.871473354231975, "grad_norm": 0.546875, "learning_rate": 3.429482581335053e-06, "loss": 0.2296, "step": 4425 }, { "epoch": 13.887147335423197, "grad_norm": 0.54296875, "learning_rate": 3.3354102134269927e-06, "loss": 0.2281, "step": 4430 }, { "epoch": 13.90282131661442, "grad_norm": 0.5625, "learning_rate": 3.2426241939873313e-06, "loss": 0.2288, "step": 4435 }, { "epoch": 13.918495297805643, "grad_norm": 0.62109375, "learning_rate": 3.151125757751083e-06, "loss": 0.2299, "step": 4440 }, { "epoch": 13.934169278996865, "grad_norm": 0.578125, "learning_rate": 3.0609161223189575e-06, "loss": 0.2238, "step": 4445 }, { "epoch": 13.949843260188088, "grad_norm": 0.58203125, "learning_rate": 2.9719964881411712e-06, "loss": 0.2243, "step": 4450 }, { "epoch": 13.96551724137931, "grad_norm": 0.890625, "learning_rate": 2.8843680385014284e-06, "loss": 0.2257, "step": 4455 }, { "epoch": 13.981191222570533, "grad_norm": 0.609375, "learning_rate": 2.798031939501222e-06, "loss": 0.2286, "step": 4460 }, { "epoch": 13.996865203761756, "grad_norm": 0.52734375, "learning_rate": 2.7129893400442807e-06, "loss": 0.2274, "step": 4465 }, { "epoch": 14.0, "eval_loss": 4.4672441482543945, "eval_runtime": 0.8024, "eval_samples_per_second": 2.492, "eval_steps_per_second": 1.246, "step": 4466 }, { "epoch": 14.012539184952978, "grad_norm": 0.58984375, "learning_rate": 2.629241371821334e-06, "loss": 0.2259, "step": 4470 }, { "epoch": 14.0282131661442, "grad_norm": 0.51953125, "learning_rate": 2.546789149294959e-06, "loss": 0.2248, "step": 4475 }, { "epoch": 14.043887147335424, "grad_norm": 0.59765625, "learning_rate": 2.4656337696848496e-06, "loss": 0.226, "step": 4480 }, { "epoch": 14.059561128526646, "grad_norm": 0.5625, "learning_rate": 2.3857763129531473e-06, "loss": 0.2239, "step": 4485 }, { "epoch": 14.075235109717868, "grad_norm": 0.51953125, "learning_rate": 2.3072178417901326e-06, "loss": 0.2277, "step": 4490 }, { "epoch": 14.090909090909092, "grad_norm": 0.57421875, "learning_rate": 2.229959401599968e-06, "loss": 0.2246, "step": 4495 }, { "epoch": 14.106583072100314, "grad_norm": 0.5546875, "learning_rate": 2.154002020486945e-06, "loss": 0.2244, "step": 4500 }, { "epoch": 14.122257053291536, "grad_norm": 0.5546875, "learning_rate": 2.0793467092416696e-06, "loss": 0.2226, "step": 4505 }, { "epoch": 14.137931034482758, "grad_norm": 0.53515625, "learning_rate": 2.005994461327698e-06, "loss": 0.2299, "step": 4510 }, { "epoch": 14.153605015673982, "grad_norm": 0.515625, "learning_rate": 1.933946252868224e-06, "loss": 0.2287, "step": 4515 }, { "epoch": 14.169278996865204, "grad_norm": 0.54296875, "learning_rate": 1.8632030426332215e-06, "loss": 0.2226, "step": 4520 }, { "epoch": 14.184952978056426, "grad_norm": 0.59765625, "learning_rate": 1.7937657720265454e-06, "loss": 0.2262, "step": 4525 }, { "epoch": 14.200626959247648, "grad_norm": 0.5703125, "learning_rate": 1.7256353650735302e-06, "loss": 0.2298, "step": 4530 }, { "epoch": 14.216300940438872, "grad_norm": 0.5859375, "learning_rate": 1.6588127284085652e-06, "loss": 0.2286, "step": 4535 }, { "epoch": 14.231974921630094, "grad_norm": 0.65234375, "learning_rate": 1.5932987512631614e-06, "loss": 0.2282, "step": 4540 }, { "epoch": 14.247648902821316, "grad_norm": 0.58203125, "learning_rate": 1.529094305453993e-06, "loss": 0.222, "step": 4545 }, { "epoch": 14.263322884012538, "grad_norm": 0.5234375, "learning_rate": 1.4662002453714074e-06, "loss": 0.2256, "step": 4550 }, { "epoch": 14.278996865203762, "grad_norm": 0.59375, "learning_rate": 1.4046174079679787e-06, "loss": 0.2266, "step": 4555 }, { "epoch": 14.294670846394984, "grad_norm": 0.59375, "learning_rate": 1.3443466127474046e-06, "loss": 0.2245, "step": 4560 }, { "epoch": 14.310344827586206, "grad_norm": 0.546875, "learning_rate": 1.285388661753595e-06, "loss": 0.2332, "step": 4565 }, { "epoch": 14.32601880877743, "grad_norm": 0.50390625, "learning_rate": 1.2277443395599886e-06, "loss": 0.2266, "step": 4570 }, { "epoch": 14.341692789968652, "grad_norm": 0.52734375, "learning_rate": 1.1714144132591199e-06, "loss": 0.2281, "step": 4575 }, { "epoch": 14.357366771159874, "grad_norm": 0.53125, "learning_rate": 1.116399632452414e-06, "loss": 0.2292, "step": 4580 }, { "epoch": 14.373040752351097, "grad_norm": 0.51171875, "learning_rate": 1.062700729240218e-06, "loss": 0.2246, "step": 4585 }, { "epoch": 14.38871473354232, "grad_norm": 0.57421875, "learning_rate": 1.0103184182120418e-06, "loss": 0.228, "step": 4590 }, { "epoch": 14.404388714733543, "grad_norm": 0.54296875, "learning_rate": 9.592533964370542e-07, "loss": 0.2257, "step": 4595 }, { "epoch": 14.420062695924765, "grad_norm": 0.51171875, "learning_rate": 9.095063434548135e-07, "loss": 0.229, "step": 4600 }, { "epoch": 14.435736677115987, "grad_norm": 0.490234375, "learning_rate": 8.61077921266229e-07, "loss": 0.2273, "step": 4605 }, { "epoch": 14.45141065830721, "grad_norm": 0.5546875, "learning_rate": 8.139687743247138e-07, "loss": 0.2256, "step": 4610 }, { "epoch": 14.467084639498433, "grad_norm": 0.640625, "learning_rate": 7.681795295276684e-07, "loss": 0.229, "step": 4615 }, { "epoch": 14.482758620689655, "grad_norm": 0.5234375, "learning_rate": 7.237107962080991e-07, "loss": 0.2248, "step": 4620 }, { "epoch": 14.498432601880877, "grad_norm": 0.54296875, "learning_rate": 6.805631661265133e-07, "loss": 0.2292, "step": 4625 }, { "epoch": 14.5141065830721, "grad_norm": 0.578125, "learning_rate": 6.387372134630587e-07, "loss": 0.2225, "step": 4630 }, { "epoch": 14.529780564263323, "grad_norm": 0.5078125, "learning_rate": 5.982334948098522e-07, "loss": 0.2291, "step": 4635 }, { "epoch": 14.545454545454545, "grad_norm": 0.5703125, "learning_rate": 5.5905254916363e-07, "loss": 0.2228, "step": 4640 }, { "epoch": 14.561128526645769, "grad_norm": 0.5546875, "learning_rate": 5.211948979184978e-07, "loss": 0.2282, "step": 4645 }, { "epoch": 14.576802507836991, "grad_norm": 0.53515625, "learning_rate": 4.846610448590804e-07, "loss": 0.2291, "step": 4650 }, { "epoch": 14.592476489028213, "grad_norm": 0.5625, "learning_rate": 4.4945147615372827e-07, "loss": 0.229, "step": 4655 }, { "epoch": 14.608150470219435, "grad_norm": 0.56640625, "learning_rate": 4.1556666034811007e-07, "loss": 0.2294, "step": 4660 }, { "epoch": 14.623824451410659, "grad_norm": 0.546875, "learning_rate": 3.8300704835896316e-07, "loss": 0.2268, "step": 4665 }, { "epoch": 14.639498432601881, "grad_norm": 0.5390625, "learning_rate": 3.517730734680869e-07, "loss": 0.2263, "step": 4670 }, { "epoch": 14.655172413793103, "grad_norm": 0.58984375, "learning_rate": 3.2186515131655823e-07, "loss": 0.2267, "step": 4675 }, { "epoch": 14.670846394984325, "grad_norm": 0.5078125, "learning_rate": 2.932836798992589e-07, "loss": 0.2281, "step": 4680 }, { "epoch": 14.68652037617555, "grad_norm": 0.52734375, "learning_rate": 2.660290395595011e-07, "loss": 0.2286, "step": 4685 }, { "epoch": 14.702194357366771, "grad_norm": 0.54296875, "learning_rate": 2.401015929840322e-07, "loss": 0.2275, "step": 4690 }, { "epoch": 14.717868338557993, "grad_norm": 0.55859375, "learning_rate": 2.155016851981717e-07, "loss": 0.229, "step": 4695 }, { "epoch": 14.733542319749215, "grad_norm": 0.50390625, "learning_rate": 1.9222964356123696e-07, "loss": 0.2284, "step": 4700 }, { "epoch": 14.74921630094044, "grad_norm": 0.56640625, "learning_rate": 1.7028577776216915e-07, "loss": 0.2328, "step": 4705 }, { "epoch": 14.764890282131661, "grad_norm": 0.59765625, "learning_rate": 1.496703798154364e-07, "loss": 0.2272, "step": 4710 }, { "epoch": 14.780564263322884, "grad_norm": 0.482421875, "learning_rate": 1.3038372405711487e-07, "loss": 0.2307, "step": 4715 }, { "epoch": 14.796238244514107, "grad_norm": 0.5234375, "learning_rate": 1.1242606714129134e-07, "loss": 0.2253, "step": 4720 }, { "epoch": 14.81191222570533, "grad_norm": 0.55859375, "learning_rate": 9.579764803658853e-08, "loss": 0.2238, "step": 4725 }, { "epoch": 14.827586206896552, "grad_norm": 0.52734375, "learning_rate": 8.049868802301187e-08, "loss": 0.2273, "step": 4730 }, { "epoch": 14.843260188087774, "grad_norm": 0.5, "learning_rate": 6.652939068899633e-08, "loss": 0.2315, "step": 4735 }, { "epoch": 14.858934169278998, "grad_norm": 0.52734375, "learning_rate": 5.388994192875307e-08, "loss": 0.2232, "step": 4740 }, { "epoch": 14.87460815047022, "grad_norm": 0.515625, "learning_rate": 4.258050993967144e-08, "loss": 0.2258, "step": 4745 }, { "epoch": 14.890282131661442, "grad_norm": 0.515625, "learning_rate": 3.260124522023178e-08, "loss": 0.2287, "step": 4750 }, { "epoch": 14.905956112852664, "grad_norm": 0.484375, "learning_rate": 2.3952280567873796e-08, "loss": 0.2208, "step": 4755 }, { "epoch": 14.921630094043888, "grad_norm": 0.515625, "learning_rate": 1.6633731077297933e-08, "loss": 0.2233, "step": 4760 }, { "epoch": 14.93730407523511, "grad_norm": 0.58203125, "learning_rate": 1.0645694138933237e-08, "loss": 0.2239, "step": 4765 }, { "epoch": 14.952978056426332, "grad_norm": 0.50390625, "learning_rate": 5.988249437627325e-09, "loss": 0.2257, "step": 4770 }, { "epoch": 14.968652037617554, "grad_norm": 0.609375, "learning_rate": 2.6614589515583377e-09, "loss": 0.2308, "step": 4775 }, { "epoch": 14.984326018808778, "grad_norm": 0.55859375, "learning_rate": 6.65366951457802e-10, "loss": 0.2262, "step": 4780 }, { "epoch": 15.0, "grad_norm": 0.5625, "learning_rate": 0.0, "loss": 0.2251, "step": 4785 }, { "epoch": 15.0, "eval_loss": 4.466753005981445, "eval_runtime": 0.7937, "eval_samples_per_second": 2.52, "eval_steps_per_second": 1.26, "step": 4785 }, { "epoch": 15.0, "step": 4785, "total_flos": 5.539666200113447e+18, "train_loss": 0.8980661474674348, "train_runtime": 27834.6576, "train_samples_per_second": 4.124, "train_steps_per_second": 0.172 } ], "logging_steps": 5, "max_steps": 4785, "num_input_tokens_seen": 0, "num_train_epochs": 15, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.539666200113447e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }