{ "best_metric": null, "best_model_checkpoint": null, "epoch": 7.072135785007072, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03536067892503536, "grad_norm": 15.248167037963867, "learning_rate": 4.800000000000001e-07, "loss": 1.6047, "step": 25 }, { "epoch": 0.07072135785007072, "grad_norm": 11.99557876586914, "learning_rate": 9.800000000000001e-07, "loss": 0.7922, "step": 50 }, { "epoch": 0.10608203677510608, "grad_norm": 10.12234878540039, "learning_rate": 1.48e-06, "loss": 0.6289, "step": 75 }, { "epoch": 0.14144271570014144, "grad_norm": 10.784132957458496, "learning_rate": 1.98e-06, "loss": 0.5627, "step": 100 }, { "epoch": 0.1768033946251768, "grad_norm": 10.604392051696777, "learning_rate": 2.4800000000000004e-06, "loss": 0.563, "step": 125 }, { "epoch": 0.21216407355021216, "grad_norm": 9.61458969116211, "learning_rate": 2.9800000000000003e-06, "loss": 0.5218, "step": 150 }, { "epoch": 0.24752475247524752, "grad_norm": 10.1797513961792, "learning_rate": 3.48e-06, "loss": 0.5316, "step": 175 }, { "epoch": 0.2828854314002829, "grad_norm": 9.834571838378906, "learning_rate": 3.980000000000001e-06, "loss": 0.5232, "step": 200 }, { "epoch": 0.31824611032531824, "grad_norm": 9.805900573730469, "learning_rate": 4.48e-06, "loss": 0.4902, "step": 225 }, { "epoch": 0.3536067892503536, "grad_norm": 10.328847885131836, "learning_rate": 4.980000000000001e-06, "loss": 0.4885, "step": 250 }, { "epoch": 0.38896746817538896, "grad_norm": 9.624419212341309, "learning_rate": 5.480000000000001e-06, "loss": 0.4615, "step": 275 }, { "epoch": 0.4243281471004243, "grad_norm": 12.492682456970215, "learning_rate": 5.98e-06, "loss": 0.4366, "step": 300 }, { "epoch": 0.4596888260254597, "grad_norm": 9.505845069885254, "learning_rate": 6.480000000000001e-06, "loss": 0.4758, "step": 325 }, { "epoch": 0.49504950495049505, "grad_norm": 8.755816459655762, "learning_rate": 6.98e-06, "loss": 0.44, "step": 350 }, { "epoch": 0.5304101838755304, "grad_norm": 9.016213417053223, "learning_rate": 7.48e-06, "loss": 0.4061, "step": 375 }, { "epoch": 0.5657708628005658, "grad_norm": 9.07496166229248, "learning_rate": 7.980000000000002e-06, "loss": 0.4171, "step": 400 }, { "epoch": 0.6011315417256011, "grad_norm": 7.877910137176514, "learning_rate": 8.48e-06, "loss": 0.4141, "step": 425 }, { "epoch": 0.6364922206506365, "grad_norm": 8.071303367614746, "learning_rate": 8.98e-06, "loss": 0.3969, "step": 450 }, { "epoch": 0.6718528995756718, "grad_norm": 8.29423713684082, "learning_rate": 9.48e-06, "loss": 0.3924, "step": 475 }, { "epoch": 0.7072135785007072, "grad_norm": 8.465646743774414, "learning_rate": 9.980000000000001e-06, "loss": 0.4029, "step": 500 }, { "epoch": 0.7425742574257426, "grad_norm": 7.924498081207275, "learning_rate": 9.946666666666667e-06, "loss": 0.379, "step": 525 }, { "epoch": 0.7779349363507779, "grad_norm": 8.063438415527344, "learning_rate": 9.891111111111113e-06, "loss": 0.3814, "step": 550 }, { "epoch": 0.8132956152758133, "grad_norm": 7.638169765472412, "learning_rate": 9.835555555555556e-06, "loss": 0.3476, "step": 575 }, { "epoch": 0.8486562942008486, "grad_norm": 8.318132400512695, "learning_rate": 9.780000000000001e-06, "loss": 0.3499, "step": 600 }, { "epoch": 0.884016973125884, "grad_norm": 7.153702735900879, "learning_rate": 9.724444444444445e-06, "loss": 0.3411, "step": 625 }, { "epoch": 0.9193776520509194, "grad_norm": 5.998073101043701, "learning_rate": 9.671111111111112e-06, "loss": 0.3206, "step": 650 }, { "epoch": 0.9547383309759547, "grad_norm": 7.312229156494141, "learning_rate": 9.615555555555558e-06, "loss": 0.3463, "step": 675 }, { "epoch": 0.9900990099009901, "grad_norm": 6.737645626068115, "learning_rate": 9.56e-06, "loss": 0.3285, "step": 700 }, { "epoch": 1.0254596888260255, "grad_norm": 6.451304912567139, "learning_rate": 9.504444444444446e-06, "loss": 0.2746, "step": 725 }, { "epoch": 1.0608203677510608, "grad_norm": 5.666613578796387, "learning_rate": 9.44888888888889e-06, "loss": 0.2455, "step": 750 }, { "epoch": 1.0961810466760962, "grad_norm": 6.165246963500977, "learning_rate": 9.393333333333334e-06, "loss": 0.2246, "step": 775 }, { "epoch": 1.1315417256011315, "grad_norm": 5.314395904541016, "learning_rate": 9.33777777777778e-06, "loss": 0.2313, "step": 800 }, { "epoch": 1.166902404526167, "grad_norm": 6.031562328338623, "learning_rate": 9.282222222222222e-06, "loss": 0.2445, "step": 825 }, { "epoch": 1.2022630834512023, "grad_norm": 8.143284797668457, "learning_rate": 9.226666666666668e-06, "loss": 0.2248, "step": 850 }, { "epoch": 1.2376237623762376, "grad_norm": 4.90247106552124, "learning_rate": 9.171111111111112e-06, "loss": 0.2296, "step": 875 }, { "epoch": 1.272984441301273, "grad_norm": 5.334079265594482, "learning_rate": 9.115555555555556e-06, "loss": 0.2252, "step": 900 }, { "epoch": 1.3083451202263083, "grad_norm": 6.497454643249512, "learning_rate": 9.060000000000001e-06, "loss": 0.2261, "step": 925 }, { "epoch": 1.3437057991513437, "grad_norm": 7.067910671234131, "learning_rate": 9.004444444444445e-06, "loss": 0.2416, "step": 950 }, { "epoch": 1.379066478076379, "grad_norm": 8.470537185668945, "learning_rate": 8.94888888888889e-06, "loss": 0.2219, "step": 975 }, { "epoch": 1.4144271570014144, "grad_norm": 4.9936747550964355, "learning_rate": 8.893333333333333e-06, "loss": 0.2147, "step": 1000 }, { "epoch": 1.4144271570014144, "eval_loss": 0.3066270351409912, "eval_runtime": 1765.706, "eval_samples_per_second": 3.048, "eval_steps_per_second": 0.191, "eval_wer": 0.23493975903614459, "step": 1000 }, { "epoch": 1.4497878359264498, "grad_norm": 7.876005172729492, "learning_rate": 8.83777777777778e-06, "loss": 0.2032, "step": 1025 }, { "epoch": 1.4851485148514851, "grad_norm": 6.1690826416015625, "learning_rate": 8.782222222222223e-06, "loss": 0.216, "step": 1050 }, { "epoch": 1.5205091937765205, "grad_norm": 5.094773769378662, "learning_rate": 8.726666666666667e-06, "loss": 0.2189, "step": 1075 }, { "epoch": 1.5558698727015559, "grad_norm": 6.5309834480285645, "learning_rate": 8.671111111111113e-06, "loss": 0.2188, "step": 1100 }, { "epoch": 1.5912305516265912, "grad_norm": 7.060642242431641, "learning_rate": 8.615555555555555e-06, "loss": 0.2233, "step": 1125 }, { "epoch": 1.6265912305516266, "grad_norm": 5.661276817321777, "learning_rate": 8.560000000000001e-06, "loss": 0.2124, "step": 1150 }, { "epoch": 1.661951909476662, "grad_norm": 5.289186477661133, "learning_rate": 8.504444444444445e-06, "loss": 0.2022, "step": 1175 }, { "epoch": 1.6973125884016973, "grad_norm": 4.7772064208984375, "learning_rate": 8.448888888888889e-06, "loss": 0.2018, "step": 1200 }, { "epoch": 1.7326732673267327, "grad_norm": 5.6891303062438965, "learning_rate": 8.393333333333335e-06, "loss": 0.2059, "step": 1225 }, { "epoch": 1.768033946251768, "grad_norm": 5.304312229156494, "learning_rate": 8.337777777777777e-06, "loss": 0.1949, "step": 1250 }, { "epoch": 1.8033946251768034, "grad_norm": 5.472725868225098, "learning_rate": 8.282222222222223e-06, "loss": 0.21, "step": 1275 }, { "epoch": 1.8387553041018387, "grad_norm": 4.937655448913574, "learning_rate": 8.226666666666667e-06, "loss": 0.2137, "step": 1300 }, { "epoch": 1.874115983026874, "grad_norm": 6.363350868225098, "learning_rate": 8.171111111111113e-06, "loss": 0.1919, "step": 1325 }, { "epoch": 1.9094766619519095, "grad_norm": 5.087652206420898, "learning_rate": 8.115555555555557e-06, "loss": 0.1783, "step": 1350 }, { "epoch": 1.9448373408769448, "grad_norm": 4.039090633392334, "learning_rate": 8.06e-06, "loss": 0.1879, "step": 1375 }, { "epoch": 1.9801980198019802, "grad_norm": 5.647103309631348, "learning_rate": 8.004444444444445e-06, "loss": 0.1904, "step": 1400 }, { "epoch": 2.0155586987270158, "grad_norm": 3.8179798126220703, "learning_rate": 7.948888888888889e-06, "loss": 0.1572, "step": 1425 }, { "epoch": 2.050919377652051, "grad_norm": 5.048496246337891, "learning_rate": 7.893333333333335e-06, "loss": 0.1087, "step": 1450 }, { "epoch": 2.0862800565770865, "grad_norm": 3.450915813446045, "learning_rate": 7.837777777777779e-06, "loss": 0.1088, "step": 1475 }, { "epoch": 2.1216407355021216, "grad_norm": 4.442863464355469, "learning_rate": 7.782222222222223e-06, "loss": 0.1049, "step": 1500 }, { "epoch": 2.157001414427157, "grad_norm": 4.130264759063721, "learning_rate": 7.726666666666667e-06, "loss": 0.1118, "step": 1525 }, { "epoch": 2.1923620933521923, "grad_norm": 3.8330652713775635, "learning_rate": 7.67111111111111e-06, "loss": 0.1068, "step": 1550 }, { "epoch": 2.227722772277228, "grad_norm": 4.3345489501953125, "learning_rate": 7.6155555555555564e-06, "loss": 0.1133, "step": 1575 }, { "epoch": 2.263083451202263, "grad_norm": 4.276752948760986, "learning_rate": 7.5600000000000005e-06, "loss": 0.1105, "step": 1600 }, { "epoch": 2.298444130127298, "grad_norm": 3.8332560062408447, "learning_rate": 7.504444444444445e-06, "loss": 0.1011, "step": 1625 }, { "epoch": 2.333804809052334, "grad_norm": 4.976287841796875, "learning_rate": 7.44888888888889e-06, "loss": 0.1151, "step": 1650 }, { "epoch": 2.3691654879773694, "grad_norm": 3.8312933444976807, "learning_rate": 7.393333333333333e-06, "loss": 0.1124, "step": 1675 }, { "epoch": 2.4045261669024045, "grad_norm": 4.4323248863220215, "learning_rate": 7.337777777777778e-06, "loss": 0.1148, "step": 1700 }, { "epoch": 2.4398868458274396, "grad_norm": 3.492539882659912, "learning_rate": 7.282222222222222e-06, "loss": 0.112, "step": 1725 }, { "epoch": 2.4752475247524752, "grad_norm": 4.846162796020508, "learning_rate": 7.226666666666667e-06, "loss": 0.1128, "step": 1750 }, { "epoch": 2.510608203677511, "grad_norm": 5.806790828704834, "learning_rate": 7.171111111111112e-06, "loss": 0.1072, "step": 1775 }, { "epoch": 2.545968882602546, "grad_norm": 3.3074755668640137, "learning_rate": 7.115555555555557e-06, "loss": 0.1067, "step": 1800 }, { "epoch": 2.581329561527581, "grad_norm": 3.70501446723938, "learning_rate": 7.06e-06, "loss": 0.1016, "step": 1825 }, { "epoch": 2.6166902404526167, "grad_norm": 2.6926796436309814, "learning_rate": 7.004444444444445e-06, "loss": 0.1018, "step": 1850 }, { "epoch": 2.6520509193776522, "grad_norm": 4.052884578704834, "learning_rate": 6.948888888888889e-06, "loss": 0.1106, "step": 1875 }, { "epoch": 2.6874115983026874, "grad_norm": 4.722410678863525, "learning_rate": 6.893333333333334e-06, "loss": 0.1133, "step": 1900 }, { "epoch": 2.7227722772277225, "grad_norm": 4.194309711456299, "learning_rate": 6.837777777777779e-06, "loss": 0.1012, "step": 1925 }, { "epoch": 2.758132956152758, "grad_norm": 3.4879953861236572, "learning_rate": 6.782222222222222e-06, "loss": 0.1091, "step": 1950 }, { "epoch": 2.7934936350777937, "grad_norm": 2.7147629261016846, "learning_rate": 6.726666666666667e-06, "loss": 0.0909, "step": 1975 }, { "epoch": 2.828854314002829, "grad_norm": 4.304833889007568, "learning_rate": 6.671111111111112e-06, "loss": 0.0989, "step": 2000 }, { "epoch": 2.828854314002829, "eval_loss": 0.2774756848812103, "eval_runtime": 1765.8044, "eval_samples_per_second": 3.047, "eval_steps_per_second": 0.191, "eval_wer": 0.20720505785518312, "step": 2000 }, { "epoch": 2.864214992927864, "grad_norm": 5.535168170928955, "learning_rate": 6.615555555555556e-06, "loss": 0.1188, "step": 2025 }, { "epoch": 2.8995756718528995, "grad_norm": 5.256195068359375, "learning_rate": 6.560000000000001e-06, "loss": 0.097, "step": 2050 }, { "epoch": 2.934936350777935, "grad_norm": 4.056241035461426, "learning_rate": 6.504444444444446e-06, "loss": 0.096, "step": 2075 }, { "epoch": 2.9702970297029703, "grad_norm": 4.049830913543701, "learning_rate": 6.448888888888889e-06, "loss": 0.1031, "step": 2100 }, { "epoch": 3.005657708628006, "grad_norm": 2.4076595306396484, "learning_rate": 6.393333333333334e-06, "loss": 0.0931, "step": 2125 }, { "epoch": 3.041018387553041, "grad_norm": 1.8493136167526245, "learning_rate": 6.3377777777777786e-06, "loss": 0.0591, "step": 2150 }, { "epoch": 3.0763790664780766, "grad_norm": 2.923140525817871, "learning_rate": 6.282222222222223e-06, "loss": 0.0498, "step": 2175 }, { "epoch": 3.1117397454031117, "grad_norm": 3.3305506706237793, "learning_rate": 6.2266666666666675e-06, "loss": 0.0538, "step": 2200 }, { "epoch": 3.1471004243281473, "grad_norm": 2.396880626678467, "learning_rate": 6.171111111111112e-06, "loss": 0.0496, "step": 2225 }, { "epoch": 3.1824611032531824, "grad_norm": 3.4060046672821045, "learning_rate": 6.1155555555555555e-06, "loss": 0.0559, "step": 2250 }, { "epoch": 3.217821782178218, "grad_norm": 4.2796478271484375, "learning_rate": 6.0600000000000004e-06, "loss": 0.0479, "step": 2275 }, { "epoch": 3.253182461103253, "grad_norm": 3.096123456954956, "learning_rate": 6.004444444444445e-06, "loss": 0.0547, "step": 2300 }, { "epoch": 3.2885431400282887, "grad_norm": 3.5626742839813232, "learning_rate": 5.948888888888889e-06, "loss": 0.058, "step": 2325 }, { "epoch": 3.323903818953324, "grad_norm": 3.2983179092407227, "learning_rate": 5.893333333333334e-06, "loss": 0.0586, "step": 2350 }, { "epoch": 3.3592644978783595, "grad_norm": 3.978823184967041, "learning_rate": 5.837777777777777e-06, "loss": 0.0554, "step": 2375 }, { "epoch": 3.3946251768033946, "grad_norm": 3.5360333919525146, "learning_rate": 5.782222222222222e-06, "loss": 0.0549, "step": 2400 }, { "epoch": 3.42998585572843, "grad_norm": 3.2348575592041016, "learning_rate": 5.726666666666667e-06, "loss": 0.0609, "step": 2425 }, { "epoch": 3.4653465346534653, "grad_norm": 4.142767906188965, "learning_rate": 5.671111111111112e-06, "loss": 0.0615, "step": 2450 }, { "epoch": 3.500707213578501, "grad_norm": 3.2305757999420166, "learning_rate": 5.615555555555556e-06, "loss": 0.0485, "step": 2475 }, { "epoch": 3.536067892503536, "grad_norm": 2.9658920764923096, "learning_rate": 5.560000000000001e-06, "loss": 0.054, "step": 2500 }, { "epoch": 3.571428571428571, "grad_norm": 3.1340267658233643, "learning_rate": 5.504444444444444e-06, "loss": 0.0633, "step": 2525 }, { "epoch": 3.6067892503536068, "grad_norm": 2.692047119140625, "learning_rate": 5.448888888888889e-06, "loss": 0.0549, "step": 2550 }, { "epoch": 3.6421499292786423, "grad_norm": 2.9761712551116943, "learning_rate": 5.393333333333334e-06, "loss": 0.0537, "step": 2575 }, { "epoch": 3.6775106082036775, "grad_norm": 2.88071346282959, "learning_rate": 5.337777777777779e-06, "loss": 0.0497, "step": 2600 }, { "epoch": 3.7128712871287126, "grad_norm": 3.6829402446746826, "learning_rate": 5.282222222222223e-06, "loss": 0.0609, "step": 2625 }, { "epoch": 3.748231966053748, "grad_norm": 4.3834614753723145, "learning_rate": 5.226666666666667e-06, "loss": 0.0567, "step": 2650 }, { "epoch": 3.783592644978784, "grad_norm": 2.6627509593963623, "learning_rate": 5.171111111111111e-06, "loss": 0.052, "step": 2675 }, { "epoch": 3.818953323903819, "grad_norm": 3.6826014518737793, "learning_rate": 5.115555555555556e-06, "loss": 0.0529, "step": 2700 }, { "epoch": 3.854314002828854, "grad_norm": 3.048292636871338, "learning_rate": 5.060000000000001e-06, "loss": 0.0617, "step": 2725 }, { "epoch": 3.8896746817538896, "grad_norm": 2.5703887939453125, "learning_rate": 5.004444444444445e-06, "loss": 0.0507, "step": 2750 }, { "epoch": 3.9250353606789252, "grad_norm": 2.8455772399902344, "learning_rate": 4.94888888888889e-06, "loss": 0.0562, "step": 2775 }, { "epoch": 3.9603960396039604, "grad_norm": 2.147115468978882, "learning_rate": 4.893333333333334e-06, "loss": 0.0479, "step": 2800 }, { "epoch": 3.9957567185289955, "grad_norm": 4.647951126098633, "learning_rate": 4.837777777777778e-06, "loss": 0.0555, "step": 2825 }, { "epoch": 4.0311173974540315, "grad_norm": 1.5333168506622314, "learning_rate": 4.7822222222222226e-06, "loss": 0.0241, "step": 2850 }, { "epoch": 4.066478076379067, "grad_norm": 2.611342668533325, "learning_rate": 4.728888888888889e-06, "loss": 0.0297, "step": 2875 }, { "epoch": 4.101838755304102, "grad_norm": 1.1682252883911133, "learning_rate": 4.673333333333333e-06, "loss": 0.028, "step": 2900 }, { "epoch": 4.137199434229137, "grad_norm": 1.971526861190796, "learning_rate": 4.617777777777778e-06, "loss": 0.0256, "step": 2925 }, { "epoch": 4.172560113154173, "grad_norm": 2.9451370239257812, "learning_rate": 4.562222222222222e-06, "loss": 0.0248, "step": 2950 }, { "epoch": 4.207920792079208, "grad_norm": 2.203016757965088, "learning_rate": 4.506666666666667e-06, "loss": 0.0254, "step": 2975 }, { "epoch": 4.243281471004243, "grad_norm": 3.0037150382995605, "learning_rate": 4.451111111111112e-06, "loss": 0.0295, "step": 3000 }, { "epoch": 4.243281471004243, "eval_loss": 0.29346707463264465, "eval_runtime": 1760.8686, "eval_samples_per_second": 3.056, "eval_steps_per_second": 0.191, "eval_wer": 0.1918962980635413, "step": 3000 }, { "epoch": 4.278642149929278, "grad_norm": 2.936363697052002, "learning_rate": 4.395555555555556e-06, "loss": 0.0235, "step": 3025 }, { "epoch": 4.314002828854314, "grad_norm": 1.909324288368225, "learning_rate": 4.34e-06, "loss": 0.027, "step": 3050 }, { "epoch": 4.3493635077793495, "grad_norm": 1.8977993726730347, "learning_rate": 4.284444444444445e-06, "loss": 0.0267, "step": 3075 }, { "epoch": 4.384724186704385, "grad_norm": 2.092729330062866, "learning_rate": 4.228888888888889e-06, "loss": 0.0277, "step": 3100 }, { "epoch": 4.42008486562942, "grad_norm": 2.539109945297241, "learning_rate": 4.173333333333334e-06, "loss": 0.0232, "step": 3125 }, { "epoch": 4.455445544554456, "grad_norm": 1.8122098445892334, "learning_rate": 4.117777777777779e-06, "loss": 0.0276, "step": 3150 }, { "epoch": 4.490806223479491, "grad_norm": 1.4347445964813232, "learning_rate": 4.062222222222223e-06, "loss": 0.021, "step": 3175 }, { "epoch": 4.526166902404526, "grad_norm": 2.03843355178833, "learning_rate": 4.006666666666667e-06, "loss": 0.0278, "step": 3200 }, { "epoch": 4.561527581329561, "grad_norm": 2.6143271923065186, "learning_rate": 3.951111111111112e-06, "loss": 0.0316, "step": 3225 }, { "epoch": 4.596888260254596, "grad_norm": 2.1741061210632324, "learning_rate": 3.895555555555556e-06, "loss": 0.0277, "step": 3250 }, { "epoch": 4.632248939179632, "grad_norm": 3.5832366943359375, "learning_rate": 3.8400000000000005e-06, "loss": 0.025, "step": 3275 }, { "epoch": 4.667609618104668, "grad_norm": 2.232801675796509, "learning_rate": 3.784444444444445e-06, "loss": 0.0282, "step": 3300 }, { "epoch": 4.702970297029703, "grad_norm": 2.736708641052246, "learning_rate": 3.728888888888889e-06, "loss": 0.0308, "step": 3325 }, { "epoch": 4.738330975954739, "grad_norm": 2.0116610527038574, "learning_rate": 3.673333333333334e-06, "loss": 0.0305, "step": 3350 }, { "epoch": 4.773691654879774, "grad_norm": 2.6088578701019287, "learning_rate": 3.617777777777778e-06, "loss": 0.025, "step": 3375 }, { "epoch": 4.809052333804809, "grad_norm": 1.400682806968689, "learning_rate": 3.5622222222222224e-06, "loss": 0.0248, "step": 3400 }, { "epoch": 4.844413012729844, "grad_norm": 1.956292748451233, "learning_rate": 3.5066666666666673e-06, "loss": 0.0258, "step": 3425 }, { "epoch": 4.879773691654879, "grad_norm": 2.2373459339141846, "learning_rate": 3.4511111111111113e-06, "loss": 0.0229, "step": 3450 }, { "epoch": 4.915134370579915, "grad_norm": 2.9303152561187744, "learning_rate": 3.3955555555555558e-06, "loss": 0.0255, "step": 3475 }, { "epoch": 4.9504950495049505, "grad_norm": 2.3725924491882324, "learning_rate": 3.3400000000000006e-06, "loss": 0.0289, "step": 3500 }, { "epoch": 4.985855728429986, "grad_norm": 2.325536012649536, "learning_rate": 3.2844444444444447e-06, "loss": 0.0277, "step": 3525 }, { "epoch": 5.021216407355022, "grad_norm": 3.026657819747925, "learning_rate": 3.228888888888889e-06, "loss": 0.0184, "step": 3550 }, { "epoch": 5.056577086280057, "grad_norm": 1.5165703296661377, "learning_rate": 3.173333333333334e-06, "loss": 0.0148, "step": 3575 }, { "epoch": 5.091937765205092, "grad_norm": 1.2629361152648926, "learning_rate": 3.117777777777778e-06, "loss": 0.0104, "step": 3600 }, { "epoch": 5.127298444130127, "grad_norm": 1.312657117843628, "learning_rate": 3.0622222222222225e-06, "loss": 0.0132, "step": 3625 }, { "epoch": 5.162659123055163, "grad_norm": 1.1759546995162964, "learning_rate": 3.0066666666666674e-06, "loss": 0.0096, "step": 3650 }, { "epoch": 5.198019801980198, "grad_norm": 0.8515748381614685, "learning_rate": 2.9511111111111114e-06, "loss": 0.0125, "step": 3675 }, { "epoch": 5.233380480905233, "grad_norm": 1.0795273780822754, "learning_rate": 2.895555555555556e-06, "loss": 0.0134, "step": 3700 }, { "epoch": 5.2687411598302685, "grad_norm": 1.2719004154205322, "learning_rate": 2.84e-06, "loss": 0.0104, "step": 3725 }, { "epoch": 5.3041018387553045, "grad_norm": 0.8354203701019287, "learning_rate": 2.784444444444445e-06, "loss": 0.0106, "step": 3750 }, { "epoch": 5.33946251768034, "grad_norm": 2.3398609161376953, "learning_rate": 2.7288888888888893e-06, "loss": 0.0126, "step": 3775 }, { "epoch": 5.374823196605375, "grad_norm": 2.567042350769043, "learning_rate": 2.6733333333333333e-06, "loss": 0.0109, "step": 3800 }, { "epoch": 5.41018387553041, "grad_norm": 1.6744190454483032, "learning_rate": 2.617777777777778e-06, "loss": 0.0117, "step": 3825 }, { "epoch": 5.445544554455446, "grad_norm": 1.0200034379959106, "learning_rate": 2.5622222222222226e-06, "loss": 0.0118, "step": 3850 }, { "epoch": 5.480905233380481, "grad_norm": 0.8212645053863525, "learning_rate": 2.5066666666666667e-06, "loss": 0.0106, "step": 3875 }, { "epoch": 5.516265912305516, "grad_norm": 0.9394924640655518, "learning_rate": 2.451111111111111e-06, "loss": 0.0103, "step": 3900 }, { "epoch": 5.551626591230551, "grad_norm": 2.0436089038848877, "learning_rate": 2.3955555555555556e-06, "loss": 0.0177, "step": 3925 }, { "epoch": 5.586987270155587, "grad_norm": 1.194536566734314, "learning_rate": 2.3400000000000005e-06, "loss": 0.0105, "step": 3950 }, { "epoch": 5.6223479490806225, "grad_norm": 1.526948094367981, "learning_rate": 2.2844444444444445e-06, "loss": 0.0097, "step": 3975 }, { "epoch": 5.657708628005658, "grad_norm": 1.8187007904052734, "learning_rate": 2.228888888888889e-06, "loss": 0.0109, "step": 4000 }, { "epoch": 5.657708628005658, "eval_loss": 0.30111825466156006, "eval_runtime": 1778.4921, "eval_samples_per_second": 3.026, "eval_steps_per_second": 0.189, "eval_wer": 0.18279056821344786, "step": 4000 }, { "epoch": 5.693069306930693, "grad_norm": 2.2238926887512207, "learning_rate": 2.1733333333333334e-06, "loss": 0.0096, "step": 4025 }, { "epoch": 5.728429985855728, "grad_norm": 1.989168405532837, "learning_rate": 2.117777777777778e-06, "loss": 0.0136, "step": 4050 }, { "epoch": 5.763790664780764, "grad_norm": 1.6722816228866577, "learning_rate": 2.0622222222222223e-06, "loss": 0.0107, "step": 4075 }, { "epoch": 5.799151343705799, "grad_norm": 1.0947668552398682, "learning_rate": 2.006666666666667e-06, "loss": 0.0083, "step": 4100 }, { "epoch": 5.834512022630834, "grad_norm": 0.5356242656707764, "learning_rate": 1.9511111111111113e-06, "loss": 0.0108, "step": 4125 }, { "epoch": 5.86987270155587, "grad_norm": 2.1374714374542236, "learning_rate": 1.8955555555555557e-06, "loss": 0.0107, "step": 4150 }, { "epoch": 5.905233380480905, "grad_norm": 0.9695949554443359, "learning_rate": 1.8400000000000002e-06, "loss": 0.0103, "step": 4175 }, { "epoch": 5.9405940594059405, "grad_norm": 1.3653473854064941, "learning_rate": 1.7844444444444444e-06, "loss": 0.0116, "step": 4200 }, { "epoch": 5.975954738330976, "grad_norm": 1.5049793720245361, "learning_rate": 1.728888888888889e-06, "loss": 0.0129, "step": 4225 }, { "epoch": 6.011315417256012, "grad_norm": 2.540705442428589, "learning_rate": 1.6733333333333335e-06, "loss": 0.0085, "step": 4250 }, { "epoch": 6.046676096181047, "grad_norm": 0.5889933705329895, "learning_rate": 1.6177777777777778e-06, "loss": 0.0043, "step": 4275 }, { "epoch": 6.082036775106082, "grad_norm": 1.6074843406677246, "learning_rate": 1.5622222222222225e-06, "loss": 0.006, "step": 4300 }, { "epoch": 6.117397454031117, "grad_norm": 0.8578146696090698, "learning_rate": 1.506666666666667e-06, "loss": 0.0054, "step": 4325 }, { "epoch": 6.152758132956153, "grad_norm": 0.6143340468406677, "learning_rate": 1.4511111111111112e-06, "loss": 0.0043, "step": 4350 }, { "epoch": 6.188118811881188, "grad_norm": 0.3723517656326294, "learning_rate": 1.3955555555555556e-06, "loss": 0.0034, "step": 4375 }, { "epoch": 6.223479490806223, "grad_norm": 1.0041955709457397, "learning_rate": 1.34e-06, "loss": 0.0059, "step": 4400 }, { "epoch": 6.258840169731259, "grad_norm": 1.2664568424224854, "learning_rate": 1.2844444444444445e-06, "loss": 0.0039, "step": 4425 }, { "epoch": 6.294200848656295, "grad_norm": 0.259546160697937, "learning_rate": 1.228888888888889e-06, "loss": 0.0051, "step": 4450 }, { "epoch": 6.32956152758133, "grad_norm": 0.9940450191497803, "learning_rate": 1.1733333333333335e-06, "loss": 0.0027, "step": 4475 }, { "epoch": 6.364922206506365, "grad_norm": 0.23329602181911469, "learning_rate": 1.117777777777778e-06, "loss": 0.0038, "step": 4500 }, { "epoch": 6.4002828854314, "grad_norm": 0.12971219420433044, "learning_rate": 1.0622222222222222e-06, "loss": 0.0035, "step": 4525 }, { "epoch": 6.435643564356436, "grad_norm": 0.4987533688545227, "learning_rate": 1.0066666666666668e-06, "loss": 0.0081, "step": 4550 }, { "epoch": 6.471004243281471, "grad_norm": 0.27808260917663574, "learning_rate": 9.511111111111111e-07, "loss": 0.0034, "step": 4575 }, { "epoch": 6.506364922206506, "grad_norm": 0.23258398473262787, "learning_rate": 8.955555555555557e-07, "loss": 0.0033, "step": 4600 }, { "epoch": 6.5417256011315414, "grad_norm": 0.3250856101512909, "learning_rate": 8.400000000000001e-07, "loss": 0.0038, "step": 4625 }, { "epoch": 6.5770862800565775, "grad_norm": 0.5452969074249268, "learning_rate": 7.844444444444445e-07, "loss": 0.0033, "step": 4650 }, { "epoch": 6.612446958981613, "grad_norm": 0.22152584791183472, "learning_rate": 7.28888888888889e-07, "loss": 0.0067, "step": 4675 }, { "epoch": 6.647807637906648, "grad_norm": 1.5061694383621216, "learning_rate": 6.733333333333334e-07, "loss": 0.0081, "step": 4700 }, { "epoch": 6.683168316831683, "grad_norm": 0.2032124400138855, "learning_rate": 6.177777777777778e-07, "loss": 0.0042, "step": 4725 }, { "epoch": 6.718528995756719, "grad_norm": 0.25647231936454773, "learning_rate": 5.622222222222223e-07, "loss": 0.0061, "step": 4750 }, { "epoch": 6.753889674681754, "grad_norm": 0.27424803376197815, "learning_rate": 5.066666666666667e-07, "loss": 0.0034, "step": 4775 }, { "epoch": 6.789250353606789, "grad_norm": 0.4183076322078705, "learning_rate": 4.511111111111111e-07, "loss": 0.004, "step": 4800 }, { "epoch": 6.824611032531824, "grad_norm": 0.40847113728523254, "learning_rate": 3.9555555555555557e-07, "loss": 0.0029, "step": 4825 }, { "epoch": 6.85997171145686, "grad_norm": 0.2560099959373474, "learning_rate": 3.4000000000000003e-07, "loss": 0.0028, "step": 4850 }, { "epoch": 6.8953323903818955, "grad_norm": 0.13006103038787842, "learning_rate": 2.844444444444445e-07, "loss": 0.0042, "step": 4875 }, { "epoch": 6.930693069306931, "grad_norm": 0.38391485810279846, "learning_rate": 2.2888888888888892e-07, "loss": 0.0029, "step": 4900 }, { "epoch": 6.966053748231966, "grad_norm": 0.4023485481739044, "learning_rate": 1.7333333333333335e-07, "loss": 0.0027, "step": 4925 }, { "epoch": 7.001414427157002, "grad_norm": 0.1441173106431961, "learning_rate": 1.1777777777777778e-07, "loss": 0.0042, "step": 4950 }, { "epoch": 7.036775106082037, "grad_norm": 0.18802694976329803, "learning_rate": 6.222222222222223e-08, "loss": 0.0033, "step": 4975 }, { "epoch": 7.072135785007072, "grad_norm": 0.14834970235824585, "learning_rate": 6.666666666666667e-09, "loss": 0.0016, "step": 5000 }, { "epoch": 7.072135785007072, "eval_loss": 0.3134021461009979, "eval_runtime": 1756.5453, "eval_samples_per_second": 3.063, "eval_steps_per_second": 0.192, "eval_wer": 0.17463915066205415, "step": 5000 }, { "epoch": 7.072135785007072, "step": 5000, "total_flos": 2.7263703386750976e+20, "train_loss": 0.12853609090298415, "train_runtime": 49407.2097, "train_samples_per_second": 3.238, "train_steps_per_second": 0.101 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 8, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.7263703386750976e+20, "train_batch_size": 16, "trial_name": null, "trial_params": null }