{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9299820466786355, "eval_steps": 500, "global_step": 51, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05745062836624776, "grad_norm": 3.1636736392974854, "learning_rate": 5e-06, "loss": 2.3037, "step": 1 }, { "epoch": 0.11490125673249552, "grad_norm": 3.3778610229492188, "learning_rate": 1e-05, "loss": 2.3337, "step": 2 }, { "epoch": 0.17235188509874327, "grad_norm": 3.0278897285461426, "learning_rate": 1.5e-05, "loss": 2.2715, "step": 3 }, { "epoch": 0.22980251346499103, "grad_norm": 2.5405051708221436, "learning_rate": 2e-05, "loss": 2.2422, "step": 4 }, { "epoch": 0.2872531418312388, "grad_norm": 1.789209246635437, "learning_rate": 2.5e-05, "loss": 2.2079, "step": 5 }, { "epoch": 0.34470377019748655, "grad_norm": 1.653626561164856, "learning_rate": 3e-05, "loss": 2.2231, "step": 6 }, { "epoch": 0.4021543985637343, "grad_norm": 2.006115436553955, "learning_rate": 3.5e-05, "loss": 2.2855, "step": 7 }, { "epoch": 0.45960502692998206, "grad_norm": 2.322628974914551, "learning_rate": 4e-05, "loss": 2.2126, "step": 8 }, { "epoch": 0.5170556552962298, "grad_norm": 2.069558620452881, "learning_rate": 4.5e-05, "loss": 2.2389, "step": 9 }, { "epoch": 0.5745062836624776, "grad_norm": 1.7339693307876587, "learning_rate": 5e-05, "loss": 2.1964, "step": 10 }, { "epoch": 0.6319569120287253, "grad_norm": 1.6608643531799316, "learning_rate": 4.992664502959351e-05, "loss": 2.2079, "step": 11 }, { "epoch": 0.6894075403949731, "grad_norm": 1.642751932144165, "learning_rate": 4.970701059450872e-05, "loss": 2.2359, "step": 12 }, { "epoch": 0.7468581687612208, "grad_norm": 1.6817108392715454, "learning_rate": 4.934238559694448e-05, "loss": 2.2229, "step": 13 }, { "epoch": 0.8043087971274686, "grad_norm": 1.6007517576217651, "learning_rate": 4.8834909801373264e-05, "loss": 2.203, "step": 14 }, { "epoch": 0.8617594254937163, "grad_norm": 1.457571029663086, "learning_rate": 4.8187561277552374e-05, "loss": 2.0994, "step": 15 }, { "epoch": 0.9192100538599641, "grad_norm": 1.486559271812439, "learning_rate": 4.740413892402639e-05, "loss": 2.1777, "step": 16 }, { "epoch": 0.9766606822262118, "grad_norm": 1.5398966073989868, "learning_rate": 4.648924017468003e-05, "loss": 2.2126, "step": 17 }, { "epoch": 1.0341113105924595, "grad_norm": 1.4850305318832397, "learning_rate": 4.5448234019167945e-05, "loss": 2.1019, "step": 18 }, { "epoch": 1.0915619389587075, "grad_norm": 1.4637010097503662, "learning_rate": 4.428722949554857e-05, "loss": 2.1119, "step": 19 }, { "epoch": 1.1490125673249552, "grad_norm": 1.3282859325408936, "learning_rate": 4.301303984001967e-05, "loss": 2.0976, "step": 20 }, { "epoch": 1.2064631956912029, "grad_norm": 1.2366828918457031, "learning_rate": 4.163314250413913e-05, "loss": 2.0682, "step": 21 }, { "epoch": 1.2639138240574506, "grad_norm": 1.3221453428268433, "learning_rate": 4.015563527416595e-05, "loss": 2.0266, "step": 22 }, { "epoch": 1.3213644524236985, "grad_norm": 1.269490122795105, "learning_rate": 3.858918875003053e-05, "loss": 2.011, "step": 23 }, { "epoch": 1.3788150807899462, "grad_norm": 1.2723491191864014, "learning_rate": 3.694299546280657e-05, "loss": 2.0503, "step": 24 }, { "epoch": 1.436265709156194, "grad_norm": 1.3157715797424316, "learning_rate": 3.5226715929283506e-05, "loss": 2.0297, "step": 25 }, { "epoch": 1.4937163375224416, "grad_norm": 1.367810606956482, "learning_rate": 3.3450421960212566e-05, "loss": 2.03, "step": 26 }, { "epoch": 1.5511669658886893, "grad_norm": 1.4925018548965454, "learning_rate": 3.162453755491655e-05, "loss": 2.0478, "step": 27 }, { "epoch": 1.608617594254937, "grad_norm": 1.350865125656128, "learning_rate": 2.975977772911671e-05, "loss": 2.1047, "step": 28 }, { "epoch": 1.666068222621185, "grad_norm": 1.3046050071716309, "learning_rate": 2.7867085634960016e-05, "loss": 2.0175, "step": 29 }, { "epoch": 1.7235188509874326, "grad_norm": 1.3713139295578003, "learning_rate": 2.595756834225089e-05, "loss": 2.0437, "step": 30 }, { "epoch": 1.7809694793536806, "grad_norm": 1.2804160118103027, "learning_rate": 2.4042431657749117e-05, "loss": 2.0573, "step": 31 }, { "epoch": 1.8384201077199283, "grad_norm": 1.257683277130127, "learning_rate": 2.2132914365039993e-05, "loss": 2.0105, "step": 32 }, { "epoch": 1.895870736086176, "grad_norm": 1.2399568557739258, "learning_rate": 2.0240222270883288e-05, "loss": 2.0335, "step": 33 }, { "epoch": 1.9533213644524237, "grad_norm": 1.1331653594970703, "learning_rate": 1.8375462445083464e-05, "loss": 2.0217, "step": 34 }, { "epoch": 2.0107719928186714, "grad_norm": 1.219581961631775, "learning_rate": 1.6549578039787436e-05, "loss": 2.0096, "step": 35 }, { "epoch": 2.068222621184919, "grad_norm": 1.239498257637024, "learning_rate": 1.4773284070716503e-05, "loss": 2.0598, "step": 36 }, { "epoch": 2.1256732495511668, "grad_norm": 1.2558954954147339, "learning_rate": 1.3057004537193423e-05, "loss": 2.0412, "step": 37 }, { "epoch": 2.183123877917415, "grad_norm": 1.0945749282836914, "learning_rate": 1.1410811249969475e-05, "loss": 1.959, "step": 38 }, { "epoch": 2.2405745062836626, "grad_norm": 1.0557724237442017, "learning_rate": 9.844364725834057e-06, "loss": 1.9716, "step": 39 }, { "epoch": 2.2980251346499103, "grad_norm": 1.1883456707000732, "learning_rate": 8.36685749586087e-06, "loss": 1.9485, "step": 40 }, { "epoch": 2.355475763016158, "grad_norm": 1.2525312900543213, "learning_rate": 6.986960159980327e-06, "loss": 2.0025, "step": 41 }, { "epoch": 2.4129263913824057, "grad_norm": 1.2137963771820068, "learning_rate": 5.712770504451426e-06, "loss": 1.9968, "step": 42 }, { "epoch": 2.4703770197486534, "grad_norm": 1.1027462482452393, "learning_rate": 4.551765980832059e-06, "loss": 2.0205, "step": 43 }, { "epoch": 2.527827648114901, "grad_norm": 1.1979150772094727, "learning_rate": 3.5107598253199758e-06, "loss": 1.9487, "step": 44 }, { "epoch": 2.585278276481149, "grad_norm": 1.2198636531829834, "learning_rate": 2.595861075973613e-06, "loss": 2.0071, "step": 45 }, { "epoch": 2.642728904847397, "grad_norm": 1.2485204935073853, "learning_rate": 1.8124387224476347e-06, "loss": 1.8711, "step": 46 }, { "epoch": 2.7001795332136447, "grad_norm": 1.2243603467941284, "learning_rate": 1.1650901986267365e-06, "loss": 1.9128, "step": 47 }, { "epoch": 2.7576301615798924, "grad_norm": 1.1391812562942505, "learning_rate": 6.576144030555259e-07, "loss": 1.9854, "step": 48 }, { "epoch": 2.81508078994614, "grad_norm": 1.2900848388671875, "learning_rate": 2.9298940549128964e-07, "loss": 1.9512, "step": 49 }, { "epoch": 2.872531418312388, "grad_norm": 1.1324858665466309, "learning_rate": 7.335497040648898e-08, "loss": 1.9581, "step": 50 }, { "epoch": 2.9299820466786355, "grad_norm": 1.2011033296585083, "learning_rate": 0.0, "loss": 1.9738, "step": 51 } ], "logging_steps": 1, "max_steps": 51, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 4, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.49530087730176e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }