{ "best_metric": 0.44272685050964355, "best_model_checkpoint": "data/Llama-31-8B_task-3_120-samples_config-4/checkpoint-214", "epoch": 46.0, "eval_steps": 500, "global_step": 253, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.18181818181818182, "grad_norm": 3.1690733432769775, "learning_rate": 1.3333333333333336e-07, "loss": 2.8174, "step": 1 }, { "epoch": 0.36363636363636365, "grad_norm": 2.785377025604248, "learning_rate": 2.666666666666667e-07, "loss": 2.5884, "step": 2 }, { "epoch": 0.7272727272727273, "grad_norm": 2.4831278324127197, "learning_rate": 5.333333333333335e-07, "loss": 2.4707, "step": 4 }, { "epoch": 0.9090909090909091, "eval_loss": 2.502408742904663, "eval_runtime": 20.9991, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 5 }, { "epoch": 1.0909090909090908, "grad_norm": 2.511209726333618, "learning_rate": 8.000000000000001e-07, "loss": 2.3921, "step": 6 }, { "epoch": 1.4545454545454546, "grad_norm": 3.1637580394744873, "learning_rate": 1.066666666666667e-06, "loss": 2.5814, "step": 8 }, { "epoch": 1.8181818181818183, "grad_norm": 2.612966775894165, "learning_rate": 1.3333333333333334e-06, "loss": 2.3847, "step": 10 }, { "epoch": 2.0, "eval_loss": 2.479367971420288, "eval_runtime": 21.0052, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 11 }, { "epoch": 2.1818181818181817, "grad_norm": 2.905508518218994, "learning_rate": 1.6000000000000001e-06, "loss": 2.6936, "step": 12 }, { "epoch": 2.5454545454545454, "grad_norm": 2.6060760021209717, "learning_rate": 1.8666666666666669e-06, "loss": 2.3757, "step": 14 }, { "epoch": 2.909090909090909, "grad_norm": 3.210188150405884, "learning_rate": 2.133333333333334e-06, "loss": 2.5822, "step": 16 }, { "epoch": 2.909090909090909, "eval_loss": 2.4506218433380127, "eval_runtime": 21.0015, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 16 }, { "epoch": 3.2727272727272725, "grad_norm": 3.861948013305664, "learning_rate": 2.4000000000000003e-06, "loss": 2.6532, "step": 18 }, { "epoch": 3.6363636363636362, "grad_norm": 2.5222151279449463, "learning_rate": 2.666666666666667e-06, "loss": 2.3743, "step": 20 }, { "epoch": 4.0, "grad_norm": 2.5924019813537598, "learning_rate": 2.9333333333333338e-06, "loss": 2.2635, "step": 22 }, { "epoch": 4.0, "eval_loss": 2.3902969360351562, "eval_runtime": 21.0176, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 22 }, { "epoch": 4.363636363636363, "grad_norm": 2.239548921585083, "learning_rate": 3.2000000000000003e-06, "loss": 2.4045, "step": 24 }, { "epoch": 4.7272727272727275, "grad_norm": 2.6937167644500732, "learning_rate": 3.4666666666666672e-06, "loss": 2.446, "step": 26 }, { "epoch": 4.909090909090909, "eval_loss": 2.3183953762054443, "eval_runtime": 21.0027, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 27 }, { "epoch": 5.090909090909091, "grad_norm": 2.26949143409729, "learning_rate": 3.7333333333333337e-06, "loss": 2.1955, "step": 28 }, { "epoch": 5.454545454545454, "grad_norm": 3.2933804988861084, "learning_rate": 4.000000000000001e-06, "loss": 2.3794, "step": 30 }, { "epoch": 5.818181818181818, "grad_norm": 3.081024169921875, "learning_rate": 4.266666666666668e-06, "loss": 2.3172, "step": 32 }, { "epoch": 6.0, "eval_loss": 2.19378399848938, "eval_runtime": 20.9996, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 33 }, { "epoch": 6.181818181818182, "grad_norm": 3.226388454437256, "learning_rate": 4.533333333333334e-06, "loss": 2.2373, "step": 34 }, { "epoch": 6.545454545454545, "grad_norm": 1.6646192073822021, "learning_rate": 4.800000000000001e-06, "loss": 2.0122, "step": 36 }, { "epoch": 6.909090909090909, "grad_norm": 2.573935031890869, "learning_rate": 5.0666666666666676e-06, "loss": 2.0582, "step": 38 }, { "epoch": 6.909090909090909, "eval_loss": 2.0559873580932617, "eval_runtime": 21.0001, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 38 }, { "epoch": 7.2727272727272725, "grad_norm": 2.028982639312744, "learning_rate": 5.333333333333334e-06, "loss": 1.9739, "step": 40 }, { "epoch": 7.636363636363637, "grad_norm": 2.90993595123291, "learning_rate": 5.600000000000001e-06, "loss": 2.0373, "step": 42 }, { "epoch": 8.0, "grad_norm": 2.5906777381896973, "learning_rate": 5.8666666666666675e-06, "loss": 1.9038, "step": 44 }, { "epoch": 8.0, "eval_loss": 1.8385127782821655, "eval_runtime": 20.9958, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 44 }, { "epoch": 8.363636363636363, "grad_norm": 2.548511266708374, "learning_rate": 6.133333333333334e-06, "loss": 1.7707, "step": 46 }, { "epoch": 8.727272727272727, "grad_norm": 2.2620716094970703, "learning_rate": 6.4000000000000006e-06, "loss": 1.7291, "step": 48 }, { "epoch": 8.909090909090908, "eval_loss": 1.625160574913025, "eval_runtime": 20.9978, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 49 }, { "epoch": 9.090909090909092, "grad_norm": 2.2526345252990723, "learning_rate": 6.666666666666667e-06, "loss": 1.607, "step": 50 }, { "epoch": 9.454545454545455, "grad_norm": 1.2636234760284424, "learning_rate": 6.9333333333333344e-06, "loss": 1.5173, "step": 52 }, { "epoch": 9.818181818181818, "grad_norm": 2.455042600631714, "learning_rate": 7.2000000000000005e-06, "loss": 1.3996, "step": 54 }, { "epoch": 10.0, "eval_loss": 1.345643162727356, "eval_runtime": 21.0044, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 55 }, { "epoch": 10.181818181818182, "grad_norm": 2.40154767036438, "learning_rate": 7.4666666666666675e-06, "loss": 1.298, "step": 56 }, { "epoch": 10.545454545454545, "grad_norm": 2.0422167778015137, "learning_rate": 7.733333333333334e-06, "loss": 1.2468, "step": 58 }, { "epoch": 10.909090909090908, "grad_norm": 2.330064058303833, "learning_rate": 8.000000000000001e-06, "loss": 1.1127, "step": 60 }, { "epoch": 10.909090909090908, "eval_loss": 1.1189473867416382, "eval_runtime": 21.0042, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 60 }, { "epoch": 11.272727272727273, "grad_norm": 1.6237120628356934, "learning_rate": 8.266666666666667e-06, "loss": 1.0346, "step": 62 }, { "epoch": 11.636363636363637, "grad_norm": 1.9251773357391357, "learning_rate": 8.533333333333335e-06, "loss": 0.9213, "step": 64 }, { "epoch": 12.0, "grad_norm": 1.3815380334854126, "learning_rate": 8.8e-06, "loss": 0.8648, "step": 66 }, { "epoch": 12.0, "eval_loss": 0.8626702427864075, "eval_runtime": 21.003, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 66 }, { "epoch": 12.363636363636363, "grad_norm": 1.8721394538879395, "learning_rate": 9.066666666666667e-06, "loss": 0.6295, "step": 68 }, { "epoch": 12.727272727272727, "grad_norm": 1.1077913045883179, "learning_rate": 9.333333333333334e-06, "loss": 0.8247, "step": 70 }, { "epoch": 12.909090909090908, "eval_loss": 0.7228104472160339, "eval_runtime": 20.9976, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 71 }, { "epoch": 13.090909090909092, "grad_norm": 0.9280027151107788, "learning_rate": 9.600000000000001e-06, "loss": 0.6643, "step": 72 }, { "epoch": 13.454545454545455, "grad_norm": 0.8919649720191956, "learning_rate": 9.866666666666668e-06, "loss": 0.5425, "step": 74 }, { "epoch": 13.818181818181818, "grad_norm": 0.8291063904762268, "learning_rate": 9.999945845889795e-06, "loss": 0.5681, "step": 76 }, { "epoch": 14.0, "eval_loss": 0.6452742218971252, "eval_runtime": 21.0019, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 77 }, { "epoch": 14.181818181818182, "grad_norm": 0.8017318844795227, "learning_rate": 9.999512620046523e-06, "loss": 0.6102, "step": 78 }, { "epoch": 14.545454545454545, "grad_norm": 0.48972490429878235, "learning_rate": 9.99864620589731e-06, "loss": 0.5325, "step": 80 }, { "epoch": 14.909090909090908, "grad_norm": 0.6467877626419067, "learning_rate": 9.99734667851357e-06, "loss": 0.4968, "step": 82 }, { "epoch": 14.909090909090908, "eval_loss": 0.6019716858863831, "eval_runtime": 21.0041, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 82 }, { "epoch": 15.272727272727273, "grad_norm": 0.4685284495353699, "learning_rate": 9.995614150494293e-06, "loss": 0.4665, "step": 84 }, { "epoch": 15.636363636363637, "grad_norm": 0.6542117595672607, "learning_rate": 9.993448771956285e-06, "loss": 0.4592, "step": 86 }, { "epoch": 16.0, "grad_norm": 0.7055411338806152, "learning_rate": 9.99085073052117e-06, "loss": 0.589, "step": 88 }, { "epoch": 16.0, "eval_loss": 0.5632085800170898, "eval_runtime": 21.0058, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 88 }, { "epoch": 16.363636363636363, "grad_norm": 0.5362458825111389, "learning_rate": 9.987820251299121e-06, "loss": 0.5109, "step": 90 }, { "epoch": 16.727272727272727, "grad_norm": 0.3406112492084503, "learning_rate": 9.984357596869369e-06, "loss": 0.3902, "step": 92 }, { "epoch": 16.90909090909091, "eval_loss": 0.5393884778022766, "eval_runtime": 21.0005, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 93 }, { "epoch": 17.09090909090909, "grad_norm": 0.5570920705795288, "learning_rate": 9.980463067257437e-06, "loss": 0.5325, "step": 94 }, { "epoch": 17.454545454545453, "grad_norm": 0.40238556265830994, "learning_rate": 9.976136999909156e-06, "loss": 0.4039, "step": 96 }, { "epoch": 17.818181818181817, "grad_norm": 0.6511011719703674, "learning_rate": 9.971379769661422e-06, "loss": 0.4795, "step": 98 }, { "epoch": 18.0, "eval_loss": 0.5253307819366455, "eval_runtime": 21.0047, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 99 }, { "epoch": 18.181818181818183, "grad_norm": 0.4080711305141449, "learning_rate": 9.966191788709716e-06, "loss": 0.4767, "step": 100 }, { "epoch": 18.545454545454547, "grad_norm": 0.35794946551322937, "learning_rate": 9.960573506572391e-06, "loss": 0.4024, "step": 102 }, { "epoch": 18.90909090909091, "grad_norm": 0.2140028029680252, "learning_rate": 9.95452541005172e-06, "loss": 0.3937, "step": 104 }, { "epoch": 18.90909090909091, "eval_loss": 0.5187709927558899, "eval_runtime": 20.9985, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 104 }, { "epoch": 19.272727272727273, "grad_norm": 0.26322704553604126, "learning_rate": 9.948048023191728e-06, "loss": 0.3845, "step": 106 }, { "epoch": 19.636363636363637, "grad_norm": 0.29932188987731934, "learning_rate": 9.941141907232766e-06, "loss": 0.5589, "step": 108 }, { "epoch": 20.0, "grad_norm": 0.3077118396759033, "learning_rate": 9.933807660562898e-06, "loss": 0.3482, "step": 110 }, { "epoch": 20.0, "eval_loss": 0.5148636698722839, "eval_runtime": 21.0007, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 110 }, { "epoch": 20.363636363636363, "grad_norm": 0.2408452033996582, "learning_rate": 9.926045918666045e-06, "loss": 0.3168, "step": 112 }, { "epoch": 20.727272727272727, "grad_norm": 0.24809201061725616, "learning_rate": 9.91785735406693e-06, "loss": 0.4633, "step": 114 }, { "epoch": 20.90909090909091, "eval_loss": 0.5076439380645752, "eval_runtime": 21.008, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 115 }, { "epoch": 21.09090909090909, "grad_norm": 0.4201977550983429, "learning_rate": 9.909242676272797e-06, "loss": 0.5529, "step": 116 }, { "epoch": 21.454545454545453, "grad_norm": 0.30744773149490356, "learning_rate": 9.90020263171194e-06, "loss": 0.3469, "step": 118 }, { "epoch": 21.818181818181817, "grad_norm": 0.2413516491651535, "learning_rate": 9.890738003669029e-06, "loss": 0.4324, "step": 120 }, { "epoch": 22.0, "eval_loss": 0.507264256477356, "eval_runtime": 20.9993, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 121 }, { "epoch": 22.181818181818183, "grad_norm": 0.291746586561203, "learning_rate": 9.880849612217238e-06, "loss": 0.325, "step": 122 }, { "epoch": 22.545454545454547, "grad_norm": 0.22690318524837494, "learning_rate": 9.870538314147194e-06, "loss": 0.3262, "step": 124 }, { "epoch": 22.90909090909091, "grad_norm": 0.28505831956863403, "learning_rate": 9.859805002892733e-06, "loss": 0.5268, "step": 126 }, { "epoch": 22.90909090909091, "eval_loss": 0.5014194846153259, "eval_runtime": 21.0084, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 126 }, { "epoch": 23.272727272727273, "grad_norm": 0.33074063062667847, "learning_rate": 9.84865060845349e-06, "loss": 0.4705, "step": 128 }, { "epoch": 23.636363636363637, "grad_norm": 0.24479380249977112, "learning_rate": 9.83707609731432e-06, "loss": 0.3475, "step": 130 }, { "epoch": 24.0, "grad_norm": 0.2692837715148926, "learning_rate": 9.825082472361558e-06, "loss": 0.3829, "step": 132 }, { "epoch": 24.0, "eval_loss": 0.49462565779685974, "eval_runtime": 21.0006, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 132 }, { "epoch": 24.363636363636363, "grad_norm": 0.2481246441602707, "learning_rate": 9.812670772796113e-06, "loss": 0.3245, "step": 134 }, { "epoch": 24.727272727272727, "grad_norm": 0.2520200312137604, "learning_rate": 9.799842074043438e-06, "loss": 0.3884, "step": 136 }, { "epoch": 24.90909090909091, "eval_loss": 0.4874545633792877, "eval_runtime": 21.0022, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 137 }, { "epoch": 25.09090909090909, "grad_norm": 0.34048882126808167, "learning_rate": 9.786597487660336e-06, "loss": 0.5025, "step": 138 }, { "epoch": 25.454545454545453, "grad_norm": 0.2133476287126541, "learning_rate": 9.77293816123866e-06, "loss": 0.3012, "step": 140 }, { "epoch": 25.818181818181817, "grad_norm": 0.3172047734260559, "learning_rate": 9.75886527830587e-06, "loss": 0.3955, "step": 142 }, { "epoch": 26.0, "eval_loss": 0.4859234094619751, "eval_runtime": 21.0019, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 143 }, { "epoch": 26.181818181818183, "grad_norm": 0.299968957901001, "learning_rate": 9.744380058222483e-06, "loss": 0.471, "step": 144 }, { "epoch": 26.545454545454547, "grad_norm": 0.22688376903533936, "learning_rate": 9.729483756076436e-06, "loss": 0.361, "step": 146 }, { "epoch": 26.90909090909091, "grad_norm": 0.20205911993980408, "learning_rate": 9.714177662574316e-06, "loss": 0.3296, "step": 148 }, { "epoch": 26.90909090909091, "eval_loss": 0.482715368270874, "eval_runtime": 21.0014, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 148 }, { "epoch": 27.272727272727273, "grad_norm": 0.2501104772090912, "learning_rate": 9.698463103929542e-06, "loss": 0.3371, "step": 150 }, { "epoch": 27.636363636363637, "grad_norm": 0.2598528265953064, "learning_rate": 9.682341441747446e-06, "loss": 0.439, "step": 152 }, { "epoch": 28.0, "grad_norm": 0.25387951731681824, "learning_rate": 9.665814072907293e-06, "loss": 0.364, "step": 154 }, { "epoch": 28.0, "eval_loss": 0.4805121123790741, "eval_runtime": 21.002, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 154 }, { "epoch": 28.363636363636363, "grad_norm": 0.2787465751171112, "learning_rate": 9.648882429441258e-06, "loss": 0.4797, "step": 156 }, { "epoch": 28.727272727272727, "grad_norm": 0.25384321808815, "learning_rate": 9.63154797841033e-06, "loss": 0.3218, "step": 158 }, { "epoch": 28.90909090909091, "eval_loss": 0.4764772951602936, "eval_runtime": 21.0046, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 159 }, { "epoch": 29.09090909090909, "grad_norm": 0.2499808818101883, "learning_rate": 9.613812221777212e-06, "loss": 0.302, "step": 160 }, { "epoch": 29.454545454545453, "grad_norm": 0.24313043057918549, "learning_rate": 9.595676696276173e-06, "loss": 0.4349, "step": 162 }, { "epoch": 29.818181818181817, "grad_norm": 0.26249364018440247, "learning_rate": 9.577142973279896e-06, "loss": 0.2995, "step": 164 }, { "epoch": 30.0, "eval_loss": 0.4727340638637543, "eval_runtime": 21.0056, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 165 }, { "epoch": 30.181818181818183, "grad_norm": 0.3164134621620178, "learning_rate": 9.55821265866333e-06, "loss": 0.3373, "step": 166 }, { "epoch": 30.545454545454547, "grad_norm": 0.23215709626674652, "learning_rate": 9.538887392664544e-06, "loss": 0.3434, "step": 168 }, { "epoch": 30.90909090909091, "grad_norm": 0.24047435820102692, "learning_rate": 9.519168849742603e-06, "loss": 0.3728, "step": 170 }, { "epoch": 30.90909090909091, "eval_loss": 0.4668484032154083, "eval_runtime": 21.0083, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 170 }, { "epoch": 31.272727272727273, "grad_norm": 0.38881421089172363, "learning_rate": 9.499058738432492e-06, "loss": 0.4329, "step": 172 }, { "epoch": 31.636363636363637, "grad_norm": 0.38587087392807007, "learning_rate": 9.478558801197065e-06, "loss": 0.3596, "step": 174 }, { "epoch": 32.0, "grad_norm": 0.2352934628725052, "learning_rate": 9.457670814276083e-06, "loss": 0.2413, "step": 176 }, { "epoch": 32.0, "eval_loss": 0.46528127789497375, "eval_runtime": 20.9994, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 176 }, { "epoch": 32.36363636363637, "grad_norm": 0.32384243607521057, "learning_rate": 9.436396587532297e-06, "loss": 0.3098, "step": 178 }, { "epoch": 32.72727272727273, "grad_norm": 0.3551454544067383, "learning_rate": 9.414737964294636e-06, "loss": 0.4141, "step": 180 }, { "epoch": 32.90909090909091, "eval_loss": 0.46264728903770447, "eval_runtime": 21.0065, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 181 }, { "epoch": 33.09090909090909, "grad_norm": 0.365192174911499, "learning_rate": 9.392696821198488e-06, "loss": 0.2879, "step": 182 }, { "epoch": 33.45454545454545, "grad_norm": 0.38976219296455383, "learning_rate": 9.370275068023097e-06, "loss": 0.3467, "step": 184 }, { "epoch": 33.81818181818182, "grad_norm": 0.3077632486820221, "learning_rate": 9.347474647526095e-06, "loss": 0.3693, "step": 186 }, { "epoch": 34.0, "eval_loss": 0.4562326967716217, "eval_runtime": 21.0019, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 187 }, { "epoch": 34.18181818181818, "grad_norm": 0.209530308842659, "learning_rate": 9.324297535275156e-06, "loss": 0.1788, "step": 188 }, { "epoch": 34.54545454545455, "grad_norm": 0.36394768953323364, "learning_rate": 9.30074573947683e-06, "loss": 0.3143, "step": 190 }, { "epoch": 34.90909090909091, "grad_norm": 0.41356202960014343, "learning_rate": 9.276821300802535e-06, "loss": 0.3666, "step": 192 }, { "epoch": 34.90909090909091, "eval_loss": 0.45260027050971985, "eval_runtime": 21.0113, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 192 }, { "epoch": 35.27272727272727, "grad_norm": 0.42078325152397156, "learning_rate": 9.25252629221175e-06, "loss": 0.468, "step": 194 }, { "epoch": 35.63636363636363, "grad_norm": 0.30819612741470337, "learning_rate": 9.227862818772392e-06, "loss": 0.1892, "step": 196 }, { "epoch": 36.0, "grad_norm": 0.3855637013912201, "learning_rate": 9.202833017478421e-06, "loss": 0.3605, "step": 198 }, { "epoch": 36.0, "eval_loss": 0.4506441354751587, "eval_runtime": 20.9992, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 198 }, { "epoch": 36.36363636363637, "grad_norm": 0.39254796504974365, "learning_rate": 9.177439057064684e-06, "loss": 0.3432, "step": 200 }, { "epoch": 36.72727272727273, "grad_norm": 0.3540774881839752, "learning_rate": 9.151683137818989e-06, "loss": 0.2923, "step": 202 }, { "epoch": 36.90909090909091, "eval_loss": 0.44742465019226074, "eval_runtime": 21.0039, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 203 }, { "epoch": 37.09090909090909, "grad_norm": 0.4445912539958954, "learning_rate": 9.125567491391476e-06, "loss": 0.3135, "step": 204 }, { "epoch": 37.45454545454545, "grad_norm": 0.39482685923576355, "learning_rate": 9.099094380601244e-06, "loss": 0.2407, "step": 206 }, { "epoch": 37.81818181818182, "grad_norm": 0.3445771634578705, "learning_rate": 9.072266099240286e-06, "loss": 0.3422, "step": 208 }, { "epoch": 38.0, "eval_loss": 0.4443068504333496, "eval_runtime": 21.0039, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 209 }, { "epoch": 38.18181818181818, "grad_norm": 0.42600157856941223, "learning_rate": 9.045084971874738e-06, "loss": 0.2673, "step": 210 }, { "epoch": 38.54545454545455, "grad_norm": 0.4967671036720276, "learning_rate": 9.017553353643479e-06, "loss": 0.3177, "step": 212 }, { "epoch": 38.90909090909091, "grad_norm": 0.39350730180740356, "learning_rate": 8.989673630054044e-06, "loss": 0.3489, "step": 214 }, { "epoch": 38.90909090909091, "eval_loss": 0.44272685050964355, "eval_runtime": 21.0007, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 214 }, { "epoch": 39.27272727272727, "grad_norm": 0.27645283937454224, "learning_rate": 8.961448216775955e-06, "loss": 0.161, "step": 216 }, { "epoch": 39.63636363636363, "grad_norm": 0.615763783454895, "learning_rate": 8.932879559431392e-06, "loss": 0.2562, "step": 218 }, { "epoch": 40.0, "grad_norm": 0.491910845041275, "learning_rate": 8.903970133383297e-06, "loss": 0.3737, "step": 220 }, { "epoch": 40.0, "eval_loss": 0.443479984998703, "eval_runtime": 21.0049, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 220 }, { "epoch": 40.36363636363637, "grad_norm": 0.47909313440322876, "learning_rate": 8.874722443520898e-06, "loss": 0.2508, "step": 222 }, { "epoch": 40.72727272727273, "grad_norm": 0.5516977906227112, "learning_rate": 8.845139024042664e-06, "loss": 0.2804, "step": 224 }, { "epoch": 40.90909090909091, "eval_loss": 0.443651407957077, "eval_runtime": 21.007, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 225 }, { "epoch": 41.09090909090909, "grad_norm": 0.5438214540481567, "learning_rate": 8.815222438236726e-06, "loss": 0.3221, "step": 226 }, { "epoch": 41.45454545454545, "grad_norm": 0.49092897772789, "learning_rate": 8.784975278258783e-06, "loss": 0.2601, "step": 228 }, { "epoch": 41.81818181818182, "grad_norm": 0.5809181332588196, "learning_rate": 8.754400164907496e-06, "loss": 0.3212, "step": 230 }, { "epoch": 42.0, "eval_loss": 0.44689202308654785, "eval_runtime": 21.001, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 231 }, { "epoch": 42.18181818181818, "grad_norm": 0.5147470831871033, "learning_rate": 8.723499747397415e-06, "loss": 0.1686, "step": 232 }, { "epoch": 42.54545454545455, "grad_norm": 0.7372314929962158, "learning_rate": 8.692276703129421e-06, "loss": 0.2733, "step": 234 }, { "epoch": 42.90909090909091, "grad_norm": 0.4814988374710083, "learning_rate": 8.660733737458751e-06, "loss": 0.2322, "step": 236 }, { "epoch": 42.90909090909091, "eval_loss": 0.4480229318141937, "eval_runtime": 21.0124, "eval_samples_per_second": 1.142, "eval_steps_per_second": 1.142, "step": 236 }, { "epoch": 43.27272727272727, "grad_norm": 0.7270358800888062, "learning_rate": 8.628873583460593e-06, "loss": 0.2312, "step": 238 }, { "epoch": 43.63636363636363, "grad_norm": 0.6802980303764343, "learning_rate": 8.596699001693257e-06, "loss": 0.2551, "step": 240 }, { "epoch": 44.0, "grad_norm": 0.6301150321960449, "learning_rate": 8.564212779959003e-06, "loss": 0.2569, "step": 242 }, { "epoch": 44.0, "eval_loss": 0.4507286548614502, "eval_runtime": 21.0008, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 242 }, { "epoch": 44.36363636363637, "grad_norm": 0.6666437983512878, "learning_rate": 8.531417733062476e-06, "loss": 0.2299, "step": 244 }, { "epoch": 44.72727272727273, "grad_norm": 0.6867101788520813, "learning_rate": 8.498316702566828e-06, "loss": 0.2501, "step": 246 }, { "epoch": 44.90909090909091, "eval_loss": 0.4543609619140625, "eval_runtime": 21.005, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 247 }, { "epoch": 45.09090909090909, "grad_norm": 0.571156919002533, "learning_rate": 8.464912556547486e-06, "loss": 0.2035, "step": 248 }, { "epoch": 45.45454545454545, "grad_norm": 0.761677086353302, "learning_rate": 8.43120818934367e-06, "loss": 0.1717, "step": 250 }, { "epoch": 45.81818181818182, "grad_norm": 0.6928288340568542, "learning_rate": 8.397206521307584e-06, "loss": 0.2247, "step": 252 }, { "epoch": 46.0, "eval_loss": 0.4641251266002655, "eval_runtime": 20.9989, "eval_samples_per_second": 1.143, "eval_steps_per_second": 1.143, "step": 253 }, { "epoch": 46.0, "step": 253, "total_flos": 3.15960483767124e+17, "train_loss": 0.7967259936888699, "train_runtime": 11647.481, "train_samples_per_second": 1.133, "train_steps_per_second": 0.064 } ], "logging_steps": 2, "max_steps": 750, "num_input_tokens_seen": 0, "num_train_epochs": 150, "save_steps": 25, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 7, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.15960483767124e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }