{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.596169193934557, "eval_steps": 10, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007980845969672785, "grad_norm": 0.7833889722824097, "learning_rate": 0.00016, "loss": 1.3081, "step": 5 }, { "epoch": 0.01596169193934557, "grad_norm": 0.6466716527938843, "learning_rate": 0.00012, "loss": 1.2368, "step": 10 }, { "epoch": 0.01596169193934557, "eval_loss": 1.2068681716918945, "eval_runtime": 20.8667, "eval_samples_per_second": 1.486, "eval_steps_per_second": 0.767, "step": 10 }, { "epoch": 0.023942537909018357, "grad_norm": 0.6432256698608398, "learning_rate": 8e-05, "loss": 1.1477, "step": 15 }, { "epoch": 0.03192338387869114, "grad_norm": 0.6714484095573425, "learning_rate": 4e-05, "loss": 1.0776, "step": 20 }, { "epoch": 0.03192338387869114, "eval_loss": 1.16266667842865, "eval_runtime": 21.285, "eval_samples_per_second": 1.456, "eval_steps_per_second": 0.752, "step": 20 }, { "epoch": 0.03990422984836393, "grad_norm": 0.6418822407722473, "learning_rate": 0.0, "loss": 1.1614, "step": 25 }, { "epoch": 0.047885075818036714, "grad_norm": 0.8091150522232056, "learning_rate": 0.000194, "loss": 1.1676, "step": 30 }, { "epoch": 0.047885075818036714, "eval_loss": 1.151872992515564, "eval_runtime": 21.3531, "eval_samples_per_second": 1.452, "eval_steps_per_second": 0.749, "step": 30 }, { "epoch": 0.055865921787709494, "grad_norm": 0.7295385003089905, "learning_rate": 0.000193, "loss": 1.1319, "step": 35 }, { "epoch": 0.06384676775738228, "grad_norm": 1.056991696357727, "learning_rate": 0.000192, "loss": 1.0188, "step": 40 }, { "epoch": 0.06384676775738228, "eval_loss": 1.1241333484649658, "eval_runtime": 21.3665, "eval_samples_per_second": 1.451, "eval_steps_per_second": 0.749, "step": 40 }, { "epoch": 0.07182761372705507, "grad_norm": 0.7033870816230774, "learning_rate": 0.000191, "loss": 1.0331, "step": 45 }, { "epoch": 0.07980845969672785, "grad_norm": 0.6337556838989258, "learning_rate": 0.00019, "loss": 1.1616, "step": 50 }, { "epoch": 0.07980845969672785, "eval_loss": 1.1174383163452148, "eval_runtime": 20.7909, "eval_samples_per_second": 1.491, "eval_steps_per_second": 0.77, "step": 50 }, { "epoch": 0.08778930566640064, "grad_norm": 0.7127214074134827, "learning_rate": 0.00018899999999999999, "loss": 1.0874, "step": 55 }, { "epoch": 0.09577015163607343, "grad_norm": 0.7299643158912659, "learning_rate": 0.000188, "loss": 1.0098, "step": 60 }, { "epoch": 0.09577015163607343, "eval_loss": 1.106464147567749, "eval_runtime": 21.2918, "eval_samples_per_second": 1.456, "eval_steps_per_second": 0.751, "step": 60 }, { "epoch": 0.10375099760574621, "grad_norm": 0.7458897829055786, "learning_rate": 0.00018700000000000002, "loss": 1.0114, "step": 65 }, { "epoch": 0.11173184357541899, "grad_norm": 0.7154062986373901, "learning_rate": 0.00018600000000000002, "loss": 1.0804, "step": 70 }, { "epoch": 0.11173184357541899, "eval_loss": 1.087841510772705, "eval_runtime": 21.2696, "eval_samples_per_second": 1.457, "eval_steps_per_second": 0.752, "step": 70 }, { "epoch": 0.11971268954509177, "grad_norm": 0.630651593208313, "learning_rate": 0.00018500000000000002, "loss": 0.988, "step": 75 }, { "epoch": 0.12769353551476456, "grad_norm": 0.6520997285842896, "learning_rate": 0.00018400000000000003, "loss": 0.9957, "step": 80 }, { "epoch": 0.12769353551476456, "eval_loss": 1.0693910121917725, "eval_runtime": 21.2801, "eval_samples_per_second": 1.457, "eval_steps_per_second": 0.752, "step": 80 }, { "epoch": 0.13567438148443736, "grad_norm": 0.612789511680603, "learning_rate": 0.000183, "loss": 1.0659, "step": 85 }, { "epoch": 0.14365522745411013, "grad_norm": 0.7308149337768555, "learning_rate": 0.000182, "loss": 1.0044, "step": 90 }, { "epoch": 0.14365522745411013, "eval_loss": 1.0618408918380737, "eval_runtime": 21.4836, "eval_samples_per_second": 1.443, "eval_steps_per_second": 0.745, "step": 90 }, { "epoch": 0.1516360734237829, "grad_norm": 0.6600955724716187, "learning_rate": 0.000181, "loss": 0.936, "step": 95 }, { "epoch": 0.1596169193934557, "grad_norm": 0.5939749479293823, "learning_rate": 0.00018, "loss": 1.0213, "step": 100 }, { "epoch": 0.1596169193934557, "eval_loss": 1.06107497215271, "eval_runtime": 20.9063, "eval_samples_per_second": 1.483, "eval_steps_per_second": 0.765, "step": 100 }, { "epoch": 0.16759776536312848, "grad_norm": 0.6475443840026855, "learning_rate": 0.00017900000000000001, "loss": 1.0099, "step": 105 }, { "epoch": 0.17557861133280128, "grad_norm": 0.6048471927642822, "learning_rate": 0.00017800000000000002, "loss": 0.9587, "step": 110 }, { "epoch": 0.17557861133280128, "eval_loss": 1.0483193397521973, "eval_runtime": 21.4083, "eval_samples_per_second": 1.448, "eval_steps_per_second": 0.747, "step": 110 }, { "epoch": 0.18355945730247406, "grad_norm": 0.5856403112411499, "learning_rate": 0.00017700000000000002, "loss": 0.9908, "step": 115 }, { "epoch": 0.19154030327214686, "grad_norm": 0.6811533570289612, "learning_rate": 0.00017600000000000002, "loss": 0.9662, "step": 120 }, { "epoch": 0.19154030327214686, "eval_loss": 1.055928349494934, "eval_runtime": 21.2973, "eval_samples_per_second": 1.456, "eval_steps_per_second": 0.751, "step": 120 }, { "epoch": 0.19952114924181963, "grad_norm": 0.5635804533958435, "learning_rate": 0.000175, "loss": 0.9637, "step": 125 }, { "epoch": 0.20750199521149243, "grad_norm": 0.6101672053337097, "learning_rate": 0.000174, "loss": 0.9846, "step": 130 }, { "epoch": 0.20750199521149243, "eval_loss": 1.0435600280761719, "eval_runtime": 21.4526, "eval_samples_per_second": 1.445, "eval_steps_per_second": 0.746, "step": 130 }, { "epoch": 0.2154828411811652, "grad_norm": 0.649726152420044, "learning_rate": 0.000173, "loss": 0.9995, "step": 135 }, { "epoch": 0.22346368715083798, "grad_norm": 0.6642662286758423, "learning_rate": 0.000172, "loss": 0.9537, "step": 140 }, { "epoch": 0.22346368715083798, "eval_loss": 1.0289674997329712, "eval_runtime": 21.0765, "eval_samples_per_second": 1.471, "eval_steps_per_second": 0.759, "step": 140 }, { "epoch": 0.23144453312051078, "grad_norm": 0.5957498550415039, "learning_rate": 0.000171, "loss": 1.0444, "step": 145 }, { "epoch": 0.23942537909018355, "grad_norm": 0.7650970816612244, "learning_rate": 0.00017, "loss": 1.0177, "step": 150 }, { "epoch": 0.23942537909018355, "eval_loss": 1.0302073955535889, "eval_runtime": 21.0268, "eval_samples_per_second": 1.474, "eval_steps_per_second": 0.761, "step": 150 }, { "epoch": 0.24740622505985635, "grad_norm": 0.6636956334114075, "learning_rate": 0.00016900000000000002, "loss": 0.9786, "step": 155 }, { "epoch": 0.2553870710295291, "grad_norm": 0.7419523596763611, "learning_rate": 0.000168, "loss": 0.9837, "step": 160 }, { "epoch": 0.2553870710295291, "eval_loss": 1.0296545028686523, "eval_runtime": 21.4194, "eval_samples_per_second": 1.447, "eval_steps_per_second": 0.747, "step": 160 }, { "epoch": 0.2633679169992019, "grad_norm": 0.5570073127746582, "learning_rate": 0.000167, "loss": 0.954, "step": 165 }, { "epoch": 0.2713487629688747, "grad_norm": 0.5050713419914246, "learning_rate": 0.000166, "loss": 1.0796, "step": 170 }, { "epoch": 0.2713487629688747, "eval_loss": 1.0314409732818604, "eval_runtime": 21.4116, "eval_samples_per_second": 1.448, "eval_steps_per_second": 0.747, "step": 170 }, { "epoch": 0.27932960893854747, "grad_norm": 0.48484355211257935, "learning_rate": 0.000165, "loss": 0.9461, "step": 175 }, { "epoch": 0.28731045490822027, "grad_norm": 0.6661683320999146, "learning_rate": 0.000164, "loss": 1.0311, "step": 180 }, { "epoch": 0.28731045490822027, "eval_loss": 1.0315138101577759, "eval_runtime": 21.2472, "eval_samples_per_second": 1.459, "eval_steps_per_second": 0.753, "step": 180 }, { "epoch": 0.29529130087789307, "grad_norm": 0.6185592412948608, "learning_rate": 0.000163, "loss": 0.9745, "step": 185 }, { "epoch": 0.3032721468475658, "grad_norm": 0.5738425850868225, "learning_rate": 0.000162, "loss": 0.9619, "step": 190 }, { "epoch": 0.3032721468475658, "eval_loss": 1.0191501379013062, "eval_runtime": 21.0898, "eval_samples_per_second": 1.47, "eval_steps_per_second": 0.759, "step": 190 }, { "epoch": 0.3112529928172386, "grad_norm": 0.6744123101234436, "learning_rate": 0.000161, "loss": 1.0097, "step": 195 }, { "epoch": 0.3192338387869114, "grad_norm": 0.7204273343086243, "learning_rate": 0.00016, "loss": 0.9916, "step": 200 }, { "epoch": 0.3192338387869114, "eval_loss": 1.0153534412384033, "eval_runtime": 21.3333, "eval_samples_per_second": 1.453, "eval_steps_per_second": 0.75, "step": 200 }, { "epoch": 0.3272146847565842, "grad_norm": 0.6031942367553711, "learning_rate": 0.00015900000000000002, "loss": 0.9673, "step": 205 }, { "epoch": 0.33519553072625696, "grad_norm": 0.5967186689376831, "learning_rate": 0.00015800000000000002, "loss": 0.9655, "step": 210 }, { "epoch": 0.33519553072625696, "eval_loss": 1.0138649940490723, "eval_runtime": 21.1862, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 210 }, { "epoch": 0.34317637669592976, "grad_norm": 0.6408796906471252, "learning_rate": 0.00015700000000000002, "loss": 0.9453, "step": 215 }, { "epoch": 0.35115722266560256, "grad_norm": 0.6557004451751709, "learning_rate": 0.00015600000000000002, "loss": 0.9536, "step": 220 }, { "epoch": 0.35115722266560256, "eval_loss": 1.0107518434524536, "eval_runtime": 21.3902, "eval_samples_per_second": 1.449, "eval_steps_per_second": 0.748, "step": 220 }, { "epoch": 0.35913806863527536, "grad_norm": 0.6112158894538879, "learning_rate": 0.000155, "loss": 0.9645, "step": 225 }, { "epoch": 0.3671189146049481, "grad_norm": 0.6465075016021729, "learning_rate": 0.000154, "loss": 1.0773, "step": 230 }, { "epoch": 0.3671189146049481, "eval_loss": 1.004480242729187, "eval_runtime": 21.1771, "eval_samples_per_second": 1.464, "eval_steps_per_second": 0.756, "step": 230 }, { "epoch": 0.3750997605746209, "grad_norm": 0.7187998294830322, "learning_rate": 0.000153, "loss": 0.9467, "step": 235 }, { "epoch": 0.3830806065442937, "grad_norm": 0.8026754260063171, "learning_rate": 0.000152, "loss": 0.9329, "step": 240 }, { "epoch": 0.3830806065442937, "eval_loss": 0.9965100288391113, "eval_runtime": 21.1786, "eval_samples_per_second": 1.464, "eval_steps_per_second": 0.755, "step": 240 }, { "epoch": 0.39106145251396646, "grad_norm": 0.6788911819458008, "learning_rate": 0.000151, "loss": 0.8683, "step": 245 }, { "epoch": 0.39904229848363926, "grad_norm": 0.6607213020324707, "learning_rate": 0.00015000000000000001, "loss": 0.9826, "step": 250 }, { "epoch": 0.39904229848363926, "eval_loss": 1.0054864883422852, "eval_runtime": 21.1094, "eval_samples_per_second": 1.469, "eval_steps_per_second": 0.758, "step": 250 }, { "epoch": 0.40702314445331206, "grad_norm": 0.6742439866065979, "learning_rate": 0.00014900000000000002, "loss": 0.9507, "step": 255 }, { "epoch": 0.41500399042298486, "grad_norm": 0.7669168710708618, "learning_rate": 0.000148, "loss": 0.9609, "step": 260 }, { "epoch": 0.41500399042298486, "eval_loss": 1.0046597719192505, "eval_runtime": 21.2545, "eval_samples_per_second": 1.459, "eval_steps_per_second": 0.753, "step": 260 }, { "epoch": 0.4229848363926576, "grad_norm": 0.6186738014221191, "learning_rate": 0.000147, "loss": 0.9875, "step": 265 }, { "epoch": 0.4309656823623304, "grad_norm": 0.633393406867981, "learning_rate": 0.000146, "loss": 0.9949, "step": 270 }, { "epoch": 0.4309656823623304, "eval_loss": 0.9924024343490601, "eval_runtime": 20.9746, "eval_samples_per_second": 1.478, "eval_steps_per_second": 0.763, "step": 270 }, { "epoch": 0.4389465283320032, "grad_norm": 0.5856119394302368, "learning_rate": 0.000145, "loss": 0.9611, "step": 275 }, { "epoch": 0.44692737430167595, "grad_norm": 0.5712947249412537, "learning_rate": 0.000144, "loss": 1.0001, "step": 280 }, { "epoch": 0.44692737430167595, "eval_loss": 0.9949964284896851, "eval_runtime": 21.5928, "eval_samples_per_second": 1.436, "eval_steps_per_second": 0.741, "step": 280 }, { "epoch": 0.45490822027134875, "grad_norm": 0.6601425409317017, "learning_rate": 0.000143, "loss": 0.9198, "step": 285 }, { "epoch": 0.46288906624102155, "grad_norm": 0.6215651631355286, "learning_rate": 0.000142, "loss": 1.0944, "step": 290 }, { "epoch": 0.46288906624102155, "eval_loss": 0.9959576725959778, "eval_runtime": 21.3626, "eval_samples_per_second": 1.451, "eval_steps_per_second": 0.749, "step": 290 }, { "epoch": 0.47086991221069435, "grad_norm": 0.6404448747634888, "learning_rate": 0.000141, "loss": 0.9556, "step": 295 }, { "epoch": 0.4788507581803671, "grad_norm": 0.6889925003051758, "learning_rate": 0.00014, "loss": 1.011, "step": 300 }, { "epoch": 0.4788507581803671, "eval_loss": 0.9947748780250549, "eval_runtime": 21.4252, "eval_samples_per_second": 1.447, "eval_steps_per_second": 0.747, "step": 300 }, { "epoch": 0.4868316041500399, "grad_norm": 0.6349315047264099, "learning_rate": 0.000139, "loss": 0.9343, "step": 305 }, { "epoch": 0.4948124501197127, "grad_norm": 0.753274142742157, "learning_rate": 0.000138, "loss": 0.9356, "step": 310 }, { "epoch": 0.4948124501197127, "eval_loss": 0.9897598624229431, "eval_runtime": 21.7041, "eval_samples_per_second": 1.428, "eval_steps_per_second": 0.737, "step": 310 }, { "epoch": 0.5027932960893855, "grad_norm": 0.6351450085639954, "learning_rate": 0.00013700000000000002, "loss": 0.9691, "step": 315 }, { "epoch": 0.5107741420590582, "grad_norm": 0.6660261154174805, "learning_rate": 0.00013600000000000003, "loss": 0.9213, "step": 320 }, { "epoch": 0.5107741420590582, "eval_loss": 0.9789547920227051, "eval_runtime": 20.997, "eval_samples_per_second": 1.476, "eval_steps_per_second": 0.762, "step": 320 }, { "epoch": 0.518754988028731, "grad_norm": 0.7038460969924927, "learning_rate": 0.00013500000000000003, "loss": 0.8654, "step": 325 }, { "epoch": 0.5267358339984038, "grad_norm": 0.6426422595977783, "learning_rate": 0.000134, "loss": 0.8611, "step": 330 }, { "epoch": 0.5267358339984038, "eval_loss": 0.9792807698249817, "eval_runtime": 20.7975, "eval_samples_per_second": 1.491, "eval_steps_per_second": 0.769, "step": 330 }, { "epoch": 0.5347166799680766, "grad_norm": 0.5722155570983887, "learning_rate": 0.000133, "loss": 0.961, "step": 335 }, { "epoch": 0.5426975259377494, "grad_norm": 0.7792693376541138, "learning_rate": 0.000132, "loss": 0.9647, "step": 340 }, { "epoch": 0.5426975259377494, "eval_loss": 0.9812614321708679, "eval_runtime": 21.6024, "eval_samples_per_second": 1.435, "eval_steps_per_second": 0.741, "step": 340 }, { "epoch": 0.5506783719074222, "grad_norm": 0.7510095834732056, "learning_rate": 0.000131, "loss": 0.817, "step": 345 }, { "epoch": 0.5586592178770949, "grad_norm": 0.6685438752174377, "learning_rate": 0.00013000000000000002, "loss": 0.8392, "step": 350 }, { "epoch": 0.5586592178770949, "eval_loss": 0.9788187742233276, "eval_runtime": 21.2885, "eval_samples_per_second": 1.456, "eval_steps_per_second": 0.752, "step": 350 }, { "epoch": 0.5666400638467678, "grad_norm": 0.6660378575325012, "learning_rate": 0.00012900000000000002, "loss": 0.9494, "step": 355 }, { "epoch": 0.5746209098164405, "grad_norm": 0.6914508938789368, "learning_rate": 0.00012800000000000002, "loss": 1.0105, "step": 360 }, { "epoch": 0.5746209098164405, "eval_loss": 0.9770519733428955, "eval_runtime": 21.0568, "eval_samples_per_second": 1.472, "eval_steps_per_second": 0.76, "step": 360 }, { "epoch": 0.5826017557861133, "grad_norm": 0.665715217590332, "learning_rate": 0.000127, "loss": 0.9159, "step": 365 }, { "epoch": 0.5905826017557861, "grad_norm": 0.686400830745697, "learning_rate": 0.000126, "loss": 1.0525, "step": 370 }, { "epoch": 0.5905826017557861, "eval_loss": 0.9769095778465271, "eval_runtime": 20.9592, "eval_samples_per_second": 1.479, "eval_steps_per_second": 0.763, "step": 370 }, { "epoch": 0.5985634477254589, "grad_norm": 0.6291290521621704, "learning_rate": 0.000125, "loss": 0.801, "step": 375 }, { "epoch": 0.6065442936951316, "grad_norm": 0.7124271988868713, "learning_rate": 0.000124, "loss": 0.8913, "step": 380 }, { "epoch": 0.6065442936951316, "eval_loss": 0.9799799919128418, "eval_runtime": 21.397, "eval_samples_per_second": 1.449, "eval_steps_per_second": 0.748, "step": 380 }, { "epoch": 0.6145251396648045, "grad_norm": 0.8505049347877502, "learning_rate": 0.000123, "loss": 0.9319, "step": 385 }, { "epoch": 0.6225059856344772, "grad_norm": 0.7308720350265503, "learning_rate": 0.000122, "loss": 0.8906, "step": 390 }, { "epoch": 0.6225059856344772, "eval_loss": 0.9712496995925903, "eval_runtime": 21.2172, "eval_samples_per_second": 1.461, "eval_steps_per_second": 0.754, "step": 390 }, { "epoch": 0.6304868316041501, "grad_norm": 0.6782912611961365, "learning_rate": 0.000121, "loss": 0.8928, "step": 395 }, { "epoch": 0.6384676775738228, "grad_norm": 0.5951765179634094, "learning_rate": 0.00012, "loss": 0.9059, "step": 400 }, { "epoch": 0.6384676775738228, "eval_loss": 0.9641667008399963, "eval_runtime": 20.89, "eval_samples_per_second": 1.484, "eval_steps_per_second": 0.766, "step": 400 }, { "epoch": 0.6464485235434956, "grad_norm": 0.7105025053024292, "learning_rate": 0.000119, "loss": 0.8986, "step": 405 }, { "epoch": 0.6544293695131684, "grad_norm": 0.835155725479126, "learning_rate": 0.000118, "loss": 0.9309, "step": 410 }, { "epoch": 0.6544293695131684, "eval_loss": 0.9708338975906372, "eval_runtime": 21.5141, "eval_samples_per_second": 1.441, "eval_steps_per_second": 0.744, "step": 410 }, { "epoch": 0.6624102154828412, "grad_norm": 0.7432790994644165, "learning_rate": 0.000117, "loss": 0.9653, "step": 415 }, { "epoch": 0.6703910614525139, "grad_norm": 0.590387761592865, "learning_rate": 0.000116, "loss": 0.8429, "step": 420 }, { "epoch": 0.6703910614525139, "eval_loss": 0.9680631160736084, "eval_runtime": 21.0468, "eval_samples_per_second": 1.473, "eval_steps_per_second": 0.76, "step": 420 }, { "epoch": 0.6783719074221868, "grad_norm": 0.6864650249481201, "learning_rate": 0.00011499999999999999, "loss": 0.9585, "step": 425 }, { "epoch": 0.6863527533918595, "grad_norm": 0.6845282912254333, "learning_rate": 0.00011399999999999999, "loss": 0.9032, "step": 430 }, { "epoch": 0.6863527533918595, "eval_loss": 0.96528559923172, "eval_runtime": 21.6521, "eval_samples_per_second": 1.432, "eval_steps_per_second": 0.739, "step": 430 }, { "epoch": 0.6943335993615323, "grad_norm": 0.717858612537384, "learning_rate": 0.000113, "loss": 0.9401, "step": 435 }, { "epoch": 0.7023144453312051, "grad_norm": 0.6566243171691895, "learning_rate": 0.00011200000000000001, "loss": 0.85, "step": 440 }, { "epoch": 0.7023144453312051, "eval_loss": 0.9651073217391968, "eval_runtime": 21.2202, "eval_samples_per_second": 1.461, "eval_steps_per_second": 0.754, "step": 440 }, { "epoch": 0.7102952913008779, "grad_norm": 0.6869224905967712, "learning_rate": 0.00011100000000000001, "loss": 0.8701, "step": 445 }, { "epoch": 0.7182761372705507, "grad_norm": 0.5897542834281921, "learning_rate": 0.00011000000000000002, "loss": 0.9531, "step": 450 }, { "epoch": 0.7182761372705507, "eval_loss": 0.9623759984970093, "eval_runtime": 21.1274, "eval_samples_per_second": 1.467, "eval_steps_per_second": 0.757, "step": 450 }, { "epoch": 0.7262569832402235, "grad_norm": 0.7018331289291382, "learning_rate": 0.000109, "loss": 0.8888, "step": 455 }, { "epoch": 0.7342378292098962, "grad_norm": 0.6487181186676025, "learning_rate": 0.00010800000000000001, "loss": 0.8862, "step": 460 }, { "epoch": 0.7342378292098962, "eval_loss": 0.9587427973747253, "eval_runtime": 21.6398, "eval_samples_per_second": 1.433, "eval_steps_per_second": 0.739, "step": 460 }, { "epoch": 0.7422186751795691, "grad_norm": 0.8121400475502014, "learning_rate": 0.00010700000000000001, "loss": 0.8856, "step": 465 }, { "epoch": 0.7501995211492418, "grad_norm": 0.6554559469223022, "learning_rate": 0.00010600000000000002, "loss": 0.8736, "step": 470 }, { "epoch": 0.7501995211492418, "eval_loss": 0.9547150731086731, "eval_runtime": 21.1236, "eval_samples_per_second": 1.468, "eval_steps_per_second": 0.757, "step": 470 }, { "epoch": 0.7581803671189146, "grad_norm": 0.7241584062576294, "learning_rate": 0.000105, "loss": 0.8002, "step": 475 }, { "epoch": 0.7661612130885874, "grad_norm": 0.7639961242675781, "learning_rate": 0.00010400000000000001, "loss": 0.93, "step": 480 }, { "epoch": 0.7661612130885874, "eval_loss": 0.9609348773956299, "eval_runtime": 21.2046, "eval_samples_per_second": 1.462, "eval_steps_per_second": 0.755, "step": 480 }, { "epoch": 0.7741420590582602, "grad_norm": 0.6694309115409851, "learning_rate": 0.00010300000000000001, "loss": 0.8696, "step": 485 }, { "epoch": 0.7821229050279329, "grad_norm": 0.6586005687713623, "learning_rate": 0.00010200000000000001, "loss": 0.9035, "step": 490 }, { "epoch": 0.7821229050279329, "eval_loss": 0.9565220475196838, "eval_runtime": 21.5471, "eval_samples_per_second": 1.439, "eval_steps_per_second": 0.743, "step": 490 }, { "epoch": 0.7901037509976058, "grad_norm": 0.6772391200065613, "learning_rate": 0.000101, "loss": 0.9016, "step": 495 }, { "epoch": 0.7980845969672785, "grad_norm": 0.7700143456459045, "learning_rate": 0.0001, "loss": 0.8871, "step": 500 }, { "epoch": 0.7980845969672785, "eval_loss": 0.9532312750816345, "eval_runtime": 21.3527, "eval_samples_per_second": 1.452, "eval_steps_per_second": 0.749, "step": 500 }, { "epoch": 0.8060654429369514, "grad_norm": 0.7088077068328857, "learning_rate": 9.900000000000001e-05, "loss": 0.9628, "step": 505 }, { "epoch": 0.8140462889066241, "grad_norm": 0.7058229446411133, "learning_rate": 9.8e-05, "loss": 0.8395, "step": 510 }, { "epoch": 0.8140462889066241, "eval_loss": 0.9453433752059937, "eval_runtime": 20.6711, "eval_samples_per_second": 1.5, "eval_steps_per_second": 0.774, "step": 510 }, { "epoch": 0.8220271348762969, "grad_norm": 0.6609935164451599, "learning_rate": 9.7e-05, "loss": 0.8693, "step": 515 }, { "epoch": 0.8300079808459697, "grad_norm": 0.8539516925811768, "learning_rate": 9.6e-05, "loss": 0.8661, "step": 520 }, { "epoch": 0.8300079808459697, "eval_loss": 0.9413149356842041, "eval_runtime": 21.4056, "eval_samples_per_second": 1.448, "eval_steps_per_second": 0.747, "step": 520 }, { "epoch": 0.8379888268156425, "grad_norm": 0.6991361975669861, "learning_rate": 9.5e-05, "loss": 1.0091, "step": 525 }, { "epoch": 0.8459696727853152, "grad_norm": 0.6568119525909424, "learning_rate": 9.4e-05, "loss": 0.8034, "step": 530 }, { "epoch": 0.8459696727853152, "eval_loss": 0.942388117313385, "eval_runtime": 21.1726, "eval_samples_per_second": 1.464, "eval_steps_per_second": 0.756, "step": 530 }, { "epoch": 0.8539505187549881, "grad_norm": 0.6941240429878235, "learning_rate": 9.300000000000001e-05, "loss": 0.8947, "step": 535 }, { "epoch": 0.8619313647246608, "grad_norm": 0.717548131942749, "learning_rate": 9.200000000000001e-05, "loss": 0.9049, "step": 540 }, { "epoch": 0.8619313647246608, "eval_loss": 0.9332796335220337, "eval_runtime": 21.1935, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 540 }, { "epoch": 0.8699122106943336, "grad_norm": 0.6851376295089722, "learning_rate": 9.1e-05, "loss": 0.9446, "step": 545 }, { "epoch": 0.8778930566640064, "grad_norm": 0.7654580473899841, "learning_rate": 9e-05, "loss": 0.8878, "step": 550 }, { "epoch": 0.8778930566640064, "eval_loss": 0.9288386106491089, "eval_runtime": 21.0235, "eval_samples_per_second": 1.475, "eval_steps_per_second": 0.761, "step": 550 }, { "epoch": 0.8858739026336792, "grad_norm": 0.6864151954650879, "learning_rate": 8.900000000000001e-05, "loss": 0.8758, "step": 555 }, { "epoch": 0.8938547486033519, "grad_norm": 0.6886463761329651, "learning_rate": 8.800000000000001e-05, "loss": 0.895, "step": 560 }, { "epoch": 0.8938547486033519, "eval_loss": 0.9302241802215576, "eval_runtime": 21.0245, "eval_samples_per_second": 1.474, "eval_steps_per_second": 0.761, "step": 560 }, { "epoch": 0.9018355945730248, "grad_norm": 0.7045288681983948, "learning_rate": 8.7e-05, "loss": 0.7896, "step": 565 }, { "epoch": 0.9098164405426975, "grad_norm": 0.8066936135292053, "learning_rate": 8.6e-05, "loss": 0.9071, "step": 570 }, { "epoch": 0.9098164405426975, "eval_loss": 0.9250284433364868, "eval_runtime": 21.0108, "eval_samples_per_second": 1.475, "eval_steps_per_second": 0.762, "step": 570 }, { "epoch": 0.9177972865123704, "grad_norm": 0.6377101540565491, "learning_rate": 8.5e-05, "loss": 0.8045, "step": 575 }, { "epoch": 0.9257781324820431, "grad_norm": 0.7454017996788025, "learning_rate": 8.4e-05, "loss": 0.9337, "step": 580 }, { "epoch": 0.9257781324820431, "eval_loss": 0.9239058494567871, "eval_runtime": 20.9525, "eval_samples_per_second": 1.48, "eval_steps_per_second": 0.764, "step": 580 }, { "epoch": 0.9337589784517158, "grad_norm": 0.6952985525131226, "learning_rate": 8.3e-05, "loss": 0.8703, "step": 585 }, { "epoch": 0.9417398244213887, "grad_norm": 0.7358236312866211, "learning_rate": 8.2e-05, "loss": 0.8793, "step": 590 }, { "epoch": 0.9417398244213887, "eval_loss": 0.9293895363807678, "eval_runtime": 20.9523, "eval_samples_per_second": 1.48, "eval_steps_per_second": 0.764, "step": 590 }, { "epoch": 0.9497206703910615, "grad_norm": 0.5817161202430725, "learning_rate": 8.1e-05, "loss": 0.8298, "step": 595 }, { "epoch": 0.9577015163607342, "grad_norm": 0.7251338362693787, "learning_rate": 8e-05, "loss": 0.8766, "step": 600 }, { "epoch": 0.9577015163607342, "eval_loss": 0.9293678998947144, "eval_runtime": 20.9457, "eval_samples_per_second": 1.48, "eval_steps_per_second": 0.764, "step": 600 }, { "epoch": 0.965682362330407, "grad_norm": 0.8278694748878479, "learning_rate": 7.900000000000001e-05, "loss": 0.7507, "step": 605 }, { "epoch": 0.9736632083000798, "grad_norm": 0.7793962955474854, "learning_rate": 7.800000000000001e-05, "loss": 0.8682, "step": 610 }, { "epoch": 0.9736632083000798, "eval_loss": 0.9281108975410461, "eval_runtime": 21.2355, "eval_samples_per_second": 1.46, "eval_steps_per_second": 0.753, "step": 610 }, { "epoch": 0.9816440542697525, "grad_norm": 0.8250753283500671, "learning_rate": 7.7e-05, "loss": 0.8196, "step": 615 }, { "epoch": 0.9896249002394254, "grad_norm": 0.7124294638633728, "learning_rate": 7.6e-05, "loss": 0.7849, "step": 620 }, { "epoch": 0.9896249002394254, "eval_loss": 0.9263650178909302, "eval_runtime": 21.1863, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 620 }, { "epoch": 0.9976057462090981, "grad_norm": 0.768418550491333, "learning_rate": 7.500000000000001e-05, "loss": 0.8589, "step": 625 }, { "epoch": 1.005586592178771, "grad_norm": 0.6237154603004456, "learning_rate": 7.4e-05, "loss": 0.8746, "step": 630 }, { "epoch": 1.005586592178771, "eval_loss": 0.9216476082801819, "eval_runtime": 21.1504, "eval_samples_per_second": 1.466, "eval_steps_per_second": 0.756, "step": 630 }, { "epoch": 1.0135674381484436, "grad_norm": 0.691764235496521, "learning_rate": 7.3e-05, "loss": 0.8455, "step": 635 }, { "epoch": 1.0215482841181165, "grad_norm": 0.6523061990737915, "learning_rate": 7.2e-05, "loss": 0.8149, "step": 640 }, { "epoch": 1.0215482841181165, "eval_loss": 0.9180006384849548, "eval_runtime": 21.3153, "eval_samples_per_second": 1.454, "eval_steps_per_second": 0.751, "step": 640 }, { "epoch": 1.0295291300877893, "grad_norm": 0.9022932648658752, "learning_rate": 7.1e-05, "loss": 0.8321, "step": 645 }, { "epoch": 1.037509976057462, "grad_norm": 0.6842184066772461, "learning_rate": 7e-05, "loss": 0.806, "step": 650 }, { "epoch": 1.037509976057462, "eval_loss": 0.9218840599060059, "eval_runtime": 21.1263, "eval_samples_per_second": 1.467, "eval_steps_per_second": 0.757, "step": 650 }, { "epoch": 1.0454908220271348, "grad_norm": 0.9547672271728516, "learning_rate": 6.9e-05, "loss": 0.8023, "step": 655 }, { "epoch": 1.0534716679968077, "grad_norm": 0.6439852714538574, "learning_rate": 6.800000000000001e-05, "loss": 0.7533, "step": 660 }, { "epoch": 1.0534716679968077, "eval_loss": 0.9165079593658447, "eval_runtime": 21.1104, "eval_samples_per_second": 1.468, "eval_steps_per_second": 0.758, "step": 660 }, { "epoch": 1.0614525139664805, "grad_norm": 0.7850003242492676, "learning_rate": 6.7e-05, "loss": 0.858, "step": 665 }, { "epoch": 1.0694333599361532, "grad_norm": 0.6285898685455322, "learning_rate": 6.6e-05, "loss": 0.7146, "step": 670 }, { "epoch": 1.0694333599361532, "eval_loss": 0.9204993844032288, "eval_runtime": 21.1527, "eval_samples_per_second": 1.466, "eval_steps_per_second": 0.756, "step": 670 }, { "epoch": 1.077414205905826, "grad_norm": 0.7350932955741882, "learning_rate": 6.500000000000001e-05, "loss": 0.8328, "step": 675 }, { "epoch": 1.085395051875499, "grad_norm": 0.7773166298866272, "learning_rate": 6.400000000000001e-05, "loss": 0.8075, "step": 680 }, { "epoch": 1.085395051875499, "eval_loss": 0.9216338992118835, "eval_runtime": 21.0323, "eval_samples_per_second": 1.474, "eval_steps_per_second": 0.761, "step": 680 }, { "epoch": 1.0933758978451715, "grad_norm": 0.9075634479522705, "learning_rate": 6.3e-05, "loss": 0.8809, "step": 685 }, { "epoch": 1.1013567438148444, "grad_norm": 0.7483291029930115, "learning_rate": 6.2e-05, "loss": 0.8302, "step": 690 }, { "epoch": 1.1013567438148444, "eval_loss": 0.9141597747802734, "eval_runtime": 20.8831, "eval_samples_per_second": 1.484, "eval_steps_per_second": 0.766, "step": 690 }, { "epoch": 1.1093375897845172, "grad_norm": 0.8029285073280334, "learning_rate": 6.1e-05, "loss": 0.7757, "step": 695 }, { "epoch": 1.1173184357541899, "grad_norm": 0.7638797760009766, "learning_rate": 6e-05, "loss": 0.8258, "step": 700 }, { "epoch": 1.1173184357541899, "eval_loss": 0.9134471416473389, "eval_runtime": 21.3635, "eval_samples_per_second": 1.451, "eval_steps_per_second": 0.749, "step": 700 }, { "epoch": 1.1252992817238627, "grad_norm": 0.8713815212249756, "learning_rate": 5.9e-05, "loss": 0.8017, "step": 705 }, { "epoch": 1.1332801276935356, "grad_norm": 0.8253017663955688, "learning_rate": 5.8e-05, "loss": 0.8873, "step": 710 }, { "epoch": 1.1332801276935356, "eval_loss": 0.91542649269104, "eval_runtime": 21.1885, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 710 }, { "epoch": 1.1412609736632082, "grad_norm": 0.8720806241035461, "learning_rate": 5.6999999999999996e-05, "loss": 0.8808, "step": 715 }, { "epoch": 1.149241819632881, "grad_norm": 0.855076253414154, "learning_rate": 5.6000000000000006e-05, "loss": 0.8285, "step": 720 }, { "epoch": 1.149241819632881, "eval_loss": 0.9051924347877502, "eval_runtime": 21.0529, "eval_samples_per_second": 1.472, "eval_steps_per_second": 0.76, "step": 720 }, { "epoch": 1.157222665602554, "grad_norm": 0.8113986849784851, "learning_rate": 5.500000000000001e-05, "loss": 0.8119, "step": 725 }, { "epoch": 1.1652035115722266, "grad_norm": 0.754554271697998, "learning_rate": 5.4000000000000005e-05, "loss": 0.7881, "step": 730 }, { "epoch": 1.1652035115722266, "eval_loss": 0.9082400798797607, "eval_runtime": 20.8583, "eval_samples_per_second": 1.486, "eval_steps_per_second": 0.767, "step": 730 }, { "epoch": 1.1731843575418994, "grad_norm": 0.9242857694625854, "learning_rate": 5.300000000000001e-05, "loss": 0.7508, "step": 735 }, { "epoch": 1.1811652035115723, "grad_norm": 0.7203059792518616, "learning_rate": 5.2000000000000004e-05, "loss": 0.8788, "step": 740 }, { "epoch": 1.1811652035115723, "eval_loss": 0.9105547666549683, "eval_runtime": 20.8797, "eval_samples_per_second": 1.485, "eval_steps_per_second": 0.766, "step": 740 }, { "epoch": 1.189146049481245, "grad_norm": 0.821240246295929, "learning_rate": 5.1000000000000006e-05, "loss": 0.767, "step": 745 }, { "epoch": 1.1971268954509178, "grad_norm": 1.0884546041488647, "learning_rate": 5e-05, "loss": 0.8585, "step": 750 }, { "epoch": 1.1971268954509178, "eval_loss": 0.9092571139335632, "eval_runtime": 21.2995, "eval_samples_per_second": 1.455, "eval_steps_per_second": 0.751, "step": 750 }, { "epoch": 1.2051077414205906, "grad_norm": 0.7183963656425476, "learning_rate": 4.9e-05, "loss": 0.8462, "step": 755 }, { "epoch": 1.2130885873902635, "grad_norm": 0.7556672692298889, "learning_rate": 4.8e-05, "loss": 0.7203, "step": 760 }, { "epoch": 1.2130885873902635, "eval_loss": 0.9125378727912903, "eval_runtime": 20.9445, "eval_samples_per_second": 1.48, "eval_steps_per_second": 0.764, "step": 760 }, { "epoch": 1.2210694333599361, "grad_norm": 0.8119292259216309, "learning_rate": 4.7e-05, "loss": 0.8294, "step": 765 }, { "epoch": 1.229050279329609, "grad_norm": 0.7207946181297302, "learning_rate": 4.600000000000001e-05, "loss": 0.8084, "step": 770 }, { "epoch": 1.229050279329609, "eval_loss": 0.9099369645118713, "eval_runtime": 20.5873, "eval_samples_per_second": 1.506, "eval_steps_per_second": 0.777, "step": 770 }, { "epoch": 1.2370311252992816, "grad_norm": 0.9262428283691406, "learning_rate": 4.5e-05, "loss": 0.7861, "step": 775 }, { "epoch": 1.2450119712689545, "grad_norm": 0.9154334664344788, "learning_rate": 4.4000000000000006e-05, "loss": 0.7298, "step": 780 }, { "epoch": 1.2450119712689545, "eval_loss": 0.9098423719406128, "eval_runtime": 21.1045, "eval_samples_per_second": 1.469, "eval_steps_per_second": 0.758, "step": 780 }, { "epoch": 1.2529928172386273, "grad_norm": 0.8377973437309265, "learning_rate": 4.3e-05, "loss": 0.7593, "step": 785 }, { "epoch": 1.2609736632083002, "grad_norm": 0.8660017848014832, "learning_rate": 4.2e-05, "loss": 0.7811, "step": 790 }, { "epoch": 1.2609736632083002, "eval_loss": 0.9046799540519714, "eval_runtime": 20.9842, "eval_samples_per_second": 1.477, "eval_steps_per_second": 0.762, "step": 790 }, { "epoch": 1.2689545091779728, "grad_norm": 0.7700662612915039, "learning_rate": 4.1e-05, "loss": 0.8493, "step": 795 }, { "epoch": 1.2769353551476457, "grad_norm": 0.8167340159416199, "learning_rate": 4e-05, "loss": 0.7774, "step": 800 }, { "epoch": 1.2769353551476457, "eval_loss": 0.8998985886573792, "eval_runtime": 21.1467, "eval_samples_per_second": 1.466, "eval_steps_per_second": 0.757, "step": 800 }, { "epoch": 1.2849162011173183, "grad_norm": 0.8162097334861755, "learning_rate": 3.9000000000000006e-05, "loss": 0.7892, "step": 805 }, { "epoch": 1.2928970470869912, "grad_norm": 0.8468261957168579, "learning_rate": 3.8e-05, "loss": 0.8324, "step": 810 }, { "epoch": 1.2928970470869912, "eval_loss": 0.8993770480155945, "eval_runtime": 20.9539, "eval_samples_per_second": 1.479, "eval_steps_per_second": 0.764, "step": 810 }, { "epoch": 1.300877893056664, "grad_norm": 0.746923565864563, "learning_rate": 3.7e-05, "loss": 0.7961, "step": 815 }, { "epoch": 1.3088587390263369, "grad_norm": 0.7284272313117981, "learning_rate": 3.6e-05, "loss": 0.6872, "step": 820 }, { "epoch": 1.3088587390263369, "eval_loss": 0.8972273468971252, "eval_runtime": 20.6969, "eval_samples_per_second": 1.498, "eval_steps_per_second": 0.773, "step": 820 }, { "epoch": 1.3168395849960095, "grad_norm": 0.8922063708305359, "learning_rate": 3.5e-05, "loss": 0.8041, "step": 825 }, { "epoch": 1.3248204309656824, "grad_norm": 0.7598963975906372, "learning_rate": 3.4000000000000007e-05, "loss": 0.7028, "step": 830 }, { "epoch": 1.3248204309656824, "eval_loss": 0.8973636031150818, "eval_runtime": 20.8738, "eval_samples_per_second": 1.485, "eval_steps_per_second": 0.767, "step": 830 }, { "epoch": 1.3328012769353552, "grad_norm": 0.7866420745849609, "learning_rate": 3.3e-05, "loss": 0.8392, "step": 835 }, { "epoch": 1.3407821229050279, "grad_norm": 0.8417834639549255, "learning_rate": 3.2000000000000005e-05, "loss": 0.7329, "step": 840 }, { "epoch": 1.3407821229050279, "eval_loss": 0.9000219106674194, "eval_runtime": 21.0159, "eval_samples_per_second": 1.475, "eval_steps_per_second": 0.761, "step": 840 }, { "epoch": 1.3487629688747007, "grad_norm": 0.8085408806800842, "learning_rate": 3.1e-05, "loss": 0.7452, "step": 845 }, { "epoch": 1.3567438148443736, "grad_norm": 0.9095738530158997, "learning_rate": 3e-05, "loss": 0.8114, "step": 850 }, { "epoch": 1.3567438148443736, "eval_loss": 0.899355411529541, "eval_runtime": 21.1951, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 850 }, { "epoch": 1.3647246608140464, "grad_norm": 0.8077012896537781, "learning_rate": 2.9e-05, "loss": 0.7842, "step": 855 }, { "epoch": 1.372705506783719, "grad_norm": 0.9387809038162231, "learning_rate": 2.8000000000000003e-05, "loss": 0.7739, "step": 860 }, { "epoch": 1.372705506783719, "eval_loss": 0.9006592631340027, "eval_runtime": 20.7294, "eval_samples_per_second": 1.495, "eval_steps_per_second": 0.772, "step": 860 }, { "epoch": 1.380686352753392, "grad_norm": 0.8230510950088501, "learning_rate": 2.7000000000000002e-05, "loss": 0.7765, "step": 865 }, { "epoch": 1.3886671987230645, "grad_norm": 0.9421499371528625, "learning_rate": 2.6000000000000002e-05, "loss": 0.7469, "step": 870 }, { "epoch": 1.3886671987230645, "eval_loss": 0.8988057374954224, "eval_runtime": 20.7795, "eval_samples_per_second": 1.492, "eval_steps_per_second": 0.77, "step": 870 }, { "epoch": 1.3966480446927374, "grad_norm": 0.8720149993896484, "learning_rate": 2.5e-05, "loss": 0.7674, "step": 875 }, { "epoch": 1.4046288906624103, "grad_norm": 0.9915176033973694, "learning_rate": 2.4e-05, "loss": 0.7925, "step": 880 }, { "epoch": 1.4046288906624103, "eval_loss": 0.8960800170898438, "eval_runtime": 21.1416, "eval_samples_per_second": 1.466, "eval_steps_per_second": 0.757, "step": 880 }, { "epoch": 1.4126097366320831, "grad_norm": 0.8962202072143555, "learning_rate": 2.3000000000000003e-05, "loss": 0.7848, "step": 885 }, { "epoch": 1.4205905826017557, "grad_norm": 0.9091677069664001, "learning_rate": 2.2000000000000003e-05, "loss": 0.8638, "step": 890 }, { "epoch": 1.4205905826017557, "eval_loss": 0.893966555595398, "eval_runtime": 21.0044, "eval_samples_per_second": 1.476, "eval_steps_per_second": 0.762, "step": 890 }, { "epoch": 1.4285714285714286, "grad_norm": 0.7839522361755371, "learning_rate": 2.1e-05, "loss": 0.8426, "step": 895 }, { "epoch": 1.4365522745411012, "grad_norm": 0.8735964894294739, "learning_rate": 2e-05, "loss": 0.852, "step": 900 }, { "epoch": 1.4365522745411012, "eval_loss": 0.8946609497070312, "eval_runtime": 20.7777, "eval_samples_per_second": 1.492, "eval_steps_per_second": 0.77, "step": 900 }, { "epoch": 1.444533120510774, "grad_norm": 0.8208412528038025, "learning_rate": 1.9e-05, "loss": 0.7983, "step": 905 }, { "epoch": 1.452513966480447, "grad_norm": 0.9176042675971985, "learning_rate": 1.8e-05, "loss": 0.8296, "step": 910 }, { "epoch": 1.452513966480447, "eval_loss": 0.8942002654075623, "eval_runtime": 21.0465, "eval_samples_per_second": 1.473, "eval_steps_per_second": 0.76, "step": 910 }, { "epoch": 1.4604948124501198, "grad_norm": 0.7592383027076721, "learning_rate": 1.7000000000000003e-05, "loss": 0.809, "step": 915 }, { "epoch": 1.4684756584197924, "grad_norm": 0.8816229701042175, "learning_rate": 1.6000000000000003e-05, "loss": 0.7502, "step": 920 }, { "epoch": 1.4684756584197924, "eval_loss": 0.8975324034690857, "eval_runtime": 21.1221, "eval_samples_per_second": 1.468, "eval_steps_per_second": 0.758, "step": 920 }, { "epoch": 1.4764565043894653, "grad_norm": 0.8397924900054932, "learning_rate": 1.5e-05, "loss": 0.791, "step": 925 }, { "epoch": 1.484437350359138, "grad_norm": 0.9116021394729614, "learning_rate": 1.4000000000000001e-05, "loss": 0.77, "step": 930 }, { "epoch": 1.484437350359138, "eval_loss": 0.8986864686012268, "eval_runtime": 21.0542, "eval_samples_per_second": 1.472, "eval_steps_per_second": 0.76, "step": 930 }, { "epoch": 1.4924181963288108, "grad_norm": 0.8789558410644531, "learning_rate": 1.3000000000000001e-05, "loss": 0.8362, "step": 935 }, { "epoch": 1.5003990422984836, "grad_norm": 1.1174862384796143, "learning_rate": 1.2e-05, "loss": 0.7821, "step": 940 }, { "epoch": 1.5003990422984836, "eval_loss": 0.8970634341239929, "eval_runtime": 20.7967, "eval_samples_per_second": 1.491, "eval_steps_per_second": 0.769, "step": 940 }, { "epoch": 1.5083798882681565, "grad_norm": 0.8152345418930054, "learning_rate": 1.1000000000000001e-05, "loss": 0.7953, "step": 945 }, { "epoch": 1.5163607342378294, "grad_norm": 0.8047603964805603, "learning_rate": 1e-05, "loss": 0.7759, "step": 950 }, { "epoch": 1.5163607342378294, "eval_loss": 0.8968480229377747, "eval_runtime": 21.2746, "eval_samples_per_second": 1.457, "eval_steps_per_second": 0.752, "step": 950 }, { "epoch": 1.524341580207502, "grad_norm": 0.8262729048728943, "learning_rate": 9e-06, "loss": 0.8128, "step": 955 }, { "epoch": 1.5323224261771746, "grad_norm": 0.8479042649269104, "learning_rate": 8.000000000000001e-06, "loss": 0.7778, "step": 960 }, { "epoch": 1.5323224261771746, "eval_loss": 0.8949146270751953, "eval_runtime": 21.0684, "eval_samples_per_second": 1.471, "eval_steps_per_second": 0.759, "step": 960 }, { "epoch": 1.5403032721468475, "grad_norm": 0.7638890147209167, "learning_rate": 7.000000000000001e-06, "loss": 0.8371, "step": 965 }, { "epoch": 1.5482841181165203, "grad_norm": 1.009861707687378, "learning_rate": 6e-06, "loss": 0.6906, "step": 970 }, { "epoch": 1.5482841181165203, "eval_loss": 0.8940587043762207, "eval_runtime": 21.4138, "eval_samples_per_second": 1.448, "eval_steps_per_second": 0.747, "step": 970 }, { "epoch": 1.5562649640861932, "grad_norm": 0.8926388025283813, "learning_rate": 5e-06, "loss": 0.7935, "step": 975 }, { "epoch": 1.564245810055866, "grad_norm": 0.7741861939430237, "learning_rate": 4.000000000000001e-06, "loss": 0.8137, "step": 980 }, { "epoch": 1.564245810055866, "eval_loss": 0.895005464553833, "eval_runtime": 20.8613, "eval_samples_per_second": 1.486, "eval_steps_per_second": 0.767, "step": 980 }, { "epoch": 1.5722266560255387, "grad_norm": 0.950386106967926, "learning_rate": 3e-06, "loss": 0.6965, "step": 985 }, { "epoch": 1.5802075019952115, "grad_norm": 0.846916675567627, "learning_rate": 2.0000000000000003e-06, "loss": 0.799, "step": 990 }, { "epoch": 1.5802075019952115, "eval_loss": 0.8942360281944275, "eval_runtime": 21.1875, "eval_samples_per_second": 1.463, "eval_steps_per_second": 0.755, "step": 990 }, { "epoch": 1.5881883479648842, "grad_norm": 0.8533262610435486, "learning_rate": 1.0000000000000002e-06, "loss": 0.753, "step": 995 }, { "epoch": 1.596169193934557, "grad_norm": 0.7492548227310181, "learning_rate": 0.0, "loss": 0.7454, "step": 1000 }, { "epoch": 1.596169193934557, "eval_loss": 0.8934003114700317, "eval_runtime": 21.313, "eval_samples_per_second": 1.455, "eval_steps_per_second": 0.751, "step": 1000 } ], "logging_steps": 5, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.292030717299424e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }