{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 11814, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012698412698412698, "grad_norm": 0.18539635837078094, "learning_rate": 0.00019917047570678858, "loss": 5.4804, "step": 50 }, { "epoch": 0.025396825396825397, "grad_norm": 0.19653430581092834, "learning_rate": 0.00019832402234636873, "loss": 5.4815, "step": 100 }, { "epoch": 0.0380952380952381, "grad_norm": 0.21583063900470734, "learning_rate": 0.0001974775689859489, "loss": 5.4752, "step": 150 }, { "epoch": 0.050793650793650794, "grad_norm": 0.659081757068634, "learning_rate": 0.00019663111562552904, "loss": 5.3388, "step": 200 }, { "epoch": 0.06349206349206349, "grad_norm": 1.1028743982315063, "learning_rate": 0.00019578466226510922, "loss": 5.0174, "step": 250 }, { "epoch": 0.0761904761904762, "grad_norm": 1.1611863374710083, "learning_rate": 0.00019493820890468936, "loss": 4.9543, "step": 300 }, { "epoch": 0.08888888888888889, "grad_norm": 1.3357131481170654, "learning_rate": 0.00019409175554426953, "loss": 4.8991, "step": 350 }, { "epoch": 0.10158730158730159, "grad_norm": 1.320380687713623, "learning_rate": 0.00019324530218384968, "loss": 4.889, "step": 400 }, { "epoch": 0.11428571428571428, "grad_norm": 1.5646241903305054, "learning_rate": 0.00019239884882342985, "loss": 4.8641, "step": 450 }, { "epoch": 0.12698412698412698, "grad_norm": 1.2670501470565796, "learning_rate": 0.00019155239546301, "loss": 4.8605, "step": 500 }, { "epoch": 0.13968253968253969, "grad_norm": 1.3732168674468994, "learning_rate": 0.00019070594210259017, "loss": 4.8206, "step": 550 }, { "epoch": 0.1523809523809524, "grad_norm": 1.7774670124053955, "learning_rate": 0.0001898594887421703, "loss": 4.8106, "step": 600 }, { "epoch": 0.16507936507936508, "grad_norm": 1.8740317821502686, "learning_rate": 0.00018901303538175048, "loss": 4.8007, "step": 650 }, { "epoch": 0.17777777777777778, "grad_norm": 1.6368571519851685, "learning_rate": 0.00018816658202133063, "loss": 4.7815, "step": 700 }, { "epoch": 0.19047619047619047, "grad_norm": 1.5557011365890503, "learning_rate": 0.0001873201286609108, "loss": 4.7716, "step": 750 }, { "epoch": 0.20317460317460317, "grad_norm": 1.6413705348968506, "learning_rate": 0.00018647367530049097, "loss": 4.767, "step": 800 }, { "epoch": 0.21587301587301588, "grad_norm": 1.6060304641723633, "learning_rate": 0.00018562722194007112, "loss": 4.7457, "step": 850 }, { "epoch": 0.22857142857142856, "grad_norm": 1.832118034362793, "learning_rate": 0.0001847807685796513, "loss": 4.739, "step": 900 }, { "epoch": 0.24126984126984127, "grad_norm": 1.7070050239562988, "learning_rate": 0.00018393431521923143, "loss": 4.7319, "step": 950 }, { "epoch": 0.25396825396825395, "grad_norm": 1.768761157989502, "learning_rate": 0.0001830878618588116, "loss": 4.7296, "step": 1000 }, { "epoch": 0.26666666666666666, "grad_norm": 1.7824410200119019, "learning_rate": 0.00018224140849839175, "loss": 4.7069, "step": 1050 }, { "epoch": 0.27936507936507937, "grad_norm": 1.8590071201324463, "learning_rate": 0.00018139495513797192, "loss": 4.6814, "step": 1100 }, { "epoch": 0.2920634920634921, "grad_norm": 1.8381578922271729, "learning_rate": 0.00018054850177755206, "loss": 4.6146, "step": 1150 }, { "epoch": 0.3047619047619048, "grad_norm": 1.8397525548934937, "learning_rate": 0.00017970204841713224, "loss": 4.5086, "step": 1200 }, { "epoch": 0.31746031746031744, "grad_norm": 1.8704237937927246, "learning_rate": 0.00017885559505671238, "loss": 4.376, "step": 1250 }, { "epoch": 0.33015873015873015, "grad_norm": 2.0715010166168213, "learning_rate": 0.00017800914169629255, "loss": 4.2503, "step": 1300 }, { "epoch": 0.34285714285714286, "grad_norm": 3.124469041824341, "learning_rate": 0.0001771626883358727, "loss": 4.157, "step": 1350 }, { "epoch": 0.35555555555555557, "grad_norm": 2.0155253410339355, "learning_rate": 0.00017631623497545287, "loss": 4.1085, "step": 1400 }, { "epoch": 0.3682539682539683, "grad_norm": 2.309039354324341, "learning_rate": 0.00017546978161503301, "loss": 4.043, "step": 1450 }, { "epoch": 0.38095238095238093, "grad_norm": 2.6701300144195557, "learning_rate": 0.0001746233282546132, "loss": 4.016, "step": 1500 }, { "epoch": 0.39365079365079364, "grad_norm": NaN, "learning_rate": 0.00017377687489419333, "loss": 3.9352, "step": 1550 }, { "epoch": 0.40634920634920635, "grad_norm": 2.3252177238464355, "learning_rate": 0.00017294735060098188, "loss": 3.9148, "step": 1600 }, { "epoch": 0.41904761904761906, "grad_norm": 2.336137533187866, "learning_rate": 0.00017210089724056205, "loss": 3.9097, "step": 1650 }, { "epoch": 0.43174603174603177, "grad_norm": 3.3141894340515137, "learning_rate": 0.0001712544438801422, "loss": 3.8494, "step": 1700 }, { "epoch": 0.4444444444444444, "grad_norm": 2.5842506885528564, "learning_rate": 0.0001704079905197224, "loss": 3.8192, "step": 1750 }, { "epoch": 0.45714285714285713, "grad_norm": 2.667503833770752, "learning_rate": 0.00016956153715930254, "loss": 3.835, "step": 1800 }, { "epoch": 0.46984126984126984, "grad_norm": 2.7759885787963867, "learning_rate": 0.0001687150837988827, "loss": 3.7916, "step": 1850 }, { "epoch": 0.48253968253968255, "grad_norm": 2.939075231552124, "learning_rate": 0.00016786863043846285, "loss": 3.7805, "step": 1900 }, { "epoch": 0.49523809523809526, "grad_norm": 3.0116183757781982, "learning_rate": 0.00016702217707804303, "loss": 3.7436, "step": 1950 }, { "epoch": 0.5079365079365079, "grad_norm": 2.9771008491516113, "learning_rate": 0.00016617572371762317, "loss": 3.7206, "step": 2000 }, { "epoch": 0.5206349206349207, "grad_norm": 3.095003843307495, "learning_rate": 0.00016532927035720334, "loss": 3.7194, "step": 2050 }, { "epoch": 0.5333333333333333, "grad_norm": 3.0788674354553223, "learning_rate": 0.0001644828169967835, "loss": 3.6632, "step": 2100 }, { "epoch": 0.546031746031746, "grad_norm": 3.467590093612671, "learning_rate": 0.00016363636363636366, "loss": 3.6919, "step": 2150 }, { "epoch": 0.5587301587301587, "grad_norm": 3.4599721431732178, "learning_rate": 0.0001627899102759438, "loss": 3.6657, "step": 2200 }, { "epoch": 0.5714285714285714, "grad_norm": 3.194898843765259, "learning_rate": 0.00016194345691552398, "loss": 3.6454, "step": 2250 }, { "epoch": 0.5841269841269842, "grad_norm": 3.5158746242523193, "learning_rate": 0.00016109700355510412, "loss": 3.6003, "step": 2300 }, { "epoch": 0.5968253968253968, "grad_norm": 3.2812819480895996, "learning_rate": 0.0001602505501946843, "loss": 3.5907, "step": 2350 }, { "epoch": 0.6095238095238096, "grad_norm": 3.5167620182037354, "learning_rate": 0.00015940409683426444, "loss": 3.5821, "step": 2400 }, { "epoch": 0.6222222222222222, "grad_norm": 3.5780296325683594, "learning_rate": 0.0001585576434738446, "loss": 3.5245, "step": 2450 }, { "epoch": 0.6349206349206349, "grad_norm": 4.31138277053833, "learning_rate": 0.00015771119011342478, "loss": 3.5197, "step": 2500 }, { "epoch": 0.6476190476190476, "grad_norm": 3.752105236053467, "learning_rate": 0.00015686473675300493, "loss": 3.5104, "step": 2550 }, { "epoch": 0.6603174603174603, "grad_norm": 3.916947603225708, "learning_rate": 0.0001560182833925851, "loss": 3.5123, "step": 2600 }, { "epoch": 0.6730158730158731, "grad_norm": 3.912555694580078, "learning_rate": 0.00015517183003216524, "loss": 3.4689, "step": 2650 }, { "epoch": 0.6857142857142857, "grad_norm": 4.15084981918335, "learning_rate": 0.00015432537667174541, "loss": 3.4572, "step": 2700 }, { "epoch": 0.6984126984126984, "grad_norm": 4.125711441040039, "learning_rate": 0.00015347892331132556, "loss": 3.4323, "step": 2750 }, { "epoch": 0.7111111111111111, "grad_norm": 3.8961021900177, "learning_rate": 0.00015263246995090573, "loss": 3.4215, "step": 2800 }, { "epoch": 0.7238095238095238, "grad_norm": 4.138737678527832, "learning_rate": 0.00015178601659048588, "loss": 3.4113, "step": 2850 }, { "epoch": 0.7365079365079366, "grad_norm": 3.9136457443237305, "learning_rate": 0.00015093956323006605, "loss": 3.4036, "step": 2900 }, { "epoch": 0.7492063492063492, "grad_norm": 4.833474636077881, "learning_rate": 0.0001500931098696462, "loss": 3.3733, "step": 2950 }, { "epoch": 0.7619047619047619, "grad_norm": 3.974536657333374, "learning_rate": 0.00014924665650922636, "loss": 3.3637, "step": 3000 }, { "epoch": 0.7746031746031746, "grad_norm": 3.9956250190734863, "learning_rate": 0.0001484002031488065, "loss": 3.3421, "step": 3050 }, { "epoch": 0.7873015873015873, "grad_norm": 4.232203483581543, "learning_rate": 0.00014755374978838668, "loss": 3.3564, "step": 3100 }, { "epoch": 0.8, "grad_norm": 4.4057207107543945, "learning_rate": 0.00014670729642796683, "loss": 3.3156, "step": 3150 }, { "epoch": 0.8126984126984127, "grad_norm": 4.205805778503418, "learning_rate": 0.000145860843067547, "loss": 3.3195, "step": 3200 }, { "epoch": 0.8253968253968254, "grad_norm": 4.056403160095215, "learning_rate": 0.00014501438970712714, "loss": 3.3083, "step": 3250 }, { "epoch": 0.8380952380952381, "grad_norm": 4.160456657409668, "learning_rate": 0.0001441679363467073, "loss": 3.2684, "step": 3300 }, { "epoch": 0.8507936507936508, "grad_norm": 4.401111602783203, "learning_rate": 0.00014332148298628746, "loss": 3.2735, "step": 3350 }, { "epoch": 0.8634920634920635, "grad_norm": 4.488119602203369, "learning_rate": 0.00014247502962586763, "loss": 3.2653, "step": 3400 }, { "epoch": 0.8761904761904762, "grad_norm": 4.22524881362915, "learning_rate": 0.00014162857626544777, "loss": 3.2396, "step": 3450 }, { "epoch": 0.8888888888888888, "grad_norm": 4.20457124710083, "learning_rate": 0.00014078212290502795, "loss": 3.2357, "step": 3500 }, { "epoch": 0.9015873015873016, "grad_norm": 4.248793601989746, "learning_rate": 0.00013993566954460812, "loss": 3.2263, "step": 3550 }, { "epoch": 0.9142857142857143, "grad_norm": 5.042767524719238, "learning_rate": 0.00013908921618418826, "loss": 3.2255, "step": 3600 }, { "epoch": 0.926984126984127, "grad_norm": 4.23060941696167, "learning_rate": 0.00013824276282376844, "loss": 3.1735, "step": 3650 }, { "epoch": 0.9396825396825397, "grad_norm": 4.368992805480957, "learning_rate": 0.00013739630946334858, "loss": 3.1896, "step": 3700 }, { "epoch": 0.9523809523809523, "grad_norm": 4.162256717681885, "learning_rate": 0.00013654985610292875, "loss": 3.1839, "step": 3750 }, { "epoch": 0.9650793650793651, "grad_norm": 4.40270471572876, "learning_rate": 0.0001357034027425089, "loss": 3.1565, "step": 3800 }, { "epoch": 0.9777777777777777, "grad_norm": 4.529857635498047, "learning_rate": 0.00013485694938208907, "loss": 3.1492, "step": 3850 }, { "epoch": 0.9904761904761905, "grad_norm": 5.210158824920654, "learning_rate": 0.00013402742508887761, "loss": 3.1567, "step": 3900 }, { "epoch": 1.003047619047619, "grad_norm": 5.2821478843688965, "learning_rate": 0.00013318097172845776, "loss": 3.1272, "step": 3950 }, { "epoch": 1.0157460317460318, "grad_norm": 4.609035015106201, "learning_rate": 0.00013233451836803793, "loss": 3.1204, "step": 4000 }, { "epoch": 1.0284444444444445, "grad_norm": 5.00595235824585, "learning_rate": 0.00013148806500761808, "loss": 3.0907, "step": 4050 }, { "epoch": 1.0411428571428571, "grad_norm": 4.449355125427246, "learning_rate": 0.00013064161164719825, "loss": 3.1063, "step": 4100 }, { "epoch": 1.0538412698412698, "grad_norm": 4.445562839508057, "learning_rate": 0.0001297951582867784, "loss": 3.097, "step": 4150 }, { "epoch": 1.0665396825396825, "grad_norm": 5.181248664855957, "learning_rate": 0.00012894870492635856, "loss": 3.0693, "step": 4200 }, { "epoch": 1.0792380952380953, "grad_norm": 4.498983383178711, "learning_rate": 0.0001281022515659387, "loss": 3.072, "step": 4250 }, { "epoch": 1.091936507936508, "grad_norm": 4.873691082000732, "learning_rate": 0.00012725579820551888, "loss": 3.0545, "step": 4300 }, { "epoch": 1.1046349206349206, "grad_norm": 4.709224700927734, "learning_rate": 0.00012640934484509903, "loss": 3.0424, "step": 4350 }, { "epoch": 1.1173333333333333, "grad_norm": 5.21176815032959, "learning_rate": 0.0001255628914846792, "loss": 3.0412, "step": 4400 }, { "epoch": 1.130031746031746, "grad_norm": 4.796786785125732, "learning_rate": 0.00012471643812425934, "loss": 3.0254, "step": 4450 }, { "epoch": 1.1427301587301588, "grad_norm": 5.506641864776611, "learning_rate": 0.00012386998476383951, "loss": 3.0268, "step": 4500 }, { "epoch": 1.1554285714285715, "grad_norm": 4.79102897644043, "learning_rate": 0.00012302353140341966, "loss": 2.9979, "step": 4550 }, { "epoch": 1.1681269841269841, "grad_norm": 5.435400009155273, "learning_rate": 0.00012217707804299983, "loss": 2.987, "step": 4600 }, { "epoch": 1.1808253968253968, "grad_norm": 4.705477714538574, "learning_rate": 0.00012133062468257999, "loss": 2.9915, "step": 4650 }, { "epoch": 1.1935238095238094, "grad_norm": 4.822847843170166, "learning_rate": 0.00012048417132216015, "loss": 2.9762, "step": 4700 }, { "epoch": 1.2062222222222223, "grad_norm": 4.770782947540283, "learning_rate": 0.0001196377179617403, "loss": 2.9744, "step": 4750 }, { "epoch": 1.218920634920635, "grad_norm": 5.113085746765137, "learning_rate": 0.00011879126460132046, "loss": 2.9761, "step": 4800 }, { "epoch": 1.2316190476190476, "grad_norm": 6.109035491943359, "learning_rate": 0.00011796174030810902, "loss": 2.934, "step": 4850 }, { "epoch": 1.2443174603174603, "grad_norm": 5.884860038757324, "learning_rate": 0.00011711528694768918, "loss": 2.9369, "step": 4900 }, { "epoch": 1.257015873015873, "grad_norm": 7.224523544311523, "learning_rate": 0.00011626883358726934, "loss": 2.9209, "step": 4950 }, { "epoch": 1.2697142857142858, "grad_norm": 5.234792232513428, "learning_rate": 0.0001154223802268495, "loss": 2.9331, "step": 5000 }, { "epoch": 1.2824126984126984, "grad_norm": 4.842894554138184, "learning_rate": 0.00011457592686642966, "loss": 2.8986, "step": 5050 }, { "epoch": 1.295111111111111, "grad_norm": 4.660989284515381, "learning_rate": 0.00011372947350600981, "loss": 2.9141, "step": 5100 }, { "epoch": 1.3078095238095238, "grad_norm": 5.343238830566406, "learning_rate": 0.00011288302014558999, "loss": 2.8895, "step": 5150 }, { "epoch": 1.3205079365079366, "grad_norm": 5.187355041503906, "learning_rate": 0.00011203656678517014, "loss": 2.892, "step": 5200 }, { "epoch": 1.3332063492063493, "grad_norm": 4.856098175048828, "learning_rate": 0.0001111901134247503, "loss": 2.8736, "step": 5250 }, { "epoch": 1.345904761904762, "grad_norm": 4.733485698699951, "learning_rate": 0.00011034366006433046, "loss": 2.8567, "step": 5300 }, { "epoch": 1.3586031746031746, "grad_norm": 4.625833034515381, "learning_rate": 0.00010949720670391062, "loss": 2.8737, "step": 5350 }, { "epoch": 1.3713015873015872, "grad_norm": 4.856983184814453, "learning_rate": 0.00010865075334349078, "loss": 2.8417, "step": 5400 }, { "epoch": 1.384, "grad_norm": 5.2909932136535645, "learning_rate": 0.00010780429998307094, "loss": 2.8346, "step": 5450 }, { "epoch": 1.3966984126984128, "grad_norm": 4.909002780914307, "learning_rate": 0.0001069578466226511, "loss": 2.8307, "step": 5500 }, { "epoch": 1.4093968253968254, "grad_norm": 4.69639778137207, "learning_rate": 0.00010611139326223125, "loss": 2.8138, "step": 5550 }, { "epoch": 1.422095238095238, "grad_norm": 4.822878837585449, "learning_rate": 0.00010526493990181141, "loss": 2.8071, "step": 5600 }, { "epoch": 1.4347936507936507, "grad_norm": 5.602210998535156, "learning_rate": 0.00010441848654139157, "loss": 2.8128, "step": 5650 }, { "epoch": 1.4474920634920636, "grad_norm": 4.855912208557129, "learning_rate": 0.00010357203318097173, "loss": 2.8066, "step": 5700 }, { "epoch": 1.4601904761904763, "grad_norm": 5.553136348724365, "learning_rate": 0.00010272557982055189, "loss": 2.7876, "step": 5750 }, { "epoch": 1.472888888888889, "grad_norm": 4.901633262634277, "learning_rate": 0.00010187912646013204, "loss": 2.7787, "step": 5800 }, { "epoch": 1.4855873015873016, "grad_norm": 6.1740217208862305, "learning_rate": 0.0001010326730997122, "loss": 2.7704, "step": 5850 }, { "epoch": 1.4982857142857142, "grad_norm": 5.393040180206299, "learning_rate": 0.00010018621973929236, "loss": 2.7753, "step": 5900 }, { "epoch": 1.5109841269841269, "grad_norm": 6.0959930419921875, "learning_rate": 9.933976637887253e-05, "loss": 2.7538, "step": 5950 }, { "epoch": 1.5236825396825395, "grad_norm": 4.659241199493408, "learning_rate": 9.849331301845269e-05, "loss": 2.7517, "step": 6000 }, { "epoch": 1.5363809523809524, "grad_norm": 5.5795087814331055, "learning_rate": 9.766378872524125e-05, "loss": 2.7409, "step": 6050 }, { "epoch": 1.549079365079365, "grad_norm": 4.83104944229126, "learning_rate": 9.681733536482141e-05, "loss": 2.7358, "step": 6100 }, { "epoch": 1.561777777777778, "grad_norm": 5.035250663757324, "learning_rate": 9.597088200440157e-05, "loss": 2.7236, "step": 6150 }, { "epoch": 1.5744761904761906, "grad_norm": 5.167687892913818, "learning_rate": 9.512442864398172e-05, "loss": 2.7232, "step": 6200 }, { "epoch": 1.5871746031746032, "grad_norm": 5.0377326011657715, "learning_rate": 9.427797528356188e-05, "loss": 2.7368, "step": 6250 }, { "epoch": 1.599873015873016, "grad_norm": 4.893152713775635, "learning_rate": 9.343152192314204e-05, "loss": 2.6973, "step": 6300 }, { "epoch": 1.6125714285714285, "grad_norm": 5.246462345123291, "learning_rate": 9.25850685627222e-05, "loss": 2.6858, "step": 6350 }, { "epoch": 1.6252698412698412, "grad_norm": 5.26235294342041, "learning_rate": 9.173861520230236e-05, "loss": 2.6816, "step": 6400 }, { "epoch": 1.6379682539682539, "grad_norm": 4.8995513916015625, "learning_rate": 9.089216184188252e-05, "loss": 2.6865, "step": 6450 }, { "epoch": 1.6506666666666665, "grad_norm": 5.598567962646484, "learning_rate": 9.004570848146267e-05, "loss": 2.6675, "step": 6500 }, { "epoch": 1.6633650793650794, "grad_norm": 5.423081874847412, "learning_rate": 8.919925512104283e-05, "loss": 2.66, "step": 6550 }, { "epoch": 1.676063492063492, "grad_norm": 4.968945026397705, "learning_rate": 8.835280176062299e-05, "loss": 2.6629, "step": 6600 }, { "epoch": 1.688761904761905, "grad_norm": 6.054278373718262, "learning_rate": 8.750634840020315e-05, "loss": 2.6784, "step": 6650 }, { "epoch": 1.7014603174603176, "grad_norm": 5.279598712921143, "learning_rate": 8.665989503978331e-05, "loss": 2.6327, "step": 6700 }, { "epoch": 1.7141587301587302, "grad_norm": 5.150700092315674, "learning_rate": 8.581344167936347e-05, "loss": 2.6394, "step": 6750 }, { "epoch": 1.7268571428571429, "grad_norm": 5.459251403808594, "learning_rate": 8.496698831894362e-05, "loss": 2.644, "step": 6800 }, { "epoch": 1.7395555555555555, "grad_norm": 5.293938159942627, "learning_rate": 8.413746402573218e-05, "loss": 2.646, "step": 6850 }, { "epoch": 1.7522539682539682, "grad_norm": 5.72529411315918, "learning_rate": 8.329101066531234e-05, "loss": 2.624, "step": 6900 }, { "epoch": 1.7649523809523808, "grad_norm": 5.739988327026367, "learning_rate": 8.24445573048925e-05, "loss": 2.6175, "step": 6950 }, { "epoch": 1.7776507936507937, "grad_norm": 5.638957500457764, "learning_rate": 8.159810394447266e-05, "loss": 2.6236, "step": 7000 }, { "epoch": 1.7903492063492064, "grad_norm": 5.7885026931762695, "learning_rate": 8.075165058405282e-05, "loss": 2.5973, "step": 7050 }, { "epoch": 1.803047619047619, "grad_norm": 5.244924545288086, "learning_rate": 7.990519722363298e-05, "loss": 2.6031, "step": 7100 }, { "epoch": 1.8157460317460319, "grad_norm": 9.06462287902832, "learning_rate": 7.905874386321313e-05, "loss": 2.5768, "step": 7150 }, { "epoch": 1.8284444444444445, "grad_norm": 5.335842132568359, "learning_rate": 7.821229050279329e-05, "loss": 2.581, "step": 7200 }, { "epoch": 1.8411428571428572, "grad_norm": 5.182567596435547, "learning_rate": 7.736583714237345e-05, "loss": 2.5754, "step": 7250 }, { "epoch": 1.8538412698412698, "grad_norm": 5.487778186798096, "learning_rate": 7.651938378195361e-05, "loss": 2.5688, "step": 7300 }, { "epoch": 1.8665396825396825, "grad_norm": 5.46382474899292, "learning_rate": 7.567293042153377e-05, "loss": 2.5632, "step": 7350 }, { "epoch": 1.8792380952380952, "grad_norm": 5.17083740234375, "learning_rate": 7.482647706111392e-05, "loss": 2.55, "step": 7400 }, { "epoch": 1.8919365079365078, "grad_norm": 5.455732345581055, "learning_rate": 7.398002370069408e-05, "loss": 2.5509, "step": 7450 }, { "epoch": 1.9046349206349207, "grad_norm": 5.0072503089904785, "learning_rate": 7.313357034027426e-05, "loss": 2.5436, "step": 7500 }, { "epoch": 1.9173333333333333, "grad_norm": 5.0145087242126465, "learning_rate": 7.228711697985441e-05, "loss": 2.5416, "step": 7550 }, { "epoch": 1.930031746031746, "grad_norm": 5.2530364990234375, "learning_rate": 7.144066361943457e-05, "loss": 2.5234, "step": 7600 }, { "epoch": 1.9427301587301589, "grad_norm": 4.886019229888916, "learning_rate": 7.059421025901474e-05, "loss": 2.5447, "step": 7650 }, { "epoch": 1.9554285714285715, "grad_norm": 5.070368766784668, "learning_rate": 6.97477568985949e-05, "loss": 2.5308, "step": 7700 }, { "epoch": 1.9681269841269842, "grad_norm": 5.158459186553955, "learning_rate": 6.890130353817506e-05, "loss": 2.5265, "step": 7750 }, { "epoch": 1.9808253968253968, "grad_norm": 5.249716281890869, "learning_rate": 6.805485017775522e-05, "loss": 2.5079, "step": 7800 }, { "epoch": 1.9935238095238095, "grad_norm": 5.3184590339660645, "learning_rate": 6.720839681733538e-05, "loss": 2.5135, "step": 7850 }, { "epoch": 2.006095238095238, "grad_norm": 4.898834705352783, "learning_rate": 6.636194345691553e-05, "loss": 2.487, "step": 7900 }, { "epoch": 2.0187936507936506, "grad_norm": 4.718045234680176, "learning_rate": 6.55154900964957e-05, "loss": 2.504, "step": 7950 }, { "epoch": 2.0314920634920637, "grad_norm": 4.977246284484863, "learning_rate": 6.466903673607585e-05, "loss": 2.4924, "step": 8000 }, { "epoch": 2.0441904761904763, "grad_norm": 5.398455619812012, "learning_rate": 6.382258337565601e-05, "loss": 2.4839, "step": 8050 }, { "epoch": 2.056888888888889, "grad_norm": 6.387637138366699, "learning_rate": 6.297613001523617e-05, "loss": 2.4833, "step": 8100 }, { "epoch": 2.0695873015873016, "grad_norm": 5.588785648345947, "learning_rate": 6.212967665481633e-05, "loss": 2.4569, "step": 8150 }, { "epoch": 2.0822857142857143, "grad_norm": 6.301563262939453, "learning_rate": 6.128322329439648e-05, "loss": 2.4748, "step": 8200 }, { "epoch": 2.094984126984127, "grad_norm": 5.7610979080200195, "learning_rate": 6.043676993397664e-05, "loss": 2.4597, "step": 8250 }, { "epoch": 2.1076825396825396, "grad_norm": 5.260268211364746, "learning_rate": 5.95903165735568e-05, "loss": 2.4581, "step": 8300 }, { "epoch": 2.1203809523809523, "grad_norm": 5.712375640869141, "learning_rate": 5.874386321313696e-05, "loss": 2.4522, "step": 8350 }, { "epoch": 2.133079365079365, "grad_norm": 6.139365196228027, "learning_rate": 5.7897409852717125e-05, "loss": 2.4588, "step": 8400 }, { "epoch": 2.145777777777778, "grad_norm": 5.294638633728027, "learning_rate": 5.705095649229728e-05, "loss": 2.4467, "step": 8450 }, { "epoch": 2.1584761904761907, "grad_norm": 6.786096572875977, "learning_rate": 5.620450313187744e-05, "loss": 2.4538, "step": 8500 }, { "epoch": 2.1711746031746033, "grad_norm": 4.95149564743042, "learning_rate": 5.53580497714576e-05, "loss": 2.4478, "step": 8550 }, { "epoch": 2.183873015873016, "grad_norm": 4.965189456939697, "learning_rate": 5.451159641103776e-05, "loss": 2.4449, "step": 8600 }, { "epoch": 2.1965714285714286, "grad_norm": 5.404216766357422, "learning_rate": 5.3665143050617916e-05, "loss": 2.4447, "step": 8650 }, { "epoch": 2.2092698412698413, "grad_norm": 5.965044021606445, "learning_rate": 5.2818689690198075e-05, "loss": 2.4337, "step": 8700 }, { "epoch": 2.221968253968254, "grad_norm": 6.933143615722656, "learning_rate": 5.197223632977823e-05, "loss": 2.4105, "step": 8750 }, { "epoch": 2.2346666666666666, "grad_norm": 4.795731067657471, "learning_rate": 5.112578296935839e-05, "loss": 2.4192, "step": 8800 }, { "epoch": 2.2473650793650792, "grad_norm": 5.221839427947998, "learning_rate": 5.027932960893855e-05, "loss": 2.4232, "step": 8850 }, { "epoch": 2.260063492063492, "grad_norm": 5.512608528137207, "learning_rate": 4.9432876248518715e-05, "loss": 2.414, "step": 8900 }, { "epoch": 2.2727619047619045, "grad_norm": 5.804533958435059, "learning_rate": 4.858642288809887e-05, "loss": 2.4088, "step": 8950 }, { "epoch": 2.2854603174603176, "grad_norm": 5.301205635070801, "learning_rate": 4.773996952767903e-05, "loss": 2.4074, "step": 9000 }, { "epoch": 2.2981587301587303, "grad_norm": 5.4529290199279785, "learning_rate": 4.689351616725919e-05, "loss": 2.3906, "step": 9050 }, { "epoch": 2.310857142857143, "grad_norm": 5.1715006828308105, "learning_rate": 4.604706280683935e-05, "loss": 2.4082, "step": 9100 }, { "epoch": 2.3235555555555556, "grad_norm": 5.739888668060303, "learning_rate": 4.5200609446419506e-05, "loss": 2.3982, "step": 9150 }, { "epoch": 2.3362539682539682, "grad_norm": 5.454600811004639, "learning_rate": 4.4354156085999664e-05, "loss": 2.3947, "step": 9200 }, { "epoch": 2.348952380952381, "grad_norm": 5.671194553375244, "learning_rate": 4.350770272557982e-05, "loss": 2.3984, "step": 9250 }, { "epoch": 2.3616507936507936, "grad_norm": 5.695377349853516, "learning_rate": 4.266124936515998e-05, "loss": 2.3732, "step": 9300 }, { "epoch": 2.374349206349206, "grad_norm": 5.454712390899658, "learning_rate": 4.181479600474014e-05, "loss": 2.3769, "step": 9350 }, { "epoch": 2.387047619047619, "grad_norm": 5.6516499519348145, "learning_rate": 4.0968342644320304e-05, "loss": 2.3788, "step": 9400 }, { "epoch": 2.399746031746032, "grad_norm": 5.617581367492676, "learning_rate": 4.012188928390046e-05, "loss": 2.3754, "step": 9450 }, { "epoch": 2.4124444444444446, "grad_norm": 5.995534420013428, "learning_rate": 3.927543592348062e-05, "loss": 2.3647, "step": 9500 }, { "epoch": 2.4251428571428573, "grad_norm": 4.861730098724365, "learning_rate": 3.842898256306078e-05, "loss": 2.3699, "step": 9550 }, { "epoch": 2.43784126984127, "grad_norm": 5.109068393707275, "learning_rate": 3.758252920264094e-05, "loss": 2.3709, "step": 9600 }, { "epoch": 2.4505396825396826, "grad_norm": 5.989678382873535, "learning_rate": 3.6736075842221096e-05, "loss": 2.3442, "step": 9650 }, { "epoch": 2.4632380952380952, "grad_norm": 5.233463764190674, "learning_rate": 3.5889622481801254e-05, "loss": 2.3567, "step": 9700 }, { "epoch": 2.475936507936508, "grad_norm": 4.930139541625977, "learning_rate": 3.504316912138141e-05, "loss": 2.3334, "step": 9750 }, { "epoch": 2.4886349206349205, "grad_norm": 5.235612392425537, "learning_rate": 3.419671576096157e-05, "loss": 2.3523, "step": 9800 }, { "epoch": 2.501333333333333, "grad_norm": 7.875730991363525, "learning_rate": 3.335026240054173e-05, "loss": 2.3508, "step": 9850 }, { "epoch": 2.514031746031746, "grad_norm": 5.479938507080078, "learning_rate": 3.2503809040121894e-05, "loss": 2.3579, "step": 9900 }, { "epoch": 2.5267301587301585, "grad_norm": 4.948204040527344, "learning_rate": 3.165735567970205e-05, "loss": 2.3475, "step": 9950 }, { "epoch": 2.5394285714285716, "grad_norm": 5.541229724884033, "learning_rate": 3.081090231928221e-05, "loss": 2.3276, "step": 10000 }, { "epoch": 2.5521269841269842, "grad_norm": 5.418817043304443, "learning_rate": 2.996444895886237e-05, "loss": 2.3433, "step": 10050 }, { "epoch": 2.564825396825397, "grad_norm": 7.228455543518066, "learning_rate": 2.9117995598442527e-05, "loss": 2.3274, "step": 10100 }, { "epoch": 2.5775238095238096, "grad_norm": 5.323376655578613, "learning_rate": 2.8271542238022686e-05, "loss": 2.3404, "step": 10150 }, { "epoch": 2.590222222222222, "grad_norm": 5.080998420715332, "learning_rate": 2.7425088877602844e-05, "loss": 2.3164, "step": 10200 }, { "epoch": 2.602920634920635, "grad_norm": 5.0400285720825195, "learning_rate": 2.6578635517183002e-05, "loss": 2.3301, "step": 10250 }, { "epoch": 2.6156190476190475, "grad_norm": 5.519168853759766, "learning_rate": 2.5732182156763164e-05, "loss": 2.3206, "step": 10300 }, { "epoch": 2.62831746031746, "grad_norm": 5.184562683105469, "learning_rate": 2.4885728796343322e-05, "loss": 2.3197, "step": 10350 }, { "epoch": 2.6410158730158733, "grad_norm": 5.173785209655762, "learning_rate": 2.403927543592348e-05, "loss": 2.3176, "step": 10400 }, { "epoch": 2.653714285714286, "grad_norm": 5.67647647857666, "learning_rate": 2.319282207550364e-05, "loss": 2.3154, "step": 10450 }, { "epoch": 2.6664126984126986, "grad_norm": 6.398087978363037, "learning_rate": 2.2346368715083797e-05, "loss": 2.3164, "step": 10500 }, { "epoch": 2.679111111111111, "grad_norm": 5.975333213806152, "learning_rate": 2.149991535466396e-05, "loss": 2.314, "step": 10550 }, { "epoch": 2.691809523809524, "grad_norm": 5.434169292449951, "learning_rate": 2.065346199424412e-05, "loss": 2.3034, "step": 10600 }, { "epoch": 2.7045079365079365, "grad_norm": 5.366811275482178, "learning_rate": 1.980700863382428e-05, "loss": 2.3161, "step": 10650 }, { "epoch": 2.717206349206349, "grad_norm": 6.124394416809082, "learning_rate": 1.8960555273404437e-05, "loss": 2.3031, "step": 10700 }, { "epoch": 2.729904761904762, "grad_norm": 6.769460201263428, "learning_rate": 1.8114101912984595e-05, "loss": 2.2893, "step": 10750 }, { "epoch": 2.7426031746031745, "grad_norm": 4.68062162399292, "learning_rate": 1.7267648552564754e-05, "loss": 2.3194, "step": 10800 }, { "epoch": 2.755301587301587, "grad_norm": 5.621355056762695, "learning_rate": 1.6421195192144915e-05, "loss": 2.3095, "step": 10850 }, { "epoch": 2.768, "grad_norm": 5.693627834320068, "learning_rate": 1.5574741831725074e-05, "loss": 2.3029, "step": 10900 }, { "epoch": 2.7806984126984124, "grad_norm": 5.223197937011719, "learning_rate": 1.4728288471305232e-05, "loss": 2.2965, "step": 10950 }, { "epoch": 2.7933968253968255, "grad_norm": 5.495180606842041, "learning_rate": 1.388183511088539e-05, "loss": 2.3037, "step": 11000 }, { "epoch": 2.806095238095238, "grad_norm": 5.025885105133057, "learning_rate": 1.303538175046555e-05, "loss": 2.2896, "step": 11050 }, { "epoch": 2.818793650793651, "grad_norm": 5.007611274719238, "learning_rate": 1.2188928390045709e-05, "loss": 2.2878, "step": 11100 }, { "epoch": 2.8314920634920635, "grad_norm": 4.786118984222412, "learning_rate": 1.1342475029625869e-05, "loss": 2.2823, "step": 11150 }, { "epoch": 2.844190476190476, "grad_norm": 5.517834663391113, "learning_rate": 1.0496021669206027e-05, "loss": 2.2796, "step": 11200 }, { "epoch": 2.856888888888889, "grad_norm": 7.0708818435668945, "learning_rate": 9.649568308786185e-06, "loss": 2.2883, "step": 11250 }, { "epoch": 2.8695873015873015, "grad_norm": 5.164283275604248, "learning_rate": 8.803114948366345e-06, "loss": 2.2921, "step": 11300 }, { "epoch": 2.8822857142857146, "grad_norm": 5.082614421844482, "learning_rate": 7.956661587946503e-06, "loss": 2.2956, "step": 11350 }, { "epoch": 2.894984126984127, "grad_norm": 5.358335018157959, "learning_rate": 7.110208227526663e-06, "loss": 2.29, "step": 11400 }, { "epoch": 2.90768253968254, "grad_norm": 4.937663555145264, "learning_rate": 6.2637548671068235e-06, "loss": 2.2912, "step": 11450 }, { "epoch": 2.9203809523809525, "grad_norm": 4.954619407653809, "learning_rate": 5.417301506686982e-06, "loss": 2.2946, "step": 11500 }, { "epoch": 2.933079365079365, "grad_norm": 5.406091690063477, "learning_rate": 4.570848146267141e-06, "loss": 2.2903, "step": 11550 }, { "epoch": 2.945777777777778, "grad_norm": 5.957233428955078, "learning_rate": 3.7243947858473e-06, "loss": 2.2808, "step": 11600 }, { "epoch": 2.9584761904761905, "grad_norm": 5.3814215660095215, "learning_rate": 2.8779414254274592e-06, "loss": 2.2918, "step": 11650 }, { "epoch": 2.971174603174603, "grad_norm": 7.456835746765137, "learning_rate": 2.0314880650076184e-06, "loss": 2.2871, "step": 11700 }, { "epoch": 2.983873015873016, "grad_norm": 5.577419757843018, "learning_rate": 1.1850347045877773e-06, "loss": 2.2791, "step": 11750 }, { "epoch": 2.9965714285714284, "grad_norm": 5.656188488006592, "learning_rate": 3.3858134416793636e-07, "loss": 2.2811, "step": 11800 }, { "epoch": 3.0, "step": 11814, "total_flos": 5952688007980032.0, "train_loss": 3.0341684326254534, "train_runtime": 5229.0251, "train_samples_per_second": 72.289, "train_steps_per_second": 2.259 } ], "logging_steps": 50, "max_steps": 11814, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5952688007980032.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }