{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.24832381425378694, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00024832381425378696, "grad_norm": 10.215656280517578, "learning_rate": 0.0, "loss": 2.4774, "num_input_tokens_seen": 262144, "step": 1 }, { "epoch": 0.0004966476285075739, "grad_norm": 11.095035552978516, "learning_rate": 1.0000000000000001e-07, "loss": 2.9956, "num_input_tokens_seen": 524288, "step": 2 }, { "epoch": 0.0007449714427613609, "grad_norm": 15.851714134216309, "learning_rate": 2.0000000000000002e-07, "loss": 3.5577, "num_input_tokens_seen": 786432, "step": 3 }, { "epoch": 0.0009932952570151478, "grad_norm": 13.334907531738281, "learning_rate": 3.0000000000000004e-07, "loss": 3.5265, "num_input_tokens_seen": 1048576, "step": 4 }, { "epoch": 0.0012416190712689348, "grad_norm": 13.653706550598145, "learning_rate": 4.0000000000000003e-07, "loss": 3.2742, "num_input_tokens_seen": 1310720, "step": 5 }, { "epoch": 0.0014899428855227217, "grad_norm": 10.820155143737793, "learning_rate": 5.000000000000001e-07, "loss": 2.7898, "num_input_tokens_seen": 1572864, "step": 6 }, { "epoch": 0.0017382666997765085, "grad_norm": 8.97152042388916, "learning_rate": 6.000000000000001e-07, "loss": 2.3872, "num_input_tokens_seen": 1835008, "step": 7 }, { "epoch": 0.0019865905140302956, "grad_norm": 8.390254020690918, "learning_rate": 7.000000000000001e-07, "loss": 2.4003, "num_input_tokens_seen": 2097152, "step": 8 }, { "epoch": 0.0022349143282840824, "grad_norm": 7.845387935638428, "learning_rate": 8.000000000000001e-07, "loss": 2.3165, "num_input_tokens_seen": 2359296, "step": 9 }, { "epoch": 0.0024832381425378696, "grad_norm": 10.649581909179688, "learning_rate": 9.000000000000001e-07, "loss": 3.193, "num_input_tokens_seen": 2621440, "step": 10 }, { "epoch": 0.0027315619567916563, "grad_norm": 9.258522033691406, "learning_rate": 1.0000000000000002e-06, "loss": 3.0098, "num_input_tokens_seen": 2883584, "step": 11 }, { "epoch": 0.0029798857710454435, "grad_norm": 8.765289306640625, "learning_rate": 1.1e-06, "loss": 2.5393, "num_input_tokens_seen": 3145728, "step": 12 }, { "epoch": 0.00322820958529923, "grad_norm": 10.003978729248047, "learning_rate": 1.2000000000000002e-06, "loss": 2.9451, "num_input_tokens_seen": 3407872, "step": 13 }, { "epoch": 0.003476533399553017, "grad_norm": 8.10649585723877, "learning_rate": 1.3e-06, "loss": 2.7237, "num_input_tokens_seen": 3670016, "step": 14 }, { "epoch": 0.003724857213806804, "grad_norm": 8.63984489440918, "learning_rate": 1.4000000000000001e-06, "loss": 2.922, "num_input_tokens_seen": 3932160, "step": 15 }, { "epoch": 0.003973181028060591, "grad_norm": 6.354504585266113, "learning_rate": 1.5e-06, "loss": 2.494, "num_input_tokens_seen": 4194304, "step": 16 }, { "epoch": 0.004221504842314378, "grad_norm": 6.988131046295166, "learning_rate": 1.6000000000000001e-06, "loss": 3.051, "num_input_tokens_seen": 4456448, "step": 17 }, { "epoch": 0.004469828656568165, "grad_norm": 6.21164608001709, "learning_rate": 1.7000000000000002e-06, "loss": 2.8812, "num_input_tokens_seen": 4718592, "step": 18 }, { "epoch": 0.0047181524708219515, "grad_norm": 7.339629173278809, "learning_rate": 1.8000000000000001e-06, "loss": 2.8418, "num_input_tokens_seen": 4980736, "step": 19 }, { "epoch": 0.004966476285075739, "grad_norm": 5.495388031005859, "learning_rate": 1.9000000000000002e-06, "loss": 2.4937, "num_input_tokens_seen": 5242880, "step": 20 }, { "epoch": 0.005214800099329526, "grad_norm": 4.960653305053711, "learning_rate": 2.0000000000000003e-06, "loss": 2.5324, "num_input_tokens_seen": 5505024, "step": 21 }, { "epoch": 0.005463123913583313, "grad_norm": 5.600866794586182, "learning_rate": 2.1000000000000002e-06, "loss": 2.9153, "num_input_tokens_seen": 5767168, "step": 22 }, { "epoch": 0.005711447727837099, "grad_norm": 4.116523265838623, "learning_rate": 2.2e-06, "loss": 2.4081, "num_input_tokens_seen": 6029312, "step": 23 }, { "epoch": 0.005959771542090887, "grad_norm": 3.069936990737915, "learning_rate": 2.3000000000000004e-06, "loss": 2.059, "num_input_tokens_seen": 6291456, "step": 24 }, { "epoch": 0.006208095356344674, "grad_norm": 5.050525188446045, "learning_rate": 2.4000000000000003e-06, "loss": 2.802, "num_input_tokens_seen": 6553600, "step": 25 }, { "epoch": 0.00645641917059846, "grad_norm": 4.871302604675293, "learning_rate": 2.5e-06, "loss": 2.5931, "num_input_tokens_seen": 6815744, "step": 26 }, { "epoch": 0.006704742984852247, "grad_norm": 3.577676773071289, "learning_rate": 2.6e-06, "loss": 2.6076, "num_input_tokens_seen": 7077888, "step": 27 }, { "epoch": 0.006953066799106034, "grad_norm": 4.384449481964111, "learning_rate": 2.7000000000000004e-06, "loss": 2.8547, "num_input_tokens_seen": 7340032, "step": 28 }, { "epoch": 0.0072013906133598215, "grad_norm": 2.9289448261260986, "learning_rate": 2.8000000000000003e-06, "loss": 2.1733, "num_input_tokens_seen": 7602176, "step": 29 }, { "epoch": 0.007449714427613608, "grad_norm": 3.34555983543396, "learning_rate": 2.9e-06, "loss": 2.4006, "num_input_tokens_seen": 7864320, "step": 30 }, { "epoch": 0.007698038241867395, "grad_norm": 2.8291356563568115, "learning_rate": 3e-06, "loss": 2.4605, "num_input_tokens_seen": 8126464, "step": 31 }, { "epoch": 0.007946362056121183, "grad_norm": 2.9072861671447754, "learning_rate": 3.1000000000000004e-06, "loss": 2.5204, "num_input_tokens_seen": 8388608, "step": 32 }, { "epoch": 0.008194685870374968, "grad_norm": 2.771606683731079, "learning_rate": 3.2000000000000003e-06, "loss": 2.4139, "num_input_tokens_seen": 8650752, "step": 33 }, { "epoch": 0.008443009684628756, "grad_norm": 2.580414295196533, "learning_rate": 3.3000000000000006e-06, "loss": 2.5677, "num_input_tokens_seen": 8912896, "step": 34 }, { "epoch": 0.008691333498882544, "grad_norm": 2.1037166118621826, "learning_rate": 3.4000000000000005e-06, "loss": 2.1773, "num_input_tokens_seen": 9175040, "step": 35 }, { "epoch": 0.00893965731313633, "grad_norm": 2.0104498863220215, "learning_rate": 3.5e-06, "loss": 2.1332, "num_input_tokens_seen": 9437184, "step": 36 }, { "epoch": 0.009187981127390117, "grad_norm": 1.954823613166809, "learning_rate": 3.6000000000000003e-06, "loss": 2.3993, "num_input_tokens_seen": 9699328, "step": 37 }, { "epoch": 0.009436304941643903, "grad_norm": 2.064880847930908, "learning_rate": 3.7e-06, "loss": 2.6064, "num_input_tokens_seen": 9961472, "step": 38 }, { "epoch": 0.00968462875589769, "grad_norm": 1.6983036994934082, "learning_rate": 3.8000000000000005e-06, "loss": 1.9788, "num_input_tokens_seen": 10223616, "step": 39 }, { "epoch": 0.009932952570151478, "grad_norm": 1.7680089473724365, "learning_rate": 3.900000000000001e-06, "loss": 2.3603, "num_input_tokens_seen": 10485760, "step": 40 }, { "epoch": 0.010181276384405264, "grad_norm": 1.5656706094741821, "learning_rate": 4.000000000000001e-06, "loss": 2.2015, "num_input_tokens_seen": 10747904, "step": 41 }, { "epoch": 0.010429600198659052, "grad_norm": 1.4727283716201782, "learning_rate": 4.1e-06, "loss": 2.0206, "num_input_tokens_seen": 11010048, "step": 42 }, { "epoch": 0.010677924012912838, "grad_norm": 1.2997132539749146, "learning_rate": 4.2000000000000004e-06, "loss": 1.7953, "num_input_tokens_seen": 11272192, "step": 43 }, { "epoch": 0.010926247827166625, "grad_norm": 1.439453363418579, "learning_rate": 4.3e-06, "loss": 2.124, "num_input_tokens_seen": 11534336, "step": 44 }, { "epoch": 0.011174571641420413, "grad_norm": 1.43881356716156, "learning_rate": 4.4e-06, "loss": 2.4577, "num_input_tokens_seen": 11796480, "step": 45 }, { "epoch": 0.011422895455674199, "grad_norm": 1.5415407419204712, "learning_rate": 4.5e-06, "loss": 2.3313, "num_input_tokens_seen": 12058624, "step": 46 }, { "epoch": 0.011671219269927986, "grad_norm": 1.2109966278076172, "learning_rate": 4.600000000000001e-06, "loss": 2.0387, "num_input_tokens_seen": 12320768, "step": 47 }, { "epoch": 0.011919543084181774, "grad_norm": 1.2937291860580444, "learning_rate": 4.7e-06, "loss": 2.3696, "num_input_tokens_seen": 12582912, "step": 48 }, { "epoch": 0.01216786689843556, "grad_norm": 1.2319666147232056, "learning_rate": 4.800000000000001e-06, "loss": 2.3702, "num_input_tokens_seen": 12845056, "step": 49 }, { "epoch": 0.012416190712689347, "grad_norm": 2.2272355556488037, "learning_rate": 4.9000000000000005e-06, "loss": 1.7134, "num_input_tokens_seen": 13107200, "step": 50 }, { "epoch": 0.012664514526943133, "grad_norm": 0.9094964265823364, "learning_rate": 5e-06, "loss": 1.8803, "num_input_tokens_seen": 13369344, "step": 51 }, { "epoch": 0.01291283834119692, "grad_norm": 1.3690983057022095, "learning_rate": 5.1e-06, "loss": 2.541, "num_input_tokens_seen": 13631488, "step": 52 }, { "epoch": 0.013161162155450708, "grad_norm": 0.8856329917907715, "learning_rate": 5.2e-06, "loss": 2.1698, "num_input_tokens_seen": 13893632, "step": 53 }, { "epoch": 0.013409485969704494, "grad_norm": 0.9775459170341492, "learning_rate": 5.300000000000001e-06, "loss": 2.0176, "num_input_tokens_seen": 14155776, "step": 54 }, { "epoch": 0.013657809783958282, "grad_norm": 0.8204602599143982, "learning_rate": 5.400000000000001e-06, "loss": 1.895, "num_input_tokens_seen": 14417920, "step": 55 }, { "epoch": 0.013906133598212068, "grad_norm": 0.9787035584449768, "learning_rate": 5.500000000000001e-06, "loss": 2.1046, "num_input_tokens_seen": 14680064, "step": 56 }, { "epoch": 0.014154457412465855, "grad_norm": 0.7377949357032776, "learning_rate": 5.600000000000001e-06, "loss": 1.9217, "num_input_tokens_seen": 14942208, "step": 57 }, { "epoch": 0.014402781226719643, "grad_norm": 0.8319304585456848, "learning_rate": 5.7e-06, "loss": 2.278, "num_input_tokens_seen": 15204352, "step": 58 }, { "epoch": 0.014651105040973429, "grad_norm": 0.8811312317848206, "learning_rate": 5.8e-06, "loss": 2.4527, "num_input_tokens_seen": 15466496, "step": 59 }, { "epoch": 0.014899428855227216, "grad_norm": 0.8537189364433289, "learning_rate": 5.9e-06, "loss": 2.0035, "num_input_tokens_seen": 15728640, "step": 60 }, { "epoch": 0.015147752669481004, "grad_norm": 0.8903886079788208, "learning_rate": 6e-06, "loss": 2.0959, "num_input_tokens_seen": 15990784, "step": 61 }, { "epoch": 0.01539607648373479, "grad_norm": 0.7955805659294128, "learning_rate": 6.1e-06, "loss": 1.9258, "num_input_tokens_seen": 16252928, "step": 62 }, { "epoch": 0.015644400297988578, "grad_norm": 0.7012743353843689, "learning_rate": 6.200000000000001e-06, "loss": 1.8984, "num_input_tokens_seen": 16515072, "step": 63 }, { "epoch": 0.015892724112242365, "grad_norm": 1.0720924139022827, "learning_rate": 6.300000000000001e-06, "loss": 2.3707, "num_input_tokens_seen": 16777216, "step": 64 }, { "epoch": 0.01614104792649615, "grad_norm": 0.7156981229782104, "learning_rate": 6.4000000000000006e-06, "loss": 2.0082, "num_input_tokens_seen": 17039360, "step": 65 }, { "epoch": 0.016389371740749937, "grad_norm": 1.379465103149414, "learning_rate": 6.5000000000000004e-06, "loss": 1.893, "num_input_tokens_seen": 17301504, "step": 66 }, { "epoch": 0.016637695555003724, "grad_norm": 0.6216750144958496, "learning_rate": 6.600000000000001e-06, "loss": 1.7876, "num_input_tokens_seen": 17563648, "step": 67 }, { "epoch": 0.016886019369257512, "grad_norm": 0.6770815253257751, "learning_rate": 6.700000000000001e-06, "loss": 1.813, "num_input_tokens_seen": 17825792, "step": 68 }, { "epoch": 0.0171343431835113, "grad_norm": 0.8068878054618835, "learning_rate": 6.800000000000001e-06, "loss": 1.9584, "num_input_tokens_seen": 18087936, "step": 69 }, { "epoch": 0.017382666997765087, "grad_norm": 0.642892062664032, "learning_rate": 6.9e-06, "loss": 1.5887, "num_input_tokens_seen": 18350080, "step": 70 }, { "epoch": 0.01763099081201887, "grad_norm": 0.6749325394630432, "learning_rate": 7e-06, "loss": 1.7739, "num_input_tokens_seen": 18612224, "step": 71 }, { "epoch": 0.01787931462627266, "grad_norm": 0.6361634135246277, "learning_rate": 7.100000000000001e-06, "loss": 1.8437, "num_input_tokens_seen": 18874368, "step": 72 }, { "epoch": 0.018127638440526447, "grad_norm": 0.8063591718673706, "learning_rate": 7.2000000000000005e-06, "loss": 1.7656, "num_input_tokens_seen": 19136512, "step": 73 }, { "epoch": 0.018375962254780234, "grad_norm": 0.608739972114563, "learning_rate": 7.3e-06, "loss": 2.0104, "num_input_tokens_seen": 19398656, "step": 74 }, { "epoch": 0.018624286069034022, "grad_norm": 0.6015883088111877, "learning_rate": 7.4e-06, "loss": 2.0805, "num_input_tokens_seen": 19660800, "step": 75 }, { "epoch": 0.018872609883287806, "grad_norm": 0.5737051367759705, "learning_rate": 7.500000000000001e-06, "loss": 1.9203, "num_input_tokens_seen": 19922944, "step": 76 }, { "epoch": 0.019120933697541594, "grad_norm": 0.5844286680221558, "learning_rate": 7.600000000000001e-06, "loss": 2.0768, "num_input_tokens_seen": 20185088, "step": 77 }, { "epoch": 0.01936925751179538, "grad_norm": 0.5958046913146973, "learning_rate": 7.7e-06, "loss": 1.5849, "num_input_tokens_seen": 20447232, "step": 78 }, { "epoch": 0.01961758132604917, "grad_norm": 0.7638349533081055, "learning_rate": 7.800000000000002e-06, "loss": 2.2179, "num_input_tokens_seen": 20709376, "step": 79 }, { "epoch": 0.019865905140302956, "grad_norm": 0.5547915697097778, "learning_rate": 7.9e-06, "loss": 1.8943, "num_input_tokens_seen": 20971520, "step": 80 }, { "epoch": 0.02011422895455674, "grad_norm": 0.45996469259262085, "learning_rate": 8.000000000000001e-06, "loss": 1.5404, "num_input_tokens_seen": 21233664, "step": 81 }, { "epoch": 0.020362552768810528, "grad_norm": 0.5060412883758545, "learning_rate": 8.1e-06, "loss": 1.7184, "num_input_tokens_seen": 21495808, "step": 82 }, { "epoch": 0.020610876583064316, "grad_norm": 0.5570970773696899, "learning_rate": 8.2e-06, "loss": 2.0638, "num_input_tokens_seen": 21757952, "step": 83 }, { "epoch": 0.020859200397318103, "grad_norm": 0.47680890560150146, "learning_rate": 8.3e-06, "loss": 1.4177, "num_input_tokens_seen": 22020096, "step": 84 }, { "epoch": 0.02110752421157189, "grad_norm": 0.5959429144859314, "learning_rate": 8.400000000000001e-06, "loss": 2.1659, "num_input_tokens_seen": 22282240, "step": 85 }, { "epoch": 0.021355848025825675, "grad_norm": 0.5277214646339417, "learning_rate": 8.5e-06, "loss": 1.8083, "num_input_tokens_seen": 22544384, "step": 86 }, { "epoch": 0.021604171840079463, "grad_norm": 0.5346829891204834, "learning_rate": 8.6e-06, "loss": 1.9323, "num_input_tokens_seen": 22806528, "step": 87 }, { "epoch": 0.02185249565433325, "grad_norm": 0.5096192359924316, "learning_rate": 8.700000000000001e-06, "loss": 1.401, "num_input_tokens_seen": 23068672, "step": 88 }, { "epoch": 0.022100819468587038, "grad_norm": 0.49696552753448486, "learning_rate": 8.8e-06, "loss": 1.4828, "num_input_tokens_seen": 23330816, "step": 89 }, { "epoch": 0.022349143282840826, "grad_norm": 0.5156976580619812, "learning_rate": 8.900000000000001e-06, "loss": 1.9016, "num_input_tokens_seen": 23592960, "step": 90 }, { "epoch": 0.02259746709709461, "grad_norm": 0.4339764714241028, "learning_rate": 9e-06, "loss": 1.3645, "num_input_tokens_seen": 23855104, "step": 91 }, { "epoch": 0.022845790911348397, "grad_norm": 0.5163502097129822, "learning_rate": 9.100000000000001e-06, "loss": 1.9044, "num_input_tokens_seen": 24117248, "step": 92 }, { "epoch": 0.023094114725602185, "grad_norm": 0.44321632385253906, "learning_rate": 9.200000000000002e-06, "loss": 1.7322, "num_input_tokens_seen": 24379392, "step": 93 }, { "epoch": 0.023342438539855973, "grad_norm": 0.5554404258728027, "learning_rate": 9.3e-06, "loss": 2.0536, "num_input_tokens_seen": 24641536, "step": 94 }, { "epoch": 0.02359076235410976, "grad_norm": 0.5148480534553528, "learning_rate": 9.4e-06, "loss": 1.9353, "num_input_tokens_seen": 24903680, "step": 95 }, { "epoch": 0.023839086168363548, "grad_norm": 0.4365761876106262, "learning_rate": 9.5e-06, "loss": 2.1485, "num_input_tokens_seen": 25165824, "step": 96 }, { "epoch": 0.024087409982617332, "grad_norm": 0.6153504252433777, "learning_rate": 9.600000000000001e-06, "loss": 2.3459, "num_input_tokens_seen": 25427968, "step": 97 }, { "epoch": 0.02433573379687112, "grad_norm": 0.7189298272132874, "learning_rate": 9.7e-06, "loss": 1.5111, "num_input_tokens_seen": 25690112, "step": 98 }, { "epoch": 0.024584057611124907, "grad_norm": 0.40012645721435547, "learning_rate": 9.800000000000001e-06, "loss": 2.0235, "num_input_tokens_seen": 25952256, "step": 99 }, { "epoch": 0.024832381425378695, "grad_norm": 1.3065471649169922, "learning_rate": 9.9e-06, "loss": 2.1165, "num_input_tokens_seen": 26214400, "step": 100 }, { "epoch": 0.025080705239632482, "grad_norm": 0.7975507378578186, "learning_rate": 1e-05, "loss": 1.722, "num_input_tokens_seen": 26476544, "step": 101 }, { "epoch": 0.025329029053886266, "grad_norm": 0.5044583678245544, "learning_rate": 9.999969538288953e-06, "loss": 2.0617, "num_input_tokens_seen": 26738688, "step": 102 }, { "epoch": 0.025577352868140054, "grad_norm": 0.6180616021156311, "learning_rate": 9.999878153526974e-06, "loss": 1.8041, "num_input_tokens_seen": 27000832, "step": 103 }, { "epoch": 0.02582567668239384, "grad_norm": 0.3513893485069275, "learning_rate": 9.999725846827562e-06, "loss": 2.0049, "num_input_tokens_seen": 27262976, "step": 104 }, { "epoch": 0.02607400049664763, "grad_norm": 0.5015475153923035, "learning_rate": 9.999512620046523e-06, "loss": 1.9308, "num_input_tokens_seen": 27525120, "step": 105 }, { "epoch": 0.026322324310901417, "grad_norm": 0.35490310192108154, "learning_rate": 9.999238475781957e-06, "loss": 1.7832, "num_input_tokens_seen": 27787264, "step": 106 }, { "epoch": 0.0265706481251552, "grad_norm": 0.4982094168663025, "learning_rate": 9.998903417374228e-06, "loss": 2.1659, "num_input_tokens_seen": 28049408, "step": 107 }, { "epoch": 0.02681897193940899, "grad_norm": 0.3601139783859253, "learning_rate": 9.998507448905917e-06, "loss": 1.8221, "num_input_tokens_seen": 28311552, "step": 108 }, { "epoch": 0.027067295753662776, "grad_norm": 0.47446349263191223, "learning_rate": 9.998050575201772e-06, "loss": 1.4861, "num_input_tokens_seen": 28573696, "step": 109 }, { "epoch": 0.027315619567916564, "grad_norm": 0.4510415196418762, "learning_rate": 9.997532801828659e-06, "loss": 1.8413, "num_input_tokens_seen": 28835840, "step": 110 }, { "epoch": 0.02756394338217035, "grad_norm": 0.34447139501571655, "learning_rate": 9.99695413509548e-06, "loss": 1.9358, "num_input_tokens_seen": 29097984, "step": 111 }, { "epoch": 0.027812267196424136, "grad_norm": 0.7382912039756775, "learning_rate": 9.996314582053106e-06, "loss": 2.0478, "num_input_tokens_seen": 29360128, "step": 112 }, { "epoch": 0.028060591010677923, "grad_norm": 0.4270409345626831, "learning_rate": 9.995614150494293e-06, "loss": 1.4107, "num_input_tokens_seen": 29622272, "step": 113 }, { "epoch": 0.02830891482493171, "grad_norm": 0.6236165761947632, "learning_rate": 9.994852848953574e-06, "loss": 1.9327, "num_input_tokens_seen": 29884416, "step": 114 }, { "epoch": 0.0285572386391855, "grad_norm": 0.44820812344551086, "learning_rate": 9.994030686707171e-06, "loss": 1.6324, "num_input_tokens_seen": 30146560, "step": 115 }, { "epoch": 0.028805562453439286, "grad_norm": 0.34471920132637024, "learning_rate": 9.993147673772869e-06, "loss": 2.0254, "num_input_tokens_seen": 30408704, "step": 116 }, { "epoch": 0.02905388626769307, "grad_norm": 0.4757576584815979, "learning_rate": 9.992203820909906e-06, "loss": 2.1909, "num_input_tokens_seen": 30670848, "step": 117 }, { "epoch": 0.029302210081946858, "grad_norm": 0.5577523112297058, "learning_rate": 9.991199139618828e-06, "loss": 1.7591, "num_input_tokens_seen": 30932992, "step": 118 }, { "epoch": 0.029550533896200645, "grad_norm": 0.4104521572589874, "learning_rate": 9.990133642141359e-06, "loss": 1.8473, "num_input_tokens_seen": 31195136, "step": 119 }, { "epoch": 0.029798857710454433, "grad_norm": 0.3846544921398163, "learning_rate": 9.989007341460251e-06, "loss": 1.786, "num_input_tokens_seen": 31457280, "step": 120 }, { "epoch": 0.03004718152470822, "grad_norm": 0.5335818529129028, "learning_rate": 9.987820251299121e-06, "loss": 2.5372, "num_input_tokens_seen": 31719424, "step": 121 }, { "epoch": 0.030295505338962008, "grad_norm": 0.5046745538711548, "learning_rate": 9.98657238612229e-06, "loss": 2.0604, "num_input_tokens_seen": 31981568, "step": 122 }, { "epoch": 0.030543829153215792, "grad_norm": 0.554063618183136, "learning_rate": 9.985263761134602e-06, "loss": 2.0795, "num_input_tokens_seen": 32243712, "step": 123 }, { "epoch": 0.03079215296746958, "grad_norm": 0.3117339611053467, "learning_rate": 9.983894392281237e-06, "loss": 1.7701, "num_input_tokens_seen": 32505856, "step": 124 }, { "epoch": 0.031040476781723367, "grad_norm": 0.42565709352493286, "learning_rate": 9.982464296247523e-06, "loss": 1.7209, "num_input_tokens_seen": 32768000, "step": 125 }, { "epoch": 0.031288800595977155, "grad_norm": 0.5742325782775879, "learning_rate": 9.980973490458728e-06, "loss": 1.5624, "num_input_tokens_seen": 33030144, "step": 126 }, { "epoch": 0.03153712441023094, "grad_norm": 0.4478677213191986, "learning_rate": 9.979421993079853e-06, "loss": 2.0519, "num_input_tokens_seen": 33292288, "step": 127 }, { "epoch": 0.03178544822448473, "grad_norm": 0.40014106035232544, "learning_rate": 9.9778098230154e-06, "loss": 1.7002, "num_input_tokens_seen": 33554432, "step": 128 }, { "epoch": 0.032033772038738514, "grad_norm": 0.37266790866851807, "learning_rate": 9.976136999909156e-06, "loss": 1.7795, "num_input_tokens_seen": 33816576, "step": 129 }, { "epoch": 0.0322820958529923, "grad_norm": 0.7558661103248596, "learning_rate": 9.974403544143942e-06, "loss": 1.5417, "num_input_tokens_seen": 34078720, "step": 130 }, { "epoch": 0.03253041966724609, "grad_norm": 0.42118939757347107, "learning_rate": 9.972609476841368e-06, "loss": 1.728, "num_input_tokens_seen": 34340864, "step": 131 }, { "epoch": 0.032778743481499874, "grad_norm": 0.45237675309181213, "learning_rate": 9.970754819861577e-06, "loss": 1.651, "num_input_tokens_seen": 34603008, "step": 132 }, { "epoch": 0.033027067295753665, "grad_norm": 0.42204225063323975, "learning_rate": 9.968839595802982e-06, "loss": 1.6179, "num_input_tokens_seen": 34865152, "step": 133 }, { "epoch": 0.03327539111000745, "grad_norm": 0.7443103790283203, "learning_rate": 9.966863828001982e-06, "loss": 2.0339, "num_input_tokens_seen": 35127296, "step": 134 }, { "epoch": 0.03352371492426124, "grad_norm": 0.46790680289268494, "learning_rate": 9.964827540532685e-06, "loss": 1.7902, "num_input_tokens_seen": 35389440, "step": 135 }, { "epoch": 0.033772038738515024, "grad_norm": 0.4609277546405792, "learning_rate": 9.962730758206612e-06, "loss": 1.7105, "num_input_tokens_seen": 35651584, "step": 136 }, { "epoch": 0.03402036255276881, "grad_norm": 0.4367188513278961, "learning_rate": 9.960573506572391e-06, "loss": 1.5516, "num_input_tokens_seen": 35913728, "step": 137 }, { "epoch": 0.0342686863670226, "grad_norm": 0.518988311290741, "learning_rate": 9.958355811915452e-06, "loss": 1.4072, "num_input_tokens_seen": 36175872, "step": 138 }, { "epoch": 0.034517010181276384, "grad_norm": 0.32017022371292114, "learning_rate": 9.95607770125771e-06, "loss": 1.7326, "num_input_tokens_seen": 36438016, "step": 139 }, { "epoch": 0.034765333995530175, "grad_norm": 0.4322156310081482, "learning_rate": 9.953739202357219e-06, "loss": 1.3485, "num_input_tokens_seen": 36700160, "step": 140 }, { "epoch": 0.03501365780978396, "grad_norm": 0.5065650939941406, "learning_rate": 9.951340343707852e-06, "loss": 1.8844, "num_input_tokens_seen": 36962304, "step": 141 }, { "epoch": 0.03526198162403774, "grad_norm": 0.46633732318878174, "learning_rate": 9.948881154538946e-06, "loss": 1.0832, "num_input_tokens_seen": 37224448, "step": 142 }, { "epoch": 0.035510305438291534, "grad_norm": 0.42399951815605164, "learning_rate": 9.946361664814942e-06, "loss": 2.0688, "num_input_tokens_seen": 37486592, "step": 143 }, { "epoch": 0.03575862925254532, "grad_norm": 0.6287701725959778, "learning_rate": 9.94378190523503e-06, "loss": 1.6954, "num_input_tokens_seen": 37748736, "step": 144 }, { "epoch": 0.03600695306679911, "grad_norm": 0.38575854897499084, "learning_rate": 9.941141907232766e-06, "loss": 1.6803, "num_input_tokens_seen": 38010880, "step": 145 }, { "epoch": 0.03625527688105289, "grad_norm": 0.5871158242225647, "learning_rate": 9.938441702975689e-06, "loss": 2.2275, "num_input_tokens_seen": 38273024, "step": 146 }, { "epoch": 0.03650360069530668, "grad_norm": 0.4878631830215454, "learning_rate": 9.93568132536494e-06, "loss": 1.9927, "num_input_tokens_seen": 38535168, "step": 147 }, { "epoch": 0.03675192450956047, "grad_norm": 0.6624669432640076, "learning_rate": 9.932860808034847e-06, "loss": 1.6445, "num_input_tokens_seen": 38797312, "step": 148 }, { "epoch": 0.03700024832381425, "grad_norm": 0.523557722568512, "learning_rate": 9.929980185352525e-06, "loss": 1.9669, "num_input_tokens_seen": 39059456, "step": 149 }, { "epoch": 0.037248572138068044, "grad_norm": 0.3871573805809021, "learning_rate": 9.927039492417452e-06, "loss": 1.57, "num_input_tokens_seen": 39321600, "step": 150 }, { "epoch": 0.03749689595232183, "grad_norm": 0.459677129983902, "learning_rate": 9.924038765061042e-06, "loss": 1.7627, "num_input_tokens_seen": 39583744, "step": 151 }, { "epoch": 0.03774521976657561, "grad_norm": 0.7199493646621704, "learning_rate": 9.92097803984621e-06, "loss": 1.6882, "num_input_tokens_seen": 39845888, "step": 152 }, { "epoch": 0.0379935435808294, "grad_norm": 0.4936988353729248, "learning_rate": 9.91785735406693e-06, "loss": 1.8768, "num_input_tokens_seen": 40108032, "step": 153 }, { "epoch": 0.03824186739508319, "grad_norm": 0.44318658113479614, "learning_rate": 9.914676745747772e-06, "loss": 1.6635, "num_input_tokens_seen": 40370176, "step": 154 }, { "epoch": 0.03849019120933698, "grad_norm": 0.438924103975296, "learning_rate": 9.911436253643445e-06, "loss": 1.6874, "num_input_tokens_seen": 40632320, "step": 155 }, { "epoch": 0.03873851502359076, "grad_norm": 0.37632086873054504, "learning_rate": 9.908135917238321e-06, "loss": 2.1325, "num_input_tokens_seen": 40894464, "step": 156 }, { "epoch": 0.03898683883784455, "grad_norm": 0.4754061698913574, "learning_rate": 9.904775776745959e-06, "loss": 1.6958, "num_input_tokens_seen": 41156608, "step": 157 }, { "epoch": 0.03923516265209834, "grad_norm": 0.629783570766449, "learning_rate": 9.901355873108611e-06, "loss": 1.6854, "num_input_tokens_seen": 41418752, "step": 158 }, { "epoch": 0.03948348646635212, "grad_norm": 0.4472093880176544, "learning_rate": 9.89787624799672e-06, "loss": 1.9023, "num_input_tokens_seen": 41680896, "step": 159 }, { "epoch": 0.03973181028060591, "grad_norm": 0.35216960310935974, "learning_rate": 9.894336943808426e-06, "loss": 1.9459, "num_input_tokens_seen": 41943040, "step": 160 }, { "epoch": 0.0399801340948597, "grad_norm": 0.34789201617240906, "learning_rate": 9.890738003669029e-06, "loss": 2.0254, "num_input_tokens_seen": 42205184, "step": 161 }, { "epoch": 0.04022845790911348, "grad_norm": 0.5008848309516907, "learning_rate": 9.887079471430481e-06, "loss": 1.7256, "num_input_tokens_seen": 42467328, "step": 162 }, { "epoch": 0.04047678172336727, "grad_norm": 0.2768523395061493, "learning_rate": 9.883361391670841e-06, "loss": 1.4073, "num_input_tokens_seen": 42729472, "step": 163 }, { "epoch": 0.040725105537621056, "grad_norm": 0.25057297945022583, "learning_rate": 9.879583809693737e-06, "loss": 1.8378, "num_input_tokens_seen": 42991616, "step": 164 }, { "epoch": 0.04097342935187485, "grad_norm": 0.34032392501831055, "learning_rate": 9.875746771527817e-06, "loss": 2.0236, "num_input_tokens_seen": 43253760, "step": 165 }, { "epoch": 0.04122175316612863, "grad_norm": 0.34935250878334045, "learning_rate": 9.871850323926178e-06, "loss": 1.25, "num_input_tokens_seen": 43515904, "step": 166 }, { "epoch": 0.041470076980382416, "grad_norm": 0.4536452889442444, "learning_rate": 9.867894514365802e-06, "loss": 1.9923, "num_input_tokens_seen": 43778048, "step": 167 }, { "epoch": 0.04171840079463621, "grad_norm": 0.38076460361480713, "learning_rate": 9.863879391046985e-06, "loss": 1.6626, "num_input_tokens_seen": 44040192, "step": 168 }, { "epoch": 0.04196672460888999, "grad_norm": 0.4140561521053314, "learning_rate": 9.859805002892733e-06, "loss": 1.7342, "num_input_tokens_seen": 44302336, "step": 169 }, { "epoch": 0.04221504842314378, "grad_norm": 0.5397564172744751, "learning_rate": 9.85567139954818e-06, "loss": 1.8625, "num_input_tokens_seen": 44564480, "step": 170 }, { "epoch": 0.042463372237397566, "grad_norm": 0.2765776216983795, "learning_rate": 9.851478631379982e-06, "loss": 1.9361, "num_input_tokens_seen": 44826624, "step": 171 }, { "epoch": 0.04271169605165135, "grad_norm": 0.3303300440311432, "learning_rate": 9.847226749475696e-06, "loss": 1.9308, "num_input_tokens_seen": 45088768, "step": 172 }, { "epoch": 0.04296001986590514, "grad_norm": 0.5077419281005859, "learning_rate": 9.842915805643156e-06, "loss": 1.8804, "num_input_tokens_seen": 45350912, "step": 173 }, { "epoch": 0.043208343680158925, "grad_norm": 0.5909125208854675, "learning_rate": 9.838545852409857e-06, "loss": 1.8626, "num_input_tokens_seen": 45613056, "step": 174 }, { "epoch": 0.04345666749441272, "grad_norm": 0.40436893701553345, "learning_rate": 9.834116943022299e-06, "loss": 1.3216, "num_input_tokens_seen": 45875200, "step": 175 }, { "epoch": 0.0437049913086665, "grad_norm": 0.2543845772743225, "learning_rate": 9.829629131445342e-06, "loss": 1.8692, "num_input_tokens_seen": 46137344, "step": 176 }, { "epoch": 0.043953315122920285, "grad_norm": 0.3154657185077667, "learning_rate": 9.825082472361558e-06, "loss": 1.8267, "num_input_tokens_seen": 46399488, "step": 177 }, { "epoch": 0.044201638937174076, "grad_norm": 0.2858025133609772, "learning_rate": 9.82047702117055e-06, "loss": 2.0557, "num_input_tokens_seen": 46661632, "step": 178 }, { "epoch": 0.04444996275142786, "grad_norm": 0.47114500403404236, "learning_rate": 9.815812833988292e-06, "loss": 1.8631, "num_input_tokens_seen": 46923776, "step": 179 }, { "epoch": 0.04469828656568165, "grad_norm": 0.3492976427078247, "learning_rate": 9.811089967646427e-06, "loss": 1.6999, "num_input_tokens_seen": 47185920, "step": 180 }, { "epoch": 0.044946610379935435, "grad_norm": 0.5230903625488281, "learning_rate": 9.806308479691595e-06, "loss": 1.9586, "num_input_tokens_seen": 47448064, "step": 181 }, { "epoch": 0.04519493419418922, "grad_norm": 0.3736109733581543, "learning_rate": 9.801468428384716e-06, "loss": 1.6724, "num_input_tokens_seen": 47710208, "step": 182 }, { "epoch": 0.04544325800844301, "grad_norm": 0.38272184133529663, "learning_rate": 9.796569872700287e-06, "loss": 2.0805, "num_input_tokens_seen": 47972352, "step": 183 }, { "epoch": 0.045691581822696795, "grad_norm": 0.3838115632534027, "learning_rate": 9.791612872325667e-06, "loss": 2.076, "num_input_tokens_seen": 48234496, "step": 184 }, { "epoch": 0.045939905636950586, "grad_norm": 0.39449867606163025, "learning_rate": 9.786597487660336e-06, "loss": 1.9271, "num_input_tokens_seen": 48496640, "step": 185 }, { "epoch": 0.04618822945120437, "grad_norm": 0.5604011416435242, "learning_rate": 9.781523779815178e-06, "loss": 2.0907, "num_input_tokens_seen": 48758784, "step": 186 }, { "epoch": 0.04643655326545816, "grad_norm": 0.756406307220459, "learning_rate": 9.776391810611719e-06, "loss": 1.8081, "num_input_tokens_seen": 49020928, "step": 187 }, { "epoch": 0.046684877079711945, "grad_norm": 0.3678906261920929, "learning_rate": 9.771201642581384e-06, "loss": 1.2397, "num_input_tokens_seen": 49283072, "step": 188 }, { "epoch": 0.04693320089396573, "grad_norm": 0.4872545599937439, "learning_rate": 9.765953338964736e-06, "loss": 1.3951, "num_input_tokens_seen": 49545216, "step": 189 }, { "epoch": 0.04718152470821952, "grad_norm": 0.46098363399505615, "learning_rate": 9.760646963710694e-06, "loss": 2.052, "num_input_tokens_seen": 49807360, "step": 190 }, { "epoch": 0.047429848522473304, "grad_norm": 0.675818681716919, "learning_rate": 9.755282581475769e-06, "loss": 1.7571, "num_input_tokens_seen": 50069504, "step": 191 }, { "epoch": 0.047678172336727095, "grad_norm": 0.6060460209846497, "learning_rate": 9.749860257623262e-06, "loss": 1.854, "num_input_tokens_seen": 50331648, "step": 192 }, { "epoch": 0.04792649615098088, "grad_norm": 0.47795215249061584, "learning_rate": 9.744380058222483e-06, "loss": 1.7057, "num_input_tokens_seen": 50593792, "step": 193 }, { "epoch": 0.048174819965234664, "grad_norm": 0.5348641872406006, "learning_rate": 9.73884205004793e-06, "loss": 1.8599, "num_input_tokens_seen": 50855936, "step": 194 }, { "epoch": 0.048423143779488455, "grad_norm": 0.4470706284046173, "learning_rate": 9.733246300578482e-06, "loss": 1.427, "num_input_tokens_seen": 51118080, "step": 195 }, { "epoch": 0.04867146759374224, "grad_norm": 0.35019198060035706, "learning_rate": 9.727592877996585e-06, "loss": 1.6146, "num_input_tokens_seen": 51380224, "step": 196 }, { "epoch": 0.04891979140799603, "grad_norm": 0.4542747139930725, "learning_rate": 9.721881851187406e-06, "loss": 1.8169, "num_input_tokens_seen": 51642368, "step": 197 }, { "epoch": 0.049168115222249814, "grad_norm": 0.5041635632514954, "learning_rate": 9.716113289738005e-06, "loss": 1.7434, "num_input_tokens_seen": 51904512, "step": 198 }, { "epoch": 0.0494164390365036, "grad_norm": 0.39864203333854675, "learning_rate": 9.710287263936485e-06, "loss": 1.5456, "num_input_tokens_seen": 52166656, "step": 199 }, { "epoch": 0.04966476285075739, "grad_norm": 0.4882585108280182, "learning_rate": 9.704403844771128e-06, "loss": 1.5535, "num_input_tokens_seen": 52428800, "step": 200 }, { "epoch": 0.049913086665011173, "grad_norm": 0.33271995186805725, "learning_rate": 9.698463103929542e-06, "loss": 1.7225, "num_input_tokens_seen": 52690944, "step": 201 }, { "epoch": 0.050161410479264965, "grad_norm": 0.5818848609924316, "learning_rate": 9.69246511379778e-06, "loss": 1.8262, "num_input_tokens_seen": 52953088, "step": 202 }, { "epoch": 0.05040973429351875, "grad_norm": 0.3519175946712494, "learning_rate": 9.68640994745946e-06, "loss": 1.8536, "num_input_tokens_seen": 53215232, "step": 203 }, { "epoch": 0.05065805810777253, "grad_norm": 0.30881020426750183, "learning_rate": 9.680297678694867e-06, "loss": 1.7819, "num_input_tokens_seen": 53477376, "step": 204 }, { "epoch": 0.050906381922026324, "grad_norm": 0.28387659788131714, "learning_rate": 9.674128381980073e-06, "loss": 1.6913, "num_input_tokens_seen": 53739520, "step": 205 }, { "epoch": 0.05115470573628011, "grad_norm": 0.42218366265296936, "learning_rate": 9.667902132486009e-06, "loss": 1.3839, "num_input_tokens_seen": 54001664, "step": 206 }, { "epoch": 0.0514030295505339, "grad_norm": 0.3479291796684265, "learning_rate": 9.661619006077562e-06, "loss": 1.7449, "num_input_tokens_seen": 54263808, "step": 207 }, { "epoch": 0.05165135336478768, "grad_norm": 0.41011831164360046, "learning_rate": 9.655279079312643e-06, "loss": 1.3836, "num_input_tokens_seen": 54525952, "step": 208 }, { "epoch": 0.05189967717904147, "grad_norm": 0.3655335605144501, "learning_rate": 9.648882429441258e-06, "loss": 2.2184, "num_input_tokens_seen": 54788096, "step": 209 }, { "epoch": 0.05214800099329526, "grad_norm": 0.4303296208381653, "learning_rate": 9.642429134404568e-06, "loss": 1.8072, "num_input_tokens_seen": 55050240, "step": 210 }, { "epoch": 0.05239632480754904, "grad_norm": 0.4718409776687622, "learning_rate": 9.635919272833938e-06, "loss": 1.7503, "num_input_tokens_seen": 55312384, "step": 211 }, { "epoch": 0.052644648621802834, "grad_norm": 0.41102349758148193, "learning_rate": 9.629352924049975e-06, "loss": 1.5696, "num_input_tokens_seen": 55574528, "step": 212 }, { "epoch": 0.05289297243605662, "grad_norm": 0.40920042991638184, "learning_rate": 9.622730168061568e-06, "loss": 1.354, "num_input_tokens_seen": 55836672, "step": 213 }, { "epoch": 0.0531412962503104, "grad_norm": 0.2790515124797821, "learning_rate": 9.616051085564905e-06, "loss": 1.6552, "num_input_tokens_seen": 56098816, "step": 214 }, { "epoch": 0.05338962006456419, "grad_norm": 0.708991289138794, "learning_rate": 9.609315757942504e-06, "loss": 1.4587, "num_input_tokens_seen": 56360960, "step": 215 }, { "epoch": 0.05363794387881798, "grad_norm": 0.5307122468948364, "learning_rate": 9.602524267262202e-06, "loss": 2.0549, "num_input_tokens_seen": 56623104, "step": 216 }, { "epoch": 0.05388626769307177, "grad_norm": 0.6132609248161316, "learning_rate": 9.595676696276173e-06, "loss": 1.8739, "num_input_tokens_seen": 56885248, "step": 217 }, { "epoch": 0.05413459150732555, "grad_norm": 0.4902276396751404, "learning_rate": 9.588773128419907e-06, "loss": 2.0105, "num_input_tokens_seen": 57147392, "step": 218 }, { "epoch": 0.054382915321579336, "grad_norm": 0.40454724431037903, "learning_rate": 9.581813647811199e-06, "loss": 1.9691, "num_input_tokens_seen": 57409536, "step": 219 }, { "epoch": 0.05463123913583313, "grad_norm": 0.5375818610191345, "learning_rate": 9.574798339249124e-06, "loss": 1.9857, "num_input_tokens_seen": 57671680, "step": 220 }, { "epoch": 0.05487956295008691, "grad_norm": 0.31554529070854187, "learning_rate": 9.567727288213005e-06, "loss": 1.7783, "num_input_tokens_seen": 57933824, "step": 221 }, { "epoch": 0.0551278867643407, "grad_norm": 0.5682997703552246, "learning_rate": 9.560600580861366e-06, "loss": 1.8144, "num_input_tokens_seen": 58195968, "step": 222 }, { "epoch": 0.05537621057859449, "grad_norm": 0.393583208322525, "learning_rate": 9.553418304030886e-06, "loss": 1.9838, "num_input_tokens_seen": 58458112, "step": 223 }, { "epoch": 0.05562453439284827, "grad_norm": 0.42172202467918396, "learning_rate": 9.546180545235344e-06, "loss": 1.7532, "num_input_tokens_seen": 58720256, "step": 224 }, { "epoch": 0.05587285820710206, "grad_norm": 0.4001232981681824, "learning_rate": 9.538887392664544e-06, "loss": 2.2716, "num_input_tokens_seen": 58982400, "step": 225 }, { "epoch": 0.056121182021355846, "grad_norm": 0.2886028289794922, "learning_rate": 9.531538935183252e-06, "loss": 1.5989, "num_input_tokens_seen": 59244544, "step": 226 }, { "epoch": 0.05636950583560964, "grad_norm": 0.49188050627708435, "learning_rate": 9.524135262330098e-06, "loss": 1.7318, "num_input_tokens_seen": 59506688, "step": 227 }, { "epoch": 0.05661782964986342, "grad_norm": 0.46622851490974426, "learning_rate": 9.516676464316505e-06, "loss": 1.4116, "num_input_tokens_seen": 59768832, "step": 228 }, { "epoch": 0.056866153464117206, "grad_norm": 0.26139262318611145, "learning_rate": 9.50916263202557e-06, "loss": 1.6799, "num_input_tokens_seen": 60030976, "step": 229 }, { "epoch": 0.057114477278371, "grad_norm": 0.4152681827545166, "learning_rate": 9.501593857010968e-06, "loss": 2.0024, "num_input_tokens_seen": 60293120, "step": 230 }, { "epoch": 0.05736280109262478, "grad_norm": 0.683701753616333, "learning_rate": 9.493970231495836e-06, "loss": 1.7395, "num_input_tokens_seen": 60555264, "step": 231 }, { "epoch": 0.05761112490687857, "grad_norm": 0.4206884503364563, "learning_rate": 9.486291848371642e-06, "loss": 1.4044, "num_input_tokens_seen": 60817408, "step": 232 }, { "epoch": 0.057859448721132356, "grad_norm": 0.3584212064743042, "learning_rate": 9.478558801197065e-06, "loss": 1.6407, "num_input_tokens_seen": 61079552, "step": 233 }, { "epoch": 0.05810777253538614, "grad_norm": 0.3988575041294098, "learning_rate": 9.470771184196842e-06, "loss": 1.8871, "num_input_tokens_seen": 61341696, "step": 234 }, { "epoch": 0.05835609634963993, "grad_norm": 0.6429465413093567, "learning_rate": 9.46292909226063e-06, "loss": 1.7096, "num_input_tokens_seen": 61603840, "step": 235 }, { "epoch": 0.058604420163893715, "grad_norm": 0.5112093091011047, "learning_rate": 9.45503262094184e-06, "loss": 1.878, "num_input_tokens_seen": 61865984, "step": 236 }, { "epoch": 0.058852743978147506, "grad_norm": 0.6004856824874878, "learning_rate": 9.44708186645649e-06, "loss": 1.73, "num_input_tokens_seen": 62128128, "step": 237 }, { "epoch": 0.05910106779240129, "grad_norm": 0.5572634339332581, "learning_rate": 9.439076925682006e-06, "loss": 1.601, "num_input_tokens_seen": 62390272, "step": 238 }, { "epoch": 0.05934939160665508, "grad_norm": 0.7834048271179199, "learning_rate": 9.431017896156074e-06, "loss": 1.8971, "num_input_tokens_seen": 62652416, "step": 239 }, { "epoch": 0.059597715420908866, "grad_norm": 0.5589991807937622, "learning_rate": 9.42290487607542e-06, "loss": 1.5294, "num_input_tokens_seen": 62914560, "step": 240 }, { "epoch": 0.05984603923516265, "grad_norm": 0.48503783345222473, "learning_rate": 9.414737964294636e-06, "loss": 1.5151, "num_input_tokens_seen": 63176704, "step": 241 }, { "epoch": 0.06009436304941644, "grad_norm": 0.39491578936576843, "learning_rate": 9.406517260324962e-06, "loss": 1.272, "num_input_tokens_seen": 63438848, "step": 242 }, { "epoch": 0.060342686863670225, "grad_norm": 0.5852888822555542, "learning_rate": 9.398242864333084e-06, "loss": 1.6314, "num_input_tokens_seen": 63700992, "step": 243 }, { "epoch": 0.060591010677924016, "grad_norm": 0.4342246949672699, "learning_rate": 9.389914877139903e-06, "loss": 1.8619, "num_input_tokens_seen": 63963136, "step": 244 }, { "epoch": 0.0608393344921778, "grad_norm": 0.45139801502227783, "learning_rate": 9.381533400219319e-06, "loss": 2.0767, "num_input_tokens_seen": 64225280, "step": 245 }, { "epoch": 0.061087658306431585, "grad_norm": 0.5253079533576965, "learning_rate": 9.37309853569698e-06, "loss": 1.7857, "num_input_tokens_seen": 64487424, "step": 246 }, { "epoch": 0.061335982120685376, "grad_norm": 0.5168160796165466, "learning_rate": 9.364610386349048e-06, "loss": 1.7157, "num_input_tokens_seen": 64749568, "step": 247 }, { "epoch": 0.06158430593493916, "grad_norm": 0.6876463890075684, "learning_rate": 9.356069055600949e-06, "loss": 1.9186, "num_input_tokens_seen": 65011712, "step": 248 }, { "epoch": 0.06183262974919295, "grad_norm": 0.5097318887710571, "learning_rate": 9.347474647526095e-06, "loss": 1.806, "num_input_tokens_seen": 65273856, "step": 249 }, { "epoch": 0.062080953563446735, "grad_norm": 0.4346969425678253, "learning_rate": 9.338827266844643e-06, "loss": 1.9837, "num_input_tokens_seen": 65536000, "step": 250 }, { "epoch": 0.06232927737770052, "grad_norm": 0.4577252268791199, "learning_rate": 9.330127018922195e-06, "loss": 1.9294, "num_input_tokens_seen": 65798144, "step": 251 }, { "epoch": 0.06257760119195431, "grad_norm": 0.5622639656066895, "learning_rate": 9.321374009768525e-06, "loss": 1.2883, "num_input_tokens_seen": 66060288, "step": 252 }, { "epoch": 0.0628259250062081, "grad_norm": 0.553583562374115, "learning_rate": 9.312568346036288e-06, "loss": 1.7474, "num_input_tokens_seen": 66322432, "step": 253 }, { "epoch": 0.06307424882046188, "grad_norm": 0.6473803520202637, "learning_rate": 9.30371013501972e-06, "loss": 1.7102, "num_input_tokens_seen": 66584576, "step": 254 }, { "epoch": 0.06332257263471566, "grad_norm": 0.6553588509559631, "learning_rate": 9.294799484653323e-06, "loss": 1.4839, "num_input_tokens_seen": 66846720, "step": 255 }, { "epoch": 0.06357089644896946, "grad_norm": 0.5461873412132263, "learning_rate": 9.285836503510562e-06, "loss": 2.0798, "num_input_tokens_seen": 67108864, "step": 256 }, { "epoch": 0.06381922026322324, "grad_norm": 0.5196068286895752, "learning_rate": 9.276821300802535e-06, "loss": 1.7274, "num_input_tokens_seen": 67371008, "step": 257 }, { "epoch": 0.06406754407747703, "grad_norm": 0.2947503924369812, "learning_rate": 9.267753986376638e-06, "loss": 1.8269, "num_input_tokens_seen": 67633152, "step": 258 }, { "epoch": 0.06431586789173081, "grad_norm": 0.41323122382164, "learning_rate": 9.25863467071524e-06, "loss": 1.3102, "num_input_tokens_seen": 67895296, "step": 259 }, { "epoch": 0.0645641917059846, "grad_norm": 0.5422347187995911, "learning_rate": 9.24946346493432e-06, "loss": 1.7149, "num_input_tokens_seen": 68157440, "step": 260 }, { "epoch": 0.0648125155202384, "grad_norm": 0.4071256220340729, "learning_rate": 9.24024048078213e-06, "loss": 1.8354, "num_input_tokens_seen": 68419584, "step": 261 }, { "epoch": 0.06506083933449218, "grad_norm": 0.31718364357948303, "learning_rate": 9.230965830637821e-06, "loss": 1.6514, "num_input_tokens_seen": 68681728, "step": 262 }, { "epoch": 0.06530916314874596, "grad_norm": 0.5854095816612244, "learning_rate": 9.221639627510076e-06, "loss": 1.1938, "num_input_tokens_seen": 68943872, "step": 263 }, { "epoch": 0.06555748696299975, "grad_norm": 0.28129681944847107, "learning_rate": 9.21226198503574e-06, "loss": 1.932, "num_input_tokens_seen": 69206016, "step": 264 }, { "epoch": 0.06580581077725355, "grad_norm": 0.39890119433403015, "learning_rate": 9.202833017478421e-06, "loss": 2.1795, "num_input_tokens_seen": 69468160, "step": 265 }, { "epoch": 0.06605413459150733, "grad_norm": 0.5349065661430359, "learning_rate": 9.193352839727122e-06, "loss": 1.7356, "num_input_tokens_seen": 69730304, "step": 266 }, { "epoch": 0.06630245840576111, "grad_norm": 0.5061510801315308, "learning_rate": 9.18382156729481e-06, "loss": 1.552, "num_input_tokens_seen": 69992448, "step": 267 }, { "epoch": 0.0665507822200149, "grad_norm": 0.5302243828773499, "learning_rate": 9.174239316317034e-06, "loss": 1.6972, "num_input_tokens_seen": 70254592, "step": 268 }, { "epoch": 0.06679910603426868, "grad_norm": 0.4849538803100586, "learning_rate": 9.164606203550498e-06, "loss": 1.8502, "num_input_tokens_seen": 70516736, "step": 269 }, { "epoch": 0.06704742984852248, "grad_norm": 0.7993111610412598, "learning_rate": 9.154922346371641e-06, "loss": 1.473, "num_input_tokens_seen": 70778880, "step": 270 }, { "epoch": 0.06729575366277626, "grad_norm": 0.2985432744026184, "learning_rate": 9.145187862775208e-06, "loss": 1.7854, "num_input_tokens_seen": 71041024, "step": 271 }, { "epoch": 0.06754407747703005, "grad_norm": 0.557628333568573, "learning_rate": 9.13540287137281e-06, "loss": 1.6575, "num_input_tokens_seen": 71303168, "step": 272 }, { "epoch": 0.06779240129128383, "grad_norm": 0.5635995268821716, "learning_rate": 9.125567491391476e-06, "loss": 1.7496, "num_input_tokens_seen": 71565312, "step": 273 }, { "epoch": 0.06804072510553762, "grad_norm": 0.5241788625717163, "learning_rate": 9.115681842672211e-06, "loss": 1.4799, "num_input_tokens_seen": 71827456, "step": 274 }, { "epoch": 0.06828904891979141, "grad_norm": 0.7073025107383728, "learning_rate": 9.10574604566852e-06, "loss": 2.0472, "num_input_tokens_seen": 72089600, "step": 275 }, { "epoch": 0.0685373727340452, "grad_norm": 0.3968837261199951, "learning_rate": 9.09576022144496e-06, "loss": 1.2316, "num_input_tokens_seen": 72351744, "step": 276 }, { "epoch": 0.06878569654829898, "grad_norm": 0.42373400926589966, "learning_rate": 9.085724491675642e-06, "loss": 1.3935, "num_input_tokens_seen": 72613888, "step": 277 }, { "epoch": 0.06903402036255277, "grad_norm": 0.3875352144241333, "learning_rate": 9.07563897864277e-06, "loss": 1.4776, "num_input_tokens_seen": 72876032, "step": 278 }, { "epoch": 0.06928234417680655, "grad_norm": 0.64447420835495, "learning_rate": 9.065503805235139e-06, "loss": 1.5065, "num_input_tokens_seen": 73138176, "step": 279 }, { "epoch": 0.06953066799106035, "grad_norm": 0.6507744789123535, "learning_rate": 9.055319094946633e-06, "loss": 1.7043, "num_input_tokens_seen": 73400320, "step": 280 }, { "epoch": 0.06977899180531413, "grad_norm": 0.6041735410690308, "learning_rate": 9.045084971874738e-06, "loss": 1.4371, "num_input_tokens_seen": 73662464, "step": 281 }, { "epoch": 0.07002731561956792, "grad_norm": 0.4756613075733185, "learning_rate": 9.03480156071901e-06, "loss": 1.5451, "num_input_tokens_seen": 73924608, "step": 282 }, { "epoch": 0.0702756394338217, "grad_norm": 0.6942863464355469, "learning_rate": 9.02446898677957e-06, "loss": 1.1622, "num_input_tokens_seen": 74186752, "step": 283 }, { "epoch": 0.07052396324807549, "grad_norm": 0.5789624452590942, "learning_rate": 9.014087375955574e-06, "loss": 2.2199, "num_input_tokens_seen": 74448896, "step": 284 }, { "epoch": 0.07077228706232928, "grad_norm": 0.5487517714500427, "learning_rate": 9.003656854743667e-06, "loss": 1.5385, "num_input_tokens_seen": 74711040, "step": 285 }, { "epoch": 0.07102061087658307, "grad_norm": 0.34135740995407104, "learning_rate": 8.993177550236464e-06, "loss": 1.5421, "num_input_tokens_seen": 74973184, "step": 286 }, { "epoch": 0.07126893469083685, "grad_norm": 0.4900202453136444, "learning_rate": 8.982649590120982e-06, "loss": 1.7945, "num_input_tokens_seen": 75235328, "step": 287 }, { "epoch": 0.07151725850509064, "grad_norm": 0.44072678685188293, "learning_rate": 8.972073102677091e-06, "loss": 1.8011, "num_input_tokens_seen": 75497472, "step": 288 }, { "epoch": 0.07176558231934442, "grad_norm": 0.4883013069629669, "learning_rate": 8.961448216775955e-06, "loss": 1.8165, "num_input_tokens_seen": 75759616, "step": 289 }, { "epoch": 0.07201390613359822, "grad_norm": 0.5138419270515442, "learning_rate": 8.950775061878453e-06, "loss": 1.5481, "num_input_tokens_seen": 76021760, "step": 290 }, { "epoch": 0.072262229947852, "grad_norm": 0.8689912557601929, "learning_rate": 8.94005376803361e-06, "loss": 1.6688, "num_input_tokens_seen": 76283904, "step": 291 }, { "epoch": 0.07251055376210579, "grad_norm": 0.6632862091064453, "learning_rate": 8.92928446587701e-06, "loss": 1.6316, "num_input_tokens_seen": 76546048, "step": 292 }, { "epoch": 0.07275887757635957, "grad_norm": 0.5113462209701538, "learning_rate": 8.9184672866292e-06, "loss": 1.8658, "num_input_tokens_seen": 76808192, "step": 293 }, { "epoch": 0.07300720139061335, "grad_norm": 0.6166042685508728, "learning_rate": 8.907602362094094e-06, "loss": 1.2848, "num_input_tokens_seen": 77070336, "step": 294 }, { "epoch": 0.07325552520486715, "grad_norm": 0.6267765164375305, "learning_rate": 8.896689824657371e-06, "loss": 1.5453, "num_input_tokens_seen": 77332480, "step": 295 }, { "epoch": 0.07350384901912094, "grad_norm": 0.5441874265670776, "learning_rate": 8.885729807284855e-06, "loss": 1.5841, "num_input_tokens_seen": 77594624, "step": 296 }, { "epoch": 0.07375217283337472, "grad_norm": 0.5888817310333252, "learning_rate": 8.874722443520898e-06, "loss": 1.9301, "num_input_tokens_seen": 77856768, "step": 297 }, { "epoch": 0.0740004966476285, "grad_norm": 0.8824830055236816, "learning_rate": 8.863667867486756e-06, "loss": 1.7341, "num_input_tokens_seen": 78118912, "step": 298 }, { "epoch": 0.07424882046188229, "grad_norm": 0.4109712541103363, "learning_rate": 8.852566213878947e-06, "loss": 1.7768, "num_input_tokens_seen": 78381056, "step": 299 }, { "epoch": 0.07449714427613609, "grad_norm": 0.4477173686027527, "learning_rate": 8.841417617967618e-06, "loss": 1.3926, "num_input_tokens_seen": 78643200, "step": 300 }, { "epoch": 0.07474546809038987, "grad_norm": 0.7767991423606873, "learning_rate": 8.83022221559489e-06, "loss": 1.3578, "num_input_tokens_seen": 78905344, "step": 301 }, { "epoch": 0.07499379190464366, "grad_norm": 0.5238236784934998, "learning_rate": 8.818980143173212e-06, "loss": 1.7547, "num_input_tokens_seen": 79167488, "step": 302 }, { "epoch": 0.07524211571889744, "grad_norm": 0.4706284999847412, "learning_rate": 8.807691537683685e-06, "loss": 2.224, "num_input_tokens_seen": 79429632, "step": 303 }, { "epoch": 0.07549043953315122, "grad_norm": 0.4068174362182617, "learning_rate": 8.796356536674404e-06, "loss": 1.8971, "num_input_tokens_seen": 79691776, "step": 304 }, { "epoch": 0.07573876334740502, "grad_norm": 0.36882147192955017, "learning_rate": 8.784975278258783e-06, "loss": 1.3618, "num_input_tokens_seen": 79953920, "step": 305 }, { "epoch": 0.0759870871616588, "grad_norm": 0.7638546228408813, "learning_rate": 8.773547901113862e-06, "loss": 1.4798, "num_input_tokens_seen": 80216064, "step": 306 }, { "epoch": 0.07623541097591259, "grad_norm": 0.7291756868362427, "learning_rate": 8.762074544478622e-06, "loss": 1.8117, "num_input_tokens_seen": 80478208, "step": 307 }, { "epoch": 0.07648373479016637, "grad_norm": 0.35935330390930176, "learning_rate": 8.750555348152299e-06, "loss": 1.7963, "num_input_tokens_seen": 80740352, "step": 308 }, { "epoch": 0.07673205860442016, "grad_norm": 0.41468292474746704, "learning_rate": 8.73899045249266e-06, "loss": 1.7334, "num_input_tokens_seen": 81002496, "step": 309 }, { "epoch": 0.07698038241867396, "grad_norm": 0.47698289155960083, "learning_rate": 8.727379998414311e-06, "loss": 1.5163, "num_input_tokens_seen": 81264640, "step": 310 }, { "epoch": 0.07722870623292774, "grad_norm": 0.3247212767601013, "learning_rate": 8.715724127386971e-06, "loss": 1.5166, "num_input_tokens_seen": 81526784, "step": 311 }, { "epoch": 0.07747703004718152, "grad_norm": 0.24329149723052979, "learning_rate": 8.70402298143375e-06, "loss": 1.1361, "num_input_tokens_seen": 81788928, "step": 312 }, { "epoch": 0.07772535386143531, "grad_norm": 0.4041454493999481, "learning_rate": 8.692276703129421e-06, "loss": 1.6928, "num_input_tokens_seen": 82051072, "step": 313 }, { "epoch": 0.0779736776756891, "grad_norm": 0.5690919160842896, "learning_rate": 8.680485435598674e-06, "loss": 1.4934, "num_input_tokens_seen": 82313216, "step": 314 }, { "epoch": 0.07822200148994289, "grad_norm": 0.5649420022964478, "learning_rate": 8.668649322514382e-06, "loss": 1.4087, "num_input_tokens_seen": 82575360, "step": 315 }, { "epoch": 0.07847032530419668, "grad_norm": 0.38145026564598083, "learning_rate": 8.656768508095853e-06, "loss": 1.9073, "num_input_tokens_seen": 82837504, "step": 316 }, { "epoch": 0.07871864911845046, "grad_norm": 0.7416847944259644, "learning_rate": 8.644843137107058e-06, "loss": 1.4074, "num_input_tokens_seen": 83099648, "step": 317 }, { "epoch": 0.07896697293270424, "grad_norm": 0.5791590213775635, "learning_rate": 8.632873354854881e-06, "loss": 1.8003, "num_input_tokens_seen": 83361792, "step": 318 }, { "epoch": 0.07921529674695803, "grad_norm": 0.5541636347770691, "learning_rate": 8.620859307187339e-06, "loss": 1.4524, "num_input_tokens_seen": 83623936, "step": 319 }, { "epoch": 0.07946362056121183, "grad_norm": 0.45152053236961365, "learning_rate": 8.608801140491811e-06, "loss": 1.9212, "num_input_tokens_seen": 83886080, "step": 320 }, { "epoch": 0.07971194437546561, "grad_norm": 0.7486585974693298, "learning_rate": 8.596699001693257e-06, "loss": 1.8341, "num_input_tokens_seen": 84148224, "step": 321 }, { "epoch": 0.0799602681897194, "grad_norm": 0.4018547832965851, "learning_rate": 8.584553038252415e-06, "loss": 1.6084, "num_input_tokens_seen": 84410368, "step": 322 }, { "epoch": 0.08020859200397318, "grad_norm": 0.5464489459991455, "learning_rate": 8.572363398164017e-06, "loss": 1.5227, "num_input_tokens_seen": 84672512, "step": 323 }, { "epoch": 0.08045691581822696, "grad_norm": 0.5198168754577637, "learning_rate": 8.560130229954985e-06, "loss": 1.6922, "num_input_tokens_seen": 84934656, "step": 324 }, { "epoch": 0.08070523963248076, "grad_norm": 0.48899322748184204, "learning_rate": 8.547853682682605e-06, "loss": 1.6412, "num_input_tokens_seen": 85196800, "step": 325 }, { "epoch": 0.08095356344673454, "grad_norm": 0.2365017831325531, "learning_rate": 8.535533905932739e-06, "loss": 1.5639, "num_input_tokens_seen": 85458944, "step": 326 }, { "epoch": 0.08120188726098833, "grad_norm": 0.4440731704235077, "learning_rate": 8.523171049817974e-06, "loss": 2.0067, "num_input_tokens_seen": 85721088, "step": 327 }, { "epoch": 0.08145021107524211, "grad_norm": 0.591474175453186, "learning_rate": 8.510765264975813e-06, "loss": 2.085, "num_input_tokens_seen": 85983232, "step": 328 }, { "epoch": 0.0816985348894959, "grad_norm": 0.8983985781669617, "learning_rate": 8.498316702566828e-06, "loss": 2.0327, "num_input_tokens_seen": 86245376, "step": 329 }, { "epoch": 0.0819468587037497, "grad_norm": 0.5473107695579529, "learning_rate": 8.485825514272824e-06, "loss": 1.3908, "num_input_tokens_seen": 86507520, "step": 330 }, { "epoch": 0.08219518251800348, "grad_norm": 0.4440311789512634, "learning_rate": 8.473291852294986e-06, "loss": 1.7376, "num_input_tokens_seen": 86769664, "step": 331 }, { "epoch": 0.08244350633225726, "grad_norm": 0.5826847553253174, "learning_rate": 8.460715869352035e-06, "loss": 1.7982, "num_input_tokens_seen": 87031808, "step": 332 }, { "epoch": 0.08269183014651105, "grad_norm": 0.4117896258831024, "learning_rate": 8.44809771867835e-06, "loss": 1.9127, "num_input_tokens_seen": 87293952, "step": 333 }, { "epoch": 0.08294015396076483, "grad_norm": 0.6492828726768494, "learning_rate": 8.435437554022116e-06, "loss": 1.4407, "num_input_tokens_seen": 87556096, "step": 334 }, { "epoch": 0.08318847777501863, "grad_norm": 0.47931548953056335, "learning_rate": 8.422735529643445e-06, "loss": 1.7312, "num_input_tokens_seen": 87818240, "step": 335 }, { "epoch": 0.08343680158927241, "grad_norm": 0.45937687158584595, "learning_rate": 8.409991800312493e-06, "loss": 1.3507, "num_input_tokens_seen": 88080384, "step": 336 }, { "epoch": 0.0836851254035262, "grad_norm": 0.34280064702033997, "learning_rate": 8.397206521307584e-06, "loss": 1.9152, "num_input_tokens_seen": 88342528, "step": 337 }, { "epoch": 0.08393344921777998, "grad_norm": 0.525225043296814, "learning_rate": 8.384379848413304e-06, "loss": 1.7714, "num_input_tokens_seen": 88604672, "step": 338 }, { "epoch": 0.08418177303203377, "grad_norm": 0.8012213706970215, "learning_rate": 8.371511937918616e-06, "loss": 1.6678, "num_input_tokens_seen": 88866816, "step": 339 }, { "epoch": 0.08443009684628756, "grad_norm": 0.49616944789886475, "learning_rate": 8.358602946614952e-06, "loss": 1.6751, "num_input_tokens_seen": 89128960, "step": 340 }, { "epoch": 0.08467842066054135, "grad_norm": 0.4202152490615845, "learning_rate": 8.345653031794292e-06, "loss": 1.595, "num_input_tokens_seen": 89391104, "step": 341 }, { "epoch": 0.08492674447479513, "grad_norm": 0.35994473099708557, "learning_rate": 8.332662351247262e-06, "loss": 1.8082, "num_input_tokens_seen": 89653248, "step": 342 }, { "epoch": 0.08517506828904892, "grad_norm": 0.3768995702266693, "learning_rate": 8.319631063261209e-06, "loss": 1.7495, "num_input_tokens_seen": 89915392, "step": 343 }, { "epoch": 0.0854233921033027, "grad_norm": 0.7302326560020447, "learning_rate": 8.30655932661826e-06, "loss": 1.335, "num_input_tokens_seen": 90177536, "step": 344 }, { "epoch": 0.0856717159175565, "grad_norm": 0.4386504888534546, "learning_rate": 8.293447300593402e-06, "loss": 1.7737, "num_input_tokens_seen": 90439680, "step": 345 }, { "epoch": 0.08592003973181028, "grad_norm": 0.7695857286453247, "learning_rate": 8.280295144952537e-06, "loss": 1.7856, "num_input_tokens_seen": 90701824, "step": 346 }, { "epoch": 0.08616836354606407, "grad_norm": 0.7740248441696167, "learning_rate": 8.267103019950529e-06, "loss": 2.0197, "num_input_tokens_seen": 90963968, "step": 347 }, { "epoch": 0.08641668736031785, "grad_norm": 3.4425835609436035, "learning_rate": 8.253871086329255e-06, "loss": 1.7337, "num_input_tokens_seen": 91226112, "step": 348 }, { "epoch": 0.08666501117457164, "grad_norm": 0.5034295916557312, "learning_rate": 8.240599505315656e-06, "loss": 1.269, "num_input_tokens_seen": 91488256, "step": 349 }, { "epoch": 0.08691333498882543, "grad_norm": 0.3939118981361389, "learning_rate": 8.227288438619754e-06, "loss": 1.7436, "num_input_tokens_seen": 91750400, "step": 350 }, { "epoch": 0.08716165880307922, "grad_norm": 0.6885347962379456, "learning_rate": 8.213938048432697e-06, "loss": 1.5946, "num_input_tokens_seen": 92012544, "step": 351 }, { "epoch": 0.087409982617333, "grad_norm": 0.4886651337146759, "learning_rate": 8.200548497424779e-06, "loss": 1.3196, "num_input_tokens_seen": 92274688, "step": 352 }, { "epoch": 0.08765830643158679, "grad_norm": 0.7113179564476013, "learning_rate": 8.18711994874345e-06, "loss": 2.1461, "num_input_tokens_seen": 92536832, "step": 353 }, { "epoch": 0.08790663024584057, "grad_norm": 0.8152286410331726, "learning_rate": 8.173652566011339e-06, "loss": 1.6922, "num_input_tokens_seen": 92798976, "step": 354 }, { "epoch": 0.08815495406009437, "grad_norm": 0.4894461929798126, "learning_rate": 8.160146513324256e-06, "loss": 1.8319, "num_input_tokens_seen": 93061120, "step": 355 }, { "epoch": 0.08840327787434815, "grad_norm": 0.5153379440307617, "learning_rate": 8.146601955249187e-06, "loss": 1.6654, "num_input_tokens_seen": 93323264, "step": 356 }, { "epoch": 0.08865160168860194, "grad_norm": 0.5455003380775452, "learning_rate": 8.133019056822303e-06, "loss": 1.8757, "num_input_tokens_seen": 93585408, "step": 357 }, { "epoch": 0.08889992550285572, "grad_norm": 0.6758735775947571, "learning_rate": 8.119397983546932e-06, "loss": 1.6944, "num_input_tokens_seen": 93847552, "step": 358 }, { "epoch": 0.0891482493171095, "grad_norm": 0.8125549554824829, "learning_rate": 8.105738901391553e-06, "loss": 1.4809, "num_input_tokens_seen": 94109696, "step": 359 }, { "epoch": 0.0893965731313633, "grad_norm": 0.5866984724998474, "learning_rate": 8.092041976787772e-06, "loss": 1.9641, "num_input_tokens_seen": 94371840, "step": 360 }, { "epoch": 0.08964489694561709, "grad_norm": 0.43463000655174255, "learning_rate": 8.078307376628292e-06, "loss": 1.5385, "num_input_tokens_seen": 94633984, "step": 361 }, { "epoch": 0.08989322075987087, "grad_norm": 0.7243896126747131, "learning_rate": 8.064535268264883e-06, "loss": 1.9196, "num_input_tokens_seen": 94896128, "step": 362 }, { "epoch": 0.09014154457412465, "grad_norm": 0.7336921095848083, "learning_rate": 8.05072581950634e-06, "loss": 1.6399, "num_input_tokens_seen": 95158272, "step": 363 }, { "epoch": 0.09038986838837844, "grad_norm": 0.4355528950691223, "learning_rate": 8.036879198616434e-06, "loss": 1.3215, "num_input_tokens_seen": 95420416, "step": 364 }, { "epoch": 0.09063819220263224, "grad_norm": 0.4627648591995239, "learning_rate": 8.022995574311876e-06, "loss": 1.7113, "num_input_tokens_seen": 95682560, "step": 365 }, { "epoch": 0.09088651601688602, "grad_norm": 1.2987405061721802, "learning_rate": 8.009075115760243e-06, "loss": 1.5487, "num_input_tokens_seen": 95944704, "step": 366 }, { "epoch": 0.0911348398311398, "grad_norm": 0.6013867855072021, "learning_rate": 7.99511799257793e-06, "loss": 1.4665, "num_input_tokens_seen": 96206848, "step": 367 }, { "epoch": 0.09138316364539359, "grad_norm": 0.7376968860626221, "learning_rate": 7.981124374828079e-06, "loss": 1.4543, "num_input_tokens_seen": 96468992, "step": 368 }, { "epoch": 0.09163148745964737, "grad_norm": 0.6412230730056763, "learning_rate": 7.967094433018508e-06, "loss": 1.2195, "num_input_tokens_seen": 96731136, "step": 369 }, { "epoch": 0.09187981127390117, "grad_norm": 0.3257865905761719, "learning_rate": 7.953028338099628e-06, "loss": 1.5656, "num_input_tokens_seen": 96993280, "step": 370 }, { "epoch": 0.09212813508815496, "grad_norm": 0.47559216618537903, "learning_rate": 7.938926261462366e-06, "loss": 1.6012, "num_input_tokens_seen": 97255424, "step": 371 }, { "epoch": 0.09237645890240874, "grad_norm": 0.45426198840141296, "learning_rate": 7.92478837493608e-06, "loss": 1.7031, "num_input_tokens_seen": 97517568, "step": 372 }, { "epoch": 0.09262478271666252, "grad_norm": 0.4942920506000519, "learning_rate": 7.910614850786448e-06, "loss": 1.6369, "num_input_tokens_seen": 97779712, "step": 373 }, { "epoch": 0.09287310653091632, "grad_norm": 0.5819427967071533, "learning_rate": 7.896405861713393e-06, "loss": 1.7998, "num_input_tokens_seen": 98041856, "step": 374 }, { "epoch": 0.0931214303451701, "grad_norm": 0.42620816826820374, "learning_rate": 7.882161580848966e-06, "loss": 1.6103, "num_input_tokens_seen": 98304000, "step": 375 }, { "epoch": 0.09336975415942389, "grad_norm": 0.6115924715995789, "learning_rate": 7.86788218175523e-06, "loss": 1.8329, "num_input_tokens_seen": 98566144, "step": 376 }, { "epoch": 0.09361807797367767, "grad_norm": 0.9807063341140747, "learning_rate": 7.85356783842216e-06, "loss": 1.3716, "num_input_tokens_seen": 98828288, "step": 377 }, { "epoch": 0.09386640178793146, "grad_norm": 0.5305372476577759, "learning_rate": 7.839218725265507e-06, "loss": 1.585, "num_input_tokens_seen": 99090432, "step": 378 }, { "epoch": 0.09411472560218526, "grad_norm": 0.7433108687400818, "learning_rate": 7.82483501712469e-06, "loss": 1.5462, "num_input_tokens_seen": 99352576, "step": 379 }, { "epoch": 0.09436304941643904, "grad_norm": 0.6521483659744263, "learning_rate": 7.810416889260653e-06, "loss": 2.1058, "num_input_tokens_seen": 99614720, "step": 380 }, { "epoch": 0.09461137323069282, "grad_norm": 0.6441096663475037, "learning_rate": 7.795964517353734e-06, "loss": 1.938, "num_input_tokens_seen": 99876864, "step": 381 }, { "epoch": 0.09485969704494661, "grad_norm": 0.4057358205318451, "learning_rate": 7.781478077501526e-06, "loss": 1.5177, "num_input_tokens_seen": 100139008, "step": 382 }, { "epoch": 0.09510802085920039, "grad_norm": 0.5994225144386292, "learning_rate": 7.76695774621672e-06, "loss": 1.6127, "num_input_tokens_seen": 100401152, "step": 383 }, { "epoch": 0.09535634467345419, "grad_norm": 0.6396727561950684, "learning_rate": 7.752403700424978e-06, "loss": 1.678, "num_input_tokens_seen": 100663296, "step": 384 }, { "epoch": 0.09560466848770797, "grad_norm": 0.647515594959259, "learning_rate": 7.737816117462752e-06, "loss": 1.7168, "num_input_tokens_seen": 100925440, "step": 385 }, { "epoch": 0.09585299230196176, "grad_norm": 0.5679896473884583, "learning_rate": 7.723195175075136e-06, "loss": 1.9103, "num_input_tokens_seen": 101187584, "step": 386 }, { "epoch": 0.09610131611621554, "grad_norm": 0.9553595185279846, "learning_rate": 7.7085410514137e-06, "loss": 1.934, "num_input_tokens_seen": 101449728, "step": 387 }, { "epoch": 0.09634963993046933, "grad_norm": 0.3686366677284241, "learning_rate": 7.693853925034316e-06, "loss": 2.0843, "num_input_tokens_seen": 101711872, "step": 388 }, { "epoch": 0.09659796374472313, "grad_norm": 0.528319239616394, "learning_rate": 7.679133974894984e-06, "loss": 2.077, "num_input_tokens_seen": 101974016, "step": 389 }, { "epoch": 0.09684628755897691, "grad_norm": 0.6407886147499084, "learning_rate": 7.66438138035365e-06, "loss": 1.634, "num_input_tokens_seen": 102236160, "step": 390 }, { "epoch": 0.0970946113732307, "grad_norm": 0.5554214715957642, "learning_rate": 7.649596321166024e-06, "loss": 1.4811, "num_input_tokens_seen": 102498304, "step": 391 }, { "epoch": 0.09734293518748448, "grad_norm": 0.5426738858222961, "learning_rate": 7.634778977483389e-06, "loss": 1.6908, "num_input_tokens_seen": 102760448, "step": 392 }, { "epoch": 0.09759125900173826, "grad_norm": 0.7456260323524475, "learning_rate": 7.619929529850397e-06, "loss": 1.8275, "num_input_tokens_seen": 103022592, "step": 393 }, { "epoch": 0.09783958281599206, "grad_norm": 0.5138922929763794, "learning_rate": 7.605048159202884e-06, "loss": 1.1156, "num_input_tokens_seen": 103284736, "step": 394 }, { "epoch": 0.09808790663024584, "grad_norm": 0.3001045882701874, "learning_rate": 7.590135046865652e-06, "loss": 1.198, "num_input_tokens_seen": 103546880, "step": 395 }, { "epoch": 0.09833623044449963, "grad_norm": 0.5880023241043091, "learning_rate": 7.575190374550272e-06, "loss": 1.4706, "num_input_tokens_seen": 103809024, "step": 396 }, { "epoch": 0.09858455425875341, "grad_norm": 0.3109273910522461, "learning_rate": 7.560214324352858e-06, "loss": 1.0782, "num_input_tokens_seen": 104071168, "step": 397 }, { "epoch": 0.0988328780730072, "grad_norm": 0.3918832540512085, "learning_rate": 7.545207078751858e-06, "loss": 1.3639, "num_input_tokens_seen": 104333312, "step": 398 }, { "epoch": 0.099081201887261, "grad_norm": 0.6184259653091431, "learning_rate": 7.530168820605819e-06, "loss": 2.0244, "num_input_tokens_seen": 104595456, "step": 399 }, { "epoch": 0.09932952570151478, "grad_norm": 0.41290906071662903, "learning_rate": 7.515099733151177e-06, "loss": 1.3445, "num_input_tokens_seen": 104857600, "step": 400 }, { "epoch": 0.09957784951576856, "grad_norm": 0.5972164273262024, "learning_rate": 7.500000000000001e-06, "loss": 1.7396, "num_input_tokens_seen": 105119744, "step": 401 }, { "epoch": 0.09982617333002235, "grad_norm": 0.49639347195625305, "learning_rate": 7.484869805137778e-06, "loss": 1.9205, "num_input_tokens_seen": 105381888, "step": 402 }, { "epoch": 0.10007449714427613, "grad_norm": 0.47210463881492615, "learning_rate": 7.469709332921155e-06, "loss": 1.434, "num_input_tokens_seen": 105644032, "step": 403 }, { "epoch": 0.10032282095852993, "grad_norm": 0.5069103837013245, "learning_rate": 7.454518768075705e-06, "loss": 1.5317, "num_input_tokens_seen": 105906176, "step": 404 }, { "epoch": 0.10057114477278371, "grad_norm": 0.6772238612174988, "learning_rate": 7.4392982956936644e-06, "loss": 1.9896, "num_input_tokens_seen": 106168320, "step": 405 }, { "epoch": 0.1008194685870375, "grad_norm": 0.5503749847412109, "learning_rate": 7.424048101231687e-06, "loss": 1.6349, "num_input_tokens_seen": 106430464, "step": 406 }, { "epoch": 0.10106779240129128, "grad_norm": 0.47426870465278625, "learning_rate": 7.408768370508577e-06, "loss": 1.221, "num_input_tokens_seen": 106692608, "step": 407 }, { "epoch": 0.10131611621554507, "grad_norm": 0.7207087874412537, "learning_rate": 7.393459289703035e-06, "loss": 1.9311, "num_input_tokens_seen": 106954752, "step": 408 }, { "epoch": 0.10156444002979886, "grad_norm": 0.6696469783782959, "learning_rate": 7.378121045351378e-06, "loss": 1.2549, "num_input_tokens_seen": 107216896, "step": 409 }, { "epoch": 0.10181276384405265, "grad_norm": 2.1489691734313965, "learning_rate": 7.362753824345271e-06, "loss": 1.8569, "num_input_tokens_seen": 107479040, "step": 410 }, { "epoch": 0.10206108765830643, "grad_norm": 0.6064687371253967, "learning_rate": 7.347357813929455e-06, "loss": 1.571, "num_input_tokens_seen": 107741184, "step": 411 }, { "epoch": 0.10230941147256022, "grad_norm": 0.651612401008606, "learning_rate": 7.3319332016994575e-06, "loss": 1.3514, "num_input_tokens_seen": 108003328, "step": 412 }, { "epoch": 0.102557735286814, "grad_norm": 0.2269715964794159, "learning_rate": 7.31648017559931e-06, "loss": 1.2449, "num_input_tokens_seen": 108265472, "step": 413 }, { "epoch": 0.1028060591010678, "grad_norm": 1.5659141540527344, "learning_rate": 7.300998923919259e-06, "loss": 1.8001, "num_input_tokens_seen": 108527616, "step": 414 }, { "epoch": 0.10305438291532158, "grad_norm": 0.7734239101409912, "learning_rate": 7.285489635293472e-06, "loss": 1.5909, "num_input_tokens_seen": 108789760, "step": 415 }, { "epoch": 0.10330270672957537, "grad_norm": 0.5222880244255066, "learning_rate": 7.269952498697734e-06, "loss": 1.7375, "num_input_tokens_seen": 109051904, "step": 416 }, { "epoch": 0.10355103054382915, "grad_norm": 0.45977625250816345, "learning_rate": 7.254387703447154e-06, "loss": 1.9382, "num_input_tokens_seen": 109314048, "step": 417 }, { "epoch": 0.10379935435808293, "grad_norm": 0.6067866683006287, "learning_rate": 7.238795439193849e-06, "loss": 1.5499, "num_input_tokens_seen": 109576192, "step": 418 }, { "epoch": 0.10404767817233673, "grad_norm": 1.0210500955581665, "learning_rate": 7.223175895924638e-06, "loss": 2.0725, "num_input_tokens_seen": 109838336, "step": 419 }, { "epoch": 0.10429600198659052, "grad_norm": 0.7185985445976257, "learning_rate": 7.207529263958727e-06, "loss": 1.9104, "num_input_tokens_seen": 110100480, "step": 420 }, { "epoch": 0.1045443258008443, "grad_norm": 0.9692136645317078, "learning_rate": 7.191855733945388e-06, "loss": 1.7128, "num_input_tokens_seen": 110362624, "step": 421 }, { "epoch": 0.10479264961509809, "grad_norm": 0.36912843585014343, "learning_rate": 7.176155496861639e-06, "loss": 1.5977, "num_input_tokens_seen": 110624768, "step": 422 }, { "epoch": 0.10504097342935187, "grad_norm": 0.7783231735229492, "learning_rate": 7.160428744009913e-06, "loss": 1.5989, "num_input_tokens_seen": 110886912, "step": 423 }, { "epoch": 0.10528929724360567, "grad_norm": 0.3686671853065491, "learning_rate": 7.1446756670157306e-06, "loss": 1.887, "num_input_tokens_seen": 111149056, "step": 424 }, { "epoch": 0.10553762105785945, "grad_norm": 0.47263607382774353, "learning_rate": 7.128896457825364e-06, "loss": 1.808, "num_input_tokens_seen": 111411200, "step": 425 }, { "epoch": 0.10578594487211324, "grad_norm": 0.543540358543396, "learning_rate": 7.113091308703498e-06, "loss": 1.7969, "num_input_tokens_seen": 111673344, "step": 426 }, { "epoch": 0.10603426868636702, "grad_norm": 0.672477126121521, "learning_rate": 7.0972604122308865e-06, "loss": 1.3285, "num_input_tokens_seen": 111935488, "step": 427 }, { "epoch": 0.1062825925006208, "grad_norm": 0.501299262046814, "learning_rate": 7.081403961302007e-06, "loss": 1.3305, "num_input_tokens_seen": 112197632, "step": 428 }, { "epoch": 0.1065309163148746, "grad_norm": 0.4477572441101074, "learning_rate": 7.06552214912271e-06, "loss": 1.1045, "num_input_tokens_seen": 112459776, "step": 429 }, { "epoch": 0.10677924012912839, "grad_norm": 0.34857606887817383, "learning_rate": 7.049615169207864e-06, "loss": 1.929, "num_input_tokens_seen": 112721920, "step": 430 }, { "epoch": 0.10702756394338217, "grad_norm": 0.6203132271766663, "learning_rate": 7.033683215379002e-06, "loss": 1.062, "num_input_tokens_seen": 112984064, "step": 431 }, { "epoch": 0.10727588775763595, "grad_norm": 0.5540516376495361, "learning_rate": 7.0177264817619514e-06, "loss": 1.7175, "num_input_tokens_seen": 113246208, "step": 432 }, { "epoch": 0.10752421157188974, "grad_norm": 0.4745628833770752, "learning_rate": 7.0017451627844765e-06, "loss": 1.8083, "num_input_tokens_seen": 113508352, "step": 433 }, { "epoch": 0.10777253538614354, "grad_norm": 0.6570994257926941, "learning_rate": 6.985739453173903e-06, "loss": 1.4396, "num_input_tokens_seen": 113770496, "step": 434 }, { "epoch": 0.10802085920039732, "grad_norm": 2.3753468990325928, "learning_rate": 6.9697095479547564e-06, "loss": 1.3002, "num_input_tokens_seen": 114032640, "step": 435 }, { "epoch": 0.1082691830146511, "grad_norm": 0.3089500367641449, "learning_rate": 6.953655642446368e-06, "loss": 1.589, "num_input_tokens_seen": 114294784, "step": 436 }, { "epoch": 0.10851750682890489, "grad_norm": 3.394044876098633, "learning_rate": 6.9375779322605154e-06, "loss": 1.1472, "num_input_tokens_seen": 114556928, "step": 437 }, { "epoch": 0.10876583064315867, "grad_norm": 0.334686279296875, "learning_rate": 6.921476613299018e-06, "loss": 1.9471, "num_input_tokens_seen": 114819072, "step": 438 }, { "epoch": 0.10901415445741247, "grad_norm": 0.7744854092597961, "learning_rate": 6.905351881751372e-06, "loss": 1.4141, "num_input_tokens_seen": 115081216, "step": 439 }, { "epoch": 0.10926247827166626, "grad_norm": 0.37837162613868713, "learning_rate": 6.889203934092337e-06, "loss": 1.6888, "num_input_tokens_seen": 115343360, "step": 440 }, { "epoch": 0.10951080208592004, "grad_norm": 0.6168341636657715, "learning_rate": 6.873032967079562e-06, "loss": 1.9082, "num_input_tokens_seen": 115605504, "step": 441 }, { "epoch": 0.10975912590017382, "grad_norm": 0.41162988543510437, "learning_rate": 6.856839177751175e-06, "loss": 1.5192, "num_input_tokens_seen": 115867648, "step": 442 }, { "epoch": 0.11000744971442761, "grad_norm": 0.5805467963218689, "learning_rate": 6.840622763423391e-06, "loss": 1.8317, "num_input_tokens_seen": 116129792, "step": 443 }, { "epoch": 0.1102557735286814, "grad_norm": 0.5233989953994751, "learning_rate": 6.824383921688098e-06, "loss": 1.273, "num_input_tokens_seen": 116391936, "step": 444 }, { "epoch": 0.11050409734293519, "grad_norm": 0.6940083503723145, "learning_rate": 6.808122850410461e-06, "loss": 1.5257, "num_input_tokens_seen": 116654080, "step": 445 }, { "epoch": 0.11075242115718897, "grad_norm": 0.680293619632721, "learning_rate": 6.7918397477265e-06, "loss": 1.7634, "num_input_tokens_seen": 116916224, "step": 446 }, { "epoch": 0.11100074497144276, "grad_norm": 0.680514931678772, "learning_rate": 6.775534812040686e-06, "loss": 1.9248, "num_input_tokens_seen": 117178368, "step": 447 }, { "epoch": 0.11124906878569654, "grad_norm": 0.6705589890480042, "learning_rate": 6.759208242023509e-06, "loss": 1.6782, "num_input_tokens_seen": 117440512, "step": 448 }, { "epoch": 0.11149739259995034, "grad_norm": 0.6143855452537537, "learning_rate": 6.7428602366090764e-06, "loss": 1.8117, "num_input_tokens_seen": 117702656, "step": 449 }, { "epoch": 0.11174571641420412, "grad_norm": 0.5517452359199524, "learning_rate": 6.7264909949926735e-06, "loss": 1.4715, "num_input_tokens_seen": 117964800, "step": 450 }, { "epoch": 0.11199404022845791, "grad_norm": 0.34088805317878723, "learning_rate": 6.710100716628345e-06, "loss": 1.6831, "num_input_tokens_seen": 118226944, "step": 451 }, { "epoch": 0.11224236404271169, "grad_norm": 0.5145544409751892, "learning_rate": 6.693689601226458e-06, "loss": 1.6246, "num_input_tokens_seen": 118489088, "step": 452 }, { "epoch": 0.11249068785696548, "grad_norm": 0.5433220267295837, "learning_rate": 6.677257848751276e-06, "loss": 1.517, "num_input_tokens_seen": 118751232, "step": 453 }, { "epoch": 0.11273901167121927, "grad_norm": 0.6025758385658264, "learning_rate": 6.6608056594185166e-06, "loss": 1.9205, "num_input_tokens_seen": 119013376, "step": 454 }, { "epoch": 0.11298733548547306, "grad_norm": 0.5882771015167236, "learning_rate": 6.644333233692917e-06, "loss": 1.6951, "num_input_tokens_seen": 119275520, "step": 455 }, { "epoch": 0.11323565929972684, "grad_norm": 0.6837276220321655, "learning_rate": 6.627840772285784e-06, "loss": 2.0879, "num_input_tokens_seen": 119537664, "step": 456 }, { "epoch": 0.11348398311398063, "grad_norm": 0.5370259284973145, "learning_rate": 6.611328476152557e-06, "loss": 1.6184, "num_input_tokens_seen": 119799808, "step": 457 }, { "epoch": 0.11373230692823441, "grad_norm": 0.4522857367992401, "learning_rate": 6.594796546490351e-06, "loss": 1.7773, "num_input_tokens_seen": 120061952, "step": 458 }, { "epoch": 0.11398063074248821, "grad_norm": 0.46354392170906067, "learning_rate": 6.578245184735513e-06, "loss": 1.5712, "num_input_tokens_seen": 120324096, "step": 459 }, { "epoch": 0.114228954556742, "grad_norm": 0.686725378036499, "learning_rate": 6.561674592561164e-06, "loss": 1.6311, "num_input_tokens_seen": 120586240, "step": 460 }, { "epoch": 0.11447727837099578, "grad_norm": 0.22250455617904663, "learning_rate": 6.545084971874738e-06, "loss": 1.7059, "num_input_tokens_seen": 120848384, "step": 461 }, { "epoch": 0.11472560218524956, "grad_norm": 0.5717688798904419, "learning_rate": 6.5284765248155295e-06, "loss": 1.9078, "num_input_tokens_seen": 121110528, "step": 462 }, { "epoch": 0.11497392599950335, "grad_norm": 0.6517221331596375, "learning_rate": 6.5118494537522235e-06, "loss": 1.923, "num_input_tokens_seen": 121372672, "step": 463 }, { "epoch": 0.11522224981375714, "grad_norm": 0.7577219009399414, "learning_rate": 6.495203961280434e-06, "loss": 1.5007, "num_input_tokens_seen": 121634816, "step": 464 }, { "epoch": 0.11547057362801093, "grad_norm": 1.1634546518325806, "learning_rate": 6.4785402502202345e-06, "loss": 1.7924, "num_input_tokens_seen": 121896960, "step": 465 }, { "epoch": 0.11571889744226471, "grad_norm": 0.5361213088035583, "learning_rate": 6.461858523613684e-06, "loss": 1.3429, "num_input_tokens_seen": 122159104, "step": 466 }, { "epoch": 0.1159672212565185, "grad_norm": 0.3826828598976135, "learning_rate": 6.445158984722358e-06, "loss": 1.4901, "num_input_tokens_seen": 122421248, "step": 467 }, { "epoch": 0.11621554507077228, "grad_norm": 0.5723513960838318, "learning_rate": 6.428441837024868e-06, "loss": 1.8145, "num_input_tokens_seen": 122683392, "step": 468 }, { "epoch": 0.11646386888502608, "grad_norm": 0.6699240803718567, "learning_rate": 6.411707284214384e-06, "loss": 1.6751, "num_input_tokens_seen": 122945536, "step": 469 }, { "epoch": 0.11671219269927986, "grad_norm": 0.35479190945625305, "learning_rate": 6.3949555301961474e-06, "loss": 1.4937, "num_input_tokens_seen": 123207680, "step": 470 }, { "epoch": 0.11696051651353365, "grad_norm": 0.7304947972297668, "learning_rate": 6.378186779084996e-06, "loss": 1.5881, "num_input_tokens_seen": 123469824, "step": 471 }, { "epoch": 0.11720884032778743, "grad_norm": 0.4155537784099579, "learning_rate": 6.361401235202872e-06, "loss": 1.5931, "num_input_tokens_seen": 123731968, "step": 472 }, { "epoch": 0.11745716414204121, "grad_norm": 0.5789989233016968, "learning_rate": 6.344599103076329e-06, "loss": 1.3299, "num_input_tokens_seen": 123994112, "step": 473 }, { "epoch": 0.11770548795629501, "grad_norm": 0.5815244913101196, "learning_rate": 6.327780587434045e-06, "loss": 1.7197, "num_input_tokens_seen": 124256256, "step": 474 }, { "epoch": 0.1179538117705488, "grad_norm": 0.7529072165489197, "learning_rate": 6.310945893204324e-06, "loss": 1.47, "num_input_tokens_seen": 124518400, "step": 475 }, { "epoch": 0.11820213558480258, "grad_norm": 0.7494714856147766, "learning_rate": 6.294095225512604e-06, "loss": 1.864, "num_input_tokens_seen": 124780544, "step": 476 }, { "epoch": 0.11845045939905637, "grad_norm": 0.3240576684474945, "learning_rate": 6.277228789678953e-06, "loss": 1.5745, "num_input_tokens_seen": 125042688, "step": 477 }, { "epoch": 0.11869878321331016, "grad_norm": 0.5607863068580627, "learning_rate": 6.26034679121557e-06, "loss": 1.9548, "num_input_tokens_seen": 125304832, "step": 478 }, { "epoch": 0.11894710702756395, "grad_norm": 0.5693123936653137, "learning_rate": 6.243449435824276e-06, "loss": 1.7745, "num_input_tokens_seen": 125566976, "step": 479 }, { "epoch": 0.11919543084181773, "grad_norm": 0.5455982685089111, "learning_rate": 6.2265369293940135e-06, "loss": 1.9634, "num_input_tokens_seen": 125829120, "step": 480 }, { "epoch": 0.11944375465607152, "grad_norm": 0.4890311658382416, "learning_rate": 6.209609477998339e-06, "loss": 2.1942, "num_input_tokens_seen": 126091264, "step": 481 }, { "epoch": 0.1196920784703253, "grad_norm": 0.5301004648208618, "learning_rate": 6.192667287892905e-06, "loss": 1.4093, "num_input_tokens_seen": 126353408, "step": 482 }, { "epoch": 0.1199404022845791, "grad_norm": 0.47906139492988586, "learning_rate": 6.17571056551295e-06, "loss": 1.7136, "num_input_tokens_seen": 126615552, "step": 483 }, { "epoch": 0.12018872609883288, "grad_norm": 0.6675156354904175, "learning_rate": 6.158739517470786e-06, "loss": 1.6023, "num_input_tokens_seen": 126877696, "step": 484 }, { "epoch": 0.12043704991308667, "grad_norm": 0.31835541129112244, "learning_rate": 6.141754350553279e-06, "loss": 1.5503, "num_input_tokens_seen": 127139840, "step": 485 }, { "epoch": 0.12068537372734045, "grad_norm": 0.7083136439323425, "learning_rate": 6.124755271719326e-06, "loss": 1.5225, "num_input_tokens_seen": 127401984, "step": 486 }, { "epoch": 0.12093369754159423, "grad_norm": 0.4737892746925354, "learning_rate": 6.107742488097338e-06, "loss": 1.4856, "num_input_tokens_seen": 127664128, "step": 487 }, { "epoch": 0.12118202135584803, "grad_norm": 0.403909832239151, "learning_rate": 6.090716206982714e-06, "loss": 1.5458, "num_input_tokens_seen": 127926272, "step": 488 }, { "epoch": 0.12143034517010182, "grad_norm": 0.6817101240158081, "learning_rate": 6.073676635835317e-06, "loss": 1.3755, "num_input_tokens_seen": 128188416, "step": 489 }, { "epoch": 0.1216786689843556, "grad_norm": 0.5082156658172607, "learning_rate": 6.056623982276945e-06, "loss": 1.8278, "num_input_tokens_seen": 128450560, "step": 490 }, { "epoch": 0.12192699279860938, "grad_norm": 0.4969983696937561, "learning_rate": 6.039558454088796e-06, "loss": 2.0705, "num_input_tokens_seen": 128712704, "step": 491 }, { "epoch": 0.12217531661286317, "grad_norm": 0.6239178776741028, "learning_rate": 6.022480259208951e-06, "loss": 1.5768, "num_input_tokens_seen": 128974848, "step": 492 }, { "epoch": 0.12242364042711697, "grad_norm": 0.5710934400558472, "learning_rate": 6.005389605729824e-06, "loss": 1.6925, "num_input_tokens_seen": 129236992, "step": 493 }, { "epoch": 0.12267196424137075, "grad_norm": 0.5775221586227417, "learning_rate": 5.988286701895631e-06, "loss": 1.924, "num_input_tokens_seen": 129499136, "step": 494 }, { "epoch": 0.12292028805562454, "grad_norm": 0.4340408146381378, "learning_rate": 5.97117175609986e-06, "loss": 1.791, "num_input_tokens_seen": 129761280, "step": 495 }, { "epoch": 0.12316861186987832, "grad_norm": 0.3268572986125946, "learning_rate": 5.954044976882725e-06, "loss": 1.7505, "num_input_tokens_seen": 130023424, "step": 496 }, { "epoch": 0.1234169356841321, "grad_norm": 0.43856051564216614, "learning_rate": 5.936906572928625e-06, "loss": 1.3188, "num_input_tokens_seen": 130285568, "step": 497 }, { "epoch": 0.1236652594983859, "grad_norm": 0.459693044424057, "learning_rate": 5.919756753063601e-06, "loss": 1.8807, "num_input_tokens_seen": 130547712, "step": 498 }, { "epoch": 0.12391358331263969, "grad_norm": 1.3583054542541504, "learning_rate": 5.902595726252801e-06, "loss": 1.5139, "num_input_tokens_seen": 130809856, "step": 499 }, { "epoch": 0.12416190712689347, "grad_norm": 0.4457927644252777, "learning_rate": 5.885423701597918e-06, "loss": 1.0955, "num_input_tokens_seen": 131072000, "step": 500 }, { "epoch": 0.12441023094114725, "grad_norm": 0.45763128995895386, "learning_rate": 5.8682408883346535e-06, "loss": 1.3018, "num_input_tokens_seen": 131334144, "step": 501 }, { "epoch": 0.12465855475540104, "grad_norm": 0.43406641483306885, "learning_rate": 5.851047495830163e-06, "loss": 1.8796, "num_input_tokens_seen": 131596288, "step": 502 }, { "epoch": 0.12490687856965484, "grad_norm": 0.6790747046470642, "learning_rate": 5.8338437335805124e-06, "loss": 1.7733, "num_input_tokens_seen": 131858432, "step": 503 }, { "epoch": 0.12515520238390862, "grad_norm": 0.469204306602478, "learning_rate": 5.816629811208112e-06, "loss": 1.7257, "num_input_tokens_seen": 132120576, "step": 504 }, { "epoch": 0.1254035261981624, "grad_norm": 0.29336878657341003, "learning_rate": 5.799405938459175e-06, "loss": 2.0604, "num_input_tokens_seen": 132382720, "step": 505 }, { "epoch": 0.1256518500124162, "grad_norm": 1.0650948286056519, "learning_rate": 5.782172325201155e-06, "loss": 1.4833, "num_input_tokens_seen": 132644864, "step": 506 }, { "epoch": 0.12590017382667, "grad_norm": 0.5020011067390442, "learning_rate": 5.764929181420191e-06, "loss": 1.9276, "num_input_tokens_seen": 132907008, "step": 507 }, { "epoch": 0.12614849764092376, "grad_norm": 0.623615562915802, "learning_rate": 5.747676717218549e-06, "loss": 1.6809, "num_input_tokens_seen": 133169152, "step": 508 }, { "epoch": 0.12639682145517755, "grad_norm": 0.6677452325820923, "learning_rate": 5.730415142812059e-06, "loss": 1.6377, "num_input_tokens_seen": 133431296, "step": 509 }, { "epoch": 0.12664514526943133, "grad_norm": 0.5016704201698303, "learning_rate": 5.7131446685275595e-06, "loss": 1.4934, "num_input_tokens_seen": 133693440, "step": 510 }, { "epoch": 0.12689346908368512, "grad_norm": 0.4135633707046509, "learning_rate": 5.695865504800328e-06, "loss": 1.6179, "num_input_tokens_seen": 133955584, "step": 511 }, { "epoch": 0.12714179289793892, "grad_norm": 0.7254384160041809, "learning_rate": 5.678577862171523e-06, "loss": 1.5534, "num_input_tokens_seen": 134217728, "step": 512 }, { "epoch": 0.1273901167121927, "grad_norm": 0.7331421971321106, "learning_rate": 5.661281951285613e-06, "loss": 1.7944, "num_input_tokens_seen": 134479872, "step": 513 }, { "epoch": 0.1276384405264465, "grad_norm": 0.6121946573257446, "learning_rate": 5.643977982887815e-06, "loss": 1.6006, "num_input_tokens_seen": 134742016, "step": 514 }, { "epoch": 0.12788676434070026, "grad_norm": 0.4421185255050659, "learning_rate": 5.626666167821522e-06, "loss": 1.5906, "num_input_tokens_seen": 135004160, "step": 515 }, { "epoch": 0.12813508815495406, "grad_norm": 0.3607276380062103, "learning_rate": 5.609346717025738e-06, "loss": 1.7194, "num_input_tokens_seen": 135266304, "step": 516 }, { "epoch": 0.12838341196920786, "grad_norm": 0.66265469789505, "learning_rate": 5.592019841532507e-06, "loss": 1.5176, "num_input_tokens_seen": 135528448, "step": 517 }, { "epoch": 0.12863173578346163, "grad_norm": 0.4719207286834717, "learning_rate": 5.5746857524643335e-06, "loss": 1.5677, "num_input_tokens_seen": 135790592, "step": 518 }, { "epoch": 0.12888005959771542, "grad_norm": 0.5282115936279297, "learning_rate": 5.557344661031628e-06, "loss": 1.6344, "num_input_tokens_seen": 136052736, "step": 519 }, { "epoch": 0.1291283834119692, "grad_norm": 0.3419008255004883, "learning_rate": 5.539996778530114e-06, "loss": 1.9051, "num_input_tokens_seen": 136314880, "step": 520 }, { "epoch": 0.129376707226223, "grad_norm": 0.7775862812995911, "learning_rate": 5.522642316338268e-06, "loss": 1.7739, "num_input_tokens_seen": 136577024, "step": 521 }, { "epoch": 0.1296250310404768, "grad_norm": 0.554291307926178, "learning_rate": 5.505281485914732e-06, "loss": 1.6473, "num_input_tokens_seen": 136839168, "step": 522 }, { "epoch": 0.12987335485473056, "grad_norm": 0.5523970127105713, "learning_rate": 5.487914498795748e-06, "loss": 1.5672, "num_input_tokens_seen": 137101312, "step": 523 }, { "epoch": 0.13012167866898436, "grad_norm": 0.4398843050003052, "learning_rate": 5.470541566592573e-06, "loss": 1.5184, "num_input_tokens_seen": 137363456, "step": 524 }, { "epoch": 0.13037000248323813, "grad_norm": 0.36770564317703247, "learning_rate": 5.453162900988902e-06, "loss": 1.272, "num_input_tokens_seen": 137625600, "step": 525 }, { "epoch": 0.13061832629749193, "grad_norm": 0.36852288246154785, "learning_rate": 5.435778713738292e-06, "loss": 1.914, "num_input_tokens_seen": 137887744, "step": 526 }, { "epoch": 0.13086665011174572, "grad_norm": 0.5804073810577393, "learning_rate": 5.41838921666158e-06, "loss": 1.3187, "num_input_tokens_seen": 138149888, "step": 527 }, { "epoch": 0.1311149739259995, "grad_norm": 0.4716220796108246, "learning_rate": 5.400994621644294e-06, "loss": 1.6562, "num_input_tokens_seen": 138412032, "step": 528 }, { "epoch": 0.1313632977402533, "grad_norm": 0.6712131500244141, "learning_rate": 5.383595140634093e-06, "loss": 1.6972, "num_input_tokens_seen": 138674176, "step": 529 }, { "epoch": 0.1316116215545071, "grad_norm": 0.6651138067245483, "learning_rate": 5.366190985638159e-06, "loss": 1.2614, "num_input_tokens_seen": 138936320, "step": 530 }, { "epoch": 0.13185994536876086, "grad_norm": 0.6065943241119385, "learning_rate": 5.348782368720627e-06, "loss": 1.8296, "num_input_tokens_seen": 139198464, "step": 531 }, { "epoch": 0.13210826918301466, "grad_norm": 0.5409244298934937, "learning_rate": 5.3313695020000026e-06, "loss": 1.6437, "num_input_tokens_seen": 139460608, "step": 532 }, { "epoch": 0.13235659299726843, "grad_norm": 0.742667555809021, "learning_rate": 5.3139525976465675e-06, "loss": 1.5862, "num_input_tokens_seen": 139722752, "step": 533 }, { "epoch": 0.13260491681152223, "grad_norm": 0.6471251845359802, "learning_rate": 5.296531867879809e-06, "loss": 1.6151, "num_input_tokens_seen": 139984896, "step": 534 }, { "epoch": 0.13285324062577603, "grad_norm": 0.643791675567627, "learning_rate": 5.27910752496582e-06, "loss": 1.7427, "num_input_tokens_seen": 140247040, "step": 535 }, { "epoch": 0.1331015644400298, "grad_norm": 0.7029093503952026, "learning_rate": 5.2616797812147205e-06, "loss": 1.6691, "num_input_tokens_seen": 140509184, "step": 536 }, { "epoch": 0.1333498882542836, "grad_norm": 0.7368614077568054, "learning_rate": 5.244248848978067e-06, "loss": 1.9763, "num_input_tokens_seen": 140771328, "step": 537 }, { "epoch": 0.13359821206853736, "grad_norm": 0.7032376527786255, "learning_rate": 5.226814940646268e-06, "loss": 1.7725, "num_input_tokens_seen": 141033472, "step": 538 }, { "epoch": 0.13384653588279116, "grad_norm": 0.47285741567611694, "learning_rate": 5.209378268645998e-06, "loss": 2.148, "num_input_tokens_seen": 141295616, "step": 539 }, { "epoch": 0.13409485969704496, "grad_norm": 0.6670664548873901, "learning_rate": 5.1919390454376e-06, "loss": 1.7662, "num_input_tokens_seen": 141557760, "step": 540 }, { "epoch": 0.13434318351129873, "grad_norm": 0.6090880632400513, "learning_rate": 5.174497483512506e-06, "loss": 1.2453, "num_input_tokens_seen": 141819904, "step": 541 }, { "epoch": 0.13459150732555253, "grad_norm": 0.48240217566490173, "learning_rate": 5.157053795390642e-06, "loss": 1.7743, "num_input_tokens_seen": 142082048, "step": 542 }, { "epoch": 0.1348398311398063, "grad_norm": 0.5195634961128235, "learning_rate": 5.139608193617846e-06, "loss": 1.8024, "num_input_tokens_seen": 142344192, "step": 543 }, { "epoch": 0.1350881549540601, "grad_norm": 0.5877946615219116, "learning_rate": 5.1221608907632665e-06, "loss": 1.6163, "num_input_tokens_seen": 142606336, "step": 544 }, { "epoch": 0.1353364787683139, "grad_norm": 0.6305245757102966, "learning_rate": 5.1047120994167855e-06, "loss": 1.5309, "num_input_tokens_seen": 142868480, "step": 545 }, { "epoch": 0.13558480258256767, "grad_norm": 0.591148853302002, "learning_rate": 5.087262032186418e-06, "loss": 1.531, "num_input_tokens_seen": 143130624, "step": 546 }, { "epoch": 0.13583312639682146, "grad_norm": 0.8955023884773254, "learning_rate": 5.069810901695727e-06, "loss": 1.7347, "num_input_tokens_seen": 143392768, "step": 547 }, { "epoch": 0.13608145021107523, "grad_norm": 0.5800215005874634, "learning_rate": 5.05235892058123e-06, "loss": 1.6724, "num_input_tokens_seen": 143654912, "step": 548 }, { "epoch": 0.13632977402532903, "grad_norm": 0.5544260144233704, "learning_rate": 5.034906301489808e-06, "loss": 1.5061, "num_input_tokens_seen": 143917056, "step": 549 }, { "epoch": 0.13657809783958283, "grad_norm": 0.5210686922073364, "learning_rate": 5.0174532570761194e-06, "loss": 1.3236, "num_input_tokens_seen": 144179200, "step": 550 }, { "epoch": 0.1368264216538366, "grad_norm": 0.39467668533325195, "learning_rate": 5e-06, "loss": 1.5716, "num_input_tokens_seen": 144441344, "step": 551 }, { "epoch": 0.1370747454680904, "grad_norm": 0.45690640807151794, "learning_rate": 4.982546742923883e-06, "loss": 1.5183, "num_input_tokens_seen": 144703488, "step": 552 }, { "epoch": 0.13732306928234417, "grad_norm": 0.34431830048561096, "learning_rate": 4.965093698510192e-06, "loss": 1.7103, "num_input_tokens_seen": 144965632, "step": 553 }, { "epoch": 0.13757139309659797, "grad_norm": 0.6095772385597229, "learning_rate": 4.9476410794187726e-06, "loss": 1.79, "num_input_tokens_seen": 145227776, "step": 554 }, { "epoch": 0.13781971691085176, "grad_norm": 0.5050289630889893, "learning_rate": 4.9301890983042744e-06, "loss": 1.6136, "num_input_tokens_seen": 145489920, "step": 555 }, { "epoch": 0.13806804072510553, "grad_norm": 0.6285063028335571, "learning_rate": 4.9127379678135825e-06, "loss": 1.5999, "num_input_tokens_seen": 145752064, "step": 556 }, { "epoch": 0.13831636453935933, "grad_norm": 0.6678712368011475, "learning_rate": 4.895287900583216e-06, "loss": 1.2344, "num_input_tokens_seen": 146014208, "step": 557 }, { "epoch": 0.1385646883536131, "grad_norm": 0.3648228347301483, "learning_rate": 4.877839109236735e-06, "loss": 1.5726, "num_input_tokens_seen": 146276352, "step": 558 }, { "epoch": 0.1388130121678669, "grad_norm": 0.5992112755775452, "learning_rate": 4.860391806382157e-06, "loss": 1.9771, "num_input_tokens_seen": 146538496, "step": 559 }, { "epoch": 0.1390613359821207, "grad_norm": 0.4710709750652313, "learning_rate": 4.842946204609359e-06, "loss": 1.8191, "num_input_tokens_seen": 146800640, "step": 560 }, { "epoch": 0.13930965979637447, "grad_norm": 0.5668407678604126, "learning_rate": 4.825502516487497e-06, "loss": 1.8406, "num_input_tokens_seen": 147062784, "step": 561 }, { "epoch": 0.13955798361062827, "grad_norm": 0.5589337944984436, "learning_rate": 4.8080609545624004e-06, "loss": 1.5411, "num_input_tokens_seen": 147324928, "step": 562 }, { "epoch": 0.13980630742488204, "grad_norm": 0.7744218707084656, "learning_rate": 4.7906217313540035e-06, "loss": 1.4392, "num_input_tokens_seen": 147587072, "step": 563 }, { "epoch": 0.14005463123913584, "grad_norm": 0.8113576769828796, "learning_rate": 4.7731850593537316e-06, "loss": 1.6712, "num_input_tokens_seen": 147849216, "step": 564 }, { "epoch": 0.14030295505338963, "grad_norm": 0.8065240979194641, "learning_rate": 4.755751151021934e-06, "loss": 1.7134, "num_input_tokens_seen": 148111360, "step": 565 }, { "epoch": 0.1405512788676434, "grad_norm": 0.6279537081718445, "learning_rate": 4.738320218785281e-06, "loss": 1.6341, "num_input_tokens_seen": 148373504, "step": 566 }, { "epoch": 0.1407996026818972, "grad_norm": 0.34301266074180603, "learning_rate": 4.720892475034181e-06, "loss": 1.7529, "num_input_tokens_seen": 148635648, "step": 567 }, { "epoch": 0.14104792649615097, "grad_norm": 0.3987272381782532, "learning_rate": 4.703468132120193e-06, "loss": 1.3504, "num_input_tokens_seen": 148897792, "step": 568 }, { "epoch": 0.14129625031040477, "grad_norm": 0.5643488168716431, "learning_rate": 4.686047402353433e-06, "loss": 1.6918, "num_input_tokens_seen": 149159936, "step": 569 }, { "epoch": 0.14154457412465857, "grad_norm": 1.1378772258758545, "learning_rate": 4.668630498000001e-06, "loss": 1.6232, "num_input_tokens_seen": 149422080, "step": 570 }, { "epoch": 0.14179289793891234, "grad_norm": 0.40794894099235535, "learning_rate": 4.651217631279374e-06, "loss": 1.2633, "num_input_tokens_seen": 149684224, "step": 571 }, { "epoch": 0.14204122175316614, "grad_norm": 0.40206679701805115, "learning_rate": 4.6338090143618435e-06, "loss": 1.7796, "num_input_tokens_seen": 149946368, "step": 572 }, { "epoch": 0.1422895455674199, "grad_norm": 0.6353849768638611, "learning_rate": 4.6164048593659076e-06, "loss": 1.5364, "num_input_tokens_seen": 150208512, "step": 573 }, { "epoch": 0.1425378693816737, "grad_norm": 0.5148465037345886, "learning_rate": 4.5990053783557066e-06, "loss": 1.5442, "num_input_tokens_seen": 150470656, "step": 574 }, { "epoch": 0.1427861931959275, "grad_norm": 0.5749762058258057, "learning_rate": 4.581610783338424e-06, "loss": 1.0381, "num_input_tokens_seen": 150732800, "step": 575 }, { "epoch": 0.14303451701018127, "grad_norm": 0.9002466201782227, "learning_rate": 4.564221286261709e-06, "loss": 1.4117, "num_input_tokens_seen": 150994944, "step": 576 }, { "epoch": 0.14328284082443507, "grad_norm": 0.3512033224105835, "learning_rate": 4.546837099011101e-06, "loss": 1.5649, "num_input_tokens_seen": 151257088, "step": 577 }, { "epoch": 0.14353116463868884, "grad_norm": 0.45879220962524414, "learning_rate": 4.529458433407429e-06, "loss": 1.4193, "num_input_tokens_seen": 151519232, "step": 578 }, { "epoch": 0.14377948845294264, "grad_norm": 0.7155461311340332, "learning_rate": 4.512085501204254e-06, "loss": 1.5547, "num_input_tokens_seen": 151781376, "step": 579 }, { "epoch": 0.14402781226719644, "grad_norm": 0.6171830892562866, "learning_rate": 4.494718514085269e-06, "loss": 1.2821, "num_input_tokens_seen": 152043520, "step": 580 }, { "epoch": 0.1442761360814502, "grad_norm": 0.5107810497283936, "learning_rate": 4.477357683661734e-06, "loss": 1.6026, "num_input_tokens_seen": 152305664, "step": 581 }, { "epoch": 0.144524459895704, "grad_norm": 0.6672670245170593, "learning_rate": 4.460003221469886e-06, "loss": 1.7716, "num_input_tokens_seen": 152567808, "step": 582 }, { "epoch": 0.14477278370995778, "grad_norm": 0.4775781035423279, "learning_rate": 4.442655338968373e-06, "loss": 1.3139, "num_input_tokens_seen": 152829952, "step": 583 }, { "epoch": 0.14502110752421157, "grad_norm": 0.665295422077179, "learning_rate": 4.425314247535668e-06, "loss": 1.2141, "num_input_tokens_seen": 153092096, "step": 584 }, { "epoch": 0.14526943133846537, "grad_norm": 0.49010828137397766, "learning_rate": 4.4079801584674955e-06, "loss": 1.3422, "num_input_tokens_seen": 153354240, "step": 585 }, { "epoch": 0.14551775515271914, "grad_norm": 0.4619687795639038, "learning_rate": 4.390653282974264e-06, "loss": 1.5652, "num_input_tokens_seen": 153616384, "step": 586 }, { "epoch": 0.14576607896697294, "grad_norm": 0.37380003929138184, "learning_rate": 4.373333832178478e-06, "loss": 1.8029, "num_input_tokens_seen": 153878528, "step": 587 }, { "epoch": 0.1460144027812267, "grad_norm": 0.5546239614486694, "learning_rate": 4.356022017112187e-06, "loss": 1.7681, "num_input_tokens_seen": 154140672, "step": 588 }, { "epoch": 0.1462627265954805, "grad_norm": 0.6969411969184875, "learning_rate": 4.3387180487143875e-06, "loss": 1.9277, "num_input_tokens_seen": 154402816, "step": 589 }, { "epoch": 0.1465110504097343, "grad_norm": 0.7840688824653625, "learning_rate": 4.321422137828479e-06, "loss": 1.6007, "num_input_tokens_seen": 154664960, "step": 590 }, { "epoch": 0.14675937422398808, "grad_norm": 0.6813507080078125, "learning_rate": 4.304134495199675e-06, "loss": 1.1834, "num_input_tokens_seen": 154927104, "step": 591 }, { "epoch": 0.14700769803824187, "grad_norm": 0.5060350298881531, "learning_rate": 4.286855331472442e-06, "loss": 1.4612, "num_input_tokens_seen": 155189248, "step": 592 }, { "epoch": 0.14725602185249564, "grad_norm": 0.5000078678131104, "learning_rate": 4.269584857187942e-06, "loss": 1.64, "num_input_tokens_seen": 155451392, "step": 593 }, { "epoch": 0.14750434566674944, "grad_norm": 0.5458803176879883, "learning_rate": 4.2523232827814534e-06, "loss": 2.1238, "num_input_tokens_seen": 155713536, "step": 594 }, { "epoch": 0.14775266948100324, "grad_norm": 1.3677194118499756, "learning_rate": 4.23507081857981e-06, "loss": 1.5499, "num_input_tokens_seen": 155975680, "step": 595 }, { "epoch": 0.148000993295257, "grad_norm": 0.5004885196685791, "learning_rate": 4.217827674798845e-06, "loss": 1.836, "num_input_tokens_seen": 156237824, "step": 596 }, { "epoch": 0.1482493171095108, "grad_norm": 1.3544896841049194, "learning_rate": 4.200594061540827e-06, "loss": 1.5202, "num_input_tokens_seen": 156499968, "step": 597 }, { "epoch": 0.14849764092376458, "grad_norm": 0.41308310627937317, "learning_rate": 4.183370188791891e-06, "loss": 1.4474, "num_input_tokens_seen": 156762112, "step": 598 }, { "epoch": 0.14874596473801838, "grad_norm": 0.5802090167999268, "learning_rate": 4.166156266419489e-06, "loss": 1.8016, "num_input_tokens_seen": 157024256, "step": 599 }, { "epoch": 0.14899428855227217, "grad_norm": 0.5909608602523804, "learning_rate": 4.148952504169839e-06, "loss": 1.5757, "num_input_tokens_seen": 157286400, "step": 600 }, { "epoch": 0.14924261236652595, "grad_norm": 0.41110658645629883, "learning_rate": 4.131759111665349e-06, "loss": 0.8013, "num_input_tokens_seen": 157548544, "step": 601 }, { "epoch": 0.14949093618077974, "grad_norm": 0.5430499911308289, "learning_rate": 4.114576298402085e-06, "loss": 1.9595, "num_input_tokens_seen": 157810688, "step": 602 }, { "epoch": 0.1497392599950335, "grad_norm": 0.7195497751235962, "learning_rate": 4.0974042737472005e-06, "loss": 1.5623, "num_input_tokens_seen": 158072832, "step": 603 }, { "epoch": 0.1499875838092873, "grad_norm": 0.4989107847213745, "learning_rate": 4.0802432469364e-06, "loss": 1.3031, "num_input_tokens_seen": 158334976, "step": 604 }, { "epoch": 0.1502359076235411, "grad_norm": 0.5433792471885681, "learning_rate": 4.063093427071376e-06, "loss": 1.5003, "num_input_tokens_seen": 158597120, "step": 605 }, { "epoch": 0.15048423143779488, "grad_norm": 0.3139537274837494, "learning_rate": 4.045955023117276e-06, "loss": 1.2943, "num_input_tokens_seen": 158859264, "step": 606 }, { "epoch": 0.15073255525204868, "grad_norm": 0.7295723557472229, "learning_rate": 4.028828243900141e-06, "loss": 1.9625, "num_input_tokens_seen": 159121408, "step": 607 }, { "epoch": 0.15098087906630245, "grad_norm": 0.7763230204582214, "learning_rate": 4.0117132981043695e-06, "loss": 1.8425, "num_input_tokens_seen": 159383552, "step": 608 }, { "epoch": 0.15122920288055625, "grad_norm": 0.5297269821166992, "learning_rate": 3.994610394270178e-06, "loss": 1.6922, "num_input_tokens_seen": 159645696, "step": 609 }, { "epoch": 0.15147752669481004, "grad_norm": 0.4328116476535797, "learning_rate": 3.977519740791049e-06, "loss": 0.9763, "num_input_tokens_seen": 159907840, "step": 610 }, { "epoch": 0.15172585050906381, "grad_norm": 0.6759671568870544, "learning_rate": 3.960441545911205e-06, "loss": 1.7297, "num_input_tokens_seen": 160169984, "step": 611 }, { "epoch": 0.1519741743233176, "grad_norm": 0.44700494408607483, "learning_rate": 3.943376017723058e-06, "loss": 1.9346, "num_input_tokens_seen": 160432128, "step": 612 }, { "epoch": 0.15222249813757138, "grad_norm": 0.8407000303268433, "learning_rate": 3.926323364164684e-06, "loss": 1.7382, "num_input_tokens_seen": 160694272, "step": 613 }, { "epoch": 0.15247082195182518, "grad_norm": 0.8016761541366577, "learning_rate": 3.909283793017289e-06, "loss": 1.7423, "num_input_tokens_seen": 160956416, "step": 614 }, { "epoch": 0.15271914576607898, "grad_norm": 0.414358526468277, "learning_rate": 3.892257511902664e-06, "loss": 1.4276, "num_input_tokens_seen": 161218560, "step": 615 }, { "epoch": 0.15296746958033275, "grad_norm": 1.0210763216018677, "learning_rate": 3.875244728280676e-06, "loss": 1.3653, "num_input_tokens_seen": 161480704, "step": 616 }, { "epoch": 0.15321579339458655, "grad_norm": 0.8063709735870361, "learning_rate": 3.8582456494467214e-06, "loss": 1.8222, "num_input_tokens_seen": 161742848, "step": 617 }, { "epoch": 0.15346411720884032, "grad_norm": 0.5376055836677551, "learning_rate": 3.841260482529215e-06, "loss": 1.7019, "num_input_tokens_seen": 162004992, "step": 618 }, { "epoch": 0.15371244102309412, "grad_norm": 0.5194859504699707, "learning_rate": 3.82428943448705e-06, "loss": 1.633, "num_input_tokens_seen": 162267136, "step": 619 }, { "epoch": 0.1539607648373479, "grad_norm": 0.5276075005531311, "learning_rate": 3.8073327121070968e-06, "loss": 1.6824, "num_input_tokens_seen": 162529280, "step": 620 }, { "epoch": 0.15420908865160168, "grad_norm": 0.7873098254203796, "learning_rate": 3.790390522001662e-06, "loss": 1.287, "num_input_tokens_seen": 162791424, "step": 621 }, { "epoch": 0.15445741246585548, "grad_norm": 0.6120622158050537, "learning_rate": 3.7734630706059873e-06, "loss": 1.8678, "num_input_tokens_seen": 163053568, "step": 622 }, { "epoch": 0.15470573628010925, "grad_norm": 0.5474843382835388, "learning_rate": 3.756550564175727e-06, "loss": 1.4913, "num_input_tokens_seen": 163315712, "step": 623 }, { "epoch": 0.15495406009436305, "grad_norm": 2.4813928604125977, "learning_rate": 3.7396532087844318e-06, "loss": 1.5975, "num_input_tokens_seen": 163577856, "step": 624 }, { "epoch": 0.15520238390861685, "grad_norm": 0.4030190408229828, "learning_rate": 3.7227712103210485e-06, "loss": 1.3417, "num_input_tokens_seen": 163840000, "step": 625 }, { "epoch": 0.15545070772287062, "grad_norm": 0.4890609085559845, "learning_rate": 3.705904774487396e-06, "loss": 1.6517, "num_input_tokens_seen": 164102144, "step": 626 }, { "epoch": 0.15569903153712442, "grad_norm": 0.4225277006626129, "learning_rate": 3.6890541067956775e-06, "loss": 1.9024, "num_input_tokens_seen": 164364288, "step": 627 }, { "epoch": 0.1559473553513782, "grad_norm": 0.3365168273448944, "learning_rate": 3.672219412565956e-06, "loss": 1.8831, "num_input_tokens_seen": 164626432, "step": 628 }, { "epoch": 0.15619567916563198, "grad_norm": 0.46206042170524597, "learning_rate": 3.655400896923672e-06, "loss": 1.5472, "num_input_tokens_seen": 164888576, "step": 629 }, { "epoch": 0.15644400297988578, "grad_norm": 0.5152633190155029, "learning_rate": 3.6385987647971287e-06, "loss": 1.9021, "num_input_tokens_seen": 165150720, "step": 630 }, { "epoch": 0.15669232679413955, "grad_norm": 0.6955782771110535, "learning_rate": 3.6218132209150047e-06, "loss": 1.5289, "num_input_tokens_seen": 165412864, "step": 631 }, { "epoch": 0.15694065060839335, "grad_norm": 0.8016681671142578, "learning_rate": 3.6050444698038547e-06, "loss": 1.2874, "num_input_tokens_seen": 165675008, "step": 632 }, { "epoch": 0.15718897442264712, "grad_norm": 0.32328078150749207, "learning_rate": 3.5882927157856175e-06, "loss": 1.9816, "num_input_tokens_seen": 165937152, "step": 633 }, { "epoch": 0.15743729823690092, "grad_norm": 0.4143851101398468, "learning_rate": 3.571558162975133e-06, "loss": 1.4726, "num_input_tokens_seen": 166199296, "step": 634 }, { "epoch": 0.15768562205115472, "grad_norm": 0.6780726909637451, "learning_rate": 3.5548410152776414e-06, "loss": 1.5232, "num_input_tokens_seen": 166461440, "step": 635 }, { "epoch": 0.1579339458654085, "grad_norm": 0.4231894314289093, "learning_rate": 3.538141476386317e-06, "loss": 1.7631, "num_input_tokens_seen": 166723584, "step": 636 }, { "epoch": 0.15818226967966229, "grad_norm": 0.4781966805458069, "learning_rate": 3.521459749779769e-06, "loss": 2.3027, "num_input_tokens_seen": 166985728, "step": 637 }, { "epoch": 0.15843059349391606, "grad_norm": 0.6974207162857056, "learning_rate": 3.5047960387195673e-06, "loss": 1.4963, "num_input_tokens_seen": 167247872, "step": 638 }, { "epoch": 0.15867891730816985, "grad_norm": 0.4340938627719879, "learning_rate": 3.488150546247778e-06, "loss": 1.6811, "num_input_tokens_seen": 167510016, "step": 639 }, { "epoch": 0.15892724112242365, "grad_norm": 0.6632753014564514, "learning_rate": 3.471523475184472e-06, "loss": 2.0226, "num_input_tokens_seen": 167772160, "step": 640 }, { "epoch": 0.15917556493667742, "grad_norm": 0.5399497747421265, "learning_rate": 3.4549150281252635e-06, "loss": 1.6728, "num_input_tokens_seen": 168034304, "step": 641 }, { "epoch": 0.15942388875093122, "grad_norm": 0.6213698387145996, "learning_rate": 3.4383254074388373e-06, "loss": 1.7282, "num_input_tokens_seen": 168296448, "step": 642 }, { "epoch": 0.159672212565185, "grad_norm": 0.8171265125274658, "learning_rate": 3.4217548152644887e-06, "loss": 1.7681, "num_input_tokens_seen": 168558592, "step": 643 }, { "epoch": 0.1599205363794388, "grad_norm": 0.36624467372894287, "learning_rate": 3.40520345350965e-06, "loss": 1.7461, "num_input_tokens_seen": 168820736, "step": 644 }, { "epoch": 0.16016886019369259, "grad_norm": 0.4452555179595947, "learning_rate": 3.3886715238474454e-06, "loss": 1.8516, "num_input_tokens_seen": 169082880, "step": 645 }, { "epoch": 0.16041718400794636, "grad_norm": 0.6282974481582642, "learning_rate": 3.372159227714218e-06, "loss": 1.589, "num_input_tokens_seen": 169345024, "step": 646 }, { "epoch": 0.16066550782220015, "grad_norm": 0.2544865012168884, "learning_rate": 3.355666766307084e-06, "loss": 1.4478, "num_input_tokens_seen": 169607168, "step": 647 }, { "epoch": 0.16091383163645392, "grad_norm": 0.7939302921295166, "learning_rate": 3.339194340581485e-06, "loss": 1.5844, "num_input_tokens_seen": 169869312, "step": 648 }, { "epoch": 0.16116215545070772, "grad_norm": 1.1076239347457886, "learning_rate": 3.322742151248726e-06, "loss": 1.8774, "num_input_tokens_seen": 170131456, "step": 649 }, { "epoch": 0.16141047926496152, "grad_norm": 0.5885612964630127, "learning_rate": 3.3063103987735433e-06, "loss": 1.8922, "num_input_tokens_seen": 170393600, "step": 650 }, { "epoch": 0.1616588030792153, "grad_norm": 0.8514074683189392, "learning_rate": 3.289899283371657e-06, "loss": 1.4842, "num_input_tokens_seen": 170655744, "step": 651 }, { "epoch": 0.1619071268934691, "grad_norm": 0.6668713092803955, "learning_rate": 3.273509005007327e-06, "loss": 1.538, "num_input_tokens_seen": 170917888, "step": 652 }, { "epoch": 0.16215545070772286, "grad_norm": 0.4702228009700775, "learning_rate": 3.2571397633909252e-06, "loss": 1.4504, "num_input_tokens_seen": 171180032, "step": 653 }, { "epoch": 0.16240377452197666, "grad_norm": 0.4500584900379181, "learning_rate": 3.2407917579764914e-06, "loss": 1.8598, "num_input_tokens_seen": 171442176, "step": 654 }, { "epoch": 0.16265209833623046, "grad_norm": 0.477428674697876, "learning_rate": 3.224465187959316e-06, "loss": 1.9269, "num_input_tokens_seen": 171704320, "step": 655 }, { "epoch": 0.16290042215048423, "grad_norm": 0.43787822127342224, "learning_rate": 3.2081602522734987e-06, "loss": 1.9365, "num_input_tokens_seen": 171966464, "step": 656 }, { "epoch": 0.16314874596473802, "grad_norm": 0.4032402038574219, "learning_rate": 3.1918771495895395e-06, "loss": 1.3458, "num_input_tokens_seen": 172228608, "step": 657 }, { "epoch": 0.1633970697789918, "grad_norm": 0.733826220035553, "learning_rate": 3.1756160783119015e-06, "loss": 1.4177, "num_input_tokens_seen": 172490752, "step": 658 }, { "epoch": 0.1636453935932456, "grad_norm": 0.34868526458740234, "learning_rate": 3.1593772365766107e-06, "loss": 1.6965, "num_input_tokens_seen": 172752896, "step": 659 }, { "epoch": 0.1638937174074994, "grad_norm": 0.8062167167663574, "learning_rate": 3.1431608222488276e-06, "loss": 1.7109, "num_input_tokens_seen": 173015040, "step": 660 }, { "epoch": 0.16414204122175316, "grad_norm": 0.6374398469924927, "learning_rate": 3.12696703292044e-06, "loss": 1.2467, "num_input_tokens_seen": 173277184, "step": 661 }, { "epoch": 0.16439036503600696, "grad_norm": 0.593471348285675, "learning_rate": 3.110796065907665e-06, "loss": 1.2338, "num_input_tokens_seen": 173539328, "step": 662 }, { "epoch": 0.16463868885026073, "grad_norm": 0.7001546621322632, "learning_rate": 3.09464811824863e-06, "loss": 1.944, "num_input_tokens_seen": 173801472, "step": 663 }, { "epoch": 0.16488701266451453, "grad_norm": 0.39311012625694275, "learning_rate": 3.078523386700982e-06, "loss": 1.8565, "num_input_tokens_seen": 174063616, "step": 664 }, { "epoch": 0.16513533647876832, "grad_norm": 0.5350921750068665, "learning_rate": 3.0624220677394854e-06, "loss": 1.7504, "num_input_tokens_seen": 174325760, "step": 665 }, { "epoch": 0.1653836602930221, "grad_norm": 0.38316506147384644, "learning_rate": 3.0463443575536324e-06, "loss": 1.6073, "num_input_tokens_seen": 174587904, "step": 666 }, { "epoch": 0.1656319841072759, "grad_norm": 0.7546458840370178, "learning_rate": 3.030290452045245e-06, "loss": 1.3026, "num_input_tokens_seen": 174850048, "step": 667 }, { "epoch": 0.16588030792152966, "grad_norm": 0.27675381302833557, "learning_rate": 3.0142605468260976e-06, "loss": 1.3484, "num_input_tokens_seen": 175112192, "step": 668 }, { "epoch": 0.16612863173578346, "grad_norm": 0.4194891154766083, "learning_rate": 2.9982548372155264e-06, "loss": 1.8019, "num_input_tokens_seen": 175374336, "step": 669 }, { "epoch": 0.16637695555003726, "grad_norm": 0.36900594830513, "learning_rate": 2.98227351823805e-06, "loss": 1.5313, "num_input_tokens_seen": 175636480, "step": 670 }, { "epoch": 0.16662527936429103, "grad_norm": 0.6106650829315186, "learning_rate": 2.966316784621e-06, "loss": 1.6545, "num_input_tokens_seen": 175898624, "step": 671 }, { "epoch": 0.16687360317854483, "grad_norm": 0.3975144624710083, "learning_rate": 2.9503848307921363e-06, "loss": 1.8634, "num_input_tokens_seen": 176160768, "step": 672 }, { "epoch": 0.1671219269927986, "grad_norm": 0.5684614777565002, "learning_rate": 2.934477850877292e-06, "loss": 1.618, "num_input_tokens_seen": 176422912, "step": 673 }, { "epoch": 0.1673702508070524, "grad_norm": 0.7491419315338135, "learning_rate": 2.918596038697995e-06, "loss": 1.8058, "num_input_tokens_seen": 176685056, "step": 674 }, { "epoch": 0.1676185746213062, "grad_norm": 0.36299964785575867, "learning_rate": 2.9027395877691143e-06, "loss": 1.6074, "num_input_tokens_seen": 176947200, "step": 675 }, { "epoch": 0.16786689843555996, "grad_norm": 1.1680899858474731, "learning_rate": 2.886908691296504e-06, "loss": 1.7054, "num_input_tokens_seen": 177209344, "step": 676 }, { "epoch": 0.16811522224981376, "grad_norm": 0.4961852729320526, "learning_rate": 2.871103542174637e-06, "loss": 1.8983, "num_input_tokens_seen": 177471488, "step": 677 }, { "epoch": 0.16836354606406753, "grad_norm": 0.42342936992645264, "learning_rate": 2.8553243329842715e-06, "loss": 1.6285, "num_input_tokens_seen": 177733632, "step": 678 }, { "epoch": 0.16861186987832133, "grad_norm": 0.6261987686157227, "learning_rate": 2.839571255990088e-06, "loss": 1.8729, "num_input_tokens_seen": 177995776, "step": 679 }, { "epoch": 0.16886019369257513, "grad_norm": 0.5163364410400391, "learning_rate": 2.8238445031383634e-06, "loss": 1.5881, "num_input_tokens_seen": 178257920, "step": 680 }, { "epoch": 0.1691085175068289, "grad_norm": 0.6004268527030945, "learning_rate": 2.8081442660546126e-06, "loss": 1.7253, "num_input_tokens_seen": 178520064, "step": 681 }, { "epoch": 0.1693568413210827, "grad_norm": 0.513282299041748, "learning_rate": 2.7924707360412743e-06, "loss": 1.3065, "num_input_tokens_seen": 178782208, "step": 682 }, { "epoch": 0.16960516513533647, "grad_norm": 0.508873701095581, "learning_rate": 2.776824104075364e-06, "loss": 1.4716, "num_input_tokens_seen": 179044352, "step": 683 }, { "epoch": 0.16985348894959026, "grad_norm": 0.43925005197525024, "learning_rate": 2.761204560806152e-06, "loss": 1.7001, "num_input_tokens_seen": 179306496, "step": 684 }, { "epoch": 0.17010181276384406, "grad_norm": 0.5974088907241821, "learning_rate": 2.7456122965528475e-06, "loss": 1.6872, "num_input_tokens_seen": 179568640, "step": 685 }, { "epoch": 0.17035013657809783, "grad_norm": 0.6642299890518188, "learning_rate": 2.7300475013022666e-06, "loss": 1.6835, "num_input_tokens_seen": 179830784, "step": 686 }, { "epoch": 0.17059846039235163, "grad_norm": 1.514657735824585, "learning_rate": 2.714510364706531e-06, "loss": 1.6302, "num_input_tokens_seen": 180092928, "step": 687 }, { "epoch": 0.1708467842066054, "grad_norm": 0.753200113773346, "learning_rate": 2.699001076080742e-06, "loss": 1.7994, "num_input_tokens_seen": 180355072, "step": 688 }, { "epoch": 0.1710951080208592, "grad_norm": 0.5442324876785278, "learning_rate": 2.683519824400693e-06, "loss": 1.1455, "num_input_tokens_seen": 180617216, "step": 689 }, { "epoch": 0.171343431835113, "grad_norm": 0.42106226086616516, "learning_rate": 2.6680667983005446e-06, "loss": 1.4338, "num_input_tokens_seen": 180879360, "step": 690 }, { "epoch": 0.17159175564936677, "grad_norm": 0.36291244626045227, "learning_rate": 2.6526421860705474e-06, "loss": 1.4588, "num_input_tokens_seen": 181141504, "step": 691 }, { "epoch": 0.17184007946362057, "grad_norm": 0.5528322458267212, "learning_rate": 2.637246175654731e-06, "loss": 1.972, "num_input_tokens_seen": 181403648, "step": 692 }, { "epoch": 0.17208840327787434, "grad_norm": 1.1730087995529175, "learning_rate": 2.6218789546486235e-06, "loss": 1.4716, "num_input_tokens_seen": 181665792, "step": 693 }, { "epoch": 0.17233672709212813, "grad_norm": 0.28138279914855957, "learning_rate": 2.6065407102969664e-06, "loss": 1.9811, "num_input_tokens_seen": 181927936, "step": 694 }, { "epoch": 0.17258505090638193, "grad_norm": 0.585191011428833, "learning_rate": 2.5912316294914232e-06, "loss": 1.413, "num_input_tokens_seen": 182190080, "step": 695 }, { "epoch": 0.1728333747206357, "grad_norm": 0.7064807415008545, "learning_rate": 2.5759518987683154e-06, "loss": 1.7552, "num_input_tokens_seen": 182452224, "step": 696 }, { "epoch": 0.1730816985348895, "grad_norm": 0.2733915448188782, "learning_rate": 2.560701704306336e-06, "loss": 1.5639, "num_input_tokens_seen": 182714368, "step": 697 }, { "epoch": 0.17333002234914327, "grad_norm": 0.37990668416023254, "learning_rate": 2.545481231924296e-06, "loss": 2.0107, "num_input_tokens_seen": 182976512, "step": 698 }, { "epoch": 0.17357834616339707, "grad_norm": 0.5895552039146423, "learning_rate": 2.5302906670788463e-06, "loss": 2.0406, "num_input_tokens_seen": 183238656, "step": 699 }, { "epoch": 0.17382666997765087, "grad_norm": 0.6895752549171448, "learning_rate": 2.5151301948622235e-06, "loss": 1.9766, "num_input_tokens_seen": 183500800, "step": 700 }, { "epoch": 0.17407499379190464, "grad_norm": 0.46613720059394836, "learning_rate": 2.5000000000000015e-06, "loss": 2.0084, "num_input_tokens_seen": 183762944, "step": 701 }, { "epoch": 0.17432331760615843, "grad_norm": 0.5650424957275391, "learning_rate": 2.484900266848825e-06, "loss": 1.7718, "num_input_tokens_seen": 184025088, "step": 702 }, { "epoch": 0.1745716414204122, "grad_norm": 0.5093001127243042, "learning_rate": 2.469831179394182e-06, "loss": 1.5874, "num_input_tokens_seen": 184287232, "step": 703 }, { "epoch": 0.174819965234666, "grad_norm": 0.7076382040977478, "learning_rate": 2.4547929212481436e-06, "loss": 1.5776, "num_input_tokens_seen": 184549376, "step": 704 }, { "epoch": 0.1750682890489198, "grad_norm": 0.7036752700805664, "learning_rate": 2.4397856756471435e-06, "loss": 1.0319, "num_input_tokens_seen": 184811520, "step": 705 }, { "epoch": 0.17531661286317357, "grad_norm": 0.6339607238769531, "learning_rate": 2.424809625449729e-06, "loss": 1.4841, "num_input_tokens_seen": 185073664, "step": 706 }, { "epoch": 0.17556493667742737, "grad_norm": 0.6772640347480774, "learning_rate": 2.40986495313435e-06, "loss": 1.398, "num_input_tokens_seen": 185335808, "step": 707 }, { "epoch": 0.17581326049168114, "grad_norm": 0.626737117767334, "learning_rate": 2.39495184079712e-06, "loss": 1.9448, "num_input_tokens_seen": 185597952, "step": 708 }, { "epoch": 0.17606158430593494, "grad_norm": 0.8285679221153259, "learning_rate": 2.380070470149605e-06, "loss": 1.596, "num_input_tokens_seen": 185860096, "step": 709 }, { "epoch": 0.17630990812018874, "grad_norm": 0.47030216455459595, "learning_rate": 2.3652210225166122e-06, "loss": 1.6729, "num_input_tokens_seen": 186122240, "step": 710 }, { "epoch": 0.1765582319344425, "grad_norm": 0.36715853214263916, "learning_rate": 2.3504036788339763e-06, "loss": 1.375, "num_input_tokens_seen": 186384384, "step": 711 }, { "epoch": 0.1768065557486963, "grad_norm": 0.5815637707710266, "learning_rate": 2.3356186196463497e-06, "loss": 1.6501, "num_input_tokens_seen": 186646528, "step": 712 }, { "epoch": 0.17705487956295007, "grad_norm": 0.5376741290092468, "learning_rate": 2.320866025105016e-06, "loss": 1.7278, "num_input_tokens_seen": 186908672, "step": 713 }, { "epoch": 0.17730320337720387, "grad_norm": 0.5446439385414124, "learning_rate": 2.3061460749656844e-06, "loss": 1.9617, "num_input_tokens_seen": 187170816, "step": 714 }, { "epoch": 0.17755152719145767, "grad_norm": 0.8022477030754089, "learning_rate": 2.2914589485863015e-06, "loss": 1.8491, "num_input_tokens_seen": 187432960, "step": 715 }, { "epoch": 0.17779985100571144, "grad_norm": 0.40645989775657654, "learning_rate": 2.2768048249248648e-06, "loss": 1.5563, "num_input_tokens_seen": 187695104, "step": 716 }, { "epoch": 0.17804817481996524, "grad_norm": 0.5382466316223145, "learning_rate": 2.2621838825372496e-06, "loss": 1.5754, "num_input_tokens_seen": 187957248, "step": 717 }, { "epoch": 0.178296498634219, "grad_norm": 0.6198011636734009, "learning_rate": 2.2475962995750224e-06, "loss": 1.8925, "num_input_tokens_seen": 188219392, "step": 718 }, { "epoch": 0.1785448224484728, "grad_norm": 0.4357577860355377, "learning_rate": 2.23304225378328e-06, "loss": 1.9635, "num_input_tokens_seen": 188481536, "step": 719 }, { "epoch": 0.1787931462627266, "grad_norm": 0.5042324662208557, "learning_rate": 2.218521922498476e-06, "loss": 1.4101, "num_input_tokens_seen": 188743680, "step": 720 }, { "epoch": 0.17904147007698037, "grad_norm": 0.5827696323394775, "learning_rate": 2.204035482646267e-06, "loss": 1.6445, "num_input_tokens_seen": 189005824, "step": 721 }, { "epoch": 0.17928979389123417, "grad_norm": 0.5384161472320557, "learning_rate": 2.1895831107393485e-06, "loss": 1.5907, "num_input_tokens_seen": 189267968, "step": 722 }, { "epoch": 0.17953811770548794, "grad_norm": 0.5630956292152405, "learning_rate": 2.175164982875311e-06, "loss": 1.4097, "num_input_tokens_seen": 189530112, "step": 723 }, { "epoch": 0.17978644151974174, "grad_norm": 0.43942683935165405, "learning_rate": 2.1607812747344955e-06, "loss": 1.393, "num_input_tokens_seen": 189792256, "step": 724 }, { "epoch": 0.18003476533399554, "grad_norm": 0.2822117209434509, "learning_rate": 2.146432161577842e-06, "loss": 1.4047, "num_input_tokens_seen": 190054400, "step": 725 }, { "epoch": 0.1802830891482493, "grad_norm": 0.7013232707977295, "learning_rate": 2.132117818244771e-06, "loss": 1.8612, "num_input_tokens_seen": 190316544, "step": 726 }, { "epoch": 0.1805314129625031, "grad_norm": 0.7166525721549988, "learning_rate": 2.1178384191510344e-06, "loss": 1.5652, "num_input_tokens_seen": 190578688, "step": 727 }, { "epoch": 0.18077973677675688, "grad_norm": 0.8852736949920654, "learning_rate": 2.103594138286607e-06, "loss": 1.8348, "num_input_tokens_seen": 190840832, "step": 728 }, { "epoch": 0.18102806059101068, "grad_norm": 0.7377064228057861, "learning_rate": 2.0893851492135536e-06, "loss": 1.5408, "num_input_tokens_seen": 191102976, "step": 729 }, { "epoch": 0.18127638440526447, "grad_norm": 0.31571272015571594, "learning_rate": 2.075211625063923e-06, "loss": 1.4516, "num_input_tokens_seen": 191365120, "step": 730 }, { "epoch": 0.18152470821951824, "grad_norm": 0.4377698004245758, "learning_rate": 2.061073738537635e-06, "loss": 1.8452, "num_input_tokens_seen": 191627264, "step": 731 }, { "epoch": 0.18177303203377204, "grad_norm": 0.6246635317802429, "learning_rate": 2.046971661900373e-06, "loss": 1.0213, "num_input_tokens_seen": 191889408, "step": 732 }, { "epoch": 0.1820213558480258, "grad_norm": 0.3734930157661438, "learning_rate": 2.0329055669814936e-06, "loss": 1.3397, "num_input_tokens_seen": 192151552, "step": 733 }, { "epoch": 0.1822696796622796, "grad_norm": 0.6088233590126038, "learning_rate": 2.0188756251719204e-06, "loss": 1.5803, "num_input_tokens_seen": 192413696, "step": 734 }, { "epoch": 0.1825180034765334, "grad_norm": 0.6044638156890869, "learning_rate": 2.0048820074220716e-06, "loss": 1.629, "num_input_tokens_seen": 192675840, "step": 735 }, { "epoch": 0.18276632729078718, "grad_norm": 0.5411587357521057, "learning_rate": 1.990924884239758e-06, "loss": 1.5093, "num_input_tokens_seen": 192937984, "step": 736 }, { "epoch": 0.18301465110504098, "grad_norm": 0.7263090014457703, "learning_rate": 1.977004425688126e-06, "loss": 1.6338, "num_input_tokens_seen": 193200128, "step": 737 }, { "epoch": 0.18326297491929475, "grad_norm": 0.5330092310905457, "learning_rate": 1.9631208013835677e-06, "loss": 1.6223, "num_input_tokens_seen": 193462272, "step": 738 }, { "epoch": 0.18351129873354854, "grad_norm": 0.8732761740684509, "learning_rate": 1.9492741804936623e-06, "loss": 1.81, "num_input_tokens_seen": 193724416, "step": 739 }, { "epoch": 0.18375962254780234, "grad_norm": 0.4876832067966461, "learning_rate": 1.9354647317351187e-06, "loss": 1.8895, "num_input_tokens_seen": 193986560, "step": 740 }, { "epoch": 0.1840079463620561, "grad_norm": 0.5221810340881348, "learning_rate": 1.9216926233717087e-06, "loss": 1.7537, "num_input_tokens_seen": 194248704, "step": 741 }, { "epoch": 0.1842562701763099, "grad_norm": 0.7166823744773865, "learning_rate": 1.90795802321223e-06, "loss": 1.6196, "num_input_tokens_seen": 194510848, "step": 742 }, { "epoch": 0.1845045939905637, "grad_norm": 0.5422106981277466, "learning_rate": 1.8942610986084487e-06, "loss": 1.7422, "num_input_tokens_seen": 194772992, "step": 743 }, { "epoch": 0.18475291780481748, "grad_norm": 0.5047271847724915, "learning_rate": 1.8806020164530702e-06, "loss": 1.1779, "num_input_tokens_seen": 195035136, "step": 744 }, { "epoch": 0.18500124161907128, "grad_norm": 0.6293126344680786, "learning_rate": 1.8669809431776991e-06, "loss": 1.923, "num_input_tokens_seen": 195297280, "step": 745 }, { "epoch": 0.18524956543332505, "grad_norm": 0.41052141785621643, "learning_rate": 1.8533980447508138e-06, "loss": 1.3557, "num_input_tokens_seen": 195559424, "step": 746 }, { "epoch": 0.18549788924757885, "grad_norm": 0.45604202151298523, "learning_rate": 1.8398534866757455e-06, "loss": 1.7005, "num_input_tokens_seen": 195821568, "step": 747 }, { "epoch": 0.18574621306183264, "grad_norm": 0.730992317199707, "learning_rate": 1.8263474339886628e-06, "loss": 2.0785, "num_input_tokens_seen": 196083712, "step": 748 }, { "epoch": 0.1859945368760864, "grad_norm": 0.5063241124153137, "learning_rate": 1.8128800512565514e-06, "loss": 2.1056, "num_input_tokens_seen": 196345856, "step": 749 }, { "epoch": 0.1862428606903402, "grad_norm": 0.2811354100704193, "learning_rate": 1.799451502575222e-06, "loss": 1.2958, "num_input_tokens_seen": 196608000, "step": 750 }, { "epoch": 0.18649118450459398, "grad_norm": 0.5314778089523315, "learning_rate": 1.7860619515673034e-06, "loss": 1.7808, "num_input_tokens_seen": 196870144, "step": 751 }, { "epoch": 0.18673950831884778, "grad_norm": 1.0328506231307983, "learning_rate": 1.7727115613802465e-06, "loss": 2.0584, "num_input_tokens_seen": 197132288, "step": 752 }, { "epoch": 0.18698783213310158, "grad_norm": 0.7677520513534546, "learning_rate": 1.7594004946843458e-06, "loss": 1.6937, "num_input_tokens_seen": 197394432, "step": 753 }, { "epoch": 0.18723615594735535, "grad_norm": 0.6427842974662781, "learning_rate": 1.746128913670746e-06, "loss": 1.8869, "num_input_tokens_seen": 197656576, "step": 754 }, { "epoch": 0.18748447976160915, "grad_norm": 1.0200343132019043, "learning_rate": 1.7328969800494727e-06, "loss": 1.796, "num_input_tokens_seen": 197918720, "step": 755 }, { "epoch": 0.18773280357586292, "grad_norm": 0.8236029148101807, "learning_rate": 1.7197048550474643e-06, "loss": 1.6866, "num_input_tokens_seen": 198180864, "step": 756 }, { "epoch": 0.18798112739011671, "grad_norm": 1.2460148334503174, "learning_rate": 1.7065526994065973e-06, "loss": 1.3428, "num_input_tokens_seen": 198443008, "step": 757 }, { "epoch": 0.1882294512043705, "grad_norm": 0.4800860285758972, "learning_rate": 1.6934406733817417e-06, "loss": 1.4296, "num_input_tokens_seen": 198705152, "step": 758 }, { "epoch": 0.18847777501862428, "grad_norm": 0.9592916965484619, "learning_rate": 1.680368936738792e-06, "loss": 1.4531, "num_input_tokens_seen": 198967296, "step": 759 }, { "epoch": 0.18872609883287808, "grad_norm": 0.6796224117279053, "learning_rate": 1.6673376487527382e-06, "loss": 1.4354, "num_input_tokens_seen": 199229440, "step": 760 }, { "epoch": 0.18897442264713185, "grad_norm": 0.6231174468994141, "learning_rate": 1.6543469682057105e-06, "loss": 1.2075, "num_input_tokens_seen": 199491584, "step": 761 }, { "epoch": 0.18922274646138565, "grad_norm": 0.7283173203468323, "learning_rate": 1.6413970533850498e-06, "loss": 1.7713, "num_input_tokens_seen": 199753728, "step": 762 }, { "epoch": 0.18947107027563945, "grad_norm": 0.6808333396911621, "learning_rate": 1.6284880620813847e-06, "loss": 1.7076, "num_input_tokens_seen": 200015872, "step": 763 }, { "epoch": 0.18971939408989322, "grad_norm": 0.7591480612754822, "learning_rate": 1.6156201515866971e-06, "loss": 1.4767, "num_input_tokens_seen": 200278016, "step": 764 }, { "epoch": 0.18996771790414702, "grad_norm": 3.4821579456329346, "learning_rate": 1.6027934786924187e-06, "loss": 1.7879, "num_input_tokens_seen": 200540160, "step": 765 }, { "epoch": 0.19021604171840079, "grad_norm": 0.5467571020126343, "learning_rate": 1.5900081996875083e-06, "loss": 1.6919, "num_input_tokens_seen": 200802304, "step": 766 }, { "epoch": 0.19046436553265458, "grad_norm": 0.4840015172958374, "learning_rate": 1.5772644703565564e-06, "loss": 1.9006, "num_input_tokens_seen": 201064448, "step": 767 }, { "epoch": 0.19071268934690838, "grad_norm": 0.4922867715358734, "learning_rate": 1.5645624459778858e-06, "loss": 1.505, "num_input_tokens_seen": 201326592, "step": 768 }, { "epoch": 0.19096101316116215, "grad_norm": 0.8669015169143677, "learning_rate": 1.551902281321651e-06, "loss": 1.5521, "num_input_tokens_seen": 201588736, "step": 769 }, { "epoch": 0.19120933697541595, "grad_norm": 0.7580779194831848, "learning_rate": 1.5392841306479667e-06, "loss": 1.5425, "num_input_tokens_seen": 201850880, "step": 770 }, { "epoch": 0.19145766078966972, "grad_norm": 0.5388744473457336, "learning_rate": 1.5267081477050132e-06, "loss": 1.6623, "num_input_tokens_seen": 202113024, "step": 771 }, { "epoch": 0.19170598460392352, "grad_norm": 0.36039191484451294, "learning_rate": 1.514174485727178e-06, "loss": 1.9833, "num_input_tokens_seen": 202375168, "step": 772 }, { "epoch": 0.19195430841817732, "grad_norm": 0.7689981460571289, "learning_rate": 1.5016832974331725e-06, "loss": 1.9371, "num_input_tokens_seen": 202637312, "step": 773 }, { "epoch": 0.1922026322324311, "grad_norm": 0.510267436504364, "learning_rate": 1.489234735024188e-06, "loss": 1.7589, "num_input_tokens_seen": 202899456, "step": 774 }, { "epoch": 0.19245095604668488, "grad_norm": 1.7721202373504639, "learning_rate": 1.4768289501820265e-06, "loss": 1.8343, "num_input_tokens_seen": 203161600, "step": 775 }, { "epoch": 0.19269927986093865, "grad_norm": 0.35707518458366394, "learning_rate": 1.4644660940672628e-06, "loss": 1.3319, "num_input_tokens_seen": 203423744, "step": 776 }, { "epoch": 0.19294760367519245, "grad_norm": 0.39053982496261597, "learning_rate": 1.4521463173173966e-06, "loss": 1.4868, "num_input_tokens_seen": 203685888, "step": 777 }, { "epoch": 0.19319592748944625, "grad_norm": 1.0973213911056519, "learning_rate": 1.4398697700450181e-06, "loss": 1.2248, "num_input_tokens_seen": 203948032, "step": 778 }, { "epoch": 0.19344425130370002, "grad_norm": 0.4421512186527252, "learning_rate": 1.4276366018359845e-06, "loss": 1.513, "num_input_tokens_seen": 204210176, "step": 779 }, { "epoch": 0.19369257511795382, "grad_norm": 0.5916955471038818, "learning_rate": 1.4154469617475864e-06, "loss": 1.5881, "num_input_tokens_seen": 204472320, "step": 780 }, { "epoch": 0.1939408989322076, "grad_norm": 0.749496340751648, "learning_rate": 1.4033009983067454e-06, "loss": 2.0316, "num_input_tokens_seen": 204734464, "step": 781 }, { "epoch": 0.1941892227464614, "grad_norm": 0.4848249554634094, "learning_rate": 1.3911988595081894e-06, "loss": 1.5584, "num_input_tokens_seen": 204996608, "step": 782 }, { "epoch": 0.19443754656071519, "grad_norm": 0.5584369897842407, "learning_rate": 1.3791406928126638e-06, "loss": 1.6737, "num_input_tokens_seen": 205258752, "step": 783 }, { "epoch": 0.19468587037496896, "grad_norm": 0.7734516263008118, "learning_rate": 1.3671266451451209e-06, "loss": 1.7325, "num_input_tokens_seen": 205520896, "step": 784 }, { "epoch": 0.19493419418922275, "grad_norm": 0.4235021471977234, "learning_rate": 1.3551568628929434e-06, "loss": 1.7884, "num_input_tokens_seen": 205783040, "step": 785 }, { "epoch": 0.19518251800347652, "grad_norm": 0.5882217288017273, "learning_rate": 1.3432314919041478e-06, "loss": 1.5367, "num_input_tokens_seen": 206045184, "step": 786 }, { "epoch": 0.19543084181773032, "grad_norm": 0.382315456867218, "learning_rate": 1.3313506774856177e-06, "loss": 1.562, "num_input_tokens_seen": 206307328, "step": 787 }, { "epoch": 0.19567916563198412, "grad_norm": 0.5313207507133484, "learning_rate": 1.3195145644013286e-06, "loss": 0.9386, "num_input_tokens_seen": 206569472, "step": 788 }, { "epoch": 0.1959274894462379, "grad_norm": 0.5512835383415222, "learning_rate": 1.3077232968705805e-06, "loss": 1.3328, "num_input_tokens_seen": 206831616, "step": 789 }, { "epoch": 0.1961758132604917, "grad_norm": 0.6115958094596863, "learning_rate": 1.2959770185662502e-06, "loss": 1.6127, "num_input_tokens_seen": 207093760, "step": 790 }, { "epoch": 0.19642413707474546, "grad_norm": 1.8349878787994385, "learning_rate": 1.2842758726130283e-06, "loss": 1.3873, "num_input_tokens_seen": 207355904, "step": 791 }, { "epoch": 0.19667246088899926, "grad_norm": 0.6167082786560059, "learning_rate": 1.2726200015856893e-06, "loss": 1.6641, "num_input_tokens_seen": 207618048, "step": 792 }, { "epoch": 0.19692078470325305, "grad_norm": 0.27391648292541504, "learning_rate": 1.2610095475073415e-06, "loss": 1.7774, "num_input_tokens_seen": 207880192, "step": 793 }, { "epoch": 0.19716910851750682, "grad_norm": 0.4807204306125641, "learning_rate": 1.2494446518477022e-06, "loss": 1.5564, "num_input_tokens_seen": 208142336, "step": 794 }, { "epoch": 0.19741743233176062, "grad_norm": 0.3716016113758087, "learning_rate": 1.2379254555213788e-06, "loss": 1.6801, "num_input_tokens_seen": 208404480, "step": 795 }, { "epoch": 0.1976657561460144, "grad_norm": 0.6262674331665039, "learning_rate": 1.22645209888614e-06, "loss": 1.6579, "num_input_tokens_seen": 208666624, "step": 796 }, { "epoch": 0.1979140799602682, "grad_norm": 0.49892279505729675, "learning_rate": 1.2150247217412186e-06, "loss": 1.4113, "num_input_tokens_seen": 208928768, "step": 797 }, { "epoch": 0.198162403774522, "grad_norm": 0.28231680393218994, "learning_rate": 1.203643463325596e-06, "loss": 1.2046, "num_input_tokens_seen": 209190912, "step": 798 }, { "epoch": 0.19841072758877576, "grad_norm": 0.39441466331481934, "learning_rate": 1.1923084623163172e-06, "loss": 1.25, "num_input_tokens_seen": 209453056, "step": 799 }, { "epoch": 0.19865905140302956, "grad_norm": 0.8825608491897583, "learning_rate": 1.1810198568267906e-06, "loss": 1.8506, "num_input_tokens_seen": 209715200, "step": 800 }, { "epoch": 0.19890737521728333, "grad_norm": 0.3593462109565735, "learning_rate": 1.1697777844051105e-06, "loss": 0.9988, "num_input_tokens_seen": 209977344, "step": 801 }, { "epoch": 0.19915569903153713, "grad_norm": 0.42397427558898926, "learning_rate": 1.1585823820323845e-06, "loss": 1.5294, "num_input_tokens_seen": 210239488, "step": 802 }, { "epoch": 0.19940402284579092, "grad_norm": 0.5862644910812378, "learning_rate": 1.1474337861210543e-06, "loss": 1.8455, "num_input_tokens_seen": 210501632, "step": 803 }, { "epoch": 0.1996523466600447, "grad_norm": 0.7144994139671326, "learning_rate": 1.136332132513245e-06, "loss": 1.6364, "num_input_tokens_seen": 210763776, "step": 804 }, { "epoch": 0.1999006704742985, "grad_norm": 0.6182805299758911, "learning_rate": 1.1252775564791023e-06, "loss": 1.533, "num_input_tokens_seen": 211025920, "step": 805 }, { "epoch": 0.20014899428855226, "grad_norm": 2.216693878173828, "learning_rate": 1.1142701927151456e-06, "loss": 1.8744, "num_input_tokens_seen": 211288064, "step": 806 }, { "epoch": 0.20039731810280606, "grad_norm": 0.4277917444705963, "learning_rate": 1.1033101753426285e-06, "loss": 1.5751, "num_input_tokens_seen": 211550208, "step": 807 }, { "epoch": 0.20064564191705986, "grad_norm": 0.5398479104042053, "learning_rate": 1.0923976379059059e-06, "loss": 1.5959, "num_input_tokens_seen": 211812352, "step": 808 }, { "epoch": 0.20089396573131363, "grad_norm": 0.38512706756591797, "learning_rate": 1.0815327133708015e-06, "loss": 1.5632, "num_input_tokens_seen": 212074496, "step": 809 }, { "epoch": 0.20114228954556743, "grad_norm": 0.28112220764160156, "learning_rate": 1.0707155341229902e-06, "loss": 1.5852, "num_input_tokens_seen": 212336640, "step": 810 }, { "epoch": 0.2013906133598212, "grad_norm": 0.3636101186275482, "learning_rate": 1.0599462319663906e-06, "loss": 1.659, "num_input_tokens_seen": 212598784, "step": 811 }, { "epoch": 0.201638937174075, "grad_norm": 0.632411003112793, "learning_rate": 1.049224938121548e-06, "loss": 1.7422, "num_input_tokens_seen": 212860928, "step": 812 }, { "epoch": 0.2018872609883288, "grad_norm": 0.4831806719303131, "learning_rate": 1.0385517832240472e-06, "loss": 1.6995, "num_input_tokens_seen": 213123072, "step": 813 }, { "epoch": 0.20213558480258256, "grad_norm": 0.6856269836425781, "learning_rate": 1.0279268973229089e-06, "loss": 1.4744, "num_input_tokens_seen": 213385216, "step": 814 }, { "epoch": 0.20238390861683636, "grad_norm": 0.706402599811554, "learning_rate": 1.0173504098790188e-06, "loss": 1.806, "num_input_tokens_seen": 213647360, "step": 815 }, { "epoch": 0.20263223243109013, "grad_norm": 0.6212208271026611, "learning_rate": 1.006822449763537e-06, "loss": 1.4372, "num_input_tokens_seen": 213909504, "step": 816 }, { "epoch": 0.20288055624534393, "grad_norm": 0.4880361258983612, "learning_rate": 9.963431452563331e-07, "loss": 1.7287, "num_input_tokens_seen": 214171648, "step": 817 }, { "epoch": 0.20312888005959773, "grad_norm": 0.5508049726486206, "learning_rate": 9.859126240444284e-07, "loss": 1.4213, "num_input_tokens_seen": 214433792, "step": 818 }, { "epoch": 0.2033772038738515, "grad_norm": 0.40056681632995605, "learning_rate": 9.7553101322043e-07, "loss": 1.6258, "num_input_tokens_seen": 214695936, "step": 819 }, { "epoch": 0.2036255276881053, "grad_norm": 0.5356477499008179, "learning_rate": 9.651984392809916e-07, "loss": 1.5275, "num_input_tokens_seen": 214958080, "step": 820 }, { "epoch": 0.20387385150235907, "grad_norm": 0.7192147374153137, "learning_rate": 9.549150281252633e-07, "loss": 1.675, "num_input_tokens_seen": 215220224, "step": 821 }, { "epoch": 0.20412217531661286, "grad_norm": 0.6497499942779541, "learning_rate": 9.446809050533679e-07, "loss": 2.0607, "num_input_tokens_seen": 215482368, "step": 822 }, { "epoch": 0.20437049913086666, "grad_norm": 0.6447398662567139, "learning_rate": 9.344961947648624e-07, "loss": 1.3855, "num_input_tokens_seen": 215744512, "step": 823 }, { "epoch": 0.20461882294512043, "grad_norm": 1.1430552005767822, "learning_rate": 9.243610213572285e-07, "loss": 1.6178, "num_input_tokens_seen": 216006656, "step": 824 }, { "epoch": 0.20486714675937423, "grad_norm": 0.6385335922241211, "learning_rate": 9.142755083243577e-07, "loss": 1.5856, "num_input_tokens_seen": 216268800, "step": 825 }, { "epoch": 0.205115470573628, "grad_norm": 1.0095096826553345, "learning_rate": 9.042397785550405e-07, "loss": 1.6267, "num_input_tokens_seen": 216530944, "step": 826 }, { "epoch": 0.2053637943878818, "grad_norm": 0.7737463712692261, "learning_rate": 8.942539543314799e-07, "loss": 1.3607, "num_input_tokens_seen": 216793088, "step": 827 }, { "epoch": 0.2056121182021356, "grad_norm": 0.6710413694381714, "learning_rate": 8.843181573277904e-07, "loss": 1.7918, "num_input_tokens_seen": 217055232, "step": 828 }, { "epoch": 0.20586044201638937, "grad_norm": 0.5380450487136841, "learning_rate": 8.744325086085248e-07, "loss": 1.8734, "num_input_tokens_seen": 217317376, "step": 829 }, { "epoch": 0.20610876583064316, "grad_norm": 0.769334614276886, "learning_rate": 8.645971286271903e-07, "loss": 1.4816, "num_input_tokens_seen": 217579520, "step": 830 }, { "epoch": 0.20635708964489693, "grad_norm": 0.49632272124290466, "learning_rate": 8.54812137224792e-07, "loss": 1.7042, "num_input_tokens_seen": 217841664, "step": 831 }, { "epoch": 0.20660541345915073, "grad_norm": 0.30258211493492126, "learning_rate": 8.450776536283594e-07, "loss": 1.5544, "num_input_tokens_seen": 218103808, "step": 832 }, { "epoch": 0.20685373727340453, "grad_norm": 0.5039138793945312, "learning_rate": 8.353937964495029e-07, "loss": 1.6497, "num_input_tokens_seen": 218365952, "step": 833 }, { "epoch": 0.2071020610876583, "grad_norm": 0.5678662657737732, "learning_rate": 8.25760683682968e-07, "loss": 1.9249, "num_input_tokens_seen": 218628096, "step": 834 }, { "epoch": 0.2073503849019121, "grad_norm": 0.7532708644866943, "learning_rate": 8.161784327051919e-07, "loss": 1.5489, "num_input_tokens_seen": 218890240, "step": 835 }, { "epoch": 0.20759870871616587, "grad_norm": 0.3733159303665161, "learning_rate": 8.066471602728804e-07, "loss": 2.1036, "num_input_tokens_seen": 219152384, "step": 836 }, { "epoch": 0.20784703253041967, "grad_norm": 0.52272629737854, "learning_rate": 7.971669825215789e-07, "loss": 1.5633, "num_input_tokens_seen": 219414528, "step": 837 }, { "epoch": 0.20809535634467347, "grad_norm": 0.6050902605056763, "learning_rate": 7.877380149642628e-07, "loss": 1.4245, "num_input_tokens_seen": 219676672, "step": 838 }, { "epoch": 0.20834368015892724, "grad_norm": 0.6302130222320557, "learning_rate": 7.783603724899258e-07, "loss": 1.3682, "num_input_tokens_seen": 219938816, "step": 839 }, { "epoch": 0.20859200397318103, "grad_norm": 0.2699492871761322, "learning_rate": 7.690341693621805e-07, "loss": 1.344, "num_input_tokens_seen": 220200960, "step": 840 }, { "epoch": 0.2088403277874348, "grad_norm": 0.7535718679428101, "learning_rate": 7.597595192178702e-07, "loss": 1.751, "num_input_tokens_seen": 220463104, "step": 841 }, { "epoch": 0.2090886516016886, "grad_norm": 0.9196535348892212, "learning_rate": 7.505365350656813e-07, "loss": 1.6558, "num_input_tokens_seen": 220725248, "step": 842 }, { "epoch": 0.2093369754159424, "grad_norm": 0.6124866604804993, "learning_rate": 7.413653292847617e-07, "loss": 1.4843, "num_input_tokens_seen": 220987392, "step": 843 }, { "epoch": 0.20958529923019617, "grad_norm": 0.830053448677063, "learning_rate": 7.322460136233622e-07, "loss": 1.583, "num_input_tokens_seen": 221249536, "step": 844 }, { "epoch": 0.20983362304444997, "grad_norm": 0.7714657783508301, "learning_rate": 7.23178699197467e-07, "loss": 1.763, "num_input_tokens_seen": 221511680, "step": 845 }, { "epoch": 0.21008194685870374, "grad_norm": 0.3930104970932007, "learning_rate": 7.141634964894389e-07, "loss": 1.381, "num_input_tokens_seen": 221773824, "step": 846 }, { "epoch": 0.21033027067295754, "grad_norm": 0.6444476246833801, "learning_rate": 7.052005153466779e-07, "loss": 1.7138, "num_input_tokens_seen": 222035968, "step": 847 }, { "epoch": 0.21057859448721133, "grad_norm": 0.7626250982284546, "learning_rate": 6.962898649802824e-07, "loss": 1.3164, "num_input_tokens_seen": 222298112, "step": 848 }, { "epoch": 0.2108269183014651, "grad_norm": 0.5761337876319885, "learning_rate": 6.874316539637127e-07, "loss": 1.4375, "num_input_tokens_seen": 222560256, "step": 849 }, { "epoch": 0.2110752421157189, "grad_norm": 0.5399318337440491, "learning_rate": 6.786259902314768e-07, "loss": 1.5317, "num_input_tokens_seen": 222822400, "step": 850 }, { "epoch": 0.21132356592997267, "grad_norm": 0.6818238496780396, "learning_rate": 6.698729810778065e-07, "loss": 1.5059, "num_input_tokens_seen": 223084544, "step": 851 }, { "epoch": 0.21157188974422647, "grad_norm": 0.5061662197113037, "learning_rate": 6.611727331553585e-07, "loss": 1.7645, "num_input_tokens_seen": 223346688, "step": 852 }, { "epoch": 0.21182021355848027, "grad_norm": 0.7583996057510376, "learning_rate": 6.52525352473905e-07, "loss": 1.7479, "num_input_tokens_seen": 223608832, "step": 853 }, { "epoch": 0.21206853737273404, "grad_norm": 0.4888990819454193, "learning_rate": 6.439309443990532e-07, "loss": 1.4111, "num_input_tokens_seen": 223870976, "step": 854 }, { "epoch": 0.21231686118698784, "grad_norm": 0.5152997374534607, "learning_rate": 6.353896136509524e-07, "loss": 1.4035, "num_input_tokens_seen": 224133120, "step": 855 }, { "epoch": 0.2125651850012416, "grad_norm": 0.8078181147575378, "learning_rate": 6.269014643030214e-07, "loss": 1.5448, "num_input_tokens_seen": 224395264, "step": 856 }, { "epoch": 0.2128135088154954, "grad_norm": 0.38689664006233215, "learning_rate": 6.184665997806832e-07, "loss": 1.5432, "num_input_tokens_seen": 224657408, "step": 857 }, { "epoch": 0.2130618326297492, "grad_norm": 0.5154587626457214, "learning_rate": 6.100851228600974e-07, "loss": 1.6356, "num_input_tokens_seen": 224919552, "step": 858 }, { "epoch": 0.21331015644400297, "grad_norm": 0.7129220366477966, "learning_rate": 6.017571356669183e-07, "loss": 1.4454, "num_input_tokens_seen": 225181696, "step": 859 }, { "epoch": 0.21355848025825677, "grad_norm": 0.4863054156303406, "learning_rate": 5.934827396750392e-07, "loss": 1.6839, "num_input_tokens_seen": 225443840, "step": 860 }, { "epoch": 0.21380680407251054, "grad_norm": 0.461911678314209, "learning_rate": 5.852620357053651e-07, "loss": 1.4632, "num_input_tokens_seen": 225705984, "step": 861 }, { "epoch": 0.21405512788676434, "grad_norm": 0.4267142415046692, "learning_rate": 5.770951239245803e-07, "loss": 1.5694, "num_input_tokens_seen": 225968128, "step": 862 }, { "epoch": 0.21430345170101814, "grad_norm": 0.4966636300086975, "learning_rate": 5.689821038439264e-07, "loss": 1.5516, "num_input_tokens_seen": 226230272, "step": 863 }, { "epoch": 0.2145517755152719, "grad_norm": 0.5140483379364014, "learning_rate": 5.609230743179939e-07, "loss": 1.3778, "num_input_tokens_seen": 226492416, "step": 864 }, { "epoch": 0.2148000993295257, "grad_norm": 0.6450092196464539, "learning_rate": 5.529181335435124e-07, "loss": 1.3893, "num_input_tokens_seen": 226754560, "step": 865 }, { "epoch": 0.21504842314377948, "grad_norm": 0.5973859429359436, "learning_rate": 5.449673790581611e-07, "loss": 1.5149, "num_input_tokens_seen": 227016704, "step": 866 }, { "epoch": 0.21529674695803327, "grad_norm": 1.7493269443511963, "learning_rate": 5.370709077393721e-07, "loss": 1.5618, "num_input_tokens_seen": 227278848, "step": 867 }, { "epoch": 0.21554507077228707, "grad_norm": 0.4318649172782898, "learning_rate": 5.292288158031595e-07, "loss": 1.1686, "num_input_tokens_seen": 227540992, "step": 868 }, { "epoch": 0.21579339458654084, "grad_norm": 0.3847927749156952, "learning_rate": 5.214411988029355e-07, "loss": 1.7999, "num_input_tokens_seen": 227803136, "step": 869 }, { "epoch": 0.21604171840079464, "grad_norm": 0.596502423286438, "learning_rate": 5.137081516283582e-07, "loss": 1.5796, "num_input_tokens_seen": 228065280, "step": 870 }, { "epoch": 0.2162900422150484, "grad_norm": 0.6176612377166748, "learning_rate": 5.06029768504166e-07, "loss": 1.7902, "num_input_tokens_seen": 228327424, "step": 871 }, { "epoch": 0.2165383660293022, "grad_norm": 0.6072035431861877, "learning_rate": 4.984061429890324e-07, "loss": 1.7899, "num_input_tokens_seen": 228589568, "step": 872 }, { "epoch": 0.216786689843556, "grad_norm": 0.3984832167625427, "learning_rate": 4.908373679744316e-07, "loss": 1.9725, "num_input_tokens_seen": 228851712, "step": 873 }, { "epoch": 0.21703501365780978, "grad_norm": 0.7739282250404358, "learning_rate": 4.833235356834959e-07, "loss": 1.8129, "num_input_tokens_seen": 229113856, "step": 874 }, { "epoch": 0.21728333747206358, "grad_norm": 1.2236608266830444, "learning_rate": 4.758647376699033e-07, "loss": 1.7323, "num_input_tokens_seen": 229376000, "step": 875 }, { "epoch": 0.21753166128631735, "grad_norm": 0.541125476360321, "learning_rate": 4.6846106481675035e-07, "loss": 1.4017, "num_input_tokens_seen": 229638144, "step": 876 }, { "epoch": 0.21777998510057114, "grad_norm": 0.6922534704208374, "learning_rate": 4.6111260733545714e-07, "loss": 1.4963, "num_input_tokens_seen": 229900288, "step": 877 }, { "epoch": 0.21802830891482494, "grad_norm": 0.5298182368278503, "learning_rate": 4.538194547646574e-07, "loss": 1.2116, "num_input_tokens_seen": 230162432, "step": 878 }, { "epoch": 0.2182766327290787, "grad_norm": 0.6427643299102783, "learning_rate": 4.4658169596911493e-07, "loss": 1.4518, "num_input_tokens_seen": 230424576, "step": 879 }, { "epoch": 0.2185249565433325, "grad_norm": 0.5809391736984253, "learning_rate": 4.3939941913863525e-07, "loss": 1.5274, "num_input_tokens_seen": 230686720, "step": 880 }, { "epoch": 0.21877328035758628, "grad_norm": 0.7898019552230835, "learning_rate": 4.322727117869951e-07, "loss": 1.4542, "num_input_tokens_seen": 230948864, "step": 881 }, { "epoch": 0.21902160417184008, "grad_norm": 0.3369694650173187, "learning_rate": 4.2520166075087635e-07, "loss": 1.7991, "num_input_tokens_seen": 231211008, "step": 882 }, { "epoch": 0.21926992798609388, "grad_norm": 0.9025521874427795, "learning_rate": 4.1818635218880186e-07, "loss": 1.5178, "num_input_tokens_seen": 231473152, "step": 883 }, { "epoch": 0.21951825180034765, "grad_norm": 0.5789931416511536, "learning_rate": 4.112268715800943e-07, "loss": 1.7304, "num_input_tokens_seen": 231735296, "step": 884 }, { "epoch": 0.21976657561460144, "grad_norm": 0.6976189613342285, "learning_rate": 4.043233037238281e-07, "loss": 1.6488, "num_input_tokens_seen": 231997440, "step": 885 }, { "epoch": 0.22001489942885522, "grad_norm": 0.49333396553993225, "learning_rate": 3.9747573273779816e-07, "loss": 1.2766, "num_input_tokens_seen": 232259584, "step": 886 }, { "epoch": 0.220263223243109, "grad_norm": 0.44112178683280945, "learning_rate": 3.90684242057498e-07, "loss": 2.2126, "num_input_tokens_seen": 232521728, "step": 887 }, { "epoch": 0.2205115470573628, "grad_norm": 0.4824700951576233, "learning_rate": 3.8394891443509554e-07, "loss": 1.4844, "num_input_tokens_seen": 232783872, "step": 888 }, { "epoch": 0.22075987087161658, "grad_norm": 0.9592105746269226, "learning_rate": 3.772698319384349e-07, "loss": 1.4548, "num_input_tokens_seen": 233046016, "step": 889 }, { "epoch": 0.22100819468587038, "grad_norm": 0.6596522927284241, "learning_rate": 3.7064707595002636e-07, "loss": 1.9353, "num_input_tokens_seen": 233308160, "step": 890 }, { "epoch": 0.22125651850012415, "grad_norm": 0.5860907435417175, "learning_rate": 3.6408072716606346e-07, "loss": 1.7116, "num_input_tokens_seen": 233570304, "step": 891 }, { "epoch": 0.22150484231437795, "grad_norm": 0.5942659974098206, "learning_rate": 3.575708655954324e-07, "loss": 1.5085, "num_input_tokens_seen": 233832448, "step": 892 }, { "epoch": 0.22175316612863175, "grad_norm": 0.45090603828430176, "learning_rate": 3.511175705587433e-07, "loss": 2.0349, "num_input_tokens_seen": 234094592, "step": 893 }, { "epoch": 0.22200148994288552, "grad_norm": 0.4501633942127228, "learning_rate": 3.4472092068735917e-07, "loss": 1.6038, "num_input_tokens_seen": 234356736, "step": 894 }, { "epoch": 0.22224981375713931, "grad_norm": 0.6762533783912659, "learning_rate": 3.3838099392243915e-07, "loss": 1.9386, "num_input_tokens_seen": 234618880, "step": 895 }, { "epoch": 0.22249813757139308, "grad_norm": 0.447121798992157, "learning_rate": 3.320978675139919e-07, "loss": 1.4892, "num_input_tokens_seen": 234881024, "step": 896 }, { "epoch": 0.22274646138564688, "grad_norm": 0.45953160524368286, "learning_rate": 3.258716180199278e-07, "loss": 1.8319, "num_input_tokens_seen": 235143168, "step": 897 }, { "epoch": 0.22299478519990068, "grad_norm": 0.6440572142601013, "learning_rate": 3.1970232130513365e-07, "loss": 1.515, "num_input_tokens_seen": 235405312, "step": 898 }, { "epoch": 0.22324310901415445, "grad_norm": 0.5581231117248535, "learning_rate": 3.135900525405428e-07, "loss": 1.4006, "num_input_tokens_seen": 235667456, "step": 899 }, { "epoch": 0.22349143282840825, "grad_norm": 0.8089503645896912, "learning_rate": 3.0753488620222037e-07, "loss": 1.9274, "num_input_tokens_seen": 235929600, "step": 900 }, { "epoch": 0.22373975664266202, "grad_norm": 0.3819998502731323, "learning_rate": 3.015368960704584e-07, "loss": 1.5894, "num_input_tokens_seen": 236191744, "step": 901 }, { "epoch": 0.22398808045691582, "grad_norm": 0.27509671449661255, "learning_rate": 2.9559615522887275e-07, "loss": 1.4556, "num_input_tokens_seen": 236453888, "step": 902 }, { "epoch": 0.22423640427116961, "grad_norm": 0.4570426344871521, "learning_rate": 2.8971273606351656e-07, "loss": 1.6122, "num_input_tokens_seen": 236716032, "step": 903 }, { "epoch": 0.22448472808542339, "grad_norm": 0.557512104511261, "learning_rate": 2.838867102619952e-07, "loss": 1.3736, "num_input_tokens_seen": 236978176, "step": 904 }, { "epoch": 0.22473305189967718, "grad_norm": 0.9721599221229553, "learning_rate": 2.7811814881259503e-07, "loss": 1.5408, "num_input_tokens_seen": 237240320, "step": 905 }, { "epoch": 0.22498137571393095, "grad_norm": 0.34721148014068604, "learning_rate": 2.724071220034158e-07, "loss": 1.3573, "num_input_tokens_seen": 237502464, "step": 906 }, { "epoch": 0.22522969952818475, "grad_norm": 0.5350072979927063, "learning_rate": 2.6675369942151864e-07, "loss": 1.7557, "num_input_tokens_seen": 237764608, "step": 907 }, { "epoch": 0.22547802334243855, "grad_norm": 0.5631807446479797, "learning_rate": 2.611579499520722e-07, "loss": 1.7951, "num_input_tokens_seen": 238026752, "step": 908 }, { "epoch": 0.22572634715669232, "grad_norm": 0.731895387172699, "learning_rate": 2.556199417775174e-07, "loss": 1.3442, "num_input_tokens_seen": 238288896, "step": 909 }, { "epoch": 0.22597467097094612, "grad_norm": 0.5491323471069336, "learning_rate": 2.5013974237673824e-07, "loss": 1.932, "num_input_tokens_seen": 238551040, "step": 910 }, { "epoch": 0.2262229947851999, "grad_norm": 0.3442050814628601, "learning_rate": 2.447174185242324e-07, "loss": 1.6905, "num_input_tokens_seen": 238813184, "step": 911 }, { "epoch": 0.22647131859945369, "grad_norm": 0.7306151390075684, "learning_rate": 2.3935303628930705e-07, "loss": 1.4804, "num_input_tokens_seen": 239075328, "step": 912 }, { "epoch": 0.22671964241370748, "grad_norm": 0.5416999459266663, "learning_rate": 2.3404666103526542e-07, "loss": 1.6624, "num_input_tokens_seen": 239337472, "step": 913 }, { "epoch": 0.22696796622796125, "grad_norm": 0.37114378809928894, "learning_rate": 2.287983574186159e-07, "loss": 1.8494, "num_input_tokens_seen": 239599616, "step": 914 }, { "epoch": 0.22721629004221505, "grad_norm": 0.6947669386863708, "learning_rate": 2.2360818938828189e-07, "loss": 1.5021, "num_input_tokens_seen": 239861760, "step": 915 }, { "epoch": 0.22746461385646882, "grad_norm": 0.5789608955383301, "learning_rate": 2.1847622018482283e-07, "loss": 1.9683, "num_input_tokens_seen": 240123904, "step": 916 }, { "epoch": 0.22771293767072262, "grad_norm": 0.6739006638526917, "learning_rate": 2.134025123396638e-07, "loss": 2.1382, "num_input_tokens_seen": 240386048, "step": 917 }, { "epoch": 0.22796126148497642, "grad_norm": 0.7561383247375488, "learning_rate": 2.083871276743338e-07, "loss": 1.423, "num_input_tokens_seen": 240648192, "step": 918 }, { "epoch": 0.2282095852992302, "grad_norm": 0.7975893616676331, "learning_rate": 2.0343012729971244e-07, "loss": 1.4102, "num_input_tokens_seen": 240910336, "step": 919 }, { "epoch": 0.228457909113484, "grad_norm": 0.2659681737422943, "learning_rate": 1.9853157161528468e-07, "loss": 1.215, "num_input_tokens_seen": 241172480, "step": 920 }, { "epoch": 0.22870623292773776, "grad_norm": 0.3225981593132019, "learning_rate": 1.9369152030840553e-07, "loss": 1.7634, "num_input_tokens_seen": 241434624, "step": 921 }, { "epoch": 0.22895455674199156, "grad_norm": 0.7171019911766052, "learning_rate": 1.8891003235357307e-07, "loss": 1.7829, "num_input_tokens_seen": 241696768, "step": 922 }, { "epoch": 0.22920288055624535, "grad_norm": 0.5440613031387329, "learning_rate": 1.841871660117095e-07, "loss": 1.3598, "num_input_tokens_seen": 241958912, "step": 923 }, { "epoch": 0.22945120437049912, "grad_norm": 0.359244167804718, "learning_rate": 1.7952297882945e-07, "loss": 1.5261, "num_input_tokens_seen": 242221056, "step": 924 }, { "epoch": 0.22969952818475292, "grad_norm": 0.7204708456993103, "learning_rate": 1.7491752763844294e-07, "loss": 1.7172, "num_input_tokens_seen": 242483200, "step": 925 }, { "epoch": 0.2299478519990067, "grad_norm": 0.4921867847442627, "learning_rate": 1.7037086855465902e-07, "loss": 1.8377, "num_input_tokens_seen": 242745344, "step": 926 }, { "epoch": 0.2301961758132605, "grad_norm": 0.8633773326873779, "learning_rate": 1.6588305697770313e-07, "loss": 1.4806, "num_input_tokens_seen": 243007488, "step": 927 }, { "epoch": 0.2304444996275143, "grad_norm": 0.5622784495353699, "learning_rate": 1.6145414759014433e-07, "loss": 1.5533, "num_input_tokens_seen": 243269632, "step": 928 }, { "epoch": 0.23069282344176806, "grad_norm": 0.5299321413040161, "learning_rate": 1.5708419435684463e-07, "loss": 1.3532, "num_input_tokens_seen": 243531776, "step": 929 }, { "epoch": 0.23094114725602186, "grad_norm": 0.6882240772247314, "learning_rate": 1.5277325052430569e-07, "loss": 1.7824, "num_input_tokens_seen": 243793920, "step": 930 }, { "epoch": 0.23118947107027563, "grad_norm": 0.431749552488327, "learning_rate": 1.4852136862001766e-07, "loss": 1.5325, "num_input_tokens_seen": 244056064, "step": 931 }, { "epoch": 0.23143779488452942, "grad_norm": 0.5995936393737793, "learning_rate": 1.4432860045182019e-07, "loss": 1.855, "num_input_tokens_seen": 244318208, "step": 932 }, { "epoch": 0.23168611869878322, "grad_norm": 0.7233447432518005, "learning_rate": 1.4019499710726913e-07, "loss": 1.6507, "num_input_tokens_seen": 244580352, "step": 933 }, { "epoch": 0.231934442513037, "grad_norm": 0.54053795337677, "learning_rate": 1.3612060895301759e-07, "loss": 1.8987, "num_input_tokens_seen": 244842496, "step": 934 }, { "epoch": 0.2321827663272908, "grad_norm": 0.3651261031627655, "learning_rate": 1.3210548563419857e-07, "loss": 1.4466, "num_input_tokens_seen": 245104640, "step": 935 }, { "epoch": 0.23243109014154456, "grad_norm": 0.597426176071167, "learning_rate": 1.2814967607382433e-07, "loss": 1.6555, "num_input_tokens_seen": 245366784, "step": 936 }, { "epoch": 0.23267941395579836, "grad_norm": 0.4557691216468811, "learning_rate": 1.2425322847218368e-07, "loss": 1.3507, "num_input_tokens_seen": 245628928, "step": 937 }, { "epoch": 0.23292773777005216, "grad_norm": 0.5856966376304626, "learning_rate": 1.2041619030626283e-07, "loss": 1.5985, "num_input_tokens_seen": 245891072, "step": 938 }, { "epoch": 0.23317606158430593, "grad_norm": 0.5499093532562256, "learning_rate": 1.166386083291604e-07, "loss": 1.9703, "num_input_tokens_seen": 246153216, "step": 939 }, { "epoch": 0.23342438539855973, "grad_norm": 0.5963953733444214, "learning_rate": 1.1292052856952063e-07, "loss": 1.5363, "num_input_tokens_seen": 246415360, "step": 940 }, { "epoch": 0.2336727092128135, "grad_norm": 0.5307541489601135, "learning_rate": 1.0926199633097156e-07, "loss": 1.2944, "num_input_tokens_seen": 246677504, "step": 941 }, { "epoch": 0.2339210330270673, "grad_norm": 0.5369905233383179, "learning_rate": 1.0566305619157502e-07, "loss": 1.933, "num_input_tokens_seen": 246939648, "step": 942 }, { "epoch": 0.2341693568413211, "grad_norm": 1.9757884740829468, "learning_rate": 1.0212375200327973e-07, "loss": 1.6174, "num_input_tokens_seen": 247201792, "step": 943 }, { "epoch": 0.23441768065557486, "grad_norm": 0.5642454028129578, "learning_rate": 9.864412689139124e-08, "loss": 1.2, "num_input_tokens_seen": 247463936, "step": 944 }, { "epoch": 0.23466600446982866, "grad_norm": 0.316659152507782, "learning_rate": 9.522422325404234e-08, "loss": 1.8678, "num_input_tokens_seen": 247726080, "step": 945 }, { "epoch": 0.23491432828408243, "grad_norm": 0.4380717873573303, "learning_rate": 9.186408276168012e-08, "loss": 1.4458, "num_input_tokens_seen": 247988224, "step": 946 }, { "epoch": 0.23516265209833623, "grad_norm": 0.5788005590438843, "learning_rate": 8.856374635655696e-08, "loss": 1.5094, "num_input_tokens_seen": 248250368, "step": 947 }, { "epoch": 0.23541097591259003, "grad_norm": 0.6096988320350647, "learning_rate": 8.53232542522292e-08, "loss": 1.5393, "num_input_tokens_seen": 248512512, "step": 948 }, { "epoch": 0.2356592997268438, "grad_norm": 0.5101311802864075, "learning_rate": 8.214264593307097e-08, "loss": 1.6412, "num_input_tokens_seen": 248774656, "step": 949 }, { "epoch": 0.2359076235410976, "grad_norm": 0.4224371314048767, "learning_rate": 7.90219601537906e-08, "loss": 1.4671, "num_input_tokens_seen": 249036800, "step": 950 }, { "epoch": 0.23615594735535136, "grad_norm": 0.5831036567687988, "learning_rate": 7.59612349389599e-08, "loss": 1.4406, "num_input_tokens_seen": 249298944, "step": 951 }, { "epoch": 0.23640427116960516, "grad_norm": 0.7039021849632263, "learning_rate": 7.296050758254958e-08, "loss": 1.7773, "num_input_tokens_seen": 249561088, "step": 952 }, { "epoch": 0.23665259498385896, "grad_norm": 0.5700446367263794, "learning_rate": 7.001981464747565e-08, "loss": 1.4874, "num_input_tokens_seen": 249823232, "step": 953 }, { "epoch": 0.23690091879811273, "grad_norm": 0.6082141995429993, "learning_rate": 6.713919196515317e-08, "loss": 1.5502, "num_input_tokens_seen": 250085376, "step": 954 }, { "epoch": 0.23714924261236653, "grad_norm": 0.5807713866233826, "learning_rate": 6.431867463506047e-08, "loss": 1.6924, "num_input_tokens_seen": 250347520, "step": 955 }, { "epoch": 0.23739756642662033, "grad_norm": 0.6795780658721924, "learning_rate": 6.15582970243117e-08, "loss": 1.6181, "num_input_tokens_seen": 250609664, "step": 956 }, { "epoch": 0.2376458902408741, "grad_norm": 0.5901414155960083, "learning_rate": 5.8858092767236084e-08, "loss": 1.6287, "num_input_tokens_seen": 250871808, "step": 957 }, { "epoch": 0.2378942140551279, "grad_norm": 0.258759081363678, "learning_rate": 5.621809476497098e-08, "loss": 1.5631, "num_input_tokens_seen": 251133952, "step": 958 }, { "epoch": 0.23814253786938167, "grad_norm": 0.793224036693573, "learning_rate": 5.363833518505834e-08, "loss": 1.6065, "num_input_tokens_seen": 251396096, "step": 959 }, { "epoch": 0.23839086168363546, "grad_norm": 0.541074275970459, "learning_rate": 5.111884546105506e-08, "loss": 1.7784, "num_input_tokens_seen": 251658240, "step": 960 }, { "epoch": 0.23863918549788926, "grad_norm": 0.6888409852981567, "learning_rate": 4.865965629214819e-08, "loss": 1.4175, "num_input_tokens_seen": 251920384, "step": 961 }, { "epoch": 0.23888750931214303, "grad_norm": 0.5229154825210571, "learning_rate": 4.626079764278202e-08, "loss": 1.7039, "num_input_tokens_seen": 252182528, "step": 962 }, { "epoch": 0.23913583312639683, "grad_norm": 0.43334439396858215, "learning_rate": 4.392229874229159e-08, "loss": 1.6586, "num_input_tokens_seen": 252444672, "step": 963 }, { "epoch": 0.2393841569406506, "grad_norm": 0.5404706597328186, "learning_rate": 4.164418808454806e-08, "loss": 1.5866, "num_input_tokens_seen": 252706816, "step": 964 }, { "epoch": 0.2396324807549044, "grad_norm": 0.31539681553840637, "learning_rate": 3.9426493427611177e-08, "loss": 1.2445, "num_input_tokens_seen": 252968960, "step": 965 }, { "epoch": 0.2398808045691582, "grad_norm": 0.767345130443573, "learning_rate": 3.726924179339009e-08, "loss": 1.6509, "num_input_tokens_seen": 253231104, "step": 966 }, { "epoch": 0.24012912838341197, "grad_norm": 0.5601321458816528, "learning_rate": 3.517245946731529e-08, "loss": 1.6218, "num_input_tokens_seen": 253493248, "step": 967 }, { "epoch": 0.24037745219766576, "grad_norm": 0.6398811936378479, "learning_rate": 3.313617199801777e-08, "loss": 1.8179, "num_input_tokens_seen": 253755392, "step": 968 }, { "epoch": 0.24062577601191953, "grad_norm": 0.3753248155117035, "learning_rate": 3.1160404197018155e-08, "loss": 1.6866, "num_input_tokens_seen": 254017536, "step": 969 }, { "epoch": 0.24087409982617333, "grad_norm": 0.7169692516326904, "learning_rate": 2.9245180138423033e-08, "loss": 1.6297, "num_input_tokens_seen": 254279680, "step": 970 }, { "epoch": 0.24112242364042713, "grad_norm": 0.501290500164032, "learning_rate": 2.7390523158633552e-08, "loss": 1.3324, "num_input_tokens_seen": 254541824, "step": 971 }, { "epoch": 0.2413707474546809, "grad_norm": 0.4676488935947418, "learning_rate": 2.5596455856058966e-08, "loss": 1.3584, "num_input_tokens_seen": 254803968, "step": 972 }, { "epoch": 0.2416190712689347, "grad_norm": 0.40131956338882446, "learning_rate": 2.386300009084408e-08, "loss": 1.0066, "num_input_tokens_seen": 255066112, "step": 973 }, { "epoch": 0.24186739508318847, "grad_norm": 0.45223063230514526, "learning_rate": 2.219017698460002e-08, "loss": 1.7419, "num_input_tokens_seen": 255328256, "step": 974 }, { "epoch": 0.24211571889744227, "grad_norm": 0.48078829050064087, "learning_rate": 2.057800692014833e-08, "loss": 1.7454, "num_input_tokens_seen": 255590400, "step": 975 }, { "epoch": 0.24236404271169606, "grad_norm": 0.3801480531692505, "learning_rate": 1.9026509541272276e-08, "loss": 1.542, "num_input_tokens_seen": 255852544, "step": 976 }, { "epoch": 0.24261236652594984, "grad_norm": 0.3696509301662445, "learning_rate": 1.753570375247815e-08, "loss": 1.8456, "num_input_tokens_seen": 256114688, "step": 977 }, { "epoch": 0.24286069034020363, "grad_norm": 0.3742285966873169, "learning_rate": 1.610560771876435e-08, "loss": 1.3322, "num_input_tokens_seen": 256376832, "step": 978 }, { "epoch": 0.2431090141544574, "grad_norm": 0.9268859624862671, "learning_rate": 1.4736238865398766e-08, "loss": 1.9955, "num_input_tokens_seen": 256638976, "step": 979 }, { "epoch": 0.2433573379687112, "grad_norm": 0.7358083128929138, "learning_rate": 1.3427613877709523e-08, "loss": 1.6458, "num_input_tokens_seen": 256901120, "step": 980 }, { "epoch": 0.243605661782965, "grad_norm": 0.4992476999759674, "learning_rate": 1.2179748700879013e-08, "loss": 1.7698, "num_input_tokens_seen": 257163264, "step": 981 }, { "epoch": 0.24385398559721877, "grad_norm": 0.7254119515419006, "learning_rate": 1.0992658539750179e-08, "loss": 2.0594, "num_input_tokens_seen": 257425408, "step": 982 }, { "epoch": 0.24410230941147257, "grad_norm": 0.30243805050849915, "learning_rate": 9.866357858642206e-09, "loss": 1.6328, "num_input_tokens_seen": 257687552, "step": 983 }, { "epoch": 0.24435063322572634, "grad_norm": 0.3717341721057892, "learning_rate": 8.800860381173448e-09, "loss": 1.7149, "num_input_tokens_seen": 257949696, "step": 984 }, { "epoch": 0.24459895703998014, "grad_norm": 0.6060426831245422, "learning_rate": 7.796179090094891e-09, "loss": 1.4753, "num_input_tokens_seen": 258211840, "step": 985 }, { "epoch": 0.24484728085423393, "grad_norm": 0.5001364350318909, "learning_rate": 6.852326227130835e-09, "loss": 1.6607, "num_input_tokens_seen": 258473984, "step": 986 }, { "epoch": 0.2450956046684877, "grad_norm": 0.41138938069343567, "learning_rate": 5.969313292830126e-09, "loss": 1.5122, "num_input_tokens_seen": 258736128, "step": 987 }, { "epoch": 0.2453439284827415, "grad_norm": 0.5986543893814087, "learning_rate": 5.147151046426824e-09, "loss": 1.3974, "num_input_tokens_seen": 258998272, "step": 988 }, { "epoch": 0.24559225229699527, "grad_norm": 0.3890846073627472, "learning_rate": 4.385849505708084e-09, "loss": 1.7548, "num_input_tokens_seen": 259260416, "step": 989 }, { "epoch": 0.24584057611124907, "grad_norm": 0.5105006694793701, "learning_rate": 3.685417946894254e-09, "loss": 1.4885, "num_input_tokens_seen": 259522560, "step": 990 }, { "epoch": 0.24608889992550287, "grad_norm": 0.5231227278709412, "learning_rate": 3.0458649045211897e-09, "loss": 1.488, "num_input_tokens_seen": 259784704, "step": 991 }, { "epoch": 0.24633722373975664, "grad_norm": 0.5039011240005493, "learning_rate": 2.4671981713420003e-09, "loss": 2.0, "num_input_tokens_seen": 260046848, "step": 992 }, { "epoch": 0.24658554755401044, "grad_norm": 0.8183209300041199, "learning_rate": 1.9494247982282386e-09, "loss": 1.5233, "num_input_tokens_seen": 260308992, "step": 993 }, { "epoch": 0.2468338713682642, "grad_norm": 0.34673362970352173, "learning_rate": 1.4925510940844157e-09, "loss": 1.2279, "num_input_tokens_seen": 260571136, "step": 994 }, { "epoch": 0.247082195182518, "grad_norm": 0.6855907440185547, "learning_rate": 1.096582625772502e-09, "loss": 1.3346, "num_input_tokens_seen": 260833280, "step": 995 }, { "epoch": 0.2473305189967718, "grad_norm": 0.5691891312599182, "learning_rate": 7.615242180436521e-10, "loss": 1.3504, "num_input_tokens_seen": 261095424, "step": 996 }, { "epoch": 0.24757884281102557, "grad_norm": 0.49807822704315186, "learning_rate": 4.87379953478806e-10, "loss": 2.1568, "num_input_tokens_seen": 261357568, "step": 997 }, { "epoch": 0.24782716662527937, "grad_norm": 0.4466933310031891, "learning_rate": 2.741531724392843e-10, "loss": 1.4098, "num_input_tokens_seen": 261619712, "step": 998 }, { "epoch": 0.24807549043953314, "grad_norm": 0.6643083691596985, "learning_rate": 1.2184647302626585e-10, "loss": 1.6152, "num_input_tokens_seen": 261881856, "step": 999 }, { "epoch": 0.24832381425378694, "grad_norm": 0.5675874352455139, "learning_rate": 3.0461711048035415e-11, "loss": 1.2485, "num_input_tokens_seen": 262144000, "step": 1000 } ], "logging_steps": 1.0, "max_steps": 1000, "num_input_tokens_seen": 262144000, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 418759311360000.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }