{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9964614295824488, "eval_steps": 500, "global_step": 2118, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0014154281670205238, "grad_norm": 52.03026133766843, "learning_rate": 0.0, "loss": 11.3104, "step": 1 }, { "epoch": 0.0028308563340410475, "grad_norm": 44.502051688198414, "learning_rate": 2.3584905660377358e-07, "loss": 11.4795, "step": 2 }, { "epoch": 0.004246284501061571, "grad_norm": 45.159850822187344, "learning_rate": 4.7169811320754717e-07, "loss": 11.6089, "step": 3 }, { "epoch": 0.005661712668082095, "grad_norm": 43.100232328671524, "learning_rate": 7.075471698113208e-07, "loss": 11.5825, "step": 4 }, { "epoch": 0.007077140835102618, "grad_norm": 42.5821886975642, "learning_rate": 9.433962264150943e-07, "loss": 11.6043, "step": 5 }, { "epoch": 0.008492569002123142, "grad_norm": 47.15198695289155, "learning_rate": 1.179245283018868e-06, "loss": 11.4062, "step": 6 }, { "epoch": 0.009907997169143666, "grad_norm": 47.21018798250648, "learning_rate": 1.4150943396226415e-06, "loss": 11.2894, "step": 7 }, { "epoch": 0.01132342533616419, "grad_norm": 52.16499613461483, "learning_rate": 1.650943396226415e-06, "loss": 11.1135, "step": 8 }, { "epoch": 0.012738853503184714, "grad_norm": 55.63144370327757, "learning_rate": 1.8867924528301887e-06, "loss": 10.9942, "step": 9 }, { "epoch": 0.014154281670205236, "grad_norm": 64.12730354425412, "learning_rate": 2.1226415094339624e-06, "loss": 10.3487, "step": 10 }, { "epoch": 0.01556970983722576, "grad_norm": 71.34125577863145, "learning_rate": 2.358490566037736e-06, "loss": 10.3151, "step": 11 }, { "epoch": 0.016985138004246284, "grad_norm": 74.0270309921208, "learning_rate": 2.5943396226415095e-06, "loss": 9.8093, "step": 12 }, { "epoch": 0.01840056617126681, "grad_norm": 103.30745846664922, "learning_rate": 2.830188679245283e-06, "loss": 6.1753, "step": 13 }, { "epoch": 0.019815994338287332, "grad_norm": 77.09348349918515, "learning_rate": 3.0660377358490567e-06, "loss": 4.7991, "step": 14 }, { "epoch": 0.021231422505307854, "grad_norm": 71.52287449239502, "learning_rate": 3.30188679245283e-06, "loss": 4.6793, "step": 15 }, { "epoch": 0.02264685067232838, "grad_norm": 46.74022190241013, "learning_rate": 3.5377358490566038e-06, "loss": 3.1627, "step": 16 }, { "epoch": 0.024062278839348902, "grad_norm": 40.45919537581584, "learning_rate": 3.7735849056603773e-06, "loss": 3.0045, "step": 17 }, { "epoch": 0.025477707006369428, "grad_norm": 8.445011797465854, "learning_rate": 4.009433962264151e-06, "loss": 1.9597, "step": 18 }, { "epoch": 0.02689313517338995, "grad_norm": 5.479013054164001, "learning_rate": 4.245283018867925e-06, "loss": 1.4691, "step": 19 }, { "epoch": 0.028308563340410473, "grad_norm": 5.440314496315961, "learning_rate": 4.481132075471698e-06, "loss": 1.7585, "step": 20 }, { "epoch": 0.029723991507430998, "grad_norm": 4.73587778702694, "learning_rate": 4.716981132075472e-06, "loss": 1.7664, "step": 21 }, { "epoch": 0.03113941967445152, "grad_norm": 3.391157126674541, "learning_rate": 4.952830188679246e-06, "loss": 1.5264, "step": 22 }, { "epoch": 0.03255484784147204, "grad_norm": 2.7846104538504726, "learning_rate": 5.188679245283019e-06, "loss": 1.6084, "step": 23 }, { "epoch": 0.03397027600849257, "grad_norm": 3.368442099627268, "learning_rate": 5.424528301886793e-06, "loss": 1.6023, "step": 24 }, { "epoch": 0.035385704175513094, "grad_norm": 1.6613773212139946, "learning_rate": 5.660377358490566e-06, "loss": 1.3373, "step": 25 }, { "epoch": 0.03680113234253362, "grad_norm": 2.4022375902835065, "learning_rate": 5.89622641509434e-06, "loss": 1.3476, "step": 26 }, { "epoch": 0.03821656050955414, "grad_norm": 1.8303381406131283, "learning_rate": 6.132075471698113e-06, "loss": 1.3573, "step": 27 }, { "epoch": 0.039631988676574664, "grad_norm": 1.0492891809918425, "learning_rate": 6.367924528301887e-06, "loss": 0.9964, "step": 28 }, { "epoch": 0.04104741684359519, "grad_norm": 1.0568449472146275, "learning_rate": 6.60377358490566e-06, "loss": 1.2754, "step": 29 }, { "epoch": 0.04246284501061571, "grad_norm": 0.8939826961932031, "learning_rate": 6.839622641509434e-06, "loss": 1.0906, "step": 30 }, { "epoch": 0.043878273177636234, "grad_norm": 0.924706231772569, "learning_rate": 7.0754716981132075e-06, "loss": 1.1488, "step": 31 }, { "epoch": 0.04529370134465676, "grad_norm": 0.9874613327771626, "learning_rate": 7.3113207547169815e-06, "loss": 0.9942, "step": 32 }, { "epoch": 0.04670912951167728, "grad_norm": 1.1039357027762495, "learning_rate": 7.547169811320755e-06, "loss": 1.2901, "step": 33 }, { "epoch": 0.048124557678697805, "grad_norm": 0.6732249168818489, "learning_rate": 7.783018867924528e-06, "loss": 1.1364, "step": 34 }, { "epoch": 0.04953998584571833, "grad_norm": 0.7389830991594677, "learning_rate": 8.018867924528302e-06, "loss": 1.0965, "step": 35 }, { "epoch": 0.050955414012738856, "grad_norm": 1.361919734305708, "learning_rate": 8.254716981132076e-06, "loss": 1.069, "step": 36 }, { "epoch": 0.052370842179759375, "grad_norm": 0.6946076673297682, "learning_rate": 8.49056603773585e-06, "loss": 0.8868, "step": 37 }, { "epoch": 0.0537862703467799, "grad_norm": 0.678223157015022, "learning_rate": 8.726415094339622e-06, "loss": 1.1229, "step": 38 }, { "epoch": 0.055201698513800426, "grad_norm": 0.6733857464549065, "learning_rate": 8.962264150943396e-06, "loss": 1.0492, "step": 39 }, { "epoch": 0.056617126680820945, "grad_norm": 0.5835930567416066, "learning_rate": 9.19811320754717e-06, "loss": 0.9794, "step": 40 }, { "epoch": 0.05803255484784147, "grad_norm": 0.5657763577994933, "learning_rate": 9.433962264150944e-06, "loss": 1.0282, "step": 41 }, { "epoch": 0.059447983014861996, "grad_norm": 0.5916133305317316, "learning_rate": 9.669811320754718e-06, "loss": 0.8039, "step": 42 }, { "epoch": 0.06086341118188252, "grad_norm": 0.6176031326803815, "learning_rate": 9.905660377358492e-06, "loss": 0.8717, "step": 43 }, { "epoch": 0.06227883934890304, "grad_norm": 0.5047080085145694, "learning_rate": 1.0141509433962266e-05, "loss": 0.9476, "step": 44 }, { "epoch": 0.06369426751592357, "grad_norm": 0.5057501627696119, "learning_rate": 1.0377358490566038e-05, "loss": 1.0082, "step": 45 }, { "epoch": 0.06510969568294409, "grad_norm": 0.4844409794803433, "learning_rate": 1.0613207547169812e-05, "loss": 0.9998, "step": 46 }, { "epoch": 0.06652512384996462, "grad_norm": 0.439835805052479, "learning_rate": 1.0849056603773586e-05, "loss": 0.7494, "step": 47 }, { "epoch": 0.06794055201698514, "grad_norm": 0.47748592314903204, "learning_rate": 1.108490566037736e-05, "loss": 0.9679, "step": 48 }, { "epoch": 0.06935598018400566, "grad_norm": 0.41694798317782045, "learning_rate": 1.1320754716981132e-05, "loss": 0.9175, "step": 49 }, { "epoch": 0.07077140835102619, "grad_norm": 0.4131031923420625, "learning_rate": 1.1556603773584906e-05, "loss": 0.8525, "step": 50 }, { "epoch": 0.07218683651804671, "grad_norm": 0.39877873003968667, "learning_rate": 1.179245283018868e-05, "loss": 0.9432, "step": 51 }, { "epoch": 0.07360226468506724, "grad_norm": 0.38778633507310184, "learning_rate": 1.2028301886792454e-05, "loss": 0.9381, "step": 52 }, { "epoch": 0.07501769285208776, "grad_norm": 0.35074905382025257, "learning_rate": 1.2264150943396227e-05, "loss": 0.7945, "step": 53 }, { "epoch": 0.07643312101910828, "grad_norm": 0.4164483097493037, "learning_rate": 1.25e-05, "loss": 0.9996, "step": 54 }, { "epoch": 0.07784854918612881, "grad_norm": 0.4634371429875181, "learning_rate": 1.2735849056603775e-05, "loss": 0.9571, "step": 55 }, { "epoch": 0.07926397735314933, "grad_norm": 0.3872343128801052, "learning_rate": 1.2971698113207547e-05, "loss": 0.8564, "step": 56 }, { "epoch": 0.08067940552016985, "grad_norm": 0.3918275587660996, "learning_rate": 1.320754716981132e-05, "loss": 0.9868, "step": 57 }, { "epoch": 0.08209483368719038, "grad_norm": 0.3454015784380163, "learning_rate": 1.3443396226415095e-05, "loss": 0.8604, "step": 58 }, { "epoch": 0.0835102618542109, "grad_norm": 0.4087333001843292, "learning_rate": 1.3679245283018869e-05, "loss": 0.9851, "step": 59 }, { "epoch": 0.08492569002123142, "grad_norm": 0.7108784390602116, "learning_rate": 1.3915094339622641e-05, "loss": 0.9191, "step": 60 }, { "epoch": 0.08634111818825195, "grad_norm": 0.3245781305656771, "learning_rate": 1.4150943396226415e-05, "loss": 0.8334, "step": 61 }, { "epoch": 0.08775654635527247, "grad_norm": 0.362401119658312, "learning_rate": 1.4386792452830189e-05, "loss": 0.9544, "step": 62 }, { "epoch": 0.08917197452229299, "grad_norm": 0.3153301984897242, "learning_rate": 1.4622641509433963e-05, "loss": 0.7678, "step": 63 }, { "epoch": 0.09058740268931352, "grad_norm": 0.3410193369934483, "learning_rate": 1.4858490566037735e-05, "loss": 0.6978, "step": 64 }, { "epoch": 0.09200283085633404, "grad_norm": 0.3327538499445616, "learning_rate": 1.509433962264151e-05, "loss": 0.8188, "step": 65 }, { "epoch": 0.09341825902335456, "grad_norm": 0.32595848738844885, "learning_rate": 1.5330188679245283e-05, "loss": 0.7869, "step": 66 }, { "epoch": 0.09483368719037509, "grad_norm": 0.3147147961761771, "learning_rate": 1.5566037735849056e-05, "loss": 0.7442, "step": 67 }, { "epoch": 0.09624911535739561, "grad_norm": 1.3369548940206364, "learning_rate": 1.580188679245283e-05, "loss": 0.8758, "step": 68 }, { "epoch": 0.09766454352441614, "grad_norm": 0.419616964657023, "learning_rate": 1.6037735849056604e-05, "loss": 0.864, "step": 69 }, { "epoch": 0.09907997169143666, "grad_norm": 0.3995185185455742, "learning_rate": 1.6273584905660376e-05, "loss": 0.837, "step": 70 }, { "epoch": 0.10049539985845718, "grad_norm": 0.3332916810605707, "learning_rate": 1.650943396226415e-05, "loss": 0.7015, "step": 71 }, { "epoch": 0.10191082802547771, "grad_norm": 0.34754244002488843, "learning_rate": 1.6745283018867924e-05, "loss": 0.7274, "step": 72 }, { "epoch": 0.10332625619249823, "grad_norm": 0.33973651313585607, "learning_rate": 1.69811320754717e-05, "loss": 0.7914, "step": 73 }, { "epoch": 0.10474168435951875, "grad_norm": 0.38442548640761615, "learning_rate": 1.7216981132075472e-05, "loss": 1.0713, "step": 74 }, { "epoch": 0.10615711252653928, "grad_norm": 0.3748707398787612, "learning_rate": 1.7452830188679244e-05, "loss": 0.729, "step": 75 }, { "epoch": 0.1075725406935598, "grad_norm": 0.33799485157775017, "learning_rate": 1.768867924528302e-05, "loss": 0.9651, "step": 76 }, { "epoch": 0.10898796886058032, "grad_norm": 0.32834750548367253, "learning_rate": 1.7924528301886792e-05, "loss": 0.7978, "step": 77 }, { "epoch": 0.11040339702760085, "grad_norm": 0.3502524463921579, "learning_rate": 1.8160377358490564e-05, "loss": 0.8221, "step": 78 }, { "epoch": 0.11181882519462137, "grad_norm": 0.3331270390715312, "learning_rate": 1.839622641509434e-05, "loss": 0.8731, "step": 79 }, { "epoch": 0.11323425336164189, "grad_norm": 0.3346584437841966, "learning_rate": 1.8632075471698112e-05, "loss": 0.8424, "step": 80 }, { "epoch": 0.11464968152866242, "grad_norm": 0.3305752792911303, "learning_rate": 1.8867924528301888e-05, "loss": 0.6816, "step": 81 }, { "epoch": 0.11606510969568294, "grad_norm": 0.30323282202609836, "learning_rate": 1.9103773584905664e-05, "loss": 0.739, "step": 82 }, { "epoch": 0.11748053786270347, "grad_norm": 0.2932445274255076, "learning_rate": 1.9339622641509436e-05, "loss": 0.7707, "step": 83 }, { "epoch": 0.11889596602972399, "grad_norm": 0.35360411323590746, "learning_rate": 1.9575471698113208e-05, "loss": 0.9276, "step": 84 }, { "epoch": 0.12031139419674451, "grad_norm": 0.31784927590295037, "learning_rate": 1.9811320754716984e-05, "loss": 0.6502, "step": 85 }, { "epoch": 0.12172682236376504, "grad_norm": 0.2930105538589784, "learning_rate": 2.0047169811320756e-05, "loss": 0.5902, "step": 86 }, { "epoch": 0.12314225053078556, "grad_norm": 0.4298747913122774, "learning_rate": 2.0283018867924532e-05, "loss": 0.8808, "step": 87 }, { "epoch": 0.12455767869780608, "grad_norm": 0.3227619460942626, "learning_rate": 2.0518867924528304e-05, "loss": 0.6742, "step": 88 }, { "epoch": 0.1259731068648266, "grad_norm": 0.32830224756849186, "learning_rate": 2.0754716981132076e-05, "loss": 0.7305, "step": 89 }, { "epoch": 0.12738853503184713, "grad_norm": 0.3508475206918051, "learning_rate": 2.0990566037735852e-05, "loss": 0.8298, "step": 90 }, { "epoch": 0.12880396319886767, "grad_norm": 0.31955241801604534, "learning_rate": 2.1226415094339624e-05, "loss": 0.7769, "step": 91 }, { "epoch": 0.13021939136588817, "grad_norm": 0.3443616665430657, "learning_rate": 2.1462264150943397e-05, "loss": 0.8659, "step": 92 }, { "epoch": 0.1316348195329087, "grad_norm": 0.36179953262070746, "learning_rate": 2.1698113207547172e-05, "loss": 0.8947, "step": 93 }, { "epoch": 0.13305024769992924, "grad_norm": 0.3362790434923789, "learning_rate": 2.1933962264150945e-05, "loss": 0.8423, "step": 94 }, { "epoch": 0.13446567586694974, "grad_norm": 0.3083875968981802, "learning_rate": 2.216981132075472e-05, "loss": 0.7886, "step": 95 }, { "epoch": 0.13588110403397027, "grad_norm": 0.32344656361127455, "learning_rate": 2.2405660377358493e-05, "loss": 0.7875, "step": 96 }, { "epoch": 0.1372965322009908, "grad_norm": 0.3165439663454565, "learning_rate": 2.2641509433962265e-05, "loss": 0.8044, "step": 97 }, { "epoch": 0.1387119603680113, "grad_norm": 0.3558619447859139, "learning_rate": 2.287735849056604e-05, "loss": 0.8377, "step": 98 }, { "epoch": 0.14012738853503184, "grad_norm": 0.3230854098787612, "learning_rate": 2.3113207547169813e-05, "loss": 0.8527, "step": 99 }, { "epoch": 0.14154281670205238, "grad_norm": 0.292154547470411, "learning_rate": 2.3349056603773585e-05, "loss": 0.6429, "step": 100 }, { "epoch": 0.14295824486907288, "grad_norm": 0.29618483220184183, "learning_rate": 2.358490566037736e-05, "loss": 0.6592, "step": 101 }, { "epoch": 0.14437367303609341, "grad_norm": 0.40512023312218853, "learning_rate": 2.3820754716981133e-05, "loss": 0.9447, "step": 102 }, { "epoch": 0.14578910120311395, "grad_norm": 0.33243783635285035, "learning_rate": 2.405660377358491e-05, "loss": 0.6862, "step": 103 }, { "epoch": 0.14720452937013448, "grad_norm": 0.2989122340388981, "learning_rate": 2.429245283018868e-05, "loss": 0.6963, "step": 104 }, { "epoch": 0.14861995753715498, "grad_norm": 0.36430675740062585, "learning_rate": 2.4528301886792453e-05, "loss": 0.8468, "step": 105 }, { "epoch": 0.15003538570417552, "grad_norm": 0.35021308710293897, "learning_rate": 2.476415094339623e-05, "loss": 0.7834, "step": 106 }, { "epoch": 0.15145081387119605, "grad_norm": 0.34224361021249144, "learning_rate": 2.5e-05, "loss": 0.7443, "step": 107 }, { "epoch": 0.15286624203821655, "grad_norm": 0.29948735447722763, "learning_rate": 2.5235849056603777e-05, "loss": 0.6854, "step": 108 }, { "epoch": 0.1542816702052371, "grad_norm": 0.3508933289889198, "learning_rate": 2.547169811320755e-05, "loss": 0.7311, "step": 109 }, { "epoch": 0.15569709837225762, "grad_norm": 0.3296823380224191, "learning_rate": 2.5707547169811325e-05, "loss": 0.7032, "step": 110 }, { "epoch": 0.15711252653927812, "grad_norm": 0.37614064574065786, "learning_rate": 2.5943396226415094e-05, "loss": 0.7134, "step": 111 }, { "epoch": 0.15852795470629866, "grad_norm": 0.31973855712990845, "learning_rate": 2.6179245283018873e-05, "loss": 0.6965, "step": 112 }, { "epoch": 0.1599433828733192, "grad_norm": 0.3251856086872114, "learning_rate": 2.641509433962264e-05, "loss": 0.7277, "step": 113 }, { "epoch": 0.1613588110403397, "grad_norm": 0.3531176610400526, "learning_rate": 2.6650943396226417e-05, "loss": 0.8766, "step": 114 }, { "epoch": 0.16277423920736023, "grad_norm": 0.35022388084220063, "learning_rate": 2.688679245283019e-05, "loss": 0.8223, "step": 115 }, { "epoch": 0.16418966737438076, "grad_norm": 0.3311722208850899, "learning_rate": 2.7122641509433965e-05, "loss": 0.6044, "step": 116 }, { "epoch": 0.16560509554140126, "grad_norm": 0.3196689616157804, "learning_rate": 2.7358490566037738e-05, "loss": 0.6474, "step": 117 }, { "epoch": 0.1670205237084218, "grad_norm": 0.3391104637918063, "learning_rate": 2.7594339622641513e-05, "loss": 0.8039, "step": 118 }, { "epoch": 0.16843595187544233, "grad_norm": 0.35406309196254343, "learning_rate": 2.7830188679245282e-05, "loss": 0.6403, "step": 119 }, { "epoch": 0.16985138004246284, "grad_norm": 0.3440331198300765, "learning_rate": 2.806603773584906e-05, "loss": 0.673, "step": 120 }, { "epoch": 0.17126680820948337, "grad_norm": 0.36841471396963354, "learning_rate": 2.830188679245283e-05, "loss": 0.7944, "step": 121 }, { "epoch": 0.1726822363765039, "grad_norm": 0.3410997940493091, "learning_rate": 2.8537735849056606e-05, "loss": 0.6465, "step": 122 }, { "epoch": 0.1740976645435244, "grad_norm": 0.3424241485506368, "learning_rate": 2.8773584905660378e-05, "loss": 0.7005, "step": 123 }, { "epoch": 0.17551309271054494, "grad_norm": 0.3087363234800392, "learning_rate": 2.9009433962264154e-05, "loss": 0.7323, "step": 124 }, { "epoch": 0.17692852087756547, "grad_norm": 0.3208379612118477, "learning_rate": 2.9245283018867926e-05, "loss": 0.715, "step": 125 }, { "epoch": 0.17834394904458598, "grad_norm": 0.3252623056179373, "learning_rate": 2.9481132075471702e-05, "loss": 0.7182, "step": 126 }, { "epoch": 0.1797593772116065, "grad_norm": 0.3260606346469109, "learning_rate": 2.971698113207547e-05, "loss": 0.6292, "step": 127 }, { "epoch": 0.18117480537862704, "grad_norm": 0.33507755461970656, "learning_rate": 2.995283018867925e-05, "loss": 0.6781, "step": 128 }, { "epoch": 0.18259023354564755, "grad_norm": 0.344152749953347, "learning_rate": 3.018867924528302e-05, "loss": 0.731, "step": 129 }, { "epoch": 0.18400566171266808, "grad_norm": 0.3790502984860035, "learning_rate": 3.0424528301886794e-05, "loss": 0.7032, "step": 130 }, { "epoch": 0.1854210898796886, "grad_norm": 0.4060176596426599, "learning_rate": 3.0660377358490567e-05, "loss": 0.7258, "step": 131 }, { "epoch": 0.18683651804670912, "grad_norm": 0.29752681279467874, "learning_rate": 3.0896226415094346e-05, "loss": 0.5724, "step": 132 }, { "epoch": 0.18825194621372965, "grad_norm": 0.4490897392705167, "learning_rate": 3.113207547169811e-05, "loss": 0.7565, "step": 133 }, { "epoch": 0.18966737438075018, "grad_norm": 0.3748295546904542, "learning_rate": 3.136792452830189e-05, "loss": 0.7491, "step": 134 }, { "epoch": 0.1910828025477707, "grad_norm": 0.4096360150169126, "learning_rate": 3.160377358490566e-05, "loss": 0.7355, "step": 135 }, { "epoch": 0.19249823071479122, "grad_norm": 0.35372872386670523, "learning_rate": 3.1839622641509435e-05, "loss": 0.6501, "step": 136 }, { "epoch": 0.19391365888181175, "grad_norm": 0.41983196647145415, "learning_rate": 3.207547169811321e-05, "loss": 0.7145, "step": 137 }, { "epoch": 0.19532908704883228, "grad_norm": 0.4100283221259522, "learning_rate": 3.2311320754716986e-05, "loss": 0.7778, "step": 138 }, { "epoch": 0.1967445152158528, "grad_norm": 0.4279192449535737, "learning_rate": 3.254716981132075e-05, "loss": 0.6512, "step": 139 }, { "epoch": 0.19815994338287332, "grad_norm": 0.37722937839286214, "learning_rate": 3.278301886792453e-05, "loss": 0.6462, "step": 140 }, { "epoch": 0.19957537154989385, "grad_norm": 0.5177263030721986, "learning_rate": 3.30188679245283e-05, "loss": 0.7195, "step": 141 }, { "epoch": 0.20099079971691436, "grad_norm": 0.3516292254214785, "learning_rate": 3.3254716981132075e-05, "loss": 0.6927, "step": 142 }, { "epoch": 0.2024062278839349, "grad_norm": 0.4369868522037781, "learning_rate": 3.349056603773585e-05, "loss": 0.6543, "step": 143 }, { "epoch": 0.20382165605095542, "grad_norm": 0.351580146475787, "learning_rate": 3.3726415094339627e-05, "loss": 0.6476, "step": 144 }, { "epoch": 0.20523708421797593, "grad_norm": 0.382859664648938, "learning_rate": 3.39622641509434e-05, "loss": 0.6577, "step": 145 }, { "epoch": 0.20665251238499646, "grad_norm": 0.389040016792054, "learning_rate": 3.419811320754717e-05, "loss": 0.6965, "step": 146 }, { "epoch": 0.208067940552017, "grad_norm": 0.30073874667213135, "learning_rate": 3.4433962264150943e-05, "loss": 0.5977, "step": 147 }, { "epoch": 0.2094833687190375, "grad_norm": 0.38732380059479576, "learning_rate": 3.466981132075472e-05, "loss": 0.6837, "step": 148 }, { "epoch": 0.21089879688605803, "grad_norm": 0.34655294600757985, "learning_rate": 3.490566037735849e-05, "loss": 0.7161, "step": 149 }, { "epoch": 0.21231422505307856, "grad_norm": 0.3474622218711454, "learning_rate": 3.514150943396227e-05, "loss": 0.6879, "step": 150 }, { "epoch": 0.21372965322009907, "grad_norm": 0.3292597346994816, "learning_rate": 3.537735849056604e-05, "loss": 0.6955, "step": 151 }, { "epoch": 0.2151450813871196, "grad_norm": 0.3383153262618983, "learning_rate": 3.561320754716981e-05, "loss": 0.7442, "step": 152 }, { "epoch": 0.21656050955414013, "grad_norm": 0.3373136910347554, "learning_rate": 3.5849056603773584e-05, "loss": 0.6368, "step": 153 }, { "epoch": 0.21797593772116064, "grad_norm": 0.32338504092281495, "learning_rate": 3.608490566037736e-05, "loss": 0.6678, "step": 154 }, { "epoch": 0.21939136588818117, "grad_norm": 0.327291405252717, "learning_rate": 3.632075471698113e-05, "loss": 0.7245, "step": 155 }, { "epoch": 0.2208067940552017, "grad_norm": 0.4079198103017352, "learning_rate": 3.655660377358491e-05, "loss": 0.758, "step": 156 }, { "epoch": 0.2222222222222222, "grad_norm": 0.3431479182310241, "learning_rate": 3.679245283018868e-05, "loss": 0.6949, "step": 157 }, { "epoch": 0.22363765038924274, "grad_norm": 0.32740527265233493, "learning_rate": 3.702830188679245e-05, "loss": 0.6967, "step": 158 }, { "epoch": 0.22505307855626328, "grad_norm": 0.355081084230637, "learning_rate": 3.7264150943396224e-05, "loss": 0.6636, "step": 159 }, { "epoch": 0.22646850672328378, "grad_norm": 0.3377921193259222, "learning_rate": 3.7500000000000003e-05, "loss": 0.6152, "step": 160 }, { "epoch": 0.2278839348903043, "grad_norm": 0.35705251222543827, "learning_rate": 3.7735849056603776e-05, "loss": 0.7044, "step": 161 }, { "epoch": 0.22929936305732485, "grad_norm": 0.2990541370982297, "learning_rate": 3.797169811320755e-05, "loss": 0.6421, "step": 162 }, { "epoch": 0.23071479122434538, "grad_norm": 0.37680162093247666, "learning_rate": 3.820754716981133e-05, "loss": 0.7236, "step": 163 }, { "epoch": 0.23213021939136588, "grad_norm": 0.3316872340797433, "learning_rate": 3.844339622641509e-05, "loss": 0.592, "step": 164 }, { "epoch": 0.23354564755838642, "grad_norm": 0.3814426738691517, "learning_rate": 3.867924528301887e-05, "loss": 0.642, "step": 165 }, { "epoch": 0.23496107572540695, "grad_norm": 0.3267572380840606, "learning_rate": 3.8915094339622644e-05, "loss": 0.7495, "step": 166 }, { "epoch": 0.23637650389242745, "grad_norm": 0.33871837867570975, "learning_rate": 3.9150943396226416e-05, "loss": 0.6928, "step": 167 }, { "epoch": 0.23779193205944799, "grad_norm": 0.323902169015648, "learning_rate": 3.938679245283019e-05, "loss": 0.8054, "step": 168 }, { "epoch": 0.23920736022646852, "grad_norm": 0.3105694760719953, "learning_rate": 3.962264150943397e-05, "loss": 0.6868, "step": 169 }, { "epoch": 0.24062278839348902, "grad_norm": 0.3282634870080322, "learning_rate": 3.985849056603774e-05, "loss": 0.6943, "step": 170 }, { "epoch": 0.24203821656050956, "grad_norm": 0.3496477577080085, "learning_rate": 4.009433962264151e-05, "loss": 0.7847, "step": 171 }, { "epoch": 0.2434536447275301, "grad_norm": 0.28731112540621906, "learning_rate": 4.0330188679245284e-05, "loss": 0.6661, "step": 172 }, { "epoch": 0.2448690728945506, "grad_norm": 0.3546905432002279, "learning_rate": 4.0566037735849064e-05, "loss": 0.7017, "step": 173 }, { "epoch": 0.24628450106157113, "grad_norm": 0.30690572266461835, "learning_rate": 4.080188679245283e-05, "loss": 0.6241, "step": 174 }, { "epoch": 0.24769992922859166, "grad_norm": 0.3264453307239247, "learning_rate": 4.103773584905661e-05, "loss": 0.7084, "step": 175 }, { "epoch": 0.24911535739561216, "grad_norm": 0.356494327898739, "learning_rate": 4.127358490566038e-05, "loss": 0.6766, "step": 176 }, { "epoch": 0.2505307855626327, "grad_norm": 0.37845684989010364, "learning_rate": 4.150943396226415e-05, "loss": 0.6764, "step": 177 }, { "epoch": 0.2519462137296532, "grad_norm": 0.3822237153839302, "learning_rate": 4.1745283018867925e-05, "loss": 0.7572, "step": 178 }, { "epoch": 0.25336164189667376, "grad_norm": 0.33180805872343516, "learning_rate": 4.1981132075471704e-05, "loss": 0.5645, "step": 179 }, { "epoch": 0.25477707006369427, "grad_norm": 0.3776125949939335, "learning_rate": 4.221698113207547e-05, "loss": 0.6428, "step": 180 }, { "epoch": 0.25619249823071477, "grad_norm": 0.34307054586410674, "learning_rate": 4.245283018867925e-05, "loss": 0.7138, "step": 181 }, { "epoch": 0.25760792639773533, "grad_norm": 0.3421864387970046, "learning_rate": 4.268867924528302e-05, "loss": 0.6872, "step": 182 }, { "epoch": 0.25902335456475584, "grad_norm": 0.3558737974385253, "learning_rate": 4.292452830188679e-05, "loss": 0.6201, "step": 183 }, { "epoch": 0.26043878273177634, "grad_norm": 0.34802732619093407, "learning_rate": 4.3160377358490565e-05, "loss": 0.7949, "step": 184 }, { "epoch": 0.2618542108987969, "grad_norm": 0.47243969103960226, "learning_rate": 4.3396226415094345e-05, "loss": 0.7231, "step": 185 }, { "epoch": 0.2632696390658174, "grad_norm": 0.48027533389765475, "learning_rate": 4.363207547169812e-05, "loss": 0.6296, "step": 186 }, { "epoch": 0.2646850672328379, "grad_norm": 0.29575230575318046, "learning_rate": 4.386792452830189e-05, "loss": 0.5118, "step": 187 }, { "epoch": 0.26610049539985847, "grad_norm": 0.45243075941702265, "learning_rate": 4.410377358490566e-05, "loss": 0.8435, "step": 188 }, { "epoch": 0.267515923566879, "grad_norm": 0.4446508772321007, "learning_rate": 4.433962264150944e-05, "loss": 0.5955, "step": 189 }, { "epoch": 0.2689313517338995, "grad_norm": 0.39098158765333574, "learning_rate": 4.4575471698113206e-05, "loss": 0.7381, "step": 190 }, { "epoch": 0.27034677990092004, "grad_norm": 0.47812821613700407, "learning_rate": 4.4811320754716985e-05, "loss": 0.686, "step": 191 }, { "epoch": 0.27176220806794055, "grad_norm": 0.4255378884479202, "learning_rate": 4.504716981132076e-05, "loss": 0.6267, "step": 192 }, { "epoch": 0.27317763623496105, "grad_norm": 0.41920925771608414, "learning_rate": 4.528301886792453e-05, "loss": 0.7695, "step": 193 }, { "epoch": 0.2745930644019816, "grad_norm": 0.4901895726580499, "learning_rate": 4.55188679245283e-05, "loss": 0.6972, "step": 194 }, { "epoch": 0.2760084925690021, "grad_norm": 0.5172747736553965, "learning_rate": 4.575471698113208e-05, "loss": 0.6524, "step": 195 }, { "epoch": 0.2774239207360226, "grad_norm": 0.32758712427375986, "learning_rate": 4.5990566037735846e-05, "loss": 0.6214, "step": 196 }, { "epoch": 0.2788393489030432, "grad_norm": 0.6629263683616758, "learning_rate": 4.6226415094339625e-05, "loss": 0.6382, "step": 197 }, { "epoch": 0.2802547770700637, "grad_norm": 0.5255514509389033, "learning_rate": 4.64622641509434e-05, "loss": 0.6253, "step": 198 }, { "epoch": 0.2816702052370842, "grad_norm": 0.4199105090028493, "learning_rate": 4.669811320754717e-05, "loss": 0.6534, "step": 199 }, { "epoch": 0.28308563340410475, "grad_norm": 0.5133025788360546, "learning_rate": 4.693396226415094e-05, "loss": 0.5992, "step": 200 }, { "epoch": 0.28450106157112526, "grad_norm": 0.3515641457954602, "learning_rate": 4.716981132075472e-05, "loss": 0.5936, "step": 201 }, { "epoch": 0.28591648973814576, "grad_norm": 0.46382284062883583, "learning_rate": 4.7405660377358494e-05, "loss": 0.683, "step": 202 }, { "epoch": 0.2873319179051663, "grad_norm": 0.5789901497369425, "learning_rate": 4.7641509433962266e-05, "loss": 0.6319, "step": 203 }, { "epoch": 0.28874734607218683, "grad_norm": 0.3643121215092036, "learning_rate": 4.787735849056604e-05, "loss": 0.5249, "step": 204 }, { "epoch": 0.29016277423920733, "grad_norm": 0.47331340234321256, "learning_rate": 4.811320754716982e-05, "loss": 0.6546, "step": 205 }, { "epoch": 0.2915782024062279, "grad_norm": 0.40955091988638687, "learning_rate": 4.834905660377358e-05, "loss": 0.7085, "step": 206 }, { "epoch": 0.2929936305732484, "grad_norm": 0.35402965766051764, "learning_rate": 4.858490566037736e-05, "loss": 0.7199, "step": 207 }, { "epoch": 0.29440905874026896, "grad_norm": 0.33550518291489584, "learning_rate": 4.8820754716981134e-05, "loss": 0.6114, "step": 208 }, { "epoch": 0.29582448690728946, "grad_norm": 0.36970155693332196, "learning_rate": 4.9056603773584906e-05, "loss": 0.5843, "step": 209 }, { "epoch": 0.29723991507430997, "grad_norm": 0.3697693128093627, "learning_rate": 4.929245283018868e-05, "loss": 0.7422, "step": 210 }, { "epoch": 0.29865534324133053, "grad_norm": 0.41500202886671717, "learning_rate": 4.952830188679246e-05, "loss": 0.639, "step": 211 }, { "epoch": 0.30007077140835103, "grad_norm": 0.2856425084141532, "learning_rate": 4.976415094339622e-05, "loss": 0.6182, "step": 212 }, { "epoch": 0.30148619957537154, "grad_norm": 0.46521888562093133, "learning_rate": 5e-05, "loss": 0.722, "step": 213 }, { "epoch": 0.3029016277423921, "grad_norm": 0.3845414037620586, "learning_rate": 4.997376705141658e-05, "loss": 0.6922, "step": 214 }, { "epoch": 0.3043170559094126, "grad_norm": 0.45351901517196663, "learning_rate": 4.994753410283316e-05, "loss": 0.6222, "step": 215 }, { "epoch": 0.3057324840764331, "grad_norm": 0.38318818305297797, "learning_rate": 4.9921301154249736e-05, "loss": 0.6177, "step": 216 }, { "epoch": 0.30714791224345367, "grad_norm": 0.4406176060172244, "learning_rate": 4.989506820566632e-05, "loss": 0.6203, "step": 217 }, { "epoch": 0.3085633404104742, "grad_norm": 0.3798385623910793, "learning_rate": 4.98688352570829e-05, "loss": 0.6833, "step": 218 }, { "epoch": 0.3099787685774947, "grad_norm": 0.4117758569188405, "learning_rate": 4.984260230849948e-05, "loss": 0.5478, "step": 219 }, { "epoch": 0.31139419674451524, "grad_norm": 0.3651670593181467, "learning_rate": 4.981636935991606e-05, "loss": 0.6085, "step": 220 }, { "epoch": 0.31280962491153574, "grad_norm": 0.3985470259616175, "learning_rate": 4.979013641133264e-05, "loss": 0.6244, "step": 221 }, { "epoch": 0.31422505307855625, "grad_norm": 0.33366417189852593, "learning_rate": 4.976390346274922e-05, "loss": 0.6859, "step": 222 }, { "epoch": 0.3156404812455768, "grad_norm": 0.40063208383608917, "learning_rate": 4.97376705141658e-05, "loss": 0.6583, "step": 223 }, { "epoch": 0.3170559094125973, "grad_norm": 0.3406033500089553, "learning_rate": 4.971143756558237e-05, "loss": 0.75, "step": 224 }, { "epoch": 0.3184713375796178, "grad_norm": 0.3167388846013373, "learning_rate": 4.968520461699895e-05, "loss": 0.6368, "step": 225 }, { "epoch": 0.3198867657466384, "grad_norm": 0.30183382248828816, "learning_rate": 4.965897166841553e-05, "loss": 0.5887, "step": 226 }, { "epoch": 0.3213021939136589, "grad_norm": 0.3410080956406333, "learning_rate": 4.963273871983211e-05, "loss": 0.6944, "step": 227 }, { "epoch": 0.3227176220806794, "grad_norm": 0.3016417065921014, "learning_rate": 4.960650577124869e-05, "loss": 0.7025, "step": 228 }, { "epoch": 0.32413305024769995, "grad_norm": 0.39533283501293887, "learning_rate": 4.958027282266527e-05, "loss": 0.7395, "step": 229 }, { "epoch": 0.32554847841472045, "grad_norm": 0.317776843840245, "learning_rate": 4.955403987408185e-05, "loss": 0.6477, "step": 230 }, { "epoch": 0.32696390658174096, "grad_norm": 2.6162825696857506, "learning_rate": 4.952780692549843e-05, "loss": 0.7805, "step": 231 }, { "epoch": 0.3283793347487615, "grad_norm": 0.38246518870773766, "learning_rate": 4.950157397691501e-05, "loss": 0.5729, "step": 232 }, { "epoch": 0.329794762915782, "grad_norm": 0.35451653588292686, "learning_rate": 4.947534102833158e-05, "loss": 0.6034, "step": 233 }, { "epoch": 0.33121019108280253, "grad_norm": 0.35793206668480076, "learning_rate": 4.944910807974817e-05, "loss": 0.6168, "step": 234 }, { "epoch": 0.3326256192498231, "grad_norm": 0.3959020643773459, "learning_rate": 4.942287513116475e-05, "loss": 0.5585, "step": 235 }, { "epoch": 0.3340410474168436, "grad_norm": 0.3805393159052667, "learning_rate": 4.939664218258133e-05, "loss": 0.6529, "step": 236 }, { "epoch": 0.3354564755838641, "grad_norm": 0.34245689486771175, "learning_rate": 4.93704092339979e-05, "loss": 0.5962, "step": 237 }, { "epoch": 0.33687190375088466, "grad_norm": 0.46119035852206347, "learning_rate": 4.934417628541448e-05, "loss": 0.6898, "step": 238 }, { "epoch": 0.33828733191790517, "grad_norm": 0.2760463725152546, "learning_rate": 4.931794333683106e-05, "loss": 0.5918, "step": 239 }, { "epoch": 0.33970276008492567, "grad_norm": 0.4212781419762082, "learning_rate": 4.929171038824764e-05, "loss": 0.8071, "step": 240 }, { "epoch": 0.34111818825194623, "grad_norm": 0.3336146067216131, "learning_rate": 4.926547743966422e-05, "loss": 0.6629, "step": 241 }, { "epoch": 0.34253361641896674, "grad_norm": 0.34986136867361556, "learning_rate": 4.92392444910808e-05, "loss": 0.6263, "step": 242 }, { "epoch": 0.34394904458598724, "grad_norm": 0.3191579021579259, "learning_rate": 4.921301154249738e-05, "loss": 0.6556, "step": 243 }, { "epoch": 0.3453644727530078, "grad_norm": 0.35217932985524025, "learning_rate": 4.918677859391396e-05, "loss": 0.6628, "step": 244 }, { "epoch": 0.3467799009200283, "grad_norm": 0.3382215322275221, "learning_rate": 4.916054564533054e-05, "loss": 0.6878, "step": 245 }, { "epoch": 0.3481953290870488, "grad_norm": 0.3192853417565939, "learning_rate": 4.913431269674712e-05, "loss": 0.6305, "step": 246 }, { "epoch": 0.34961075725406937, "grad_norm": 0.39592306170954034, "learning_rate": 4.910807974816369e-05, "loss": 0.6313, "step": 247 }, { "epoch": 0.3510261854210899, "grad_norm": 0.3717422268582204, "learning_rate": 4.908184679958028e-05, "loss": 0.6787, "step": 248 }, { "epoch": 0.3524416135881104, "grad_norm": 0.4697879198033925, "learning_rate": 4.905561385099686e-05, "loss": 0.6593, "step": 249 }, { "epoch": 0.35385704175513094, "grad_norm": 0.3145225755984823, "learning_rate": 4.902938090241343e-05, "loss": 0.559, "step": 250 }, { "epoch": 0.35527246992215145, "grad_norm": 0.4055147305953671, "learning_rate": 4.900314795383001e-05, "loss": 0.6411, "step": 251 }, { "epoch": 0.35668789808917195, "grad_norm": 0.3296941362737166, "learning_rate": 4.897691500524659e-05, "loss": 0.6256, "step": 252 }, { "epoch": 0.3581033262561925, "grad_norm": 0.4307575396162028, "learning_rate": 4.895068205666317e-05, "loss": 0.5943, "step": 253 }, { "epoch": 0.359518754423213, "grad_norm": 0.3602007816421482, "learning_rate": 4.8924449108079753e-05, "loss": 0.6412, "step": 254 }, { "epoch": 0.3609341825902335, "grad_norm": 0.32293453159004193, "learning_rate": 4.889821615949633e-05, "loss": 0.5814, "step": 255 }, { "epoch": 0.3623496107572541, "grad_norm": 0.4395377605765523, "learning_rate": 4.887198321091291e-05, "loss": 0.6761, "step": 256 }, { "epoch": 0.3637650389242746, "grad_norm": 0.30063842475465097, "learning_rate": 4.884575026232949e-05, "loss": 0.6224, "step": 257 }, { "epoch": 0.3651804670912951, "grad_norm": 0.4466468804355786, "learning_rate": 4.881951731374607e-05, "loss": 0.6249, "step": 258 }, { "epoch": 0.36659589525831565, "grad_norm": 0.4402685935470065, "learning_rate": 4.879328436516265e-05, "loss": 0.6335, "step": 259 }, { "epoch": 0.36801132342533616, "grad_norm": 0.4103451796715846, "learning_rate": 4.876705141657922e-05, "loss": 0.694, "step": 260 }, { "epoch": 0.36942675159235666, "grad_norm": 0.45194536777403443, "learning_rate": 4.87408184679958e-05, "loss": 0.64, "step": 261 }, { "epoch": 0.3708421797593772, "grad_norm": 0.3933446116945307, "learning_rate": 4.871458551941239e-05, "loss": 0.6434, "step": 262 }, { "epoch": 0.3722576079263977, "grad_norm": 0.48408491103735724, "learning_rate": 4.868835257082896e-05, "loss": 0.668, "step": 263 }, { "epoch": 0.37367303609341823, "grad_norm": 0.3422922106651585, "learning_rate": 4.866211962224554e-05, "loss": 0.6148, "step": 264 }, { "epoch": 0.3750884642604388, "grad_norm": 0.47594899577183286, "learning_rate": 4.863588667366212e-05, "loss": 0.6714, "step": 265 }, { "epoch": 0.3765038924274593, "grad_norm": 0.40685951284647204, "learning_rate": 4.86096537250787e-05, "loss": 0.636, "step": 266 }, { "epoch": 0.37791932059447986, "grad_norm": 0.332440654967862, "learning_rate": 4.8583420776495284e-05, "loss": 0.6513, "step": 267 }, { "epoch": 0.37933474876150036, "grad_norm": 0.652696489413694, "learning_rate": 4.855718782791186e-05, "loss": 0.6459, "step": 268 }, { "epoch": 0.38075017692852087, "grad_norm": 0.3123837110864311, "learning_rate": 4.853095487932844e-05, "loss": 0.6556, "step": 269 }, { "epoch": 0.3821656050955414, "grad_norm": 0.4288519383694323, "learning_rate": 4.850472193074502e-05, "loss": 0.6684, "step": 270 }, { "epoch": 0.38358103326256193, "grad_norm": 0.48238162709886434, "learning_rate": 4.84784889821616e-05, "loss": 0.7107, "step": 271 }, { "epoch": 0.38499646142958244, "grad_norm": 0.38376189241758835, "learning_rate": 4.845225603357818e-05, "loss": 0.6473, "step": 272 }, { "epoch": 0.386411889596603, "grad_norm": 0.7127505698452409, "learning_rate": 4.842602308499475e-05, "loss": 0.6139, "step": 273 }, { "epoch": 0.3878273177636235, "grad_norm": 0.3692518181743927, "learning_rate": 4.839979013641133e-05, "loss": 0.677, "step": 274 }, { "epoch": 0.389242745930644, "grad_norm": 0.38701071771235995, "learning_rate": 4.837355718782791e-05, "loss": 0.6924, "step": 275 }, { "epoch": 0.39065817409766457, "grad_norm": 1.725951353723549, "learning_rate": 4.834732423924449e-05, "loss": 0.6087, "step": 276 }, { "epoch": 0.3920736022646851, "grad_norm": 1.942438208713734, "learning_rate": 4.832109129066107e-05, "loss": 0.7028, "step": 277 }, { "epoch": 0.3934890304317056, "grad_norm": 1.3706349023564903, "learning_rate": 4.829485834207765e-05, "loss": 0.6555, "step": 278 }, { "epoch": 0.39490445859872614, "grad_norm": 0.6987371799804422, "learning_rate": 4.8268625393494233e-05, "loss": 0.6366, "step": 279 }, { "epoch": 0.39631988676574664, "grad_norm": 0.48038293423554523, "learning_rate": 4.8242392444910814e-05, "loss": 0.6256, "step": 280 }, { "epoch": 0.39773531493276715, "grad_norm": 0.41106735186436566, "learning_rate": 4.821615949632739e-05, "loss": 0.6777, "step": 281 }, { "epoch": 0.3991507430997877, "grad_norm": 0.5165628841938925, "learning_rate": 4.818992654774397e-05, "loss": 0.6373, "step": 282 }, { "epoch": 0.4005661712668082, "grad_norm": 2.4701073204434203, "learning_rate": 4.816369359916055e-05, "loss": 0.656, "step": 283 }, { "epoch": 0.4019815994338287, "grad_norm": 0.5737260562005873, "learning_rate": 4.813746065057713e-05, "loss": 0.6091, "step": 284 }, { "epoch": 0.4033970276008493, "grad_norm": 0.3502815963568765, "learning_rate": 4.811122770199371e-05, "loss": 0.571, "step": 285 }, { "epoch": 0.4048124557678698, "grad_norm": 6.354647535609901, "learning_rate": 4.808499475341028e-05, "loss": 0.6529, "step": 286 }, { "epoch": 0.4062278839348903, "grad_norm": 0.6715967469992172, "learning_rate": 4.805876180482686e-05, "loss": 0.6772, "step": 287 }, { "epoch": 0.40764331210191085, "grad_norm": 0.4037300470693345, "learning_rate": 4.803252885624344e-05, "loss": 0.5825, "step": 288 }, { "epoch": 0.40905874026893135, "grad_norm": 0.6251534477877752, "learning_rate": 4.800629590766002e-05, "loss": 0.6897, "step": 289 }, { "epoch": 0.41047416843595186, "grad_norm": 0.5295316359819032, "learning_rate": 4.79800629590766e-05, "loss": 0.6591, "step": 290 }, { "epoch": 0.4118895966029724, "grad_norm": 0.43307943598387333, "learning_rate": 4.795383001049318e-05, "loss": 0.6226, "step": 291 }, { "epoch": 0.4133050247699929, "grad_norm": 1.2645789414674928, "learning_rate": 4.7927597061909764e-05, "loss": 0.5863, "step": 292 }, { "epoch": 0.41472045293701343, "grad_norm": 0.43201269228704425, "learning_rate": 4.7901364113326344e-05, "loss": 0.5843, "step": 293 }, { "epoch": 0.416135881104034, "grad_norm": 0.40349127042570915, "learning_rate": 4.787513116474292e-05, "loss": 0.549, "step": 294 }, { "epoch": 0.4175513092710545, "grad_norm": 0.6492196187297071, "learning_rate": 4.78488982161595e-05, "loss": 0.6521, "step": 295 }, { "epoch": 0.418966737438075, "grad_norm": 0.5361960066185653, "learning_rate": 4.782266526757608e-05, "loss": 0.6529, "step": 296 }, { "epoch": 0.42038216560509556, "grad_norm": 1.9127877440573802, "learning_rate": 4.779643231899266e-05, "loss": 0.6502, "step": 297 }, { "epoch": 0.42179759377211606, "grad_norm": 0.4987202952653217, "learning_rate": 4.777019937040924e-05, "loss": 0.5868, "step": 298 }, { "epoch": 0.42321302193913657, "grad_norm": 0.42834190595019034, "learning_rate": 4.774396642182581e-05, "loss": 0.588, "step": 299 }, { "epoch": 0.42462845010615713, "grad_norm": 0.37598238951488705, "learning_rate": 4.771773347324239e-05, "loss": 0.63, "step": 300 }, { "epoch": 0.42604387827317763, "grad_norm": 0.4023043136893649, "learning_rate": 4.769150052465897e-05, "loss": 0.563, "step": 301 }, { "epoch": 0.42745930644019814, "grad_norm": 1.8869422299730276, "learning_rate": 4.766526757607555e-05, "loss": 0.5939, "step": 302 }, { "epoch": 0.4288747346072187, "grad_norm": 0.4570024354376153, "learning_rate": 4.763903462749213e-05, "loss": 0.5991, "step": 303 }, { "epoch": 0.4302901627742392, "grad_norm": 0.33510204236751395, "learning_rate": 4.7612801678908713e-05, "loss": 0.5857, "step": 304 }, { "epoch": 0.4317055909412597, "grad_norm": 4.13866712753077, "learning_rate": 4.7586568730325294e-05, "loss": 0.6918, "step": 305 }, { "epoch": 0.43312101910828027, "grad_norm": 0.566238838007585, "learning_rate": 4.7560335781741874e-05, "loss": 0.6613, "step": 306 }, { "epoch": 0.4345364472753008, "grad_norm": 0.4204911407406948, "learning_rate": 4.753410283315845e-05, "loss": 0.6573, "step": 307 }, { "epoch": 0.4359518754423213, "grad_norm": 0.4687083421971458, "learning_rate": 4.750786988457503e-05, "loss": 0.6571, "step": 308 }, { "epoch": 0.43736730360934184, "grad_norm": 5.055944341640214, "learning_rate": 4.748163693599161e-05, "loss": 0.6597, "step": 309 }, { "epoch": 0.43878273177636234, "grad_norm": 0.5275330144202897, "learning_rate": 4.745540398740819e-05, "loss": 0.7025, "step": 310 }, { "epoch": 0.44019815994338285, "grad_norm": 0.38739607411426685, "learning_rate": 4.742917103882477e-05, "loss": 0.6323, "step": 311 }, { "epoch": 0.4416135881104034, "grad_norm": 1.5067287407388463, "learning_rate": 4.740293809024134e-05, "loss": 0.6731, "step": 312 }, { "epoch": 0.4430290162774239, "grad_norm": 0.4590664765317656, "learning_rate": 4.737670514165792e-05, "loss": 0.6059, "step": 313 }, { "epoch": 0.4444444444444444, "grad_norm": 0.38594803494368446, "learning_rate": 4.73504721930745e-05, "loss": 0.6469, "step": 314 }, { "epoch": 0.445859872611465, "grad_norm": 1.6681766507590514, "learning_rate": 4.732423924449108e-05, "loss": 0.6949, "step": 315 }, { "epoch": 0.4472753007784855, "grad_norm": 2.2024847217066563, "learning_rate": 4.7298006295907657e-05, "loss": 0.6696, "step": 316 }, { "epoch": 0.448690728945506, "grad_norm": 0.6485924362900419, "learning_rate": 4.7271773347324244e-05, "loss": 0.6551, "step": 317 }, { "epoch": 0.45010615711252655, "grad_norm": 0.32656948375873524, "learning_rate": 4.7245540398740824e-05, "loss": 0.5412, "step": 318 }, { "epoch": 0.45152158527954706, "grad_norm": 0.525034119738257, "learning_rate": 4.7219307450157404e-05, "loss": 0.7114, "step": 319 }, { "epoch": 0.45293701344656756, "grad_norm": 0.3910634973163426, "learning_rate": 4.719307450157398e-05, "loss": 0.6183, "step": 320 }, { "epoch": 0.4543524416135881, "grad_norm": 0.6158827475228706, "learning_rate": 4.716684155299056e-05, "loss": 0.7741, "step": 321 }, { "epoch": 0.4557678697806086, "grad_norm": 0.4390670439291706, "learning_rate": 4.714060860440714e-05, "loss": 0.6353, "step": 322 }, { "epoch": 0.45718329794762913, "grad_norm": 0.3634832232181815, "learning_rate": 4.711437565582372e-05, "loss": 0.6317, "step": 323 }, { "epoch": 0.4585987261146497, "grad_norm": 0.33099299386363096, "learning_rate": 4.70881427072403e-05, "loss": 0.5928, "step": 324 }, { "epoch": 0.4600141542816702, "grad_norm": 0.7283175564058771, "learning_rate": 4.706190975865687e-05, "loss": 0.5132, "step": 325 }, { "epoch": 0.46142958244869076, "grad_norm": 0.39663831734790206, "learning_rate": 4.703567681007345e-05, "loss": 0.5949, "step": 326 }, { "epoch": 0.46284501061571126, "grad_norm": 0.33617445280552166, "learning_rate": 4.700944386149003e-05, "loss": 0.67, "step": 327 }, { "epoch": 0.46426043878273177, "grad_norm": 0.3790314173603713, "learning_rate": 4.698321091290661e-05, "loss": 0.6469, "step": 328 }, { "epoch": 0.4656758669497523, "grad_norm": 0.3641500603632123, "learning_rate": 4.695697796432319e-05, "loss": 0.6129, "step": 329 }, { "epoch": 0.46709129511677283, "grad_norm": 0.38523640099062106, "learning_rate": 4.693074501573977e-05, "loss": 0.6352, "step": 330 }, { "epoch": 0.46850672328379334, "grad_norm": 0.31533765441818223, "learning_rate": 4.6904512067156354e-05, "loss": 0.5071, "step": 331 }, { "epoch": 0.4699221514508139, "grad_norm": 0.3607326185265803, "learning_rate": 4.6878279118572934e-05, "loss": 0.5643, "step": 332 }, { "epoch": 0.4713375796178344, "grad_norm": 0.3340538630754626, "learning_rate": 4.685204616998951e-05, "loss": 0.6602, "step": 333 }, { "epoch": 0.4727530077848549, "grad_norm": 0.3653458517506279, "learning_rate": 4.682581322140609e-05, "loss": 0.5338, "step": 334 }, { "epoch": 0.47416843595187547, "grad_norm": 1.0688271736428423, "learning_rate": 4.679958027282267e-05, "loss": 0.6009, "step": 335 }, { "epoch": 0.47558386411889597, "grad_norm": 0.3809122272722356, "learning_rate": 4.677334732423925e-05, "loss": 0.6423, "step": 336 }, { "epoch": 0.4769992922859165, "grad_norm": 0.39627182381703874, "learning_rate": 4.674711437565583e-05, "loss": 0.6301, "step": 337 }, { "epoch": 0.47841472045293704, "grad_norm": 0.38449241857885996, "learning_rate": 4.67208814270724e-05, "loss": 0.7192, "step": 338 }, { "epoch": 0.47983014861995754, "grad_norm": 0.35204634256821776, "learning_rate": 4.669464847848898e-05, "loss": 0.5858, "step": 339 }, { "epoch": 0.48124557678697805, "grad_norm": 0.3608851500498845, "learning_rate": 4.666841552990556e-05, "loss": 0.6616, "step": 340 }, { "epoch": 0.4826610049539986, "grad_norm": 2.044217701010847, "learning_rate": 4.664218258132214e-05, "loss": 0.6232, "step": 341 }, { "epoch": 0.4840764331210191, "grad_norm": 0.48159872669244885, "learning_rate": 4.661594963273872e-05, "loss": 0.6532, "step": 342 }, { "epoch": 0.4854918612880396, "grad_norm": 0.3124760248273583, "learning_rate": 4.65897166841553e-05, "loss": 0.6304, "step": 343 }, { "epoch": 0.4869072894550602, "grad_norm": 1.1543392767277225, "learning_rate": 4.656348373557188e-05, "loss": 0.5875, "step": 344 }, { "epoch": 0.4883227176220807, "grad_norm": 0.3944218685357484, "learning_rate": 4.6537250786988465e-05, "loss": 0.6112, "step": 345 }, { "epoch": 0.4897381457891012, "grad_norm": 0.3257422007112123, "learning_rate": 4.651101783840504e-05, "loss": 0.5728, "step": 346 }, { "epoch": 0.49115357395612175, "grad_norm": 1.1660986725076619, "learning_rate": 4.648478488982162e-05, "loss": 0.668, "step": 347 }, { "epoch": 0.49256900212314225, "grad_norm": 0.37555157866666583, "learning_rate": 4.64585519412382e-05, "loss": 0.6376, "step": 348 }, { "epoch": 0.49398443029016276, "grad_norm": 0.37256619809919633, "learning_rate": 4.643231899265478e-05, "loss": 0.6579, "step": 349 }, { "epoch": 0.4953998584571833, "grad_norm": 0.36137278553859004, "learning_rate": 4.640608604407136e-05, "loss": 0.6346, "step": 350 }, { "epoch": 0.4968152866242038, "grad_norm": 0.3116835508059875, "learning_rate": 4.637985309548793e-05, "loss": 0.6401, "step": 351 }, { "epoch": 0.4982307147912243, "grad_norm": 0.3651551236969446, "learning_rate": 4.635362014690451e-05, "loss": 0.5816, "step": 352 }, { "epoch": 0.4996461429582449, "grad_norm": 0.3207905146349543, "learning_rate": 4.632738719832109e-05, "loss": 0.576, "step": 353 }, { "epoch": 0.5010615711252654, "grad_norm": 0.3633417389596548, "learning_rate": 4.6301154249737674e-05, "loss": 0.5614, "step": 354 }, { "epoch": 0.502476999292286, "grad_norm": 0.34577140815932816, "learning_rate": 4.627492130115425e-05, "loss": 0.6194, "step": 355 }, { "epoch": 0.5038924274593064, "grad_norm": 0.6817177742810275, "learning_rate": 4.624868835257083e-05, "loss": 0.5554, "step": 356 }, { "epoch": 0.505307855626327, "grad_norm": 0.33098759185554166, "learning_rate": 4.622245540398741e-05, "loss": 0.6662, "step": 357 }, { "epoch": 0.5067232837933475, "grad_norm": 0.3091638749967989, "learning_rate": 4.619622245540399e-05, "loss": 0.5937, "step": 358 }, { "epoch": 0.508138711960368, "grad_norm": 0.33686096205760013, "learning_rate": 4.616998950682057e-05, "loss": 0.6425, "step": 359 }, { "epoch": 0.5095541401273885, "grad_norm": 0.4555661086258636, "learning_rate": 4.614375655823715e-05, "loss": 0.6441, "step": 360 }, { "epoch": 0.5109695682944091, "grad_norm": 0.3597460163123963, "learning_rate": 4.611752360965373e-05, "loss": 0.62, "step": 361 }, { "epoch": 0.5123849964614295, "grad_norm": 0.3252882049035588, "learning_rate": 4.609129066107031e-05, "loss": 0.6298, "step": 362 }, { "epoch": 0.5138004246284501, "grad_norm": 0.27357662277889927, "learning_rate": 4.606505771248689e-05, "loss": 0.6079, "step": 363 }, { "epoch": 0.5152158527954707, "grad_norm": 0.34345308801248725, "learning_rate": 4.603882476390346e-05, "loss": 0.6715, "step": 364 }, { "epoch": 0.5166312809624911, "grad_norm": 1.1812651580481537, "learning_rate": 4.601259181532004e-05, "loss": 0.6276, "step": 365 }, { "epoch": 0.5180467091295117, "grad_norm": 0.3433024108167062, "learning_rate": 4.5986358866736623e-05, "loss": 0.5972, "step": 366 }, { "epoch": 0.5194621372965322, "grad_norm": 0.7489416791007744, "learning_rate": 4.5960125918153204e-05, "loss": 0.5834, "step": 367 }, { "epoch": 0.5208775654635527, "grad_norm": 1.7901825570043197, "learning_rate": 4.5933892969569784e-05, "loss": 0.7262, "step": 368 }, { "epoch": 0.5222929936305732, "grad_norm": 0.35535801236874504, "learning_rate": 4.590766002098636e-05, "loss": 0.6818, "step": 369 }, { "epoch": 0.5237084217975938, "grad_norm": 0.32103386113441407, "learning_rate": 4.588142707240294e-05, "loss": 0.5779, "step": 370 }, { "epoch": 0.5251238499646143, "grad_norm": 0.3412242384849581, "learning_rate": 4.585519412381952e-05, "loss": 0.6121, "step": 371 }, { "epoch": 0.5265392781316348, "grad_norm": 0.3032846386207085, "learning_rate": 4.58289611752361e-05, "loss": 0.5789, "step": 372 }, { "epoch": 0.5279547062986554, "grad_norm": 0.35028746399522914, "learning_rate": 4.580272822665268e-05, "loss": 0.6587, "step": 373 }, { "epoch": 0.5293701344656758, "grad_norm": 0.3475580934754737, "learning_rate": 4.577649527806926e-05, "loss": 0.6348, "step": 374 }, { "epoch": 0.5307855626326964, "grad_norm": 0.2858824540747321, "learning_rate": 4.575026232948584e-05, "loss": 0.5532, "step": 375 }, { "epoch": 0.5322009907997169, "grad_norm": 0.3775123937334319, "learning_rate": 4.572402938090242e-05, "loss": 0.7839, "step": 376 }, { "epoch": 0.5336164189667374, "grad_norm": 1.821003071814801, "learning_rate": 4.569779643231899e-05, "loss": 0.7244, "step": 377 }, { "epoch": 0.535031847133758, "grad_norm": 0.36372118605519194, "learning_rate": 4.567156348373557e-05, "loss": 0.5677, "step": 378 }, { "epoch": 0.5364472753007785, "grad_norm": 0.3392417180576534, "learning_rate": 4.5645330535152154e-05, "loss": 0.6502, "step": 379 }, { "epoch": 0.537862703467799, "grad_norm": 0.32558658253497313, "learning_rate": 4.5619097586568734e-05, "loss": 0.607, "step": 380 }, { "epoch": 0.5392781316348195, "grad_norm": 0.3057510804204223, "learning_rate": 4.5592864637985314e-05, "loss": 0.5875, "step": 381 }, { "epoch": 0.5406935598018401, "grad_norm": 1.4205434623418252, "learning_rate": 4.556663168940189e-05, "loss": 0.6474, "step": 382 }, { "epoch": 0.5421089879688605, "grad_norm": 0.43334996565829886, "learning_rate": 4.554039874081847e-05, "loss": 0.6331, "step": 383 }, { "epoch": 0.5435244161358811, "grad_norm": 0.2893938355609154, "learning_rate": 4.551416579223505e-05, "loss": 0.6066, "step": 384 }, { "epoch": 0.5449398443029017, "grad_norm": 0.385254742728004, "learning_rate": 4.548793284365163e-05, "loss": 0.561, "step": 385 }, { "epoch": 0.5463552724699221, "grad_norm": 0.4067804940921093, "learning_rate": 4.546169989506821e-05, "loss": 0.6397, "step": 386 }, { "epoch": 0.5477707006369427, "grad_norm": 0.3634827637755242, "learning_rate": 4.543546694648479e-05, "loss": 0.5953, "step": 387 }, { "epoch": 0.5491861288039632, "grad_norm": 0.4063791990642425, "learning_rate": 4.540923399790137e-05, "loss": 0.6715, "step": 388 }, { "epoch": 0.5506015569709837, "grad_norm": 0.36506068471085257, "learning_rate": 4.538300104931795e-05, "loss": 0.5811, "step": 389 }, { "epoch": 0.5520169851380042, "grad_norm": 0.3857615944899794, "learning_rate": 4.535676810073452e-05, "loss": 0.5902, "step": 390 }, { "epoch": 0.5534324133050248, "grad_norm": 0.3591246927876168, "learning_rate": 4.5330535152151103e-05, "loss": 0.5877, "step": 391 }, { "epoch": 0.5548478414720452, "grad_norm": 0.32793131919774665, "learning_rate": 4.5304302203567684e-05, "loss": 0.5179, "step": 392 }, { "epoch": 0.5562632696390658, "grad_norm": 0.41862081925025835, "learning_rate": 4.5278069254984264e-05, "loss": 0.5812, "step": 393 }, { "epoch": 0.5576786978060864, "grad_norm": 0.37153219913252544, "learning_rate": 4.5251836306400844e-05, "loss": 0.6806, "step": 394 }, { "epoch": 0.5590941259731068, "grad_norm": 0.3899375923227849, "learning_rate": 4.522560335781742e-05, "loss": 0.6399, "step": 395 }, { "epoch": 0.5605095541401274, "grad_norm": 0.2931470563734505, "learning_rate": 4.5199370409234e-05, "loss": 0.6244, "step": 396 }, { "epoch": 0.5619249823071479, "grad_norm": 3.92179635800638, "learning_rate": 4.517313746065058e-05, "loss": 0.7752, "step": 397 }, { "epoch": 0.5633404104741684, "grad_norm": 0.5167593678135474, "learning_rate": 4.514690451206716e-05, "loss": 0.7345, "step": 398 }, { "epoch": 0.564755838641189, "grad_norm": 0.3106438677475814, "learning_rate": 4.512067156348373e-05, "loss": 0.5603, "step": 399 }, { "epoch": 0.5661712668082095, "grad_norm": 4.58393490326734, "learning_rate": 4.509443861490032e-05, "loss": 0.6884, "step": 400 }, { "epoch": 0.56758669497523, "grad_norm": 0.6102958460821443, "learning_rate": 4.50682056663169e-05, "loss": 0.6524, "step": 401 }, { "epoch": 0.5690021231422505, "grad_norm": 0.3886188068845869, "learning_rate": 4.504197271773348e-05, "loss": 0.5546, "step": 402 }, { "epoch": 0.5704175513092711, "grad_norm": 0.4812854857575041, "learning_rate": 4.501573976915005e-05, "loss": 0.6143, "step": 403 }, { "epoch": 0.5718329794762915, "grad_norm": 0.5122348712154553, "learning_rate": 4.4989506820566634e-05, "loss": 0.6373, "step": 404 }, { "epoch": 0.5732484076433121, "grad_norm": 0.5776634035946314, "learning_rate": 4.4963273871983214e-05, "loss": 0.7066, "step": 405 }, { "epoch": 0.5746638358103326, "grad_norm": 0.32137764365447785, "learning_rate": 4.4937040923399794e-05, "loss": 0.5542, "step": 406 }, { "epoch": 0.5760792639773531, "grad_norm": 0.5479916180781855, "learning_rate": 4.4910807974816374e-05, "loss": 0.6616, "step": 407 }, { "epoch": 0.5774946921443737, "grad_norm": 0.4386841207404295, "learning_rate": 4.488457502623295e-05, "loss": 0.5806, "step": 408 }, { "epoch": 0.5789101203113942, "grad_norm": 0.2996291147474398, "learning_rate": 4.485834207764953e-05, "loss": 0.6317, "step": 409 }, { "epoch": 0.5803255484784147, "grad_norm": 0.423626648128835, "learning_rate": 4.483210912906611e-05, "loss": 0.5113, "step": 410 }, { "epoch": 0.5817409766454352, "grad_norm": 0.37350221595148414, "learning_rate": 4.480587618048269e-05, "loss": 0.5596, "step": 411 }, { "epoch": 0.5831564048124558, "grad_norm": 0.31258204461663813, "learning_rate": 4.477964323189926e-05, "loss": 0.561, "step": 412 }, { "epoch": 0.5845718329794763, "grad_norm": 0.42453118149870256, "learning_rate": 4.475341028331584e-05, "loss": 0.6642, "step": 413 }, { "epoch": 0.5859872611464968, "grad_norm": 0.3307633002928578, "learning_rate": 4.472717733473243e-05, "loss": 0.6141, "step": 414 }, { "epoch": 0.5874026893135174, "grad_norm": 0.32803157792209847, "learning_rate": 4.470094438614901e-05, "loss": 0.4943, "step": 415 }, { "epoch": 0.5888181174805379, "grad_norm": 0.2952618534431419, "learning_rate": 4.4674711437565583e-05, "loss": 0.5624, "step": 416 }, { "epoch": 0.5902335456475584, "grad_norm": 0.32571174942874237, "learning_rate": 4.4648478488982164e-05, "loss": 0.5794, "step": 417 }, { "epoch": 0.5916489738145789, "grad_norm": 0.3121943905539143, "learning_rate": 4.4622245540398744e-05, "loss": 0.5624, "step": 418 }, { "epoch": 0.5930644019815995, "grad_norm": 0.31577278636640205, "learning_rate": 4.4596012591815324e-05, "loss": 0.6004, "step": 419 }, { "epoch": 0.5944798301486199, "grad_norm": 0.4489394982468287, "learning_rate": 4.4569779643231905e-05, "loss": 0.6909, "step": 420 }, { "epoch": 0.5958952583156405, "grad_norm": 1.6471017152087755, "learning_rate": 4.454354669464848e-05, "loss": 0.7002, "step": 421 }, { "epoch": 0.5973106864826611, "grad_norm": 0.3439433119149739, "learning_rate": 4.451731374606506e-05, "loss": 0.492, "step": 422 }, { "epoch": 0.5987261146496815, "grad_norm": 0.3266215336524973, "learning_rate": 4.449108079748164e-05, "loss": 0.6881, "step": 423 }, { "epoch": 0.6001415428167021, "grad_norm": 0.2864170199424571, "learning_rate": 4.446484784889822e-05, "loss": 0.6268, "step": 424 }, { "epoch": 0.6015569709837226, "grad_norm": 0.3395106503408427, "learning_rate": 4.443861490031479e-05, "loss": 0.5892, "step": 425 }, { "epoch": 0.6029723991507431, "grad_norm": 0.3257927040748985, "learning_rate": 4.441238195173137e-05, "loss": 0.5565, "step": 426 }, { "epoch": 0.6043878273177636, "grad_norm": 0.26864176496804787, "learning_rate": 4.438614900314795e-05, "loss": 0.5998, "step": 427 }, { "epoch": 0.6058032554847842, "grad_norm": 0.298872589737985, "learning_rate": 4.435991605456454e-05, "loss": 0.5388, "step": 428 }, { "epoch": 0.6072186836518046, "grad_norm": 0.3285567835993635, "learning_rate": 4.4333683105981114e-05, "loss": 0.5981, "step": 429 }, { "epoch": 0.6086341118188252, "grad_norm": 0.2935869880085779, "learning_rate": 4.4307450157397694e-05, "loss": 0.6392, "step": 430 }, { "epoch": 0.6100495399858458, "grad_norm": 0.311961309731244, "learning_rate": 4.4281217208814274e-05, "loss": 0.5346, "step": 431 }, { "epoch": 0.6114649681528662, "grad_norm": 0.2937715983397586, "learning_rate": 4.4254984260230854e-05, "loss": 0.513, "step": 432 }, { "epoch": 0.6128803963198868, "grad_norm": 0.3086091462694762, "learning_rate": 4.4228751311647435e-05, "loss": 0.6774, "step": 433 }, { "epoch": 0.6142958244869073, "grad_norm": 0.3137400042943637, "learning_rate": 4.420251836306401e-05, "loss": 0.6156, "step": 434 }, { "epoch": 0.6157112526539278, "grad_norm": 0.301249268154309, "learning_rate": 4.417628541448059e-05, "loss": 0.6278, "step": 435 }, { "epoch": 0.6171266808209483, "grad_norm": 1.2953455962994163, "learning_rate": 4.415005246589717e-05, "loss": 0.6142, "step": 436 }, { "epoch": 0.6185421089879689, "grad_norm": 0.3825945056357451, "learning_rate": 4.412381951731375e-05, "loss": 0.5328, "step": 437 }, { "epoch": 0.6199575371549894, "grad_norm": 0.34362025288757764, "learning_rate": 4.409758656873032e-05, "loss": 0.5534, "step": 438 }, { "epoch": 0.6213729653220099, "grad_norm": 0.3020577660807664, "learning_rate": 4.40713536201469e-05, "loss": 0.6801, "step": 439 }, { "epoch": 0.6227883934890305, "grad_norm": 0.3773050366602719, "learning_rate": 4.404512067156348e-05, "loss": 0.5966, "step": 440 }, { "epoch": 0.6242038216560509, "grad_norm": 3.7089677265124594, "learning_rate": 4.401888772298007e-05, "loss": 0.8189, "step": 441 }, { "epoch": 0.6256192498230715, "grad_norm": 0.4274436279207135, "learning_rate": 4.3992654774396644e-05, "loss": 0.514, "step": 442 }, { "epoch": 0.627034677990092, "grad_norm": 0.36160488766019305, "learning_rate": 4.3966421825813224e-05, "loss": 0.5859, "step": 443 }, { "epoch": 0.6284501061571125, "grad_norm": 0.3227055099211922, "learning_rate": 4.3940188877229804e-05, "loss": 0.5393, "step": 444 }, { "epoch": 0.6298655343241331, "grad_norm": 0.6724594497481464, "learning_rate": 4.3913955928646385e-05, "loss": 0.6873, "step": 445 }, { "epoch": 0.6312809624911536, "grad_norm": 0.3723175424196791, "learning_rate": 4.3887722980062965e-05, "loss": 0.5801, "step": 446 }, { "epoch": 0.6326963906581741, "grad_norm": 0.30694921528390495, "learning_rate": 4.386149003147954e-05, "loss": 0.6157, "step": 447 }, { "epoch": 0.6341118188251946, "grad_norm": 0.4225504956814101, "learning_rate": 4.383525708289612e-05, "loss": 0.5958, "step": 448 }, { "epoch": 0.6355272469922152, "grad_norm": 0.3464379091821917, "learning_rate": 4.38090241343127e-05, "loss": 0.6272, "step": 449 }, { "epoch": 0.6369426751592356, "grad_norm": 0.29965769968222034, "learning_rate": 4.378279118572928e-05, "loss": 0.6077, "step": 450 }, { "epoch": 0.6383581033262562, "grad_norm": 0.3576403182854974, "learning_rate": 4.375655823714585e-05, "loss": 0.6889, "step": 451 }, { "epoch": 0.6397735314932768, "grad_norm": 0.3809053842975852, "learning_rate": 4.373032528856243e-05, "loss": 0.6196, "step": 452 }, { "epoch": 0.6411889596602972, "grad_norm": 0.2609993967316393, "learning_rate": 4.370409233997901e-05, "loss": 0.5372, "step": 453 }, { "epoch": 0.6426043878273178, "grad_norm": 0.33683959705079014, "learning_rate": 4.3677859391395594e-05, "loss": 0.6254, "step": 454 }, { "epoch": 0.6440198159943383, "grad_norm": 0.33712328791082974, "learning_rate": 4.3651626442812174e-05, "loss": 0.6412, "step": 455 }, { "epoch": 0.6454352441613588, "grad_norm": 0.29273668833002314, "learning_rate": 4.3625393494228754e-05, "loss": 0.6813, "step": 456 }, { "epoch": 0.6468506723283793, "grad_norm": 0.2884106844348947, "learning_rate": 4.3599160545645335e-05, "loss": 0.6326, "step": 457 }, { "epoch": 0.6482661004953999, "grad_norm": 0.3008543789946977, "learning_rate": 4.3572927597061915e-05, "loss": 0.5453, "step": 458 }, { "epoch": 0.6496815286624203, "grad_norm": 10.35738969499521, "learning_rate": 4.3546694648478495e-05, "loss": 1.025, "step": 459 }, { "epoch": 0.6510969568294409, "grad_norm": 0.3714989087251277, "learning_rate": 4.352046169989507e-05, "loss": 0.6139, "step": 460 }, { "epoch": 0.6525123849964615, "grad_norm": 0.2980789729774163, "learning_rate": 4.349422875131165e-05, "loss": 0.537, "step": 461 }, { "epoch": 0.6539278131634819, "grad_norm": 0.3303567136210746, "learning_rate": 4.346799580272823e-05, "loss": 0.6871, "step": 462 }, { "epoch": 0.6553432413305025, "grad_norm": 0.30196589524687256, "learning_rate": 4.344176285414481e-05, "loss": 0.5793, "step": 463 }, { "epoch": 0.656758669497523, "grad_norm": 0.45188015017130356, "learning_rate": 4.341552990556138e-05, "loss": 0.482, "step": 464 }, { "epoch": 0.6581740976645435, "grad_norm": 0.3025850006337929, "learning_rate": 4.338929695697796e-05, "loss": 0.6142, "step": 465 }, { "epoch": 0.659589525831564, "grad_norm": 0.5015898326623891, "learning_rate": 4.3363064008394544e-05, "loss": 0.6674, "step": 466 }, { "epoch": 0.6610049539985846, "grad_norm": 0.3145802398243886, "learning_rate": 4.3336831059811124e-05, "loss": 0.581, "step": 467 }, { "epoch": 0.6624203821656051, "grad_norm": 0.27586547369610387, "learning_rate": 4.3310598111227704e-05, "loss": 0.6641, "step": 468 }, { "epoch": 0.6638358103326256, "grad_norm": 0.33843075842140036, "learning_rate": 4.3284365162644284e-05, "loss": 0.6853, "step": 469 }, { "epoch": 0.6652512384996462, "grad_norm": 0.2765663953850667, "learning_rate": 4.3258132214060865e-05, "loss": 0.6237, "step": 470 }, { "epoch": 0.6666666666666666, "grad_norm": 0.2898526449010812, "learning_rate": 4.3231899265477445e-05, "loss": 0.5708, "step": 471 }, { "epoch": 0.6680820948336872, "grad_norm": 0.27564496200589295, "learning_rate": 4.3205666316894025e-05, "loss": 0.5657, "step": 472 }, { "epoch": 0.6694975230007078, "grad_norm": 0.25175800000637805, "learning_rate": 4.31794333683106e-05, "loss": 0.6057, "step": 473 }, { "epoch": 0.6709129511677282, "grad_norm": 0.322554869858225, "learning_rate": 4.315320041972718e-05, "loss": 0.6661, "step": 474 }, { "epoch": 0.6723283793347488, "grad_norm": 0.28101804468399305, "learning_rate": 4.312696747114376e-05, "loss": 0.5861, "step": 475 }, { "epoch": 0.6737438075017693, "grad_norm": 0.29168395505013195, "learning_rate": 4.310073452256034e-05, "loss": 0.6473, "step": 476 }, { "epoch": 0.6751592356687898, "grad_norm": 0.26749351847887926, "learning_rate": 4.307450157397691e-05, "loss": 0.5958, "step": 477 }, { "epoch": 0.6765746638358103, "grad_norm": 0.3135420128834056, "learning_rate": 4.304826862539349e-05, "loss": 0.5514, "step": 478 }, { "epoch": 0.6779900920028309, "grad_norm": 0.25355647122491026, "learning_rate": 4.3022035676810074e-05, "loss": 0.5944, "step": 479 }, { "epoch": 0.6794055201698513, "grad_norm": 0.27688273093419147, "learning_rate": 4.2995802728226654e-05, "loss": 0.589, "step": 480 }, { "epoch": 0.6808209483368719, "grad_norm": 0.2964986987389644, "learning_rate": 4.2969569779643234e-05, "loss": 0.5189, "step": 481 }, { "epoch": 0.6822363765038925, "grad_norm": 0.3004168997919821, "learning_rate": 4.294333683105981e-05, "loss": 0.6814, "step": 482 }, { "epoch": 0.6836518046709129, "grad_norm": 0.2717886113590949, "learning_rate": 4.2917103882476395e-05, "loss": 0.5767, "step": 483 }, { "epoch": 0.6850672328379335, "grad_norm": 0.2539767960016265, "learning_rate": 4.2890870933892975e-05, "loss": 0.6011, "step": 484 }, { "epoch": 0.686482661004954, "grad_norm": 0.7440050555777202, "learning_rate": 4.2864637985309555e-05, "loss": 0.5521, "step": 485 }, { "epoch": 0.6878980891719745, "grad_norm": 0.29647620149339093, "learning_rate": 4.283840503672613e-05, "loss": 0.6721, "step": 486 }, { "epoch": 0.689313517338995, "grad_norm": 0.2963743861972195, "learning_rate": 4.281217208814271e-05, "loss": 0.6564, "step": 487 }, { "epoch": 0.6907289455060156, "grad_norm": 0.30060259079097623, "learning_rate": 4.278593913955929e-05, "loss": 0.6195, "step": 488 }, { "epoch": 0.692144373673036, "grad_norm": 0.25412311784703495, "learning_rate": 4.275970619097587e-05, "loss": 0.5647, "step": 489 }, { "epoch": 0.6935598018400566, "grad_norm": 0.35278688480809256, "learning_rate": 4.273347324239245e-05, "loss": 0.5898, "step": 490 }, { "epoch": 0.6949752300070772, "grad_norm": 0.2862876117817953, "learning_rate": 4.2707240293809024e-05, "loss": 0.591, "step": 491 }, { "epoch": 0.6963906581740976, "grad_norm": 0.2818896648532148, "learning_rate": 4.2681007345225604e-05, "loss": 0.5724, "step": 492 }, { "epoch": 0.6978060863411182, "grad_norm": 0.28125410229184444, "learning_rate": 4.2654774396642184e-05, "loss": 0.6174, "step": 493 }, { "epoch": 0.6992215145081387, "grad_norm": 0.2744705379358727, "learning_rate": 4.2628541448058764e-05, "loss": 0.5892, "step": 494 }, { "epoch": 0.7006369426751592, "grad_norm": 0.2741434669225602, "learning_rate": 4.260230849947534e-05, "loss": 0.5604, "step": 495 }, { "epoch": 0.7020523708421798, "grad_norm": 0.32668510179495774, "learning_rate": 4.257607555089192e-05, "loss": 0.5718, "step": 496 }, { "epoch": 0.7034677990092003, "grad_norm": 0.3047620022933404, "learning_rate": 4.2549842602308505e-05, "loss": 0.5691, "step": 497 }, { "epoch": 0.7048832271762208, "grad_norm": 0.28241567675152757, "learning_rate": 4.2523609653725086e-05, "loss": 0.6505, "step": 498 }, { "epoch": 0.7062986553432413, "grad_norm": 0.3014000437347817, "learning_rate": 4.249737670514166e-05, "loss": 0.5782, "step": 499 }, { "epoch": 0.7077140835102619, "grad_norm": 0.23099688650678965, "learning_rate": 4.247114375655824e-05, "loss": 0.5292, "step": 500 }, { "epoch": 0.7091295116772823, "grad_norm": 0.288512470003529, "learning_rate": 4.244491080797482e-05, "loss": 0.5417, "step": 501 }, { "epoch": 0.7105449398443029, "grad_norm": 0.2890211026356785, "learning_rate": 4.24186778593914e-05, "loss": 0.6717, "step": 502 }, { "epoch": 0.7119603680113235, "grad_norm": 0.2603112959698144, "learning_rate": 4.239244491080798e-05, "loss": 0.5458, "step": 503 }, { "epoch": 0.7133757961783439, "grad_norm": 0.2755082295336734, "learning_rate": 4.2366211962224554e-05, "loss": 0.5793, "step": 504 }, { "epoch": 0.7147912243453645, "grad_norm": 0.2754622975959251, "learning_rate": 4.2339979013641134e-05, "loss": 0.5343, "step": 505 }, { "epoch": 0.716206652512385, "grad_norm": 0.27011894670220055, "learning_rate": 4.2313746065057714e-05, "loss": 0.6084, "step": 506 }, { "epoch": 0.7176220806794055, "grad_norm": 0.2696120852707313, "learning_rate": 4.2287513116474295e-05, "loss": 0.6192, "step": 507 }, { "epoch": 0.719037508846426, "grad_norm": 0.24868838929558423, "learning_rate": 4.226128016789087e-05, "loss": 0.5834, "step": 508 }, { "epoch": 0.7204529370134466, "grad_norm": 0.2799818580148587, "learning_rate": 4.223504721930745e-05, "loss": 0.619, "step": 509 }, { "epoch": 0.721868365180467, "grad_norm": 0.24896768145718445, "learning_rate": 4.2208814270724035e-05, "loss": 0.4383, "step": 510 }, { "epoch": 0.7232837933474876, "grad_norm": 2.784964837059265, "learning_rate": 4.2182581322140616e-05, "loss": 0.6158, "step": 511 }, { "epoch": 0.7246992215145082, "grad_norm": 0.3288428985349652, "learning_rate": 4.215634837355719e-05, "loss": 0.5824, "step": 512 }, { "epoch": 0.7261146496815286, "grad_norm": 0.270381517889225, "learning_rate": 4.213011542497377e-05, "loss": 0.6237, "step": 513 }, { "epoch": 0.7275300778485492, "grad_norm": 0.2887367100459008, "learning_rate": 4.210388247639035e-05, "loss": 0.5857, "step": 514 }, { "epoch": 0.7289455060155697, "grad_norm": 0.31003369966092637, "learning_rate": 4.207764952780693e-05, "loss": 0.5976, "step": 515 }, { "epoch": 0.7303609341825902, "grad_norm": 0.28351506632352835, "learning_rate": 4.205141657922351e-05, "loss": 0.52, "step": 516 }, { "epoch": 0.7317763623496107, "grad_norm": 0.3356356489905326, "learning_rate": 4.2025183630640084e-05, "loss": 0.6782, "step": 517 }, { "epoch": 0.7331917905166313, "grad_norm": 0.2765878293984698, "learning_rate": 4.1998950682056664e-05, "loss": 0.6152, "step": 518 }, { "epoch": 0.7346072186836518, "grad_norm": 0.26499313822264126, "learning_rate": 4.1972717733473244e-05, "loss": 0.5279, "step": 519 }, { "epoch": 0.7360226468506723, "grad_norm": 0.27460211792497674, "learning_rate": 4.1946484784889825e-05, "loss": 0.6255, "step": 520 }, { "epoch": 0.7374380750176929, "grad_norm": 0.2554505704777475, "learning_rate": 4.19202518363064e-05, "loss": 0.5178, "step": 521 }, { "epoch": 0.7388535031847133, "grad_norm": 0.27446639448887195, "learning_rate": 4.189401888772298e-05, "loss": 0.5982, "step": 522 }, { "epoch": 0.7402689313517339, "grad_norm": 0.2753277870564774, "learning_rate": 4.186778593913956e-05, "loss": 0.5425, "step": 523 }, { "epoch": 0.7416843595187544, "grad_norm": 0.30388507743331217, "learning_rate": 4.1841552990556146e-05, "loss": 0.6095, "step": 524 }, { "epoch": 0.7430997876857749, "grad_norm": 0.2674547695054725, "learning_rate": 4.181532004197272e-05, "loss": 0.6144, "step": 525 }, { "epoch": 0.7445152158527955, "grad_norm": 0.25908756166992924, "learning_rate": 4.17890870933893e-05, "loss": 0.5348, "step": 526 }, { "epoch": 0.745930644019816, "grad_norm": 0.25834605700842117, "learning_rate": 4.176285414480588e-05, "loss": 0.4905, "step": 527 }, { "epoch": 0.7473460721868365, "grad_norm": 0.2604563358719648, "learning_rate": 4.173662119622246e-05, "loss": 0.6299, "step": 528 }, { "epoch": 0.748761500353857, "grad_norm": 0.2884185006269154, "learning_rate": 4.171038824763904e-05, "loss": 0.6399, "step": 529 }, { "epoch": 0.7501769285208776, "grad_norm": 0.27330464240404045, "learning_rate": 4.1684155299055614e-05, "loss": 0.5452, "step": 530 }, { "epoch": 0.7515923566878981, "grad_norm": 1.3715142138570997, "learning_rate": 4.1657922350472194e-05, "loss": 0.6225, "step": 531 }, { "epoch": 0.7530077848549186, "grad_norm": 0.3543237791515518, "learning_rate": 4.1631689401888775e-05, "loss": 0.6815, "step": 532 }, { "epoch": 0.7544232130219392, "grad_norm": 0.3221700313412834, "learning_rate": 4.1605456453305355e-05, "loss": 0.629, "step": 533 }, { "epoch": 0.7558386411889597, "grad_norm": 0.30392541264587125, "learning_rate": 4.157922350472193e-05, "loss": 0.5199, "step": 534 }, { "epoch": 0.7572540693559802, "grad_norm": 0.3029339931387272, "learning_rate": 4.155299055613851e-05, "loss": 0.5197, "step": 535 }, { "epoch": 0.7586694975230007, "grad_norm": 0.31621698004884125, "learning_rate": 4.152675760755509e-05, "loss": 0.5867, "step": 536 }, { "epoch": 0.7600849256900213, "grad_norm": 0.28814019849199607, "learning_rate": 4.150052465897167e-05, "loss": 0.6364, "step": 537 }, { "epoch": 0.7615003538570417, "grad_norm": 0.29237179383094253, "learning_rate": 4.147429171038825e-05, "loss": 0.5636, "step": 538 }, { "epoch": 0.7629157820240623, "grad_norm": 0.3642930281551312, "learning_rate": 4.144805876180483e-05, "loss": 0.692, "step": 539 }, { "epoch": 0.7643312101910829, "grad_norm": 0.2816381389682949, "learning_rate": 4.142182581322141e-05, "loss": 0.6313, "step": 540 }, { "epoch": 0.7657466383581033, "grad_norm": 0.31808906679687576, "learning_rate": 4.139559286463799e-05, "loss": 0.5918, "step": 541 }, { "epoch": 0.7671620665251239, "grad_norm": 0.27256916513618773, "learning_rate": 4.136935991605457e-05, "loss": 0.5621, "step": 542 }, { "epoch": 0.7685774946921444, "grad_norm": 0.2607254484094606, "learning_rate": 4.1343126967471144e-05, "loss": 0.5182, "step": 543 }, { "epoch": 0.7699929228591649, "grad_norm": 0.2829216229335715, "learning_rate": 4.1316894018887724e-05, "loss": 0.6379, "step": 544 }, { "epoch": 0.7714083510261854, "grad_norm": 0.2880790487930277, "learning_rate": 4.1290661070304305e-05, "loss": 0.6227, "step": 545 }, { "epoch": 0.772823779193206, "grad_norm": 0.2996132822785243, "learning_rate": 4.1264428121720885e-05, "loss": 0.6055, "step": 546 }, { "epoch": 0.7742392073602264, "grad_norm": 0.2538777986121411, "learning_rate": 4.123819517313746e-05, "loss": 0.4933, "step": 547 }, { "epoch": 0.775654635527247, "grad_norm": 0.2946301022847751, "learning_rate": 4.121196222455404e-05, "loss": 0.5704, "step": 548 }, { "epoch": 0.7770700636942676, "grad_norm": 0.2601702246669579, "learning_rate": 4.118572927597062e-05, "loss": 0.5303, "step": 549 }, { "epoch": 0.778485491861288, "grad_norm": 0.28467177701497776, "learning_rate": 4.11594963273872e-05, "loss": 0.5815, "step": 550 }, { "epoch": 0.7799009200283086, "grad_norm": 0.30475496699597565, "learning_rate": 4.113326337880378e-05, "loss": 0.623, "step": 551 }, { "epoch": 0.7813163481953291, "grad_norm": 0.29736546055647334, "learning_rate": 4.110703043022036e-05, "loss": 0.6027, "step": 552 }, { "epoch": 0.7827317763623496, "grad_norm": 0.25963842574091467, "learning_rate": 4.108079748163694e-05, "loss": 0.544, "step": 553 }, { "epoch": 0.7841472045293701, "grad_norm": 0.2882644562548918, "learning_rate": 4.105456453305352e-05, "loss": 0.6175, "step": 554 }, { "epoch": 0.7855626326963907, "grad_norm": 0.6580202281062054, "learning_rate": 4.10283315844701e-05, "loss": 0.6881, "step": 555 }, { "epoch": 0.7869780608634112, "grad_norm": 0.2510772787710439, "learning_rate": 4.1002098635886674e-05, "loss": 0.5427, "step": 556 }, { "epoch": 0.7883934890304317, "grad_norm": 0.25333209936456746, "learning_rate": 4.0975865687303255e-05, "loss": 0.5314, "step": 557 }, { "epoch": 0.7898089171974523, "grad_norm": 0.3399760693466158, "learning_rate": 4.0949632738719835e-05, "loss": 0.5766, "step": 558 }, { "epoch": 0.7912243453644727, "grad_norm": 0.2545404542414338, "learning_rate": 4.0923399790136415e-05, "loss": 0.605, "step": 559 }, { "epoch": 0.7926397735314933, "grad_norm": 0.2689807888410843, "learning_rate": 4.089716684155299e-05, "loss": 0.6298, "step": 560 }, { "epoch": 0.7940552016985138, "grad_norm": 0.26979275459836133, "learning_rate": 4.087093389296957e-05, "loss": 0.6411, "step": 561 }, { "epoch": 0.7954706298655343, "grad_norm": 0.26477102023358506, "learning_rate": 4.084470094438615e-05, "loss": 0.6496, "step": 562 }, { "epoch": 0.7968860580325549, "grad_norm": 0.2842035921406642, "learning_rate": 4.081846799580273e-05, "loss": 0.6734, "step": 563 }, { "epoch": 0.7983014861995754, "grad_norm": 0.2759384254763282, "learning_rate": 4.079223504721931e-05, "loss": 0.5948, "step": 564 }, { "epoch": 0.7997169143665959, "grad_norm": 0.29312620721071925, "learning_rate": 4.076600209863588e-05, "loss": 0.6588, "step": 565 }, { "epoch": 0.8011323425336164, "grad_norm": 0.279881273844713, "learning_rate": 4.073976915005247e-05, "loss": 0.5988, "step": 566 }, { "epoch": 0.802547770700637, "grad_norm": 0.24491092162063974, "learning_rate": 4.071353620146905e-05, "loss": 0.5586, "step": 567 }, { "epoch": 0.8039631988676574, "grad_norm": 0.24392190118225973, "learning_rate": 4.068730325288563e-05, "loss": 0.6342, "step": 568 }, { "epoch": 0.805378627034678, "grad_norm": 0.2582997869107178, "learning_rate": 4.0661070304302204e-05, "loss": 0.6197, "step": 569 }, { "epoch": 0.8067940552016986, "grad_norm": 0.24443958952182887, "learning_rate": 4.0634837355718785e-05, "loss": 0.5583, "step": 570 }, { "epoch": 0.808209483368719, "grad_norm": 0.25103295210737114, "learning_rate": 4.0608604407135365e-05, "loss": 0.5813, "step": 571 }, { "epoch": 0.8096249115357396, "grad_norm": 0.2587091249984754, "learning_rate": 4.0582371458551945e-05, "loss": 0.6203, "step": 572 }, { "epoch": 0.8110403397027601, "grad_norm": 5.077427791117972, "learning_rate": 4.055613850996852e-05, "loss": 0.571, "step": 573 }, { "epoch": 0.8124557678697806, "grad_norm": 0.267569113524592, "learning_rate": 4.05299055613851e-05, "loss": 0.5514, "step": 574 }, { "epoch": 0.8138711960368011, "grad_norm": 0.25344246892144145, "learning_rate": 4.050367261280168e-05, "loss": 0.5189, "step": 575 }, { "epoch": 0.8152866242038217, "grad_norm": 0.26036664546568666, "learning_rate": 4.047743966421826e-05, "loss": 0.6321, "step": 576 }, { "epoch": 0.8167020523708421, "grad_norm": 0.30472024610966936, "learning_rate": 4.045120671563484e-05, "loss": 0.6286, "step": 577 }, { "epoch": 0.8181174805378627, "grad_norm": 0.3009657072585352, "learning_rate": 4.0424973767051413e-05, "loss": 0.5791, "step": 578 }, { "epoch": 0.8195329087048833, "grad_norm": 0.30448223273011926, "learning_rate": 4.0398740818467994e-05, "loss": 0.6528, "step": 579 }, { "epoch": 0.8209483368719037, "grad_norm": 0.2947969249489438, "learning_rate": 4.037250786988458e-05, "loss": 0.6334, "step": 580 }, { "epoch": 0.8223637650389243, "grad_norm": 0.24462892435405953, "learning_rate": 4.034627492130116e-05, "loss": 0.5516, "step": 581 }, { "epoch": 0.8237791932059448, "grad_norm": 0.3055832851870209, "learning_rate": 4.0320041972717735e-05, "loss": 0.6258, "step": 582 }, { "epoch": 0.8251946213729653, "grad_norm": 0.24441620573166245, "learning_rate": 4.0293809024134315e-05, "loss": 0.5825, "step": 583 }, { "epoch": 0.8266100495399858, "grad_norm": 0.24212270024095364, "learning_rate": 4.0267576075550895e-05, "loss": 0.5374, "step": 584 }, { "epoch": 0.8280254777070064, "grad_norm": 0.2892849365436293, "learning_rate": 4.0241343126967475e-05, "loss": 0.5259, "step": 585 }, { "epoch": 0.8294409058740269, "grad_norm": 0.26206893040947127, "learning_rate": 4.021511017838405e-05, "loss": 0.5325, "step": 586 }, { "epoch": 0.8308563340410474, "grad_norm": 0.2574461087293804, "learning_rate": 4.018887722980063e-05, "loss": 0.5218, "step": 587 }, { "epoch": 0.832271762208068, "grad_norm": 0.28681708948939577, "learning_rate": 4.016264428121721e-05, "loss": 0.6362, "step": 588 }, { "epoch": 0.8336871903750884, "grad_norm": 1.0318753511372296, "learning_rate": 4.013641133263379e-05, "loss": 0.6171, "step": 589 }, { "epoch": 0.835102618542109, "grad_norm": 0.34847852929225803, "learning_rate": 4.011017838405037e-05, "loss": 0.6329, "step": 590 }, { "epoch": 0.8365180467091295, "grad_norm": 0.24148552486467043, "learning_rate": 4.0083945435466944e-05, "loss": 0.5311, "step": 591 }, { "epoch": 0.83793347487615, "grad_norm": 0.3498402864926876, "learning_rate": 4.0057712486883524e-05, "loss": 0.709, "step": 592 }, { "epoch": 0.8393489030431706, "grad_norm": 0.252828599776942, "learning_rate": 4.003147953830011e-05, "loss": 0.5973, "step": 593 }, { "epoch": 0.8407643312101911, "grad_norm": 0.2708337319292238, "learning_rate": 4.000524658971669e-05, "loss": 0.5746, "step": 594 }, { "epoch": 0.8421797593772116, "grad_norm": 0.26006848312181075, "learning_rate": 3.9979013641133265e-05, "loss": 0.6047, "step": 595 }, { "epoch": 0.8435951875442321, "grad_norm": 0.25011185374558303, "learning_rate": 3.9952780692549845e-05, "loss": 0.5338, "step": 596 }, { "epoch": 0.8450106157112527, "grad_norm": 0.32906968439296075, "learning_rate": 3.9926547743966425e-05, "loss": 0.6514, "step": 597 }, { "epoch": 0.8464260438782731, "grad_norm": 0.25228064747265583, "learning_rate": 3.9900314795383006e-05, "loss": 0.5676, "step": 598 }, { "epoch": 0.8478414720452937, "grad_norm": 0.31326927533470245, "learning_rate": 3.987408184679958e-05, "loss": 0.5229, "step": 599 }, { "epoch": 0.8492569002123143, "grad_norm": 0.2435322870011402, "learning_rate": 3.984784889821616e-05, "loss": 0.5208, "step": 600 }, { "epoch": 0.8506723283793347, "grad_norm": 0.33182164176895124, "learning_rate": 3.982161594963274e-05, "loss": 0.6113, "step": 601 }, { "epoch": 0.8520877565463553, "grad_norm": 0.28453071974121313, "learning_rate": 3.979538300104932e-05, "loss": 0.5691, "step": 602 }, { "epoch": 0.8535031847133758, "grad_norm": 0.3218593621452372, "learning_rate": 3.97691500524659e-05, "loss": 0.5606, "step": 603 }, { "epoch": 0.8549186128803963, "grad_norm": 0.3040945875286931, "learning_rate": 3.9742917103882474e-05, "loss": 0.5704, "step": 604 }, { "epoch": 0.8563340410474168, "grad_norm": 0.2683282837210314, "learning_rate": 3.9716684155299054e-05, "loss": 0.5483, "step": 605 }, { "epoch": 0.8577494692144374, "grad_norm": 0.25839227614777754, "learning_rate": 3.9690451206715634e-05, "loss": 0.4989, "step": 606 }, { "epoch": 0.8591648973814578, "grad_norm": 0.26956974149348745, "learning_rate": 3.966421825813222e-05, "loss": 0.64, "step": 607 }, { "epoch": 0.8605803255484784, "grad_norm": 0.2672956498419619, "learning_rate": 3.9637985309548795e-05, "loss": 0.5982, "step": 608 }, { "epoch": 0.861995753715499, "grad_norm": 0.25016570426383566, "learning_rate": 3.9611752360965375e-05, "loss": 0.6098, "step": 609 }, { "epoch": 0.8634111818825194, "grad_norm": 0.27107103773859054, "learning_rate": 3.9585519412381956e-05, "loss": 0.6166, "step": 610 }, { "epoch": 0.86482661004954, "grad_norm": 0.4151753982663599, "learning_rate": 3.9559286463798536e-05, "loss": 0.5807, "step": 611 }, { "epoch": 0.8662420382165605, "grad_norm": 0.23976821200260148, "learning_rate": 3.953305351521511e-05, "loss": 0.5509, "step": 612 }, { "epoch": 0.867657466383581, "grad_norm": 0.3520835351925965, "learning_rate": 3.950682056663169e-05, "loss": 0.5085, "step": 613 }, { "epoch": 0.8690728945506015, "grad_norm": 0.2696270539363782, "learning_rate": 3.948058761804827e-05, "loss": 0.6259, "step": 614 }, { "epoch": 0.8704883227176221, "grad_norm": 0.2566798072335374, "learning_rate": 3.945435466946485e-05, "loss": 0.5034, "step": 615 }, { "epoch": 0.8719037508846426, "grad_norm": 0.27860917796336493, "learning_rate": 3.942812172088143e-05, "loss": 0.5738, "step": 616 }, { "epoch": 0.8733191790516631, "grad_norm": 0.26826393817331345, "learning_rate": 3.9401888772298004e-05, "loss": 0.6201, "step": 617 }, { "epoch": 0.8747346072186837, "grad_norm": 0.27074655141154563, "learning_rate": 3.9375655823714584e-05, "loss": 0.5748, "step": 618 }, { "epoch": 0.8761500353857041, "grad_norm": 0.25848566297252273, "learning_rate": 3.9349422875131165e-05, "loss": 0.567, "step": 619 }, { "epoch": 0.8775654635527247, "grad_norm": 0.2538167120734919, "learning_rate": 3.9323189926547745e-05, "loss": 0.645, "step": 620 }, { "epoch": 0.8789808917197452, "grad_norm": 0.25591843624105864, "learning_rate": 3.9296956977964325e-05, "loss": 0.5171, "step": 621 }, { "epoch": 0.8803963198867657, "grad_norm": 0.2711546089590064, "learning_rate": 3.9270724029380905e-05, "loss": 0.6277, "step": 622 }, { "epoch": 0.8818117480537863, "grad_norm": 0.2566505626483008, "learning_rate": 3.9244491080797486e-05, "loss": 0.5329, "step": 623 }, { "epoch": 0.8832271762208068, "grad_norm": 0.27195831622640293, "learning_rate": 3.9218258132214066e-05, "loss": 0.6002, "step": 624 }, { "epoch": 0.8846426043878273, "grad_norm": 0.239594292957346, "learning_rate": 3.9192025183630646e-05, "loss": 0.6301, "step": 625 }, { "epoch": 0.8860580325548478, "grad_norm": 0.21543925331090563, "learning_rate": 3.916579223504722e-05, "loss": 0.4855, "step": 626 }, { "epoch": 0.8874734607218684, "grad_norm": 0.2627020572877223, "learning_rate": 3.91395592864638e-05, "loss": 0.536, "step": 627 }, { "epoch": 0.8888888888888888, "grad_norm": 0.25994843629610076, "learning_rate": 3.911332633788038e-05, "loss": 0.5694, "step": 628 }, { "epoch": 0.8903043170559094, "grad_norm": 0.2596772324229409, "learning_rate": 3.908709338929696e-05, "loss": 0.5503, "step": 629 }, { "epoch": 0.89171974522293, "grad_norm": 0.24003235792566835, "learning_rate": 3.9060860440713534e-05, "loss": 0.5838, "step": 630 }, { "epoch": 0.8931351733899504, "grad_norm": 0.22872137738038825, "learning_rate": 3.9034627492130114e-05, "loss": 0.4768, "step": 631 }, { "epoch": 0.894550601556971, "grad_norm": 0.26240114327499175, "learning_rate": 3.9008394543546695e-05, "loss": 0.6221, "step": 632 }, { "epoch": 0.8959660297239915, "grad_norm": 0.2736984168600782, "learning_rate": 3.8982161594963275e-05, "loss": 0.646, "step": 633 }, { "epoch": 0.897381457891012, "grad_norm": 0.24510404040546305, "learning_rate": 3.8955928646379855e-05, "loss": 0.6028, "step": 634 }, { "epoch": 0.8987968860580325, "grad_norm": 0.2487820797852633, "learning_rate": 3.8929695697796436e-05, "loss": 0.4868, "step": 635 }, { "epoch": 0.9002123142250531, "grad_norm": 0.3011312076480095, "learning_rate": 3.8903462749213016e-05, "loss": 0.6963, "step": 636 }, { "epoch": 0.9016277423920736, "grad_norm": 0.2703438446586986, "learning_rate": 3.8877229800629596e-05, "loss": 0.6708, "step": 637 }, { "epoch": 0.9030431705590941, "grad_norm": 0.27901463098331925, "learning_rate": 3.8850996852046176e-05, "loss": 0.6576, "step": 638 }, { "epoch": 0.9044585987261147, "grad_norm": 0.30297040426188737, "learning_rate": 3.882476390346275e-05, "loss": 0.6201, "step": 639 }, { "epoch": 0.9058740268931351, "grad_norm": 0.2410564866480961, "learning_rate": 3.879853095487933e-05, "loss": 0.5523, "step": 640 }, { "epoch": 0.9072894550601557, "grad_norm": 0.30255449339120893, "learning_rate": 3.877229800629591e-05, "loss": 0.5466, "step": 641 }, { "epoch": 0.9087048832271762, "grad_norm": 0.6032084870416943, "learning_rate": 3.874606505771249e-05, "loss": 0.5904, "step": 642 }, { "epoch": 0.9101203113941967, "grad_norm": 0.256190445638519, "learning_rate": 3.8719832109129064e-05, "loss": 0.5878, "step": 643 }, { "epoch": 0.9115357395612173, "grad_norm": 0.3102660753676702, "learning_rate": 3.8693599160545645e-05, "loss": 0.6893, "step": 644 }, { "epoch": 0.9129511677282378, "grad_norm": 0.28928475943481113, "learning_rate": 3.8667366211962225e-05, "loss": 0.5205, "step": 645 }, { "epoch": 0.9143665958952583, "grad_norm": 2.5782261650324085, "learning_rate": 3.8641133263378805e-05, "loss": 0.8237, "step": 646 }, { "epoch": 0.9157820240622788, "grad_norm": 0.3375425332462325, "learning_rate": 3.8614900314795385e-05, "loss": 0.6284, "step": 647 }, { "epoch": 0.9171974522292994, "grad_norm": 0.5450089954920853, "learning_rate": 3.858866736621196e-05, "loss": 0.6052, "step": 648 }, { "epoch": 0.9186128803963199, "grad_norm": 0.29482791588888185, "learning_rate": 3.8562434417628546e-05, "loss": 0.5513, "step": 649 }, { "epoch": 0.9200283085633404, "grad_norm": 0.2790147345264433, "learning_rate": 3.8536201469045126e-05, "loss": 0.5789, "step": 650 }, { "epoch": 0.921443736730361, "grad_norm": 0.2799735113368812, "learning_rate": 3.8509968520461707e-05, "loss": 0.5998, "step": 651 }, { "epoch": 0.9228591648973815, "grad_norm": 0.29273992863088466, "learning_rate": 3.848373557187828e-05, "loss": 0.5695, "step": 652 }, { "epoch": 0.924274593064402, "grad_norm": 1.3412545528101212, "learning_rate": 3.845750262329486e-05, "loss": 0.6271, "step": 653 }, { "epoch": 0.9256900212314225, "grad_norm": 0.2960936597898938, "learning_rate": 3.843126967471144e-05, "loss": 0.5124, "step": 654 }, { "epoch": 0.9271054493984431, "grad_norm": 0.24973177405132, "learning_rate": 3.840503672612802e-05, "loss": 0.5026, "step": 655 }, { "epoch": 0.9285208775654635, "grad_norm": 0.31183159855056186, "learning_rate": 3.8378803777544594e-05, "loss": 0.5358, "step": 656 }, { "epoch": 0.9299363057324841, "grad_norm": 0.32824518111338075, "learning_rate": 3.8352570828961175e-05, "loss": 0.5609, "step": 657 }, { "epoch": 0.9313517338995047, "grad_norm": 0.2714867436973302, "learning_rate": 3.8326337880377755e-05, "loss": 0.499, "step": 658 }, { "epoch": 0.9327671620665251, "grad_norm": 0.28085284253709075, "learning_rate": 3.8300104931794335e-05, "loss": 0.5869, "step": 659 }, { "epoch": 0.9341825902335457, "grad_norm": 0.2870910451588877, "learning_rate": 3.8273871983210916e-05, "loss": 0.5582, "step": 660 }, { "epoch": 0.9355980184005662, "grad_norm": 0.2688685777279877, "learning_rate": 3.824763903462749e-05, "loss": 0.5481, "step": 661 }, { "epoch": 0.9370134465675867, "grad_norm": 0.2633867881477559, "learning_rate": 3.8221406086044076e-05, "loss": 0.4945, "step": 662 }, { "epoch": 0.9384288747346072, "grad_norm": 0.24924008904882905, "learning_rate": 3.8195173137460656e-05, "loss": 0.5321, "step": 663 }, { "epoch": 0.9398443029016278, "grad_norm": 0.26365210458344096, "learning_rate": 3.816894018887724e-05, "loss": 0.6119, "step": 664 }, { "epoch": 0.9412597310686482, "grad_norm": 0.25688659153777077, "learning_rate": 3.814270724029381e-05, "loss": 0.5305, "step": 665 }, { "epoch": 0.9426751592356688, "grad_norm": 0.2689108075584822, "learning_rate": 3.811647429171039e-05, "loss": 0.4616, "step": 666 }, { "epoch": 0.9440905874026894, "grad_norm": 0.2753729809225918, "learning_rate": 3.809024134312697e-05, "loss": 0.538, "step": 667 }, { "epoch": 0.9455060155697098, "grad_norm": 0.24398271598476046, "learning_rate": 3.806400839454355e-05, "loss": 0.5212, "step": 668 }, { "epoch": 0.9469214437367304, "grad_norm": 0.5451156284839687, "learning_rate": 3.8037775445960125e-05, "loss": 0.5724, "step": 669 }, { "epoch": 0.9483368719037509, "grad_norm": 0.25713924672797955, "learning_rate": 3.8011542497376705e-05, "loss": 0.5829, "step": 670 }, { "epoch": 0.9497523000707714, "grad_norm": 0.30912466280078726, "learning_rate": 3.7985309548793285e-05, "loss": 0.515, "step": 671 }, { "epoch": 0.9511677282377919, "grad_norm": 0.28775548891760677, "learning_rate": 3.7959076600209865e-05, "loss": 0.6289, "step": 672 }, { "epoch": 0.9525831564048125, "grad_norm": 0.2498263173619407, "learning_rate": 3.7932843651626446e-05, "loss": 0.6028, "step": 673 }, { "epoch": 0.953998584571833, "grad_norm": 0.25277311684116194, "learning_rate": 3.790661070304302e-05, "loss": 0.5237, "step": 674 }, { "epoch": 0.9554140127388535, "grad_norm": 0.2739873998144345, "learning_rate": 3.78803777544596e-05, "loss": 0.5346, "step": 675 }, { "epoch": 0.9568294409058741, "grad_norm": 0.29695038936503027, "learning_rate": 3.7854144805876187e-05, "loss": 0.6129, "step": 676 }, { "epoch": 0.9582448690728945, "grad_norm": 0.2592036420138258, "learning_rate": 3.782791185729277e-05, "loss": 0.5731, "step": 677 }, { "epoch": 0.9596602972399151, "grad_norm": 0.2558871446271218, "learning_rate": 3.780167890870934e-05, "loss": 0.5979, "step": 678 }, { "epoch": 0.9610757254069356, "grad_norm": 0.2982843354022497, "learning_rate": 3.777544596012592e-05, "loss": 0.5924, "step": 679 }, { "epoch": 0.9624911535739561, "grad_norm": 0.2488352289399196, "learning_rate": 3.77492130115425e-05, "loss": 0.5526, "step": 680 }, { "epoch": 0.9639065817409767, "grad_norm": 0.25611812584464694, "learning_rate": 3.772298006295908e-05, "loss": 0.6042, "step": 681 }, { "epoch": 0.9653220099079972, "grad_norm": 0.29555532650943567, "learning_rate": 3.7696747114375655e-05, "loss": 0.5963, "step": 682 }, { "epoch": 0.9667374380750177, "grad_norm": 0.25428893954509546, "learning_rate": 3.7670514165792235e-05, "loss": 0.5021, "step": 683 }, { "epoch": 0.9681528662420382, "grad_norm": 0.2244747782441335, "learning_rate": 3.7644281217208815e-05, "loss": 0.5129, "step": 684 }, { "epoch": 0.9695682944090588, "grad_norm": 0.2539476280706618, "learning_rate": 3.7618048268625396e-05, "loss": 0.6208, "step": 685 }, { "epoch": 0.9709837225760792, "grad_norm": 0.26689201349616526, "learning_rate": 3.7591815320041976e-05, "loss": 0.4655, "step": 686 }, { "epoch": 0.9723991507430998, "grad_norm": 0.2665073304122392, "learning_rate": 3.756558237145855e-05, "loss": 0.5653, "step": 687 }, { "epoch": 0.9738145789101204, "grad_norm": 0.2857822474503496, "learning_rate": 3.753934942287513e-05, "loss": 0.5386, "step": 688 }, { "epoch": 0.9752300070771408, "grad_norm": 0.2924353589789488, "learning_rate": 3.751311647429171e-05, "loss": 0.5909, "step": 689 }, { "epoch": 0.9766454352441614, "grad_norm": 0.2873977701123475, "learning_rate": 3.74868835257083e-05, "loss": 0.6711, "step": 690 }, { "epoch": 0.9780608634111819, "grad_norm": 0.2960710414329434, "learning_rate": 3.746065057712487e-05, "loss": 0.5039, "step": 691 }, { "epoch": 0.9794762915782024, "grad_norm": 0.2807847179526359, "learning_rate": 3.743441762854145e-05, "loss": 0.5993, "step": 692 }, { "epoch": 0.9808917197452229, "grad_norm": 0.23634958473891998, "learning_rate": 3.740818467995803e-05, "loss": 0.4978, "step": 693 }, { "epoch": 0.9823071479122435, "grad_norm": 0.2765985645444167, "learning_rate": 3.738195173137461e-05, "loss": 0.669, "step": 694 }, { "epoch": 0.9837225760792639, "grad_norm": 0.26167974900207414, "learning_rate": 3.7355718782791185e-05, "loss": 0.504, "step": 695 }, { "epoch": 0.9851380042462845, "grad_norm": 0.31832168867659044, "learning_rate": 3.7329485834207765e-05, "loss": 0.6157, "step": 696 }, { "epoch": 0.9865534324133051, "grad_norm": 0.27502165919598376, "learning_rate": 3.7303252885624345e-05, "loss": 0.6101, "step": 697 }, { "epoch": 0.9879688605803255, "grad_norm": 0.30620856960200665, "learning_rate": 3.7277019937040926e-05, "loss": 0.6019, "step": 698 }, { "epoch": 0.9893842887473461, "grad_norm": 0.24607762964278576, "learning_rate": 3.7250786988457506e-05, "loss": 0.5015, "step": 699 }, { "epoch": 0.9907997169143666, "grad_norm": 0.2911017746963043, "learning_rate": 3.722455403987408e-05, "loss": 0.5343, "step": 700 }, { "epoch": 0.9922151450813871, "grad_norm": 0.31548858358379145, "learning_rate": 3.719832109129066e-05, "loss": 0.5551, "step": 701 }, { "epoch": 0.9936305732484076, "grad_norm": 0.3039085464002624, "learning_rate": 3.717208814270724e-05, "loss": 0.6619, "step": 702 }, { "epoch": 0.9950460014154282, "grad_norm": 0.24692888851369393, "learning_rate": 3.714585519412382e-05, "loss": 0.5174, "step": 703 }, { "epoch": 0.9964614295824487, "grad_norm": 0.29659026256453436, "learning_rate": 3.71196222455404e-05, "loss": 0.4911, "step": 704 }, { "epoch": 0.9978768577494692, "grad_norm": 0.302397588200181, "learning_rate": 3.709338929695698e-05, "loss": 0.5383, "step": 705 }, { "epoch": 0.9992922859164898, "grad_norm": 0.2613748044843372, "learning_rate": 3.706715634837356e-05, "loss": 0.5833, "step": 706 }, { "epoch": 1.0, "grad_norm": 0.2613748044843372, "learning_rate": 3.704092339979014e-05, "loss": 0.5796, "step": 707 }, { "epoch": 1.0014154281670204, "grad_norm": 0.5528101354566592, "learning_rate": 3.7014690451206715e-05, "loss": 0.4508, "step": 708 }, { "epoch": 1.0028308563340411, "grad_norm": 0.288786666348198, "learning_rate": 3.6988457502623295e-05, "loss": 0.485, "step": 709 }, { "epoch": 1.0042462845010616, "grad_norm": 0.7619106987569315, "learning_rate": 3.6962224554039876e-05, "loss": 0.4902, "step": 710 }, { "epoch": 1.005661712668082, "grad_norm": 0.2996892202706279, "learning_rate": 3.6935991605456456e-05, "loss": 0.3895, "step": 711 }, { "epoch": 1.0070771408351027, "grad_norm": 0.29613902550473165, "learning_rate": 3.6909758656873036e-05, "loss": 0.4934, "step": 712 }, { "epoch": 1.0084925690021231, "grad_norm": 0.31978838237376206, "learning_rate": 3.688352570828961e-05, "loss": 0.4632, "step": 713 }, { "epoch": 1.0099079971691436, "grad_norm": 0.3409684419101744, "learning_rate": 3.685729275970619e-05, "loss": 0.5109, "step": 714 }, { "epoch": 1.0113234253361643, "grad_norm": 0.31456340994022647, "learning_rate": 3.683105981112277e-05, "loss": 0.5287, "step": 715 }, { "epoch": 1.0127388535031847, "grad_norm": 0.32103586566695375, "learning_rate": 3.680482686253935e-05, "loss": 0.5167, "step": 716 }, { "epoch": 1.0141542816702052, "grad_norm": 0.3320968062517569, "learning_rate": 3.677859391395593e-05, "loss": 0.5082, "step": 717 }, { "epoch": 1.0155697098372258, "grad_norm": 0.30584543091711014, "learning_rate": 3.675236096537251e-05, "loss": 0.524, "step": 718 }, { "epoch": 1.0169851380042463, "grad_norm": 0.3138045748342782, "learning_rate": 3.672612801678909e-05, "loss": 0.4462, "step": 719 }, { "epoch": 1.0184005661712667, "grad_norm": 0.2745307135839983, "learning_rate": 3.669989506820567e-05, "loss": 0.5087, "step": 720 }, { "epoch": 1.0198159943382874, "grad_norm": 5.258650257167126, "learning_rate": 3.6673662119622245e-05, "loss": 0.4828, "step": 721 }, { "epoch": 1.0212314225053079, "grad_norm": 0.3287645006651341, "learning_rate": 3.6647429171038825e-05, "loss": 0.4432, "step": 722 }, { "epoch": 1.0226468506723283, "grad_norm": 0.31292657982725264, "learning_rate": 3.6621196222455406e-05, "loss": 0.4675, "step": 723 }, { "epoch": 1.024062278839349, "grad_norm": 0.2847961461690093, "learning_rate": 3.6594963273871986e-05, "loss": 0.4833, "step": 724 }, { "epoch": 1.0254777070063694, "grad_norm": 0.31120964842943327, "learning_rate": 3.6568730325288566e-05, "loss": 0.5254, "step": 725 }, { "epoch": 1.0268931351733899, "grad_norm": 0.258099839637945, "learning_rate": 3.654249737670514e-05, "loss": 0.4168, "step": 726 }, { "epoch": 1.0283085633404105, "grad_norm": 0.28589238757128593, "learning_rate": 3.651626442812172e-05, "loss": 0.3825, "step": 727 }, { "epoch": 1.029723991507431, "grad_norm": 0.2575331936520694, "learning_rate": 3.64900314795383e-05, "loss": 0.4914, "step": 728 }, { "epoch": 1.0311394196744514, "grad_norm": 0.30838547918171205, "learning_rate": 3.646379853095488e-05, "loss": 0.42, "step": 729 }, { "epoch": 1.0325548478414721, "grad_norm": 0.31046089073138644, "learning_rate": 3.643756558237146e-05, "loss": 0.4155, "step": 730 }, { "epoch": 1.0339702760084926, "grad_norm": 0.26747997987601974, "learning_rate": 3.641133263378804e-05, "loss": 0.5073, "step": 731 }, { "epoch": 1.035385704175513, "grad_norm": 0.30602021395282875, "learning_rate": 3.638509968520462e-05, "loss": 0.5294, "step": 732 }, { "epoch": 1.0368011323425337, "grad_norm": 0.2916135518120182, "learning_rate": 3.63588667366212e-05, "loss": 0.4692, "step": 733 }, { "epoch": 1.0382165605095541, "grad_norm": 0.26730632113660724, "learning_rate": 3.6332633788037775e-05, "loss": 0.441, "step": 734 }, { "epoch": 1.0396319886765746, "grad_norm": 0.25696358589681917, "learning_rate": 3.6306400839454356e-05, "loss": 0.4457, "step": 735 }, { "epoch": 1.0410474168435953, "grad_norm": 0.2990631968890318, "learning_rate": 3.6280167890870936e-05, "loss": 0.4836, "step": 736 }, { "epoch": 1.0424628450106157, "grad_norm": 0.28660601960737864, "learning_rate": 3.6253934942287516e-05, "loss": 0.4691, "step": 737 }, { "epoch": 1.0438782731776362, "grad_norm": 0.24380734366700835, "learning_rate": 3.6227701993704097e-05, "loss": 0.4592, "step": 738 }, { "epoch": 1.0452937013446568, "grad_norm": 0.25241122343037414, "learning_rate": 3.620146904512067e-05, "loss": 0.4531, "step": 739 }, { "epoch": 1.0467091295116773, "grad_norm": 0.2535564968999766, "learning_rate": 3.617523609653725e-05, "loss": 0.5335, "step": 740 }, { "epoch": 1.0481245576786977, "grad_norm": 0.25979506943880126, "learning_rate": 3.614900314795383e-05, "loss": 0.4713, "step": 741 }, { "epoch": 1.0495399858457184, "grad_norm": 0.2664669724250493, "learning_rate": 3.612277019937041e-05, "loss": 0.4686, "step": 742 }, { "epoch": 1.0509554140127388, "grad_norm": 0.2509227853228027, "learning_rate": 3.609653725078699e-05, "loss": 0.5015, "step": 743 }, { "epoch": 1.0523708421797593, "grad_norm": 0.24602973422666952, "learning_rate": 3.6070304302203565e-05, "loss": 0.5268, "step": 744 }, { "epoch": 1.05378627034678, "grad_norm": 0.2441351012508443, "learning_rate": 3.604407135362015e-05, "loss": 0.4722, "step": 745 }, { "epoch": 1.0552016985138004, "grad_norm": 0.2618929433054825, "learning_rate": 3.601783840503673e-05, "loss": 0.5075, "step": 746 }, { "epoch": 1.0566171266808209, "grad_norm": 0.23628916616707335, "learning_rate": 3.599160545645331e-05, "loss": 0.4584, "step": 747 }, { "epoch": 1.0580325548478415, "grad_norm": 0.22777912078546816, "learning_rate": 3.5965372507869886e-05, "loss": 0.4912, "step": 748 }, { "epoch": 1.059447983014862, "grad_norm": 0.22856290218831565, "learning_rate": 3.5939139559286466e-05, "loss": 0.4486, "step": 749 }, { "epoch": 1.0608634111818824, "grad_norm": 0.23825245216080632, "learning_rate": 3.5912906610703046e-05, "loss": 0.4592, "step": 750 }, { "epoch": 1.062278839348903, "grad_norm": 0.2408459349082231, "learning_rate": 3.588667366211963e-05, "loss": 0.4928, "step": 751 }, { "epoch": 1.0636942675159236, "grad_norm": 0.24863045510811296, "learning_rate": 3.58604407135362e-05, "loss": 0.4711, "step": 752 }, { "epoch": 1.065109695682944, "grad_norm": 0.235408066450741, "learning_rate": 3.583420776495278e-05, "loss": 0.4264, "step": 753 }, { "epoch": 1.0665251238499647, "grad_norm": 0.22142286921184834, "learning_rate": 3.580797481636936e-05, "loss": 0.4651, "step": 754 }, { "epoch": 1.0679405520169851, "grad_norm": 0.2446926882206817, "learning_rate": 3.578174186778594e-05, "loss": 0.4873, "step": 755 }, { "epoch": 1.0693559801840056, "grad_norm": 0.24042796503093813, "learning_rate": 3.575550891920252e-05, "loss": 0.4189, "step": 756 }, { "epoch": 1.0707714083510262, "grad_norm": 0.22711997762409242, "learning_rate": 3.5729275970619095e-05, "loss": 0.4473, "step": 757 }, { "epoch": 1.0721868365180467, "grad_norm": 0.23545924865106554, "learning_rate": 3.5703043022035675e-05, "loss": 0.5082, "step": 758 }, { "epoch": 1.0736022646850671, "grad_norm": 0.23555821624394915, "learning_rate": 3.567681007345226e-05, "loss": 0.5034, "step": 759 }, { "epoch": 1.0750176928520878, "grad_norm": 0.31042957046084124, "learning_rate": 3.565057712486884e-05, "loss": 0.5632, "step": 760 }, { "epoch": 1.0764331210191083, "grad_norm": 0.23561315854681714, "learning_rate": 3.5624344176285416e-05, "loss": 0.4329, "step": 761 }, { "epoch": 1.0778485491861287, "grad_norm": 4.634358749996599, "learning_rate": 3.5598111227701996e-05, "loss": 0.8723, "step": 762 }, { "epoch": 1.0792639773531494, "grad_norm": 0.24825196084343662, "learning_rate": 3.5571878279118577e-05, "loss": 0.4512, "step": 763 }, { "epoch": 1.0806794055201698, "grad_norm": 0.23154206128644247, "learning_rate": 3.554564533053516e-05, "loss": 0.5075, "step": 764 }, { "epoch": 1.0820948336871903, "grad_norm": 0.22375260713189787, "learning_rate": 3.551941238195173e-05, "loss": 0.5031, "step": 765 }, { "epoch": 1.083510261854211, "grad_norm": 0.2658012519094841, "learning_rate": 3.549317943336831e-05, "loss": 0.4712, "step": 766 }, { "epoch": 1.0849256900212314, "grad_norm": 0.21384773833507273, "learning_rate": 3.546694648478489e-05, "loss": 0.4179, "step": 767 }, { "epoch": 1.0863411181882519, "grad_norm": 0.25518402808963636, "learning_rate": 3.544071353620147e-05, "loss": 0.5349, "step": 768 }, { "epoch": 1.0877565463552725, "grad_norm": 0.23346259858577614, "learning_rate": 3.541448058761805e-05, "loss": 0.4919, "step": 769 }, { "epoch": 1.089171974522293, "grad_norm": 0.2654448042799127, "learning_rate": 3.5388247639034625e-05, "loss": 0.4578, "step": 770 }, { "epoch": 1.0905874026893134, "grad_norm": 0.23663475395121936, "learning_rate": 3.5362014690451205e-05, "loss": 0.4876, "step": 771 }, { "epoch": 1.092002830856334, "grad_norm": 0.28682779756077115, "learning_rate": 3.5335781741867786e-05, "loss": 0.4672, "step": 772 }, { "epoch": 1.0934182590233545, "grad_norm": 0.23900495511199596, "learning_rate": 3.530954879328437e-05, "loss": 0.4966, "step": 773 }, { "epoch": 1.094833687190375, "grad_norm": 0.22906300377082478, "learning_rate": 3.5283315844700946e-05, "loss": 0.4555, "step": 774 }, { "epoch": 1.0962491153573957, "grad_norm": 0.28952521918138424, "learning_rate": 3.5257082896117526e-05, "loss": 0.4936, "step": 775 }, { "epoch": 1.0976645435244161, "grad_norm": 0.25159046633454313, "learning_rate": 3.523084994753411e-05, "loss": 0.5253, "step": 776 }, { "epoch": 1.0990799716914366, "grad_norm": 0.22215431109988767, "learning_rate": 3.520461699895069e-05, "loss": 0.4271, "step": 777 }, { "epoch": 1.1004953998584572, "grad_norm": 0.23733069585816344, "learning_rate": 3.517838405036726e-05, "loss": 0.4476, "step": 778 }, { "epoch": 1.1019108280254777, "grad_norm": 0.22833326061820136, "learning_rate": 3.515215110178384e-05, "loss": 0.4446, "step": 779 }, { "epoch": 1.1033262561924981, "grad_norm": 0.9783242275899222, "learning_rate": 3.512591815320042e-05, "loss": 0.5131, "step": 780 }, { "epoch": 1.1047416843595188, "grad_norm": 0.2520699915905762, "learning_rate": 3.5099685204617e-05, "loss": 0.5245, "step": 781 }, { "epoch": 1.1061571125265393, "grad_norm": 0.2540029796902151, "learning_rate": 3.507345225603358e-05, "loss": 0.4551, "step": 782 }, { "epoch": 1.1075725406935597, "grad_norm": 0.24500481650416778, "learning_rate": 3.5047219307450155e-05, "loss": 0.4209, "step": 783 }, { "epoch": 1.1089879688605804, "grad_norm": 0.24689125325205152, "learning_rate": 3.5020986358866735e-05, "loss": 0.4828, "step": 784 }, { "epoch": 1.1104033970276008, "grad_norm": 0.26383620785800466, "learning_rate": 3.4994753410283316e-05, "loss": 0.4429, "step": 785 }, { "epoch": 1.1118188251946213, "grad_norm": 0.24930702957956494, "learning_rate": 3.4968520461699896e-05, "loss": 0.5144, "step": 786 }, { "epoch": 1.113234253361642, "grad_norm": 0.323610419343545, "learning_rate": 3.4942287513116476e-05, "loss": 0.4868, "step": 787 }, { "epoch": 1.1146496815286624, "grad_norm": 0.24451259912991474, "learning_rate": 3.4916054564533057e-05, "loss": 0.5213, "step": 788 }, { "epoch": 1.1160651096956828, "grad_norm": 26.067316043931317, "learning_rate": 3.488982161594964e-05, "loss": 0.6876, "step": 789 }, { "epoch": 1.1174805378627035, "grad_norm": 0.28236163541057463, "learning_rate": 3.486358866736622e-05, "loss": 0.51, "step": 790 }, { "epoch": 1.118895966029724, "grad_norm": 2.266655130223754, "learning_rate": 3.483735571878279e-05, "loss": 0.446, "step": 791 }, { "epoch": 1.1203113941967444, "grad_norm": 0.2963042982811455, "learning_rate": 3.481112277019937e-05, "loss": 0.4626, "step": 792 }, { "epoch": 1.121726822363765, "grad_norm": 0.23579236543255203, "learning_rate": 3.478488982161595e-05, "loss": 0.4657, "step": 793 }, { "epoch": 1.1231422505307855, "grad_norm": 0.3116708551094102, "learning_rate": 3.475865687303253e-05, "loss": 0.5268, "step": 794 }, { "epoch": 1.124557678697806, "grad_norm": 0.25446445484842806, "learning_rate": 3.473242392444911e-05, "loss": 0.4926, "step": 795 }, { "epoch": 1.1259731068648267, "grad_norm": 0.27850892606942623, "learning_rate": 3.4706190975865685e-05, "loss": 0.5049, "step": 796 }, { "epoch": 1.127388535031847, "grad_norm": 0.25055938561632096, "learning_rate": 3.4679958027282266e-05, "loss": 0.4825, "step": 797 }, { "epoch": 1.1288039631988678, "grad_norm": 0.22638791452627927, "learning_rate": 3.4653725078698846e-05, "loss": 0.4704, "step": 798 }, { "epoch": 1.1302193913658882, "grad_norm": 0.26913458804456786, "learning_rate": 3.4627492130115426e-05, "loss": 0.5425, "step": 799 }, { "epoch": 1.1316348195329087, "grad_norm": 0.2580916567814921, "learning_rate": 3.4601259181532006e-05, "loss": 0.4642, "step": 800 }, { "epoch": 1.1330502476999293, "grad_norm": 0.23641384169641477, "learning_rate": 3.457502623294859e-05, "loss": 0.4672, "step": 801 }, { "epoch": 1.1344656758669498, "grad_norm": 0.2494168779349966, "learning_rate": 3.454879328436517e-05, "loss": 0.4502, "step": 802 }, { "epoch": 1.1358811040339702, "grad_norm": 0.22721047026479216, "learning_rate": 3.452256033578175e-05, "loss": 0.4378, "step": 803 }, { "epoch": 1.137296532200991, "grad_norm": 0.285321833786431, "learning_rate": 3.449632738719832e-05, "loss": 0.4712, "step": 804 }, { "epoch": 1.1387119603680114, "grad_norm": 0.2676136691460945, "learning_rate": 3.44700944386149e-05, "loss": 0.4391, "step": 805 }, { "epoch": 1.1401273885350318, "grad_norm": 0.26345274892177956, "learning_rate": 3.444386149003148e-05, "loss": 0.574, "step": 806 }, { "epoch": 1.1415428167020525, "grad_norm": 0.25711450287192117, "learning_rate": 3.441762854144806e-05, "loss": 0.4732, "step": 807 }, { "epoch": 1.142958244869073, "grad_norm": 3.871966915539433, "learning_rate": 3.439139559286464e-05, "loss": 0.5585, "step": 808 }, { "epoch": 1.1443736730360934, "grad_norm": 0.2885361898507915, "learning_rate": 3.4365162644281215e-05, "loss": 0.4955, "step": 809 }, { "epoch": 1.145789101203114, "grad_norm": 0.27866300020903967, "learning_rate": 3.4338929695697796e-05, "loss": 0.4305, "step": 810 }, { "epoch": 1.1472045293701345, "grad_norm": 0.23500417618489086, "learning_rate": 3.4312696747114376e-05, "loss": 0.8119, "step": 811 }, { "epoch": 1.148619957537155, "grad_norm": 6.094213492209398, "learning_rate": 3.4286463798530956e-05, "loss": 0.4879, "step": 812 }, { "epoch": 1.1500353857041756, "grad_norm": 0.2876658250108991, "learning_rate": 3.4260230849947537e-05, "loss": 0.4099, "step": 813 }, { "epoch": 1.151450813871196, "grad_norm": 0.27228640641162283, "learning_rate": 3.423399790136412e-05, "loss": 0.5456, "step": 814 }, { "epoch": 1.1528662420382165, "grad_norm": 0.23578573565103855, "learning_rate": 3.42077649527807e-05, "loss": 0.5166, "step": 815 }, { "epoch": 1.1542816702052372, "grad_norm": 0.29346102651122435, "learning_rate": 3.418153200419728e-05, "loss": 0.5555, "step": 816 }, { "epoch": 1.1556970983722576, "grad_norm": 0.3214253003176066, "learning_rate": 3.415529905561385e-05, "loss": 0.5404, "step": 817 }, { "epoch": 1.157112526539278, "grad_norm": 0.23422729928452454, "learning_rate": 3.412906610703043e-05, "loss": 0.4641, "step": 818 }, { "epoch": 1.1585279547062988, "grad_norm": 0.2511446359869489, "learning_rate": 3.410283315844701e-05, "loss": 0.5066, "step": 819 }, { "epoch": 1.1599433828733192, "grad_norm": 0.27588936308706186, "learning_rate": 3.407660020986359e-05, "loss": 0.5084, "step": 820 }, { "epoch": 1.1613588110403397, "grad_norm": 0.23047802744875695, "learning_rate": 3.405036726128017e-05, "loss": 0.4677, "step": 821 }, { "epoch": 1.1627742392073603, "grad_norm": 0.2830873368697821, "learning_rate": 3.4024134312696746e-05, "loss": 0.4899, "step": 822 }, { "epoch": 1.1641896673743808, "grad_norm": 0.2464140175950677, "learning_rate": 3.3997901364113326e-05, "loss": 0.4852, "step": 823 }, { "epoch": 1.1656050955414012, "grad_norm": 0.28234016547070406, "learning_rate": 3.3971668415529906e-05, "loss": 0.5077, "step": 824 }, { "epoch": 1.167020523708422, "grad_norm": 0.2865470126601044, "learning_rate": 3.3945435466946486e-05, "loss": 0.4079, "step": 825 }, { "epoch": 1.1684359518754424, "grad_norm": 0.24870979695020073, "learning_rate": 3.391920251836307e-05, "loss": 0.5259, "step": 826 }, { "epoch": 1.1698513800424628, "grad_norm": 0.2655929806781905, "learning_rate": 3.389296956977964e-05, "loss": 0.4316, "step": 827 }, { "epoch": 1.1712668082094835, "grad_norm": 0.2465051182015438, "learning_rate": 3.386673662119623e-05, "loss": 0.504, "step": 828 }, { "epoch": 1.172682236376504, "grad_norm": 0.25728893232852484, "learning_rate": 3.384050367261281e-05, "loss": 0.4545, "step": 829 }, { "epoch": 1.1740976645435244, "grad_norm": 0.2534788672603512, "learning_rate": 3.381427072402938e-05, "loss": 0.4468, "step": 830 }, { "epoch": 1.175513092710545, "grad_norm": 0.2702950684859484, "learning_rate": 3.378803777544596e-05, "loss": 0.4836, "step": 831 }, { "epoch": 1.1769285208775655, "grad_norm": 2.1650581544740706, "learning_rate": 3.376180482686254e-05, "loss": 0.5522, "step": 832 }, { "epoch": 1.178343949044586, "grad_norm": 0.26212934230121615, "learning_rate": 3.373557187827912e-05, "loss": 0.4916, "step": 833 }, { "epoch": 1.1797593772116066, "grad_norm": 0.23992354272423078, "learning_rate": 3.37093389296957e-05, "loss": 0.4496, "step": 834 }, { "epoch": 1.181174805378627, "grad_norm": 0.24347908114752023, "learning_rate": 3.3683105981112276e-05, "loss": 0.4423, "step": 835 }, { "epoch": 1.1825902335456475, "grad_norm": 0.2481739432773745, "learning_rate": 3.3656873032528856e-05, "loss": 0.4634, "step": 836 }, { "epoch": 1.1840056617126682, "grad_norm": 0.23592862661803066, "learning_rate": 3.3630640083945436e-05, "loss": 0.5092, "step": 837 }, { "epoch": 1.1854210898796886, "grad_norm": 0.26981634557028705, "learning_rate": 3.3604407135362017e-05, "loss": 0.5233, "step": 838 }, { "epoch": 1.186836518046709, "grad_norm": 0.27743377500140626, "learning_rate": 3.35781741867786e-05, "loss": 0.5167, "step": 839 }, { "epoch": 1.1882519462137298, "grad_norm": 0.8530392452926353, "learning_rate": 3.355194123819517e-05, "loss": 0.4858, "step": 840 }, { "epoch": 1.1896673743807502, "grad_norm": 0.2478597033408157, "learning_rate": 3.352570828961175e-05, "loss": 0.4535, "step": 841 }, { "epoch": 1.1910828025477707, "grad_norm": 0.22109567674510855, "learning_rate": 3.349947534102834e-05, "loss": 0.4773, "step": 842 }, { "epoch": 1.1924982307147913, "grad_norm": 0.24990807485377547, "learning_rate": 3.347324239244491e-05, "loss": 0.5202, "step": 843 }, { "epoch": 1.1939136588818118, "grad_norm": 0.2233711639059363, "learning_rate": 3.344700944386149e-05, "loss": 0.4579, "step": 844 }, { "epoch": 1.1953290870488322, "grad_norm": 0.219961031890485, "learning_rate": 3.342077649527807e-05, "loss": 0.4246, "step": 845 }, { "epoch": 1.196744515215853, "grad_norm": 0.5028905087093073, "learning_rate": 3.339454354669465e-05, "loss": 0.5419, "step": 846 }, { "epoch": 1.1981599433828733, "grad_norm": 0.2578553709131576, "learning_rate": 3.336831059811123e-05, "loss": 0.4634, "step": 847 }, { "epoch": 1.1995753715498938, "grad_norm": 0.2312399763446513, "learning_rate": 3.3342077649527806e-05, "loss": 0.478, "step": 848 }, { "epoch": 1.2009907997169145, "grad_norm": 0.2637322321499473, "learning_rate": 3.3315844700944386e-05, "loss": 0.57, "step": 849 }, { "epoch": 1.202406227883935, "grad_norm": 0.24732656083357885, "learning_rate": 3.3289611752360966e-05, "loss": 0.4796, "step": 850 }, { "epoch": 1.2038216560509554, "grad_norm": 0.21999558816702827, "learning_rate": 3.326337880377755e-05, "loss": 0.4194, "step": 851 }, { "epoch": 1.205237084217976, "grad_norm": 0.22634734957293223, "learning_rate": 3.323714585519413e-05, "loss": 0.4247, "step": 852 }, { "epoch": 1.2066525123849965, "grad_norm": 0.21455915933962133, "learning_rate": 3.32109129066107e-05, "loss": 0.4582, "step": 853 }, { "epoch": 1.208067940552017, "grad_norm": 0.23973577008088373, "learning_rate": 3.318467995802728e-05, "loss": 0.529, "step": 854 }, { "epoch": 1.2094833687190376, "grad_norm": 1.4575856656096844, "learning_rate": 3.315844700944386e-05, "loss": 0.5281, "step": 855 }, { "epoch": 1.210898796886058, "grad_norm": 0.22286056311078778, "learning_rate": 3.313221406086044e-05, "loss": 0.5208, "step": 856 }, { "epoch": 1.2123142250530785, "grad_norm": 0.24242130053002622, "learning_rate": 3.310598111227702e-05, "loss": 0.4632, "step": 857 }, { "epoch": 1.2137296532200992, "grad_norm": 0.2191878726682969, "learning_rate": 3.30797481636936e-05, "loss": 0.4902, "step": 858 }, { "epoch": 1.2151450813871196, "grad_norm": 0.22845347926773774, "learning_rate": 3.305351521511018e-05, "loss": 0.4516, "step": 859 }, { "epoch": 1.21656050955414, "grad_norm": 0.22736406821613941, "learning_rate": 3.302728226652676e-05, "loss": 0.5536, "step": 860 }, { "epoch": 1.2179759377211608, "grad_norm": 0.27494329647314425, "learning_rate": 3.3001049317943336e-05, "loss": 0.4804, "step": 861 }, { "epoch": 1.2193913658881812, "grad_norm": 0.22810142235797856, "learning_rate": 3.2974816369359916e-05, "loss": 0.428, "step": 862 }, { "epoch": 1.2208067940552016, "grad_norm": 0.23307785071101525, "learning_rate": 3.29485834207765e-05, "loss": 0.4578, "step": 863 }, { "epoch": 1.2222222222222223, "grad_norm": 0.2452984529709648, "learning_rate": 3.292235047219308e-05, "loss": 0.5095, "step": 864 }, { "epoch": 1.2236376503892428, "grad_norm": 0.23636401383941374, "learning_rate": 3.289611752360966e-05, "loss": 0.4137, "step": 865 }, { "epoch": 1.2250530785562632, "grad_norm": 0.22493101535805382, "learning_rate": 3.286988457502623e-05, "loss": 0.4343, "step": 866 }, { "epoch": 1.226468506723284, "grad_norm": 0.2615979917519968, "learning_rate": 3.284365162644281e-05, "loss": 0.4999, "step": 867 }, { "epoch": 1.2278839348903043, "grad_norm": 0.26507045747378327, "learning_rate": 3.281741867785939e-05, "loss": 0.4466, "step": 868 }, { "epoch": 1.2292993630573248, "grad_norm": 0.2570384535246848, "learning_rate": 3.279118572927597e-05, "loss": 0.5331, "step": 869 }, { "epoch": 1.2307147912243455, "grad_norm": 0.29062338170307533, "learning_rate": 3.276495278069255e-05, "loss": 0.505, "step": 870 }, { "epoch": 1.232130219391366, "grad_norm": 0.26476425087429617, "learning_rate": 3.273871983210913e-05, "loss": 0.425, "step": 871 }, { "epoch": 1.2335456475583864, "grad_norm": 0.2477692770292864, "learning_rate": 3.271248688352571e-05, "loss": 0.5162, "step": 872 }, { "epoch": 1.234961075725407, "grad_norm": 0.27488613410288, "learning_rate": 3.268625393494229e-05, "loss": 0.4596, "step": 873 }, { "epoch": 1.2363765038924275, "grad_norm": 0.23168934163955254, "learning_rate": 3.2660020986358866e-05, "loss": 0.4323, "step": 874 }, { "epoch": 1.237791932059448, "grad_norm": 0.24431079242225726, "learning_rate": 3.2633788037775447e-05, "loss": 0.4508, "step": 875 }, { "epoch": 1.2392073602264686, "grad_norm": 0.25823573038102476, "learning_rate": 3.260755508919203e-05, "loss": 0.4996, "step": 876 }, { "epoch": 1.240622788393489, "grad_norm": 0.27358168249697484, "learning_rate": 3.258132214060861e-05, "loss": 0.4859, "step": 877 }, { "epoch": 1.2420382165605095, "grad_norm": 0.2626978965439729, "learning_rate": 3.255508919202519e-05, "loss": 0.4217, "step": 878 }, { "epoch": 1.2434536447275302, "grad_norm": 0.26935438304307635, "learning_rate": 3.252885624344176e-05, "loss": 0.4972, "step": 879 }, { "epoch": 1.2448690728945506, "grad_norm": 0.29258459138262827, "learning_rate": 3.250262329485834e-05, "loss": 0.4704, "step": 880 }, { "epoch": 1.246284501061571, "grad_norm": 0.279869756393473, "learning_rate": 3.247639034627492e-05, "loss": 0.5122, "step": 881 }, { "epoch": 1.2476999292285917, "grad_norm": 0.24949248682073935, "learning_rate": 3.24501573976915e-05, "loss": 0.4552, "step": 882 }, { "epoch": 1.2491153573956122, "grad_norm": 0.2592993293922595, "learning_rate": 3.242392444910808e-05, "loss": 0.4835, "step": 883 }, { "epoch": 1.2505307855626326, "grad_norm": 0.2531961962401023, "learning_rate": 3.239769150052466e-05, "loss": 0.4653, "step": 884 }, { "epoch": 1.2519462137296533, "grad_norm": 0.24363061117308296, "learning_rate": 3.237145855194124e-05, "loss": 0.8861, "step": 885 }, { "epoch": 1.2533616418966738, "grad_norm": 5.495245862133001, "learning_rate": 3.234522560335782e-05, "loss": 0.489, "step": 886 }, { "epoch": 1.2547770700636942, "grad_norm": 0.2795802637952873, "learning_rate": 3.2318992654774396e-05, "loss": 0.4847, "step": 887 }, { "epoch": 1.2561924982307149, "grad_norm": 0.27057315035050855, "learning_rate": 3.229275970619098e-05, "loss": 0.4453, "step": 888 }, { "epoch": 1.2576079263977353, "grad_norm": 0.23049854015007026, "learning_rate": 3.226652675760756e-05, "loss": 0.4268, "step": 889 }, { "epoch": 1.2590233545647558, "grad_norm": 0.286146313618306, "learning_rate": 3.224029380902414e-05, "loss": 0.4644, "step": 890 }, { "epoch": 1.2604387827317765, "grad_norm": 0.24410307044469992, "learning_rate": 3.221406086044072e-05, "loss": 0.4784, "step": 891 }, { "epoch": 1.261854210898797, "grad_norm": 0.23806816934844402, "learning_rate": 3.218782791185729e-05, "loss": 0.5156, "step": 892 }, { "epoch": 1.2632696390658174, "grad_norm": 0.23357027315842244, "learning_rate": 3.216159496327387e-05, "loss": 0.4456, "step": 893 }, { "epoch": 1.264685067232838, "grad_norm": 0.2566041334719668, "learning_rate": 3.213536201469045e-05, "loss": 0.4236, "step": 894 }, { "epoch": 1.2661004953998585, "grad_norm": 0.24594304892172075, "learning_rate": 3.210912906610703e-05, "loss": 0.4315, "step": 895 }, { "epoch": 1.267515923566879, "grad_norm": 0.23237072851985374, "learning_rate": 3.2082896117523605e-05, "loss": 0.4882, "step": 896 }, { "epoch": 1.2689313517338996, "grad_norm": 0.2716965951301799, "learning_rate": 3.205666316894019e-05, "loss": 0.4854, "step": 897 }, { "epoch": 1.27034677990092, "grad_norm": 0.24686150409198165, "learning_rate": 3.203043022035677e-05, "loss": 0.5459, "step": 898 }, { "epoch": 1.2717622080679405, "grad_norm": 0.23380918880880155, "learning_rate": 3.200419727177335e-05, "loss": 0.4041, "step": 899 }, { "epoch": 1.2731776362349612, "grad_norm": 0.22071068762301357, "learning_rate": 3.1977964323189927e-05, "loss": 0.434, "step": 900 }, { "epoch": 1.2745930644019816, "grad_norm": 0.2585145407637671, "learning_rate": 3.195173137460651e-05, "loss": 0.3878, "step": 901 }, { "epoch": 1.276008492569002, "grad_norm": 0.23289141730231533, "learning_rate": 3.192549842602309e-05, "loss": 0.5144, "step": 902 }, { "epoch": 1.2774239207360227, "grad_norm": 0.23222966956027472, "learning_rate": 3.189926547743967e-05, "loss": 0.4388, "step": 903 }, { "epoch": 1.2788393489030432, "grad_norm": 0.22261881527619265, "learning_rate": 3.187303252885625e-05, "loss": 0.4779, "step": 904 }, { "epoch": 1.2802547770700636, "grad_norm": 0.23633660841854076, "learning_rate": 3.184679958027282e-05, "loss": 0.5065, "step": 905 }, { "epoch": 1.2816702052370843, "grad_norm": 0.2563369762371776, "learning_rate": 3.18205666316894e-05, "loss": 0.4675, "step": 906 }, { "epoch": 1.2830856334041048, "grad_norm": 0.24241008063485786, "learning_rate": 3.179433368310598e-05, "loss": 0.526, "step": 907 }, { "epoch": 1.2845010615711252, "grad_norm": 0.2438974850078751, "learning_rate": 3.176810073452256e-05, "loss": 0.4372, "step": 908 }, { "epoch": 1.2859164897381459, "grad_norm": 0.2484499708071917, "learning_rate": 3.1741867785939136e-05, "loss": 0.4471, "step": 909 }, { "epoch": 1.2873319179051663, "grad_norm": 0.22020624012488788, "learning_rate": 3.1715634837355716e-05, "loss": 0.4817, "step": 910 }, { "epoch": 1.2887473460721868, "grad_norm": 0.2079438907389061, "learning_rate": 3.16894018887723e-05, "loss": 0.4036, "step": 911 }, { "epoch": 1.2901627742392074, "grad_norm": 0.23809027919871859, "learning_rate": 3.166316894018888e-05, "loss": 0.4757, "step": 912 }, { "epoch": 1.291578202406228, "grad_norm": 0.2534537581446485, "learning_rate": 3.163693599160546e-05, "loss": 0.4632, "step": 913 }, { "epoch": 1.2929936305732483, "grad_norm": 0.2508591501784022, "learning_rate": 3.161070304302204e-05, "loss": 0.5236, "step": 914 }, { "epoch": 1.294409058740269, "grad_norm": 0.2692799394554751, "learning_rate": 3.158447009443862e-05, "loss": 0.5129, "step": 915 }, { "epoch": 1.2958244869072895, "grad_norm": 0.26762393032776954, "learning_rate": 3.15582371458552e-05, "loss": 0.4533, "step": 916 }, { "epoch": 1.29723991507431, "grad_norm": 0.22352913686126197, "learning_rate": 3.153200419727178e-05, "loss": 0.5367, "step": 917 }, { "epoch": 1.2986553432413306, "grad_norm": 0.2702914668199128, "learning_rate": 3.150577124868835e-05, "loss": 0.4394, "step": 918 }, { "epoch": 1.300070771408351, "grad_norm": 0.26063754811266, "learning_rate": 3.147953830010493e-05, "loss": 0.4127, "step": 919 }, { "epoch": 1.3014861995753715, "grad_norm": 0.24585300821822167, "learning_rate": 3.145330535152151e-05, "loss": 0.4451, "step": 920 }, { "epoch": 1.3029016277423922, "grad_norm": 0.268034212777459, "learning_rate": 3.142707240293809e-05, "loss": 0.467, "step": 921 }, { "epoch": 1.3043170559094126, "grad_norm": 0.2409662370057274, "learning_rate": 3.140083945435467e-05, "loss": 0.4499, "step": 922 }, { "epoch": 1.305732484076433, "grad_norm": 0.28243546373141387, "learning_rate": 3.1374606505771246e-05, "loss": 0.5331, "step": 923 }, { "epoch": 1.3071479122434537, "grad_norm": 0.25855346685649155, "learning_rate": 3.1348373557187826e-05, "loss": 0.5168, "step": 924 }, { "epoch": 1.3085633404104742, "grad_norm": 0.28182218038496243, "learning_rate": 3.132214060860441e-05, "loss": 0.4565, "step": 925 }, { "epoch": 1.3099787685774946, "grad_norm": 0.2403385267177625, "learning_rate": 3.129590766002099e-05, "loss": 0.5043, "step": 926 }, { "epoch": 1.3113941967445153, "grad_norm": 0.27408089069596053, "learning_rate": 3.126967471143757e-05, "loss": 0.5451, "step": 927 }, { "epoch": 1.3128096249115357, "grad_norm": 0.2646909795264679, "learning_rate": 3.124344176285415e-05, "loss": 0.479, "step": 928 }, { "epoch": 1.3142250530785562, "grad_norm": 0.2305069678068314, "learning_rate": 3.121720881427073e-05, "loss": 0.5296, "step": 929 }, { "epoch": 1.3156404812455769, "grad_norm": 0.2744177117718372, "learning_rate": 3.119097586568731e-05, "loss": 0.4754, "step": 930 }, { "epoch": 1.3170559094125973, "grad_norm": 0.6499533802326067, "learning_rate": 3.116474291710388e-05, "loss": 0.5306, "step": 931 }, { "epoch": 1.3184713375796178, "grad_norm": 0.2303504782797021, "learning_rate": 3.113850996852046e-05, "loss": 0.4832, "step": 932 }, { "epoch": 1.3198867657466384, "grad_norm": 0.2615164107491973, "learning_rate": 3.111227701993704e-05, "loss": 0.5244, "step": 933 }, { "epoch": 1.3213021939136589, "grad_norm": 0.2959878077888004, "learning_rate": 3.108604407135362e-05, "loss": 0.4774, "step": 934 }, { "epoch": 1.3227176220806793, "grad_norm": 0.23964789845458753, "learning_rate": 3.10598111227702e-05, "loss": 0.4766, "step": 935 }, { "epoch": 1.3241330502477, "grad_norm": 0.21690181913877962, "learning_rate": 3.1033578174186776e-05, "loss": 0.419, "step": 936 }, { "epoch": 1.3255484784147205, "grad_norm": 0.6426389076756095, "learning_rate": 3.1007345225603356e-05, "loss": 0.4507, "step": 937 }, { "epoch": 1.326963906581741, "grad_norm": 0.263147646977239, "learning_rate": 3.098111227701994e-05, "loss": 0.4573, "step": 938 }, { "epoch": 1.3283793347487616, "grad_norm": 0.24757143586337804, "learning_rate": 3.095487932843652e-05, "loss": 0.5759, "step": 939 }, { "epoch": 1.329794762915782, "grad_norm": 0.25257042616743186, "learning_rate": 3.09286463798531e-05, "loss": 0.4862, "step": 940 }, { "epoch": 1.3312101910828025, "grad_norm": 0.24251280132899075, "learning_rate": 3.090241343126968e-05, "loss": 0.4896, "step": 941 }, { "epoch": 1.3326256192498231, "grad_norm": 0.25064062263912584, "learning_rate": 3.087618048268626e-05, "loss": 0.489, "step": 942 }, { "epoch": 1.3340410474168436, "grad_norm": 0.24516768410214754, "learning_rate": 3.084994753410284e-05, "loss": 0.4174, "step": 943 }, { "epoch": 1.335456475583864, "grad_norm": 0.25861482096183286, "learning_rate": 3.082371458551941e-05, "loss": 0.4574, "step": 944 }, { "epoch": 1.3368719037508847, "grad_norm": 6.68702234250407, "learning_rate": 3.079748163693599e-05, "loss": 0.4732, "step": 945 }, { "epoch": 1.3382873319179052, "grad_norm": 0.25068169445356625, "learning_rate": 3.077124868835257e-05, "loss": 0.4789, "step": 946 }, { "epoch": 1.3397027600849256, "grad_norm": 0.25093075289474387, "learning_rate": 3.074501573976915e-05, "loss": 0.4568, "step": 947 }, { "epoch": 1.3411181882519463, "grad_norm": 0.28246974745034786, "learning_rate": 3.071878279118573e-05, "loss": 0.4825, "step": 948 }, { "epoch": 1.3425336164189667, "grad_norm": 0.21166319183056573, "learning_rate": 3.0692549842602306e-05, "loss": 0.439, "step": 949 }, { "epoch": 1.3439490445859872, "grad_norm": 1.0385631184142639, "learning_rate": 3.0666316894018887e-05, "loss": 0.4454, "step": 950 }, { "epoch": 1.3453644727530079, "grad_norm": 0.25777032972003616, "learning_rate": 3.064008394543547e-05, "loss": 0.5057, "step": 951 }, { "epoch": 1.3467799009200283, "grad_norm": 0.26692860775734006, "learning_rate": 3.061385099685205e-05, "loss": 0.481, "step": 952 }, { "epoch": 1.3481953290870488, "grad_norm": 0.24088223472029782, "learning_rate": 3.058761804826863e-05, "loss": 0.5184, "step": 953 }, { "epoch": 1.3496107572540694, "grad_norm": 0.2247620489296305, "learning_rate": 3.056138509968521e-05, "loss": 0.5462, "step": 954 }, { "epoch": 1.3510261854210899, "grad_norm": 0.23526403150601521, "learning_rate": 3.053515215110179e-05, "loss": 0.472, "step": 955 }, { "epoch": 1.3524416135881103, "grad_norm": 0.2318641777138758, "learning_rate": 3.0508919202518365e-05, "loss": 0.446, "step": 956 }, { "epoch": 1.353857041755131, "grad_norm": 0.24399478839109545, "learning_rate": 3.0482686253934945e-05, "loss": 0.45, "step": 957 }, { "epoch": 1.3552724699221514, "grad_norm": 0.2467593178269623, "learning_rate": 3.0456453305351522e-05, "loss": 0.5054, "step": 958 }, { "epoch": 1.356687898089172, "grad_norm": 0.2417281084040149, "learning_rate": 3.0430220356768102e-05, "loss": 0.4866, "step": 959 }, { "epoch": 1.3581033262561926, "grad_norm": 0.24492624076347055, "learning_rate": 3.040398740818468e-05, "loss": 0.5484, "step": 960 }, { "epoch": 1.359518754423213, "grad_norm": 0.2666390176436146, "learning_rate": 3.037775445960126e-05, "loss": 0.513, "step": 961 }, { "epoch": 1.3609341825902335, "grad_norm": 0.2619780519186626, "learning_rate": 3.035152151101784e-05, "loss": 0.4835, "step": 962 }, { "epoch": 1.3623496107572541, "grad_norm": 0.2753506150378985, "learning_rate": 3.0325288562434417e-05, "loss": 0.4276, "step": 963 }, { "epoch": 1.3637650389242746, "grad_norm": 0.23838022166368394, "learning_rate": 3.0299055613850997e-05, "loss": 0.4849, "step": 964 }, { "epoch": 1.365180467091295, "grad_norm": 0.24285680424504225, "learning_rate": 3.0272822665267574e-05, "loss": 0.4538, "step": 965 }, { "epoch": 1.3665958952583157, "grad_norm": 4.7723592644763535, "learning_rate": 3.024658971668416e-05, "loss": 0.4491, "step": 966 }, { "epoch": 1.3680113234253362, "grad_norm": 0.24476982720764057, "learning_rate": 3.0220356768100738e-05, "loss": 0.4339, "step": 967 }, { "epoch": 1.3694267515923566, "grad_norm": 0.2959167527473618, "learning_rate": 3.0194123819517318e-05, "loss": 0.4118, "step": 968 }, { "epoch": 1.3708421797593773, "grad_norm": 0.2408274889934171, "learning_rate": 3.0167890870933895e-05, "loss": 0.405, "step": 969 }, { "epoch": 1.3722576079263977, "grad_norm": 0.23977225860357196, "learning_rate": 3.0141657922350475e-05, "loss": 0.4566, "step": 970 }, { "epoch": 1.3736730360934182, "grad_norm": 0.258233874390108, "learning_rate": 3.0115424973767052e-05, "loss": 0.4709, "step": 971 }, { "epoch": 1.3750884642604388, "grad_norm": 0.23298288179195542, "learning_rate": 3.0089192025183633e-05, "loss": 0.4883, "step": 972 }, { "epoch": 1.3765038924274593, "grad_norm": 0.25732371520926234, "learning_rate": 3.0062959076600213e-05, "loss": 0.4587, "step": 973 }, { "epoch": 1.3779193205944797, "grad_norm": 0.2770166948552172, "learning_rate": 3.003672612801679e-05, "loss": 0.4446, "step": 974 }, { "epoch": 1.3793347487615004, "grad_norm": 0.25686220374398416, "learning_rate": 3.001049317943337e-05, "loss": 0.4659, "step": 975 }, { "epoch": 1.3807501769285209, "grad_norm": 0.2748242222470758, "learning_rate": 2.9984260230849947e-05, "loss": 0.5211, "step": 976 }, { "epoch": 1.3821656050955413, "grad_norm": 0.24194184131657517, "learning_rate": 2.9958027282266527e-05, "loss": 0.4752, "step": 977 }, { "epoch": 1.383581033262562, "grad_norm": 0.2725631825825404, "learning_rate": 2.9931794333683104e-05, "loss": 0.4234, "step": 978 }, { "epoch": 1.3849964614295824, "grad_norm": 0.23748217806062075, "learning_rate": 2.9905561385099684e-05, "loss": 0.464, "step": 979 }, { "epoch": 1.3864118895966029, "grad_norm": 0.2309228037376953, "learning_rate": 2.9879328436516268e-05, "loss": 0.4827, "step": 980 }, { "epoch": 1.3878273177636236, "grad_norm": 0.2891815389720354, "learning_rate": 2.985309548793285e-05, "loss": 0.4465, "step": 981 }, { "epoch": 1.389242745930644, "grad_norm": 0.22958479608149038, "learning_rate": 2.9826862539349425e-05, "loss": 0.5058, "step": 982 }, { "epoch": 1.3906581740976645, "grad_norm": 0.24514929972489777, "learning_rate": 2.9800629590766006e-05, "loss": 0.5549, "step": 983 }, { "epoch": 1.3920736022646851, "grad_norm": 1.069141050088208, "learning_rate": 2.9774396642182582e-05, "loss": 0.5151, "step": 984 }, { "epoch": 1.3934890304317056, "grad_norm": 0.24065132318214158, "learning_rate": 2.9748163693599163e-05, "loss": 0.5095, "step": 985 }, { "epoch": 1.394904458598726, "grad_norm": 0.29294351959468184, "learning_rate": 2.9721930745015743e-05, "loss": 0.4606, "step": 986 }, { "epoch": 1.3963198867657467, "grad_norm": 0.2700738054926797, "learning_rate": 2.969569779643232e-05, "loss": 0.4666, "step": 987 }, { "epoch": 1.3977353149327671, "grad_norm": 0.24780694955297938, "learning_rate": 2.96694648478489e-05, "loss": 0.4336, "step": 988 }, { "epoch": 1.3991507430997876, "grad_norm": 0.2825839661544209, "learning_rate": 2.9643231899265477e-05, "loss": 0.511, "step": 989 }, { "epoch": 1.4005661712668083, "grad_norm": 0.31985767764921325, "learning_rate": 2.9616998950682057e-05, "loss": 0.4769, "step": 990 }, { "epoch": 1.4019815994338287, "grad_norm": 0.21515488179858627, "learning_rate": 2.9590766002098634e-05, "loss": 0.4841, "step": 991 }, { "epoch": 1.4033970276008492, "grad_norm": 0.2999937500375315, "learning_rate": 2.9564533053515215e-05, "loss": 0.5267, "step": 992 }, { "epoch": 1.4048124557678698, "grad_norm": 0.27110098463243326, "learning_rate": 2.9538300104931795e-05, "loss": 0.4391, "step": 993 }, { "epoch": 1.4062278839348903, "grad_norm": 0.2632120517796395, "learning_rate": 2.951206715634838e-05, "loss": 0.453, "step": 994 }, { "epoch": 1.4076433121019107, "grad_norm": 0.2738208188659536, "learning_rate": 2.9485834207764955e-05, "loss": 0.4775, "step": 995 }, { "epoch": 1.4090587402689314, "grad_norm": 0.258769719684058, "learning_rate": 2.9459601259181536e-05, "loss": 0.5675, "step": 996 }, { "epoch": 1.4104741684359519, "grad_norm": 0.2513014251809082, "learning_rate": 2.9433368310598113e-05, "loss": 0.5778, "step": 997 }, { "epoch": 1.4118895966029723, "grad_norm": 0.2615702887534967, "learning_rate": 2.9407135362014693e-05, "loss": 0.4508, "step": 998 }, { "epoch": 1.413305024769993, "grad_norm": 0.2630569521513353, "learning_rate": 2.9380902413431273e-05, "loss": 0.4531, "step": 999 }, { "epoch": 1.4147204529370134, "grad_norm": 0.24297232603525354, "learning_rate": 2.935466946484785e-05, "loss": 0.4656, "step": 1000 }, { "epoch": 1.4161358811040339, "grad_norm": 0.23366469373399976, "learning_rate": 2.932843651626443e-05, "loss": 0.4308, "step": 1001 }, { "epoch": 1.4175513092710545, "grad_norm": 0.23892561947726176, "learning_rate": 2.9302203567681007e-05, "loss": 0.4375, "step": 1002 }, { "epoch": 1.418966737438075, "grad_norm": 0.23195340218843138, "learning_rate": 2.9275970619097587e-05, "loss": 0.444, "step": 1003 }, { "epoch": 1.4203821656050954, "grad_norm": 0.22948701966121285, "learning_rate": 2.9249737670514164e-05, "loss": 0.4848, "step": 1004 }, { "epoch": 1.4217975937721161, "grad_norm": 0.24785624514905413, "learning_rate": 2.9223504721930745e-05, "loss": 0.4634, "step": 1005 }, { "epoch": 1.4232130219391366, "grad_norm": 0.2796816145500579, "learning_rate": 2.9197271773347325e-05, "loss": 0.5248, "step": 1006 }, { "epoch": 1.424628450106157, "grad_norm": 0.21395483176765387, "learning_rate": 2.9171038824763902e-05, "loss": 0.4028, "step": 1007 }, { "epoch": 1.4260438782731777, "grad_norm": 0.23564260855313562, "learning_rate": 2.9144805876180486e-05, "loss": 0.5343, "step": 1008 }, { "epoch": 1.4274593064401981, "grad_norm": 0.26481931465222386, "learning_rate": 2.9118572927597066e-05, "loss": 0.4962, "step": 1009 }, { "epoch": 1.4288747346072186, "grad_norm": 0.26619192924819485, "learning_rate": 2.9092339979013643e-05, "loss": 0.5139, "step": 1010 }, { "epoch": 1.4302901627742393, "grad_norm": 0.25387021436623847, "learning_rate": 2.9066107030430223e-05, "loss": 0.5136, "step": 1011 }, { "epoch": 1.4317055909412597, "grad_norm": 0.24429440511933043, "learning_rate": 2.9039874081846803e-05, "loss": 0.4366, "step": 1012 }, { "epoch": 1.4331210191082802, "grad_norm": 0.22053476848614068, "learning_rate": 2.901364113326338e-05, "loss": 0.4985, "step": 1013 }, { "epoch": 1.4345364472753008, "grad_norm": 0.44885317266562963, "learning_rate": 2.898740818467996e-05, "loss": 0.4722, "step": 1014 }, { "epoch": 1.4359518754423213, "grad_norm": 0.2647577301470883, "learning_rate": 2.8961175236096537e-05, "loss": 0.5957, "step": 1015 }, { "epoch": 1.4373673036093417, "grad_norm": 0.25300983944384264, "learning_rate": 2.8934942287513118e-05, "loss": 0.5029, "step": 1016 }, { "epoch": 1.4387827317763624, "grad_norm": 0.3069886239763433, "learning_rate": 2.8908709338929695e-05, "loss": 0.4781, "step": 1017 }, { "epoch": 1.4401981599433828, "grad_norm": 0.24029825627591023, "learning_rate": 2.8882476390346275e-05, "loss": 0.4628, "step": 1018 }, { "epoch": 1.4416135881104033, "grad_norm": 0.2542260271787419, "learning_rate": 2.8856243441762855e-05, "loss": 0.5957, "step": 1019 }, { "epoch": 1.443029016277424, "grad_norm": 0.2659087195888366, "learning_rate": 2.8830010493179432e-05, "loss": 0.4687, "step": 1020 }, { "epoch": 1.4444444444444444, "grad_norm": 0.22703447924437756, "learning_rate": 2.8803777544596016e-05, "loss": 0.4701, "step": 1021 }, { "epoch": 1.4458598726114649, "grad_norm": 0.23932999856039933, "learning_rate": 2.8777544596012596e-05, "loss": 0.4962, "step": 1022 }, { "epoch": 1.4472753007784855, "grad_norm": 0.23705026394966283, "learning_rate": 2.8751311647429173e-05, "loss": 0.4994, "step": 1023 }, { "epoch": 1.448690728945506, "grad_norm": 0.23924281655893787, "learning_rate": 2.8725078698845753e-05, "loss": 0.4897, "step": 1024 }, { "epoch": 1.4501061571125264, "grad_norm": 0.24534844390507993, "learning_rate": 2.8698845750262333e-05, "loss": 0.4633, "step": 1025 }, { "epoch": 1.451521585279547, "grad_norm": 0.24647975510130424, "learning_rate": 2.867261280167891e-05, "loss": 0.5761, "step": 1026 }, { "epoch": 1.4529370134465676, "grad_norm": 0.25902949327635205, "learning_rate": 2.864637985309549e-05, "loss": 0.5124, "step": 1027 }, { "epoch": 1.454352441613588, "grad_norm": 0.24227778031818434, "learning_rate": 2.8620146904512068e-05, "loss": 0.4764, "step": 1028 }, { "epoch": 1.4557678697806087, "grad_norm": 0.23574028432437744, "learning_rate": 2.8593913955928648e-05, "loss": 0.4763, "step": 1029 }, { "epoch": 1.4571832979476291, "grad_norm": 0.255082791321367, "learning_rate": 2.8567681007345225e-05, "loss": 0.4469, "step": 1030 }, { "epoch": 1.4585987261146496, "grad_norm": 0.2337134120806889, "learning_rate": 2.8541448058761805e-05, "loss": 0.4831, "step": 1031 }, { "epoch": 1.4600141542816703, "grad_norm": 0.2373703793885165, "learning_rate": 2.8515215110178385e-05, "loss": 0.5018, "step": 1032 }, { "epoch": 1.4614295824486907, "grad_norm": 0.23276457945691267, "learning_rate": 2.8488982161594962e-05, "loss": 0.4405, "step": 1033 }, { "epoch": 1.4628450106157111, "grad_norm": 0.25045501124926023, "learning_rate": 2.8462749213011542e-05, "loss": 0.4732, "step": 1034 }, { "epoch": 1.4642604387827318, "grad_norm": 0.25464013273882563, "learning_rate": 2.8436516264428126e-05, "loss": 0.4849, "step": 1035 }, { "epoch": 1.4656758669497523, "grad_norm": 0.22818989277388887, "learning_rate": 2.8410283315844703e-05, "loss": 0.4076, "step": 1036 }, { "epoch": 1.4670912951167727, "grad_norm": 0.22555792664130817, "learning_rate": 2.8384050367261283e-05, "loss": 0.4359, "step": 1037 }, { "epoch": 1.4685067232837934, "grad_norm": 0.21960333308218966, "learning_rate": 2.8357817418677864e-05, "loss": 0.47, "step": 1038 }, { "epoch": 1.4699221514508138, "grad_norm": 0.22431300388479583, "learning_rate": 2.833158447009444e-05, "loss": 0.4458, "step": 1039 }, { "epoch": 1.4713375796178343, "grad_norm": 0.22822883270470412, "learning_rate": 2.830535152151102e-05, "loss": 0.5535, "step": 1040 }, { "epoch": 1.472753007784855, "grad_norm": 0.25040042431079407, "learning_rate": 2.8279118572927598e-05, "loss": 0.4972, "step": 1041 }, { "epoch": 1.4741684359518754, "grad_norm": 0.24527285015339936, "learning_rate": 2.8252885624344178e-05, "loss": 0.4816, "step": 1042 }, { "epoch": 1.4755838641188959, "grad_norm": 0.21545250579454714, "learning_rate": 2.8226652675760755e-05, "loss": 0.4822, "step": 1043 }, { "epoch": 1.4769992922859165, "grad_norm": 0.21160614593382832, "learning_rate": 2.8200419727177335e-05, "loss": 0.4696, "step": 1044 }, { "epoch": 1.478414720452937, "grad_norm": 6.954437350346113, "learning_rate": 2.8174186778593915e-05, "loss": 0.4696, "step": 1045 }, { "epoch": 1.4798301486199574, "grad_norm": 0.2446422497343303, "learning_rate": 2.8147953830010492e-05, "loss": 0.4642, "step": 1046 }, { "epoch": 1.481245576786978, "grad_norm": 0.23710815310012034, "learning_rate": 2.8121720881427073e-05, "loss": 0.4564, "step": 1047 }, { "epoch": 1.4826610049539986, "grad_norm": 0.21642519714321273, "learning_rate": 2.809548793284365e-05, "loss": 0.467, "step": 1048 }, { "epoch": 1.484076433121019, "grad_norm": 0.24533444903258622, "learning_rate": 2.8069254984260233e-05, "loss": 0.4857, "step": 1049 }, { "epoch": 1.4854918612880397, "grad_norm": 0.24166298624425117, "learning_rate": 2.8043022035676813e-05, "loss": 0.4672, "step": 1050 }, { "epoch": 1.4869072894550601, "grad_norm": 0.21749550059484396, "learning_rate": 2.8016789087093394e-05, "loss": 0.4719, "step": 1051 }, { "epoch": 1.4883227176220806, "grad_norm": 0.26678528308771254, "learning_rate": 2.799055613850997e-05, "loss": 0.4649, "step": 1052 }, { "epoch": 1.4897381457891012, "grad_norm": 0.21661033809145605, "learning_rate": 2.796432318992655e-05, "loss": 0.4224, "step": 1053 }, { "epoch": 1.4911535739561217, "grad_norm": 0.2313532912014695, "learning_rate": 2.7938090241343128e-05, "loss": 0.5097, "step": 1054 }, { "epoch": 1.4925690021231421, "grad_norm": 0.26205168168462356, "learning_rate": 2.7911857292759708e-05, "loss": 0.5469, "step": 1055 }, { "epoch": 1.4939844302901628, "grad_norm": 0.4863580049870588, "learning_rate": 2.7885624344176285e-05, "loss": 0.429, "step": 1056 }, { "epoch": 1.4953998584571833, "grad_norm": 0.23631813461640477, "learning_rate": 2.7859391395592865e-05, "loss": 0.4717, "step": 1057 }, { "epoch": 1.4968152866242037, "grad_norm": 0.2629048815627727, "learning_rate": 2.7833158447009446e-05, "loss": 0.4474, "step": 1058 }, { "epoch": 1.4982307147912244, "grad_norm": 0.23677824354152793, "learning_rate": 2.7806925498426022e-05, "loss": 0.4815, "step": 1059 }, { "epoch": 1.4996461429582448, "grad_norm": 0.24762130191336798, "learning_rate": 2.7780692549842603e-05, "loss": 0.44, "step": 1060 }, { "epoch": 1.5010615711252653, "grad_norm": 0.24988799078497606, "learning_rate": 2.775445960125918e-05, "loss": 0.3898, "step": 1061 }, { "epoch": 1.502476999292286, "grad_norm": 0.2368243984498134, "learning_rate": 2.772822665267576e-05, "loss": 0.448, "step": 1062 }, { "epoch": 1.5038924274593064, "grad_norm": 0.2725939583573723, "learning_rate": 2.7701993704092344e-05, "loss": 0.4815, "step": 1063 }, { "epoch": 1.5053078556263269, "grad_norm": 0.22491600948340684, "learning_rate": 2.7675760755508924e-05, "loss": 0.4645, "step": 1064 }, { "epoch": 1.5067232837933475, "grad_norm": 0.24749866709382495, "learning_rate": 2.76495278069255e-05, "loss": 0.4113, "step": 1065 }, { "epoch": 1.508138711960368, "grad_norm": 0.2665066221199876, "learning_rate": 2.762329485834208e-05, "loss": 0.5052, "step": 1066 }, { "epoch": 1.5095541401273884, "grad_norm": 0.24199054657747365, "learning_rate": 2.7597061909758658e-05, "loss": 0.4706, "step": 1067 }, { "epoch": 1.510969568294409, "grad_norm": 0.21771253763658477, "learning_rate": 2.7570828961175238e-05, "loss": 0.4613, "step": 1068 }, { "epoch": 1.5123849964614295, "grad_norm": 0.2237890153752989, "learning_rate": 2.7544596012591815e-05, "loss": 0.436, "step": 1069 }, { "epoch": 1.51380042462845, "grad_norm": 0.22977916156772293, "learning_rate": 2.7518363064008395e-05, "loss": 0.4764, "step": 1070 }, { "epoch": 1.5152158527954707, "grad_norm": 0.2436197119328285, "learning_rate": 2.7492130115424976e-05, "loss": 0.4191, "step": 1071 }, { "epoch": 1.5166312809624911, "grad_norm": 0.2814116794689585, "learning_rate": 2.7465897166841553e-05, "loss": 0.4331, "step": 1072 }, { "epoch": 1.5180467091295116, "grad_norm": 0.25998996932832125, "learning_rate": 2.7439664218258133e-05, "loss": 0.4788, "step": 1073 }, { "epoch": 1.5194621372965322, "grad_norm": 0.23739549641148952, "learning_rate": 2.741343126967471e-05, "loss": 0.4872, "step": 1074 }, { "epoch": 1.5208775654635527, "grad_norm": 0.28943485657274004, "learning_rate": 2.738719832109129e-05, "loss": 0.4178, "step": 1075 }, { "epoch": 1.5222929936305731, "grad_norm": 0.20996237549839084, "learning_rate": 2.7360965372507867e-05, "loss": 0.4111, "step": 1076 }, { "epoch": 1.5237084217975938, "grad_norm": 0.23327937342431862, "learning_rate": 2.7334732423924454e-05, "loss": 0.5006, "step": 1077 }, { "epoch": 1.5251238499646143, "grad_norm": 0.26823642539957526, "learning_rate": 2.730849947534103e-05, "loss": 0.4652, "step": 1078 }, { "epoch": 1.5265392781316347, "grad_norm": 0.21072129577399698, "learning_rate": 2.728226652675761e-05, "loss": 0.4129, "step": 1079 }, { "epoch": 1.5279547062986554, "grad_norm": 0.22274933741818195, "learning_rate": 2.7256033578174188e-05, "loss": 0.4531, "step": 1080 }, { "epoch": 1.5293701344656758, "grad_norm": 0.4986936028511287, "learning_rate": 2.722980062959077e-05, "loss": 0.4597, "step": 1081 }, { "epoch": 1.5307855626326963, "grad_norm": 0.25019227363506485, "learning_rate": 2.7203567681007345e-05, "loss": 0.4921, "step": 1082 }, { "epoch": 1.532200990799717, "grad_norm": 0.21779980593438472, "learning_rate": 2.7177334732423926e-05, "loss": 0.4393, "step": 1083 }, { "epoch": 1.5336164189667374, "grad_norm": 0.24912965629233305, "learning_rate": 2.7151101783840506e-05, "loss": 0.433, "step": 1084 }, { "epoch": 1.5350318471337578, "grad_norm": 0.41623656096204564, "learning_rate": 2.7124868835257083e-05, "loss": 0.516, "step": 1085 }, { "epoch": 1.5364472753007785, "grad_norm": 0.2687253995199058, "learning_rate": 2.7098635886673663e-05, "loss": 0.5026, "step": 1086 }, { "epoch": 1.537862703467799, "grad_norm": 0.22337167247678047, "learning_rate": 2.707240293809024e-05, "loss": 0.4452, "step": 1087 }, { "epoch": 1.5392781316348194, "grad_norm": 0.24670114978050278, "learning_rate": 2.704616998950682e-05, "loss": 0.544, "step": 1088 }, { "epoch": 1.54069355980184, "grad_norm": 0.23666084857328787, "learning_rate": 2.7019937040923397e-05, "loss": 0.4499, "step": 1089 }, { "epoch": 1.5421089879688605, "grad_norm": 0.21971236961761928, "learning_rate": 2.6993704092339977e-05, "loss": 0.4844, "step": 1090 }, { "epoch": 1.543524416135881, "grad_norm": 0.23803059606252222, "learning_rate": 2.696747114375656e-05, "loss": 0.4869, "step": 1091 }, { "epoch": 1.5449398443029017, "grad_norm": 0.24184454986545373, "learning_rate": 2.694123819517314e-05, "loss": 0.4538, "step": 1092 }, { "epoch": 1.546355272469922, "grad_norm": 0.2432655023422465, "learning_rate": 2.6915005246589718e-05, "loss": 0.5182, "step": 1093 }, { "epoch": 1.5477707006369426, "grad_norm": 0.21884280908716727, "learning_rate": 2.68887722980063e-05, "loss": 0.4508, "step": 1094 }, { "epoch": 1.5491861288039632, "grad_norm": 0.23837889034609053, "learning_rate": 2.686253934942288e-05, "loss": 0.4026, "step": 1095 }, { "epoch": 1.5506015569709837, "grad_norm": 2.186577418971484, "learning_rate": 2.6836306400839456e-05, "loss": 0.4374, "step": 1096 }, { "epoch": 1.5520169851380041, "grad_norm": 0.2301787401222238, "learning_rate": 2.6810073452256036e-05, "loss": 0.4811, "step": 1097 }, { "epoch": 1.5534324133050248, "grad_norm": 0.26506009481521925, "learning_rate": 2.6783840503672613e-05, "loss": 0.5027, "step": 1098 }, { "epoch": 1.5548478414720452, "grad_norm": 0.22334070900295735, "learning_rate": 2.6757607555089193e-05, "loss": 0.4562, "step": 1099 }, { "epoch": 1.5562632696390657, "grad_norm": 0.23509770488005835, "learning_rate": 2.673137460650577e-05, "loss": 0.5018, "step": 1100 }, { "epoch": 1.5576786978060864, "grad_norm": 0.25822797097500144, "learning_rate": 2.670514165792235e-05, "loss": 0.4649, "step": 1101 }, { "epoch": 1.5590941259731068, "grad_norm": 0.22020847666976612, "learning_rate": 2.6678908709338927e-05, "loss": 0.5247, "step": 1102 }, { "epoch": 1.5605095541401273, "grad_norm": 0.2352135371428812, "learning_rate": 2.6652675760755508e-05, "loss": 0.5146, "step": 1103 }, { "epoch": 1.561924982307148, "grad_norm": 0.2489630883187179, "learning_rate": 2.662644281217209e-05, "loss": 0.481, "step": 1104 }, { "epoch": 1.5633404104741684, "grad_norm": 0.25105003951133575, "learning_rate": 2.660020986358867e-05, "loss": 0.4849, "step": 1105 }, { "epoch": 1.5647558386411888, "grad_norm": 0.22146346229154182, "learning_rate": 2.657397691500525e-05, "loss": 0.514, "step": 1106 }, { "epoch": 1.5661712668082095, "grad_norm": 0.24361463581563225, "learning_rate": 2.654774396642183e-05, "loss": 0.5323, "step": 1107 }, { "epoch": 1.56758669497523, "grad_norm": 0.269911412108188, "learning_rate": 2.652151101783841e-05, "loss": 0.4628, "step": 1108 }, { "epoch": 1.5690021231422504, "grad_norm": 0.2537121060671178, "learning_rate": 2.6495278069254986e-05, "loss": 0.4818, "step": 1109 }, { "epoch": 1.570417551309271, "grad_norm": 0.25966850377325174, "learning_rate": 2.6469045120671566e-05, "loss": 0.5097, "step": 1110 }, { "epoch": 1.5718329794762915, "grad_norm": 0.265144191902071, "learning_rate": 2.6442812172088143e-05, "loss": 0.4998, "step": 1111 }, { "epoch": 1.573248407643312, "grad_norm": 0.26052675948490334, "learning_rate": 2.6416579223504723e-05, "loss": 0.469, "step": 1112 }, { "epoch": 1.5746638358103326, "grad_norm": 1.5639646666796965, "learning_rate": 2.63903462749213e-05, "loss": 0.4874, "step": 1113 }, { "epoch": 1.576079263977353, "grad_norm": 0.34737557088415505, "learning_rate": 2.636411332633788e-05, "loss": 0.4283, "step": 1114 }, { "epoch": 1.5774946921443735, "grad_norm": 0.25563683675844123, "learning_rate": 2.6337880377754457e-05, "loss": 0.4203, "step": 1115 }, { "epoch": 1.5789101203113942, "grad_norm": 0.28441359753020323, "learning_rate": 2.6311647429171038e-05, "loss": 0.468, "step": 1116 }, { "epoch": 1.5803255484784147, "grad_norm": 0.372909035094879, "learning_rate": 2.6285414480587618e-05, "loss": 0.5886, "step": 1117 }, { "epoch": 1.5817409766454351, "grad_norm": 1.5462598794149647, "learning_rate": 2.6259181532004202e-05, "loss": 0.5186, "step": 1118 }, { "epoch": 1.5831564048124558, "grad_norm": 0.2600177084809281, "learning_rate": 2.623294858342078e-05, "loss": 0.3898, "step": 1119 }, { "epoch": 1.5845718329794765, "grad_norm": 0.29616234684449666, "learning_rate": 2.620671563483736e-05, "loss": 0.4848, "step": 1120 }, { "epoch": 1.5859872611464967, "grad_norm": 0.25246015830438095, "learning_rate": 2.618048268625394e-05, "loss": 0.5138, "step": 1121 }, { "epoch": 1.5874026893135174, "grad_norm": 0.2592928418343624, "learning_rate": 2.6154249737670516e-05, "loss": 0.5177, "step": 1122 }, { "epoch": 1.588818117480538, "grad_norm": 1.343675437936197, "learning_rate": 2.6128016789087096e-05, "loss": 0.4652, "step": 1123 }, { "epoch": 1.5902335456475583, "grad_norm": 0.3378412630785343, "learning_rate": 2.6101783840503673e-05, "loss": 0.4741, "step": 1124 }, { "epoch": 1.591648973814579, "grad_norm": 0.23591041247632422, "learning_rate": 2.6075550891920254e-05, "loss": 0.4027, "step": 1125 }, { "epoch": 1.5930644019815996, "grad_norm": 0.27025424322925645, "learning_rate": 2.604931794333683e-05, "loss": 0.5317, "step": 1126 }, { "epoch": 1.5944798301486198, "grad_norm": 0.3032366021165955, "learning_rate": 2.602308499475341e-05, "loss": 0.4353, "step": 1127 }, { "epoch": 1.5958952583156405, "grad_norm": 0.2380386156268189, "learning_rate": 2.599685204616999e-05, "loss": 0.4725, "step": 1128 }, { "epoch": 1.5973106864826612, "grad_norm": 0.2861736326134302, "learning_rate": 2.5970619097586568e-05, "loss": 0.4417, "step": 1129 }, { "epoch": 1.5987261146496814, "grad_norm": 0.26368405652585286, "learning_rate": 2.5944386149003148e-05, "loss": 0.4391, "step": 1130 }, { "epoch": 1.600141542816702, "grad_norm": 0.2276153908320552, "learning_rate": 2.5918153200419725e-05, "loss": 0.4493, "step": 1131 }, { "epoch": 1.6015569709837227, "grad_norm": 0.22999463566162967, "learning_rate": 2.589192025183631e-05, "loss": 0.5438, "step": 1132 }, { "epoch": 1.602972399150743, "grad_norm": 0.2809859072713602, "learning_rate": 2.586568730325289e-05, "loss": 0.5141, "step": 1133 }, { "epoch": 1.6043878273177636, "grad_norm": 0.2396152170770102, "learning_rate": 2.583945435466947e-05, "loss": 0.5947, "step": 1134 }, { "epoch": 1.6058032554847843, "grad_norm": 0.2931573292499133, "learning_rate": 2.5813221406086046e-05, "loss": 0.4498, "step": 1135 }, { "epoch": 1.6072186836518045, "grad_norm": 0.2492431057897274, "learning_rate": 2.5786988457502627e-05, "loss": 0.4594, "step": 1136 }, { "epoch": 1.6086341118188252, "grad_norm": 0.22849622509328704, "learning_rate": 2.5760755508919203e-05, "loss": 0.4161, "step": 1137 }, { "epoch": 1.6100495399858459, "grad_norm": 0.2276611441960024, "learning_rate": 2.5734522560335784e-05, "loss": 0.4717, "step": 1138 }, { "epoch": 1.611464968152866, "grad_norm": 0.24910642514475684, "learning_rate": 2.570828961175236e-05, "loss": 0.5203, "step": 1139 }, { "epoch": 1.6128803963198868, "grad_norm": 0.2607001585160038, "learning_rate": 2.568205666316894e-05, "loss": 0.5007, "step": 1140 }, { "epoch": 1.6142958244869074, "grad_norm": 0.22409171708653328, "learning_rate": 2.565582371458552e-05, "loss": 0.4343, "step": 1141 }, { "epoch": 1.6157112526539277, "grad_norm": 0.22361825047458853, "learning_rate": 2.5629590766002098e-05, "loss": 0.518, "step": 1142 }, { "epoch": 1.6171266808209483, "grad_norm": 0.23193326257443647, "learning_rate": 2.560335781741868e-05, "loss": 0.4865, "step": 1143 }, { "epoch": 1.618542108987969, "grad_norm": 0.23639779385117354, "learning_rate": 2.5577124868835255e-05, "loss": 0.4892, "step": 1144 }, { "epoch": 1.6199575371549892, "grad_norm": 0.2439972587338885, "learning_rate": 2.5550891920251836e-05, "loss": 0.5131, "step": 1145 }, { "epoch": 1.62137296532201, "grad_norm": 0.23805120948621156, "learning_rate": 2.552465897166842e-05, "loss": 0.5121, "step": 1146 }, { "epoch": 1.6227883934890306, "grad_norm": 0.21467323810324876, "learning_rate": 2.5498426023085e-05, "loss": 0.4598, "step": 1147 }, { "epoch": 1.6242038216560508, "grad_norm": 0.22227022281283504, "learning_rate": 2.5472193074501576e-05, "loss": 0.4353, "step": 1148 }, { "epoch": 1.6256192498230715, "grad_norm": 0.22915765706021562, "learning_rate": 2.5445960125918157e-05, "loss": 0.5105, "step": 1149 }, { "epoch": 1.6270346779900922, "grad_norm": 0.25397114301762364, "learning_rate": 2.5419727177334734e-05, "loss": 0.4626, "step": 1150 }, { "epoch": 1.6284501061571124, "grad_norm": 0.22184273484318087, "learning_rate": 2.5393494228751314e-05, "loss": 0.4462, "step": 1151 }, { "epoch": 1.629865534324133, "grad_norm": 0.2089350680438182, "learning_rate": 2.536726128016789e-05, "loss": 0.3759, "step": 1152 }, { "epoch": 1.6312809624911537, "grad_norm": 0.24870576324892082, "learning_rate": 2.534102833158447e-05, "loss": 0.4802, "step": 1153 }, { "epoch": 1.632696390658174, "grad_norm": 0.21126461355901588, "learning_rate": 2.531479538300105e-05, "loss": 0.4249, "step": 1154 }, { "epoch": 1.6341118188251946, "grad_norm": 0.2486402703677725, "learning_rate": 2.5288562434417628e-05, "loss": 0.4362, "step": 1155 }, { "epoch": 1.6355272469922153, "grad_norm": 0.2152020707436181, "learning_rate": 2.526232948583421e-05, "loss": 0.4393, "step": 1156 }, { "epoch": 1.6369426751592355, "grad_norm": 0.22050316835622147, "learning_rate": 2.5236096537250785e-05, "loss": 0.4635, "step": 1157 }, { "epoch": 1.6383581033262562, "grad_norm": 0.23009695935320004, "learning_rate": 2.5209863588667366e-05, "loss": 0.4453, "step": 1158 }, { "epoch": 1.6397735314932769, "grad_norm": 0.25731684547250727, "learning_rate": 2.5183630640083943e-05, "loss": 0.4021, "step": 1159 }, { "epoch": 1.641188959660297, "grad_norm": 0.2206489665591597, "learning_rate": 2.515739769150053e-05, "loss": 0.5104, "step": 1160 }, { "epoch": 1.6426043878273178, "grad_norm": 0.252288856938944, "learning_rate": 2.5131164742917107e-05, "loss": 0.4646, "step": 1161 }, { "epoch": 1.6440198159943384, "grad_norm": 0.25452492357715273, "learning_rate": 2.5104931794333687e-05, "loss": 0.449, "step": 1162 }, { "epoch": 1.6454352441613587, "grad_norm": 0.24120366995017042, "learning_rate": 2.5078698845750264e-05, "loss": 0.4649, "step": 1163 }, { "epoch": 1.6468506723283793, "grad_norm": 0.2279786516010407, "learning_rate": 2.5052465897166844e-05, "loss": 0.4609, "step": 1164 }, { "epoch": 1.6482661004954, "grad_norm": 0.19460609034904436, "learning_rate": 2.502623294858342e-05, "loss": 0.4021, "step": 1165 }, { "epoch": 1.6496815286624202, "grad_norm": 0.23745682422534709, "learning_rate": 2.5e-05, "loss": 0.4729, "step": 1166 }, { "epoch": 1.651096956829441, "grad_norm": 0.22632870172314187, "learning_rate": 2.497376705141658e-05, "loss": 0.4038, "step": 1167 }, { "epoch": 1.6525123849964616, "grad_norm": 0.21060671269418166, "learning_rate": 2.494753410283316e-05, "loss": 0.4152, "step": 1168 }, { "epoch": 1.6539278131634818, "grad_norm": 0.2144278010114158, "learning_rate": 2.492130115424974e-05, "loss": 0.491, "step": 1169 }, { "epoch": 1.6553432413305025, "grad_norm": 0.23930362218821835, "learning_rate": 2.489506820566632e-05, "loss": 0.5087, "step": 1170 }, { "epoch": 1.6567586694975232, "grad_norm": 0.23480237512635063, "learning_rate": 2.48688352570829e-05, "loss": 0.5187, "step": 1171 }, { "epoch": 1.6581740976645434, "grad_norm": 0.24535990295367607, "learning_rate": 2.4842602308499476e-05, "loss": 0.4854, "step": 1172 }, { "epoch": 1.659589525831564, "grad_norm": 0.21706667220912929, "learning_rate": 2.4816369359916056e-05, "loss": 0.4305, "step": 1173 }, { "epoch": 1.6610049539985847, "grad_norm": 0.21570842708240046, "learning_rate": 2.4790136411332633e-05, "loss": 0.4588, "step": 1174 }, { "epoch": 1.662420382165605, "grad_norm": 0.24620836293827908, "learning_rate": 2.4763903462749214e-05, "loss": 0.5459, "step": 1175 }, { "epoch": 1.6638358103326256, "grad_norm": 0.23227318833395602, "learning_rate": 2.473767051416579e-05, "loss": 0.4533, "step": 1176 }, { "epoch": 1.6652512384996463, "grad_norm": 0.2420577256811438, "learning_rate": 2.4711437565582374e-05, "loss": 0.5202, "step": 1177 }, { "epoch": 1.6666666666666665, "grad_norm": 0.24918517663693857, "learning_rate": 2.468520461699895e-05, "loss": 0.4424, "step": 1178 }, { "epoch": 1.6680820948336872, "grad_norm": 0.22057124481546828, "learning_rate": 2.465897166841553e-05, "loss": 0.4245, "step": 1179 }, { "epoch": 1.6694975230007079, "grad_norm": 0.2399440923022933, "learning_rate": 2.463273871983211e-05, "loss": 0.4362, "step": 1180 }, { "epoch": 1.670912951167728, "grad_norm": 4.334965775067043, "learning_rate": 2.460650577124869e-05, "loss": 0.4193, "step": 1181 }, { "epoch": 1.6723283793347488, "grad_norm": 0.2672420532047751, "learning_rate": 2.458027282266527e-05, "loss": 0.4602, "step": 1182 }, { "epoch": 1.6737438075017694, "grad_norm": 0.45886783224845246, "learning_rate": 2.4554039874081846e-05, "loss": 0.4929, "step": 1183 }, { "epoch": 1.6751592356687897, "grad_norm": 0.23253089832864093, "learning_rate": 2.452780692549843e-05, "loss": 0.5029, "step": 1184 }, { "epoch": 1.6765746638358103, "grad_norm": 0.2754881453715846, "learning_rate": 2.4501573976915006e-05, "loss": 0.4858, "step": 1185 }, { "epoch": 1.677990092002831, "grad_norm": 0.24718432903788826, "learning_rate": 2.4475341028331587e-05, "loss": 0.4406, "step": 1186 }, { "epoch": 1.6794055201698512, "grad_norm": 0.24486533742290445, "learning_rate": 2.4449108079748163e-05, "loss": 0.4639, "step": 1187 }, { "epoch": 1.680820948336872, "grad_norm": 0.27303563278355925, "learning_rate": 2.4422875131164744e-05, "loss": 0.4587, "step": 1188 }, { "epoch": 1.6822363765038926, "grad_norm": 0.23682103443569977, "learning_rate": 2.4396642182581324e-05, "loss": 0.4478, "step": 1189 }, { "epoch": 1.6836518046709128, "grad_norm": 0.22594744221435362, "learning_rate": 2.43704092339979e-05, "loss": 0.4854, "step": 1190 }, { "epoch": 1.6850672328379335, "grad_norm": 0.2734834603871489, "learning_rate": 2.434417628541448e-05, "loss": 0.4582, "step": 1191 }, { "epoch": 1.6864826610049541, "grad_norm": 0.21788822338250274, "learning_rate": 2.431794333683106e-05, "loss": 0.3986, "step": 1192 }, { "epoch": 1.6878980891719744, "grad_norm": 0.238195100497809, "learning_rate": 2.4291710388247642e-05, "loss": 0.4712, "step": 1193 }, { "epoch": 1.689313517338995, "grad_norm": 0.2787782727643181, "learning_rate": 2.426547743966422e-05, "loss": 0.4726, "step": 1194 }, { "epoch": 1.6907289455060157, "grad_norm": 0.4416723084952574, "learning_rate": 2.42392444910808e-05, "loss": 0.4401, "step": 1195 }, { "epoch": 1.692144373673036, "grad_norm": 0.2419495654272071, "learning_rate": 2.4213011542497376e-05, "loss": 0.4695, "step": 1196 }, { "epoch": 1.6935598018400566, "grad_norm": 0.26693116473266787, "learning_rate": 2.4186778593913956e-05, "loss": 0.4934, "step": 1197 }, { "epoch": 1.6949752300070773, "grad_norm": 0.22799379548924234, "learning_rate": 2.4160545645330536e-05, "loss": 0.4356, "step": 1198 }, { "epoch": 1.6963906581740975, "grad_norm": 0.9900211172083577, "learning_rate": 2.4134312696747117e-05, "loss": 0.4798, "step": 1199 }, { "epoch": 1.6978060863411182, "grad_norm": 0.23148916684473117, "learning_rate": 2.4108079748163694e-05, "loss": 0.4515, "step": 1200 }, { "epoch": 1.6992215145081389, "grad_norm": 0.22874042461258706, "learning_rate": 2.4081846799580274e-05, "loss": 0.4527, "step": 1201 }, { "epoch": 1.700636942675159, "grad_norm": 0.24211185463412316, "learning_rate": 2.4055613850996854e-05, "loss": 0.445, "step": 1202 }, { "epoch": 1.7020523708421798, "grad_norm": 0.30431255476741226, "learning_rate": 2.402938090241343e-05, "loss": 0.5107, "step": 1203 }, { "epoch": 1.7034677990092004, "grad_norm": 0.22113554817941933, "learning_rate": 2.400314795383001e-05, "loss": 0.45, "step": 1204 }, { "epoch": 1.7048832271762207, "grad_norm": 0.24032698117750861, "learning_rate": 2.397691500524659e-05, "loss": 0.5043, "step": 1205 }, { "epoch": 1.7062986553432413, "grad_norm": 0.23870364857077842, "learning_rate": 2.3950682056663172e-05, "loss": 0.5209, "step": 1206 }, { "epoch": 1.707714083510262, "grad_norm": 0.2302688871180308, "learning_rate": 2.392444910807975e-05, "loss": 0.4242, "step": 1207 }, { "epoch": 1.7091295116772822, "grad_norm": 0.2436155876922659, "learning_rate": 2.389821615949633e-05, "loss": 0.4704, "step": 1208 }, { "epoch": 1.710544939844303, "grad_norm": 0.4273326272781385, "learning_rate": 2.3871983210912906e-05, "loss": 0.4401, "step": 1209 }, { "epoch": 1.7119603680113236, "grad_norm": 0.268965312167453, "learning_rate": 2.3845750262329486e-05, "loss": 0.4998, "step": 1210 }, { "epoch": 1.7133757961783438, "grad_norm": 0.21968584742302194, "learning_rate": 2.3819517313746067e-05, "loss": 0.4093, "step": 1211 }, { "epoch": 1.7147912243453645, "grad_norm": 0.23577769866318082, "learning_rate": 2.3793284365162647e-05, "loss": 0.4447, "step": 1212 }, { "epoch": 1.7162066525123851, "grad_norm": 0.22990195035734406, "learning_rate": 2.3767051416579224e-05, "loss": 0.5292, "step": 1213 }, { "epoch": 1.7176220806794054, "grad_norm": 0.21474090925988665, "learning_rate": 2.3740818467995804e-05, "loss": 0.503, "step": 1214 }, { "epoch": 1.719037508846426, "grad_norm": 0.24605939907412877, "learning_rate": 2.3714585519412384e-05, "loss": 0.4326, "step": 1215 }, { "epoch": 1.7204529370134467, "grad_norm": 0.24308945674870652, "learning_rate": 2.368835257082896e-05, "loss": 0.3943, "step": 1216 }, { "epoch": 1.721868365180467, "grad_norm": 0.21001902737119993, "learning_rate": 2.366211962224554e-05, "loss": 0.5185, "step": 1217 }, { "epoch": 1.7232837933474876, "grad_norm": 0.22899520312784166, "learning_rate": 2.3635886673662122e-05, "loss": 0.4642, "step": 1218 }, { "epoch": 1.7246992215145083, "grad_norm": 0.23724020514959368, "learning_rate": 2.3609653725078702e-05, "loss": 0.4431, "step": 1219 }, { "epoch": 1.7261146496815285, "grad_norm": 0.2263807462898294, "learning_rate": 2.358342077649528e-05, "loss": 0.5352, "step": 1220 }, { "epoch": 1.7275300778485492, "grad_norm": 0.2230790085454061, "learning_rate": 2.355718782791186e-05, "loss": 0.481, "step": 1221 }, { "epoch": 1.7289455060155698, "grad_norm": 0.23095257603790728, "learning_rate": 2.3530954879328436e-05, "loss": 0.4367, "step": 1222 }, { "epoch": 1.73036093418259, "grad_norm": 0.22096408952247884, "learning_rate": 2.3504721930745016e-05, "loss": 0.4688, "step": 1223 }, { "epoch": 1.7317763623496107, "grad_norm": 0.21323951749813636, "learning_rate": 2.3478488982161593e-05, "loss": 0.4884, "step": 1224 }, { "epoch": 1.7331917905166314, "grad_norm": 0.21045169434522537, "learning_rate": 2.3452256033578177e-05, "loss": 0.4512, "step": 1225 }, { "epoch": 1.7346072186836516, "grad_norm": 0.22474956769797635, "learning_rate": 2.3426023084994754e-05, "loss": 0.531, "step": 1226 }, { "epoch": 1.7360226468506723, "grad_norm": 0.24318820018692644, "learning_rate": 2.3399790136411334e-05, "loss": 0.4697, "step": 1227 }, { "epoch": 1.737438075017693, "grad_norm": 0.22215833754925915, "learning_rate": 2.3373557187827915e-05, "loss": 0.4926, "step": 1228 }, { "epoch": 1.7388535031847132, "grad_norm": 0.21453240070070215, "learning_rate": 2.334732423924449e-05, "loss": 0.4158, "step": 1229 }, { "epoch": 1.7402689313517339, "grad_norm": 0.2586416867995017, "learning_rate": 2.332109129066107e-05, "loss": 0.4979, "step": 1230 }, { "epoch": 1.7416843595187546, "grad_norm": 0.24659631712596117, "learning_rate": 2.329485834207765e-05, "loss": 0.5521, "step": 1231 }, { "epoch": 1.7430997876857748, "grad_norm": 0.32417069581572344, "learning_rate": 2.3268625393494232e-05, "loss": 0.4195, "step": 1232 }, { "epoch": 1.7445152158527955, "grad_norm": 0.23039000450013308, "learning_rate": 2.324239244491081e-05, "loss": 0.5154, "step": 1233 }, { "epoch": 1.7459306440198161, "grad_norm": 0.23344012631501052, "learning_rate": 2.321615949632739e-05, "loss": 0.4377, "step": 1234 }, { "epoch": 1.7473460721868364, "grad_norm": 0.21715813767080286, "learning_rate": 2.3189926547743966e-05, "loss": 0.4044, "step": 1235 }, { "epoch": 1.748761500353857, "grad_norm": 0.2725802380164536, "learning_rate": 2.3163693599160547e-05, "loss": 0.4771, "step": 1236 }, { "epoch": 1.7501769285208777, "grad_norm": 0.23461308933319772, "learning_rate": 2.3137460650577124e-05, "loss": 0.5712, "step": 1237 }, { "epoch": 1.7515923566878981, "grad_norm": 0.21422574478415968, "learning_rate": 2.3111227701993704e-05, "loss": 0.4413, "step": 1238 }, { "epoch": 1.7530077848549186, "grad_norm": 0.27760237803106597, "learning_rate": 2.3084994753410284e-05, "loss": 0.4901, "step": 1239 }, { "epoch": 1.7544232130219393, "grad_norm": 4.188952617257135, "learning_rate": 2.3058761804826864e-05, "loss": 0.5375, "step": 1240 }, { "epoch": 1.7558386411889597, "grad_norm": 0.2480906746200484, "learning_rate": 2.3032528856243445e-05, "loss": 0.483, "step": 1241 }, { "epoch": 1.7572540693559802, "grad_norm": 0.22531099081716, "learning_rate": 2.300629590766002e-05, "loss": 0.5066, "step": 1242 }, { "epoch": 1.7586694975230008, "grad_norm": 0.21940727426318465, "learning_rate": 2.2980062959076602e-05, "loss": 0.4087, "step": 1243 }, { "epoch": 1.7600849256900213, "grad_norm": 0.19554804756344746, "learning_rate": 2.295383001049318e-05, "loss": 0.3825, "step": 1244 }, { "epoch": 1.7615003538570417, "grad_norm": 0.22562606588810402, "learning_rate": 2.292759706190976e-05, "loss": 0.4439, "step": 1245 }, { "epoch": 1.7629157820240624, "grad_norm": 0.2271682488662497, "learning_rate": 2.290136411332634e-05, "loss": 0.4196, "step": 1246 }, { "epoch": 1.7643312101910829, "grad_norm": 0.21455820735860226, "learning_rate": 2.287513116474292e-05, "loss": 0.4322, "step": 1247 }, { "epoch": 1.7657466383581033, "grad_norm": 0.22651953039208592, "learning_rate": 2.2848898216159496e-05, "loss": 0.4682, "step": 1248 }, { "epoch": 1.767162066525124, "grad_norm": 0.24921888820516835, "learning_rate": 2.2822665267576077e-05, "loss": 0.4786, "step": 1249 }, { "epoch": 1.7685774946921444, "grad_norm": 0.2351290162709047, "learning_rate": 2.2796432318992657e-05, "loss": 0.4688, "step": 1250 }, { "epoch": 1.7699929228591649, "grad_norm": 0.22097668014081087, "learning_rate": 2.2770199370409234e-05, "loss": 0.4237, "step": 1251 }, { "epoch": 1.7714083510261855, "grad_norm": 0.24303531282271149, "learning_rate": 2.2743966421825814e-05, "loss": 0.4993, "step": 1252 }, { "epoch": 1.772823779193206, "grad_norm": 0.999217635929671, "learning_rate": 2.2717733473242395e-05, "loss": 0.4566, "step": 1253 }, { "epoch": 1.7742392073602264, "grad_norm": 0.23661072479165118, "learning_rate": 2.2691500524658975e-05, "loss": 0.4344, "step": 1254 }, { "epoch": 1.7756546355272471, "grad_norm": 0.2238140247175545, "learning_rate": 2.2665267576075552e-05, "loss": 0.4854, "step": 1255 }, { "epoch": 1.7770700636942676, "grad_norm": 0.23846918971265546, "learning_rate": 2.2639034627492132e-05, "loss": 0.4918, "step": 1256 }, { "epoch": 1.778485491861288, "grad_norm": 0.27010281272058134, "learning_rate": 2.261280167890871e-05, "loss": 0.4618, "step": 1257 }, { "epoch": 1.7799009200283087, "grad_norm": 0.2402855254748168, "learning_rate": 2.258656873032529e-05, "loss": 0.4816, "step": 1258 }, { "epoch": 1.7813163481953291, "grad_norm": 0.284861020752327, "learning_rate": 2.2560335781741866e-05, "loss": 0.4537, "step": 1259 }, { "epoch": 1.7827317763623496, "grad_norm": 0.23162852747166598, "learning_rate": 2.253410283315845e-05, "loss": 0.5254, "step": 1260 }, { "epoch": 1.7841472045293703, "grad_norm": 0.23010356275368837, "learning_rate": 2.2507869884575027e-05, "loss": 0.4629, "step": 1261 }, { "epoch": 1.7855626326963907, "grad_norm": 0.26477440200537117, "learning_rate": 2.2481636935991607e-05, "loss": 0.431, "step": 1262 }, { "epoch": 1.7869780608634112, "grad_norm": 0.2469472463475338, "learning_rate": 2.2455403987408187e-05, "loss": 0.5095, "step": 1263 }, { "epoch": 1.7883934890304318, "grad_norm": 0.250854568912954, "learning_rate": 2.2429171038824764e-05, "loss": 0.4695, "step": 1264 }, { "epoch": 1.7898089171974523, "grad_norm": 0.2358553336427738, "learning_rate": 2.2402938090241344e-05, "loss": 0.4718, "step": 1265 }, { "epoch": 1.7912243453644727, "grad_norm": 0.26092941743814313, "learning_rate": 2.237670514165792e-05, "loss": 0.4844, "step": 1266 }, { "epoch": 1.7926397735314934, "grad_norm": 0.25364777428866075, "learning_rate": 2.2350472193074505e-05, "loss": 0.5053, "step": 1267 }, { "epoch": 1.7940552016985138, "grad_norm": 0.22195077907811386, "learning_rate": 2.2324239244491082e-05, "loss": 0.4421, "step": 1268 }, { "epoch": 1.7954706298655343, "grad_norm": 1.4410343829585255, "learning_rate": 2.2298006295907662e-05, "loss": 0.4146, "step": 1269 }, { "epoch": 1.796886058032555, "grad_norm": 0.25129972037159265, "learning_rate": 2.227177334732424e-05, "loss": 0.483, "step": 1270 }, { "epoch": 1.7983014861995754, "grad_norm": 0.26142130013103143, "learning_rate": 2.224554039874082e-05, "loss": 0.4594, "step": 1271 }, { "epoch": 1.7997169143665959, "grad_norm": 0.23266793598351823, "learning_rate": 2.2219307450157396e-05, "loss": 0.491, "step": 1272 }, { "epoch": 1.8011323425336165, "grad_norm": 0.2303010389050869, "learning_rate": 2.2193074501573977e-05, "loss": 0.4874, "step": 1273 }, { "epoch": 1.802547770700637, "grad_norm": 0.23118206006279696, "learning_rate": 2.2166841552990557e-05, "loss": 0.3986, "step": 1274 }, { "epoch": 1.8039631988676574, "grad_norm": 0.2143231238326485, "learning_rate": 2.2140608604407137e-05, "loss": 0.3811, "step": 1275 }, { "epoch": 1.805378627034678, "grad_norm": 0.22724689078735588, "learning_rate": 2.2114375655823717e-05, "loss": 0.4208, "step": 1276 }, { "epoch": 1.8067940552016986, "grad_norm": 2.0654374438327734, "learning_rate": 2.2088142707240294e-05, "loss": 0.5874, "step": 1277 }, { "epoch": 1.808209483368719, "grad_norm": 0.2409280392188967, "learning_rate": 2.2061909758656875e-05, "loss": 0.5285, "step": 1278 }, { "epoch": 1.8096249115357397, "grad_norm": 0.23334400572166863, "learning_rate": 2.203567681007345e-05, "loss": 0.4063, "step": 1279 }, { "epoch": 1.8110403397027601, "grad_norm": 0.22984388021071223, "learning_rate": 2.2009443861490035e-05, "loss": 0.4671, "step": 1280 }, { "epoch": 1.8124557678697806, "grad_norm": 0.2301151153313693, "learning_rate": 2.1983210912906612e-05, "loss": 0.5002, "step": 1281 }, { "epoch": 1.8138711960368012, "grad_norm": 0.24246653508800442, "learning_rate": 2.1956977964323192e-05, "loss": 0.4151, "step": 1282 }, { "epoch": 1.8152866242038217, "grad_norm": 0.2525452240180046, "learning_rate": 2.193074501573977e-05, "loss": 0.4274, "step": 1283 }, { "epoch": 1.8167020523708421, "grad_norm": 0.20804041175775848, "learning_rate": 2.190451206715635e-05, "loss": 0.4412, "step": 1284 }, { "epoch": 1.8181174805378628, "grad_norm": 0.23209003341106552, "learning_rate": 2.1878279118572926e-05, "loss": 0.4785, "step": 1285 }, { "epoch": 1.8195329087048833, "grad_norm": 0.21506235966409293, "learning_rate": 2.1852046169989507e-05, "loss": 0.4901, "step": 1286 }, { "epoch": 1.8209483368719037, "grad_norm": 0.24201084469050718, "learning_rate": 2.1825813221406087e-05, "loss": 0.4653, "step": 1287 }, { "epoch": 1.8223637650389244, "grad_norm": 0.23237298583987862, "learning_rate": 2.1799580272822667e-05, "loss": 0.4521, "step": 1288 }, { "epoch": 1.8237791932059448, "grad_norm": 0.23184894299524061, "learning_rate": 2.1773347324239248e-05, "loss": 0.4141, "step": 1289 }, { "epoch": 1.8251946213729653, "grad_norm": 0.21915908624117594, "learning_rate": 2.1747114375655824e-05, "loss": 0.4833, "step": 1290 }, { "epoch": 1.826610049539986, "grad_norm": 0.2204582698480696, "learning_rate": 2.1720881427072405e-05, "loss": 0.4646, "step": 1291 }, { "epoch": 1.8280254777070064, "grad_norm": 0.23857430239630453, "learning_rate": 2.169464847848898e-05, "loss": 0.4334, "step": 1292 }, { "epoch": 1.8294409058740269, "grad_norm": 0.2626478642231881, "learning_rate": 2.1668415529905562e-05, "loss": 0.4823, "step": 1293 }, { "epoch": 1.8308563340410475, "grad_norm": 0.22467020564400583, "learning_rate": 2.1642182581322142e-05, "loss": 0.5265, "step": 1294 }, { "epoch": 1.832271762208068, "grad_norm": 0.24570130352348415, "learning_rate": 2.1615949632738722e-05, "loss": 0.4235, "step": 1295 }, { "epoch": 1.8336871903750884, "grad_norm": 0.23356731544612377, "learning_rate": 2.15897166841553e-05, "loss": 0.503, "step": 1296 }, { "epoch": 1.835102618542109, "grad_norm": 0.25095911275324184, "learning_rate": 2.156348373557188e-05, "loss": 0.5299, "step": 1297 }, { "epoch": 1.8365180467091295, "grad_norm": 0.31733420723462563, "learning_rate": 2.1537250786988457e-05, "loss": 0.5588, "step": 1298 }, { "epoch": 1.83793347487615, "grad_norm": 0.23939961643806795, "learning_rate": 2.1511017838405037e-05, "loss": 0.4863, "step": 1299 }, { "epoch": 1.8393489030431707, "grad_norm": 0.2229451447526651, "learning_rate": 2.1484784889821617e-05, "loss": 0.4864, "step": 1300 }, { "epoch": 1.8407643312101911, "grad_norm": 0.28005231708013956, "learning_rate": 2.1458551941238197e-05, "loss": 0.4793, "step": 1301 }, { "epoch": 1.8421797593772116, "grad_norm": 0.2416939594691491, "learning_rate": 2.1432318992654778e-05, "loss": 0.4641, "step": 1302 }, { "epoch": 1.8435951875442322, "grad_norm": 0.22685599373710225, "learning_rate": 2.1406086044071355e-05, "loss": 0.4525, "step": 1303 }, { "epoch": 1.8450106157112527, "grad_norm": 0.22150306708586515, "learning_rate": 2.1379853095487935e-05, "loss": 0.4576, "step": 1304 }, { "epoch": 1.8464260438782731, "grad_norm": 0.23965364622580573, "learning_rate": 2.1353620146904512e-05, "loss": 0.4441, "step": 1305 }, { "epoch": 1.8478414720452938, "grad_norm": 0.24247715310259696, "learning_rate": 2.1327387198321092e-05, "loss": 0.4975, "step": 1306 }, { "epoch": 1.8492569002123143, "grad_norm": 0.22511524026533378, "learning_rate": 2.130115424973767e-05, "loss": 0.5094, "step": 1307 }, { "epoch": 1.8506723283793347, "grad_norm": 0.24407240925469395, "learning_rate": 2.1274921301154253e-05, "loss": 0.4551, "step": 1308 }, { "epoch": 1.8520877565463554, "grad_norm": 0.24306659852597814, "learning_rate": 2.124868835257083e-05, "loss": 0.4386, "step": 1309 }, { "epoch": 1.8535031847133758, "grad_norm": 0.2241048925204749, "learning_rate": 2.122245540398741e-05, "loss": 0.471, "step": 1310 }, { "epoch": 1.8549186128803963, "grad_norm": 0.22335342460079466, "learning_rate": 2.119622245540399e-05, "loss": 0.4297, "step": 1311 }, { "epoch": 1.856334041047417, "grad_norm": 0.22466883866462156, "learning_rate": 2.1169989506820567e-05, "loss": 0.4116, "step": 1312 }, { "epoch": 1.8577494692144374, "grad_norm": 0.22804012366044188, "learning_rate": 2.1143756558237147e-05, "loss": 0.4921, "step": 1313 }, { "epoch": 1.8591648973814578, "grad_norm": 0.2513473537546202, "learning_rate": 2.1117523609653724e-05, "loss": 0.4047, "step": 1314 }, { "epoch": 1.8605803255484785, "grad_norm": 0.2084112605966311, "learning_rate": 2.1091290661070308e-05, "loss": 0.4821, "step": 1315 }, { "epoch": 1.861995753715499, "grad_norm": 0.23646819837085292, "learning_rate": 2.1065057712486885e-05, "loss": 0.3978, "step": 1316 }, { "epoch": 1.8634111818825194, "grad_norm": 0.22261204824879074, "learning_rate": 2.1038824763903465e-05, "loss": 0.4947, "step": 1317 }, { "epoch": 1.86482661004954, "grad_norm": 0.2437405611108718, "learning_rate": 2.1012591815320042e-05, "loss": 0.439, "step": 1318 }, { "epoch": 1.8662420382165605, "grad_norm": 0.23344188200332183, "learning_rate": 2.0986358866736622e-05, "loss": 0.4871, "step": 1319 }, { "epoch": 1.867657466383581, "grad_norm": 0.23982921750771513, "learning_rate": 2.09601259181532e-05, "loss": 0.482, "step": 1320 }, { "epoch": 1.8690728945506017, "grad_norm": 0.23790242563138025, "learning_rate": 2.093389296956978e-05, "loss": 0.4202, "step": 1321 }, { "epoch": 1.870488322717622, "grad_norm": 0.23424229708006536, "learning_rate": 2.090766002098636e-05, "loss": 0.4865, "step": 1322 }, { "epoch": 1.8719037508846426, "grad_norm": 0.22239898962936439, "learning_rate": 2.088142707240294e-05, "loss": 0.4049, "step": 1323 }, { "epoch": 1.8733191790516632, "grad_norm": 0.19757565268413543, "learning_rate": 2.085519412381952e-05, "loss": 0.4318, "step": 1324 }, { "epoch": 1.8747346072186837, "grad_norm": 0.24461105569164185, "learning_rate": 2.0828961175236097e-05, "loss": 0.4631, "step": 1325 }, { "epoch": 1.8761500353857041, "grad_norm": 0.5947746440626941, "learning_rate": 2.0802728226652677e-05, "loss": 0.4549, "step": 1326 }, { "epoch": 1.8775654635527248, "grad_norm": 0.22133962167412313, "learning_rate": 2.0776495278069254e-05, "loss": 0.5018, "step": 1327 }, { "epoch": 1.8789808917197452, "grad_norm": 0.2233899656516572, "learning_rate": 2.0750262329485835e-05, "loss": 0.4883, "step": 1328 }, { "epoch": 1.8803963198867657, "grad_norm": 0.2444529353766216, "learning_rate": 2.0724029380902415e-05, "loss": 0.5194, "step": 1329 }, { "epoch": 1.8818117480537864, "grad_norm": 0.25965188711877873, "learning_rate": 2.0697796432318995e-05, "loss": 0.4585, "step": 1330 }, { "epoch": 1.8832271762208068, "grad_norm": 0.21130420022445334, "learning_rate": 2.0671563483735572e-05, "loss": 0.5233, "step": 1331 }, { "epoch": 1.8846426043878273, "grad_norm": 0.24423899208067895, "learning_rate": 2.0645330535152152e-05, "loss": 0.5032, "step": 1332 }, { "epoch": 1.886058032554848, "grad_norm": 0.23473966436217122, "learning_rate": 2.061909758656873e-05, "loss": 0.444, "step": 1333 }, { "epoch": 1.8874734607218684, "grad_norm": 0.23024212934980773, "learning_rate": 2.059286463798531e-05, "loss": 0.4864, "step": 1334 }, { "epoch": 1.8888888888888888, "grad_norm": 0.23297536760980386, "learning_rate": 2.056663168940189e-05, "loss": 0.3823, "step": 1335 }, { "epoch": 1.8903043170559095, "grad_norm": 0.2381046335808105, "learning_rate": 2.054039874081847e-05, "loss": 0.4893, "step": 1336 }, { "epoch": 1.89171974522293, "grad_norm": 0.24897499927700323, "learning_rate": 2.051416579223505e-05, "loss": 0.5544, "step": 1337 }, { "epoch": 1.8931351733899504, "grad_norm": 0.25742996674526, "learning_rate": 2.0487932843651627e-05, "loss": 0.4581, "step": 1338 }, { "epoch": 1.894550601556971, "grad_norm": 0.22376579236901967, "learning_rate": 2.0461699895068208e-05, "loss": 0.4679, "step": 1339 }, { "epoch": 1.8959660297239915, "grad_norm": 0.22104742077146222, "learning_rate": 2.0435466946484784e-05, "loss": 0.5172, "step": 1340 }, { "epoch": 1.897381457891012, "grad_norm": 0.2800835936243694, "learning_rate": 2.0409233997901365e-05, "loss": 0.4541, "step": 1341 }, { "epoch": 1.8987968860580327, "grad_norm": 0.23180196456243227, "learning_rate": 2.038300104931794e-05, "loss": 0.4398, "step": 1342 }, { "epoch": 1.900212314225053, "grad_norm": 0.21488819834194242, "learning_rate": 2.0356768100734525e-05, "loss": 0.4557, "step": 1343 }, { "epoch": 1.9016277423920736, "grad_norm": 0.24372869965856336, "learning_rate": 2.0330535152151102e-05, "loss": 0.4794, "step": 1344 }, { "epoch": 1.9030431705590942, "grad_norm": 0.2366213579699343, "learning_rate": 2.0304302203567683e-05, "loss": 0.3981, "step": 1345 }, { "epoch": 1.9044585987261147, "grad_norm": 0.3648752073142994, "learning_rate": 2.027806925498426e-05, "loss": 0.534, "step": 1346 }, { "epoch": 1.9058740268931351, "grad_norm": 0.2197325729465737, "learning_rate": 2.025183630640084e-05, "loss": 0.3927, "step": 1347 }, { "epoch": 1.9072894550601558, "grad_norm": 0.231081310750387, "learning_rate": 2.022560335781742e-05, "loss": 0.4873, "step": 1348 }, { "epoch": 1.9087048832271762, "grad_norm": 0.24082762445235403, "learning_rate": 2.0199370409233997e-05, "loss": 0.4328, "step": 1349 }, { "epoch": 1.9101203113941967, "grad_norm": 0.20825514018587932, "learning_rate": 2.017313746065058e-05, "loss": 0.4976, "step": 1350 }, { "epoch": 1.9115357395612174, "grad_norm": 0.21894111983154066, "learning_rate": 2.0146904512067157e-05, "loss": 0.4424, "step": 1351 }, { "epoch": 1.9129511677282378, "grad_norm": 0.2286478651129217, "learning_rate": 2.0120671563483738e-05, "loss": 0.4343, "step": 1352 }, { "epoch": 1.9143665958952583, "grad_norm": 0.2204646971643326, "learning_rate": 2.0094438614900315e-05, "loss": 0.4452, "step": 1353 }, { "epoch": 1.915782024062279, "grad_norm": 0.23638163886103097, "learning_rate": 2.0068205666316895e-05, "loss": 0.482, "step": 1354 }, { "epoch": 1.9171974522292994, "grad_norm": 0.20878654613258196, "learning_rate": 2.0041972717733472e-05, "loss": 0.424, "step": 1355 }, { "epoch": 1.9186128803963198, "grad_norm": 0.2098932184443533, "learning_rate": 2.0015739769150055e-05, "loss": 0.4047, "step": 1356 }, { "epoch": 1.9200283085633405, "grad_norm": 0.20399735990780962, "learning_rate": 1.9989506820566632e-05, "loss": 0.4147, "step": 1357 }, { "epoch": 1.921443736730361, "grad_norm": 0.22943335349547167, "learning_rate": 1.9963273871983213e-05, "loss": 0.477, "step": 1358 }, { "epoch": 1.9228591648973814, "grad_norm": 0.2228970758057448, "learning_rate": 1.993704092339979e-05, "loss": 0.4254, "step": 1359 }, { "epoch": 1.924274593064402, "grad_norm": 0.2107497813441486, "learning_rate": 1.991080797481637e-05, "loss": 0.4092, "step": 1360 }, { "epoch": 1.9256900212314225, "grad_norm": 0.19967419194846606, "learning_rate": 1.988457502623295e-05, "loss": 0.4836, "step": 1361 }, { "epoch": 1.927105449398443, "grad_norm": 0.23592377920831498, "learning_rate": 1.9858342077649527e-05, "loss": 0.4821, "step": 1362 }, { "epoch": 1.9285208775654636, "grad_norm": 0.23650322697474802, "learning_rate": 1.983210912906611e-05, "loss": 0.4715, "step": 1363 }, { "epoch": 1.929936305732484, "grad_norm": 0.2185083326059402, "learning_rate": 1.9805876180482688e-05, "loss": 0.5157, "step": 1364 }, { "epoch": 1.9313517338995045, "grad_norm": 0.2296974244832122, "learning_rate": 1.9779643231899268e-05, "loss": 0.4627, "step": 1365 }, { "epoch": 1.9327671620665252, "grad_norm": 0.21126190043837798, "learning_rate": 1.9753410283315845e-05, "loss": 0.4603, "step": 1366 }, { "epoch": 1.9341825902335457, "grad_norm": 0.21423948829296396, "learning_rate": 1.9727177334732425e-05, "loss": 0.4744, "step": 1367 }, { "epoch": 1.9355980184005661, "grad_norm": 0.21810060321266686, "learning_rate": 1.9700944386149002e-05, "loss": 0.4797, "step": 1368 }, { "epoch": 1.9370134465675868, "grad_norm": 0.2828545367941199, "learning_rate": 1.9674711437565582e-05, "loss": 0.4612, "step": 1369 }, { "epoch": 1.9384288747346072, "grad_norm": 0.2165297001551884, "learning_rate": 1.9648478488982163e-05, "loss": 0.4409, "step": 1370 }, { "epoch": 1.9398443029016277, "grad_norm": 0.24431415598675452, "learning_rate": 1.9622245540398743e-05, "loss": 0.4572, "step": 1371 }, { "epoch": 1.9412597310686484, "grad_norm": 0.4086000321654116, "learning_rate": 1.9596012591815323e-05, "loss": 0.4953, "step": 1372 }, { "epoch": 1.9426751592356688, "grad_norm": 0.22847856039481462, "learning_rate": 1.95697796432319e-05, "loss": 0.3886, "step": 1373 }, { "epoch": 1.9440905874026893, "grad_norm": 0.20902762981225703, "learning_rate": 1.954354669464848e-05, "loss": 0.4709, "step": 1374 }, { "epoch": 1.94550601556971, "grad_norm": 0.2424786257148823, "learning_rate": 1.9517313746065057e-05, "loss": 0.5191, "step": 1375 }, { "epoch": 1.9469214437367304, "grad_norm": 0.2159429193506879, "learning_rate": 1.9491080797481637e-05, "loss": 0.4453, "step": 1376 }, { "epoch": 1.9483368719037508, "grad_norm": 0.23477092809891625, "learning_rate": 1.9464847848898218e-05, "loss": 0.4891, "step": 1377 }, { "epoch": 1.9497523000707715, "grad_norm": 0.22425815180240935, "learning_rate": 1.9438614900314798e-05, "loss": 0.4888, "step": 1378 }, { "epoch": 1.951167728237792, "grad_norm": 0.45474270956547114, "learning_rate": 1.9412381951731375e-05, "loss": 0.5187, "step": 1379 }, { "epoch": 1.9525831564048124, "grad_norm": 0.2538581966186155, "learning_rate": 1.9386149003147955e-05, "loss": 0.4714, "step": 1380 }, { "epoch": 1.953998584571833, "grad_norm": 0.22480122774302955, "learning_rate": 1.9359916054564532e-05, "loss": 0.4594, "step": 1381 }, { "epoch": 1.9554140127388535, "grad_norm": 0.22940865698452123, "learning_rate": 1.9333683105981112e-05, "loss": 0.4497, "step": 1382 }, { "epoch": 1.956829440905874, "grad_norm": 0.2297486104164606, "learning_rate": 1.9307450157397693e-05, "loss": 0.4814, "step": 1383 }, { "epoch": 1.9582448690728946, "grad_norm": 0.2374799661012413, "learning_rate": 1.9281217208814273e-05, "loss": 0.4425, "step": 1384 }, { "epoch": 1.959660297239915, "grad_norm": 0.2517317396047779, "learning_rate": 1.9254984260230853e-05, "loss": 0.4487, "step": 1385 }, { "epoch": 1.9610757254069355, "grad_norm": 0.21629554380819382, "learning_rate": 1.922875131164743e-05, "loss": 0.4222, "step": 1386 }, { "epoch": 1.9624911535739562, "grad_norm": 0.21894295567177233, "learning_rate": 1.920251836306401e-05, "loss": 0.5109, "step": 1387 }, { "epoch": 1.9639065817409767, "grad_norm": 0.25551379680235153, "learning_rate": 1.9176285414480587e-05, "loss": 0.5169, "step": 1388 }, { "epoch": 1.965322009907997, "grad_norm": 0.22701226384061032, "learning_rate": 1.9150052465897168e-05, "loss": 0.4238, "step": 1389 }, { "epoch": 1.9667374380750178, "grad_norm": 0.21764512077540232, "learning_rate": 1.9123819517313745e-05, "loss": 0.483, "step": 1390 }, { "epoch": 1.9681528662420382, "grad_norm": 0.2379062720137337, "learning_rate": 1.9097586568730328e-05, "loss": 0.4301, "step": 1391 }, { "epoch": 1.9695682944090587, "grad_norm": 0.21706551907158597, "learning_rate": 1.9071353620146905e-05, "loss": 0.4435, "step": 1392 }, { "epoch": 1.9709837225760793, "grad_norm": 0.2681923527504643, "learning_rate": 1.9045120671563485e-05, "loss": 0.5055, "step": 1393 }, { "epoch": 1.9723991507430998, "grad_norm": 0.24764910859697023, "learning_rate": 1.9018887722980062e-05, "loss": 0.5379, "step": 1394 }, { "epoch": 1.9738145789101202, "grad_norm": 0.24028803996244996, "learning_rate": 1.8992654774396643e-05, "loss": 0.4422, "step": 1395 }, { "epoch": 1.975230007077141, "grad_norm": 0.2667908175923755, "learning_rate": 1.8966421825813223e-05, "loss": 0.4601, "step": 1396 }, { "epoch": 1.9766454352441614, "grad_norm": 0.1983331591989112, "learning_rate": 1.89401888772298e-05, "loss": 0.3583, "step": 1397 }, { "epoch": 1.9780608634111818, "grad_norm": 0.2129299178458175, "learning_rate": 1.8913955928646383e-05, "loss": 0.5322, "step": 1398 }, { "epoch": 1.9794762915782025, "grad_norm": 0.24494279276071143, "learning_rate": 1.888772298006296e-05, "loss": 0.4284, "step": 1399 }, { "epoch": 1.980891719745223, "grad_norm": 0.2346234778429022, "learning_rate": 1.886149003147954e-05, "loss": 0.5095, "step": 1400 }, { "epoch": 1.9823071479122434, "grad_norm": 0.22148149677042864, "learning_rate": 1.8835257082896118e-05, "loss": 0.4191, "step": 1401 }, { "epoch": 1.983722576079264, "grad_norm": 0.20471789202971866, "learning_rate": 1.8809024134312698e-05, "loss": 0.4225, "step": 1402 }, { "epoch": 1.9851380042462845, "grad_norm": 0.24568010662632564, "learning_rate": 1.8782791185729275e-05, "loss": 0.4981, "step": 1403 }, { "epoch": 1.986553432413305, "grad_norm": 0.21930415253021862, "learning_rate": 1.8756558237145855e-05, "loss": 0.4752, "step": 1404 }, { "epoch": 1.9879688605803256, "grad_norm": 0.251898694328717, "learning_rate": 1.8730325288562435e-05, "loss": 0.4316, "step": 1405 }, { "epoch": 1.989384288747346, "grad_norm": 0.2176564270865631, "learning_rate": 1.8704092339979016e-05, "loss": 0.5356, "step": 1406 }, { "epoch": 1.9907997169143665, "grad_norm": 0.2330020823912981, "learning_rate": 1.8677859391395592e-05, "loss": 0.4502, "step": 1407 }, { "epoch": 1.9922151450813872, "grad_norm": 0.2357008471649503, "learning_rate": 1.8651626442812173e-05, "loss": 0.463, "step": 1408 }, { "epoch": 1.9936305732484076, "grad_norm": 0.2167498961975368, "learning_rate": 1.8625393494228753e-05, "loss": 0.4825, "step": 1409 }, { "epoch": 1.995046001415428, "grad_norm": 0.3680365703435767, "learning_rate": 1.859916054564533e-05, "loss": 0.4899, "step": 1410 }, { "epoch": 1.9964614295824488, "grad_norm": 0.2261187086530332, "learning_rate": 1.857292759706191e-05, "loss": 0.5036, "step": 1411 }, { "epoch": 1.9978768577494692, "grad_norm": 0.22055517468425753, "learning_rate": 1.854669464847849e-05, "loss": 0.5127, "step": 1412 }, { "epoch": 1.9992922859164897, "grad_norm": 0.24038986520047975, "learning_rate": 1.852046169989507e-05, "loss": 0.4774, "step": 1413 }, { "epoch": 2.0, "grad_norm": 0.3649991056816497, "learning_rate": 1.8494228751311648e-05, "loss": 0.4991, "step": 1414 }, { "epoch": 2.0014154281670207, "grad_norm": 0.4166693991610525, "learning_rate": 1.8467995802728228e-05, "loss": 0.4337, "step": 1415 }, { "epoch": 2.002830856334041, "grad_norm": 0.260666322974337, "learning_rate": 1.8441762854144805e-05, "loss": 0.3922, "step": 1416 }, { "epoch": 2.0042462845010616, "grad_norm": 0.29238039510242636, "learning_rate": 1.8415529905561385e-05, "loss": 0.3747, "step": 1417 }, { "epoch": 2.0056617126680822, "grad_norm": 0.3564915371981773, "learning_rate": 1.8389296956977965e-05, "loss": 0.382, "step": 1418 }, { "epoch": 2.0070771408351025, "grad_norm": 0.26783850124835984, "learning_rate": 1.8363064008394546e-05, "loss": 0.4175, "step": 1419 }, { "epoch": 2.008492569002123, "grad_norm": 0.25248301392487915, "learning_rate": 1.8336831059811123e-05, "loss": 0.3667, "step": 1420 }, { "epoch": 2.009907997169144, "grad_norm": 0.2970158100432694, "learning_rate": 1.8310598111227703e-05, "loss": 0.4218, "step": 1421 }, { "epoch": 2.011323425336164, "grad_norm": 0.32058082497326085, "learning_rate": 1.8284365162644283e-05, "loss": 0.3682, "step": 1422 }, { "epoch": 2.0127388535031847, "grad_norm": 0.258438347594369, "learning_rate": 1.825813221406086e-05, "loss": 0.3814, "step": 1423 }, { "epoch": 2.0141542816702054, "grad_norm": 0.2309440049259677, "learning_rate": 1.823189926547744e-05, "loss": 0.3613, "step": 1424 }, { "epoch": 2.0155697098372256, "grad_norm": 0.23711843903773477, "learning_rate": 1.820566631689402e-05, "loss": 0.3638, "step": 1425 }, { "epoch": 2.0169851380042463, "grad_norm": 0.30154935042770303, "learning_rate": 1.81794333683106e-05, "loss": 0.3911, "step": 1426 }, { "epoch": 2.018400566171267, "grad_norm": 0.24985753759283674, "learning_rate": 1.8153200419727178e-05, "loss": 0.3859, "step": 1427 }, { "epoch": 2.019815994338287, "grad_norm": 0.21822783879615154, "learning_rate": 1.8126967471143758e-05, "loss": 0.3692, "step": 1428 }, { "epoch": 2.021231422505308, "grad_norm": 0.23574623808166204, "learning_rate": 1.8100734522560335e-05, "loss": 0.3694, "step": 1429 }, { "epoch": 2.0226468506723285, "grad_norm": 0.2656054636640407, "learning_rate": 1.8074501573976915e-05, "loss": 0.3394, "step": 1430 }, { "epoch": 2.0240622788393487, "grad_norm": 0.2603011508150877, "learning_rate": 1.8048268625393496e-05, "loss": 0.3267, "step": 1431 }, { "epoch": 2.0254777070063694, "grad_norm": 0.23364246069398123, "learning_rate": 1.8022035676810076e-05, "loss": 0.3429, "step": 1432 }, { "epoch": 2.02689313517339, "grad_norm": 0.2491023446129911, "learning_rate": 1.7995802728226656e-05, "loss": 0.4053, "step": 1433 }, { "epoch": 2.0283085633404103, "grad_norm": 0.24128596635121036, "learning_rate": 1.7969569779643233e-05, "loss": 0.3407, "step": 1434 }, { "epoch": 2.029723991507431, "grad_norm": 0.2803188728012592, "learning_rate": 1.7943336831059813e-05, "loss": 0.3737, "step": 1435 }, { "epoch": 2.0311394196744517, "grad_norm": 0.27015953132865195, "learning_rate": 1.791710388247639e-05, "loss": 0.3921, "step": 1436 }, { "epoch": 2.032554847841472, "grad_norm": 0.25020458517438576, "learning_rate": 1.789087093389297e-05, "loss": 0.3706, "step": 1437 }, { "epoch": 2.0339702760084926, "grad_norm": 0.2640785254176496, "learning_rate": 1.7864637985309547e-05, "loss": 0.3618, "step": 1438 }, { "epoch": 2.0353857041755132, "grad_norm": 0.240523144913415, "learning_rate": 1.783840503672613e-05, "loss": 0.332, "step": 1439 }, { "epoch": 2.0368011323425335, "grad_norm": 0.23611315437615582, "learning_rate": 1.7812172088142708e-05, "loss": 0.2785, "step": 1440 }, { "epoch": 2.038216560509554, "grad_norm": 0.2252806592035128, "learning_rate": 1.7785939139559288e-05, "loss": 0.3577, "step": 1441 }, { "epoch": 2.039631988676575, "grad_norm": 0.24882838329271678, "learning_rate": 1.7759706190975865e-05, "loss": 0.4204, "step": 1442 }, { "epoch": 2.041047416843595, "grad_norm": 0.2107192080934422, "learning_rate": 1.7733473242392445e-05, "loss": 0.3904, "step": 1443 }, { "epoch": 2.0424628450106157, "grad_norm": 0.21057384801951756, "learning_rate": 1.7707240293809026e-05, "loss": 0.3695, "step": 1444 }, { "epoch": 2.0438782731776364, "grad_norm": 0.2165214812310043, "learning_rate": 1.7681007345225603e-05, "loss": 0.3465, "step": 1445 }, { "epoch": 2.0452937013446566, "grad_norm": 0.21451815510889052, "learning_rate": 1.7654774396642186e-05, "loss": 0.3473, "step": 1446 }, { "epoch": 2.0467091295116773, "grad_norm": 0.2255237074033376, "learning_rate": 1.7628541448058763e-05, "loss": 0.3915, "step": 1447 }, { "epoch": 2.048124557678698, "grad_norm": 0.21102539596402395, "learning_rate": 1.7602308499475343e-05, "loss": 0.4007, "step": 1448 }, { "epoch": 2.049539985845718, "grad_norm": 0.23126500561761867, "learning_rate": 1.757607555089192e-05, "loss": 0.3449, "step": 1449 }, { "epoch": 2.050955414012739, "grad_norm": 0.21701265916378112, "learning_rate": 1.75498426023085e-05, "loss": 0.3613, "step": 1450 }, { "epoch": 2.0523708421797595, "grad_norm": 0.22711923189993327, "learning_rate": 1.7523609653725078e-05, "loss": 0.401, "step": 1451 }, { "epoch": 2.0537862703467797, "grad_norm": 0.20861442180382758, "learning_rate": 1.7497376705141658e-05, "loss": 0.3655, "step": 1452 }, { "epoch": 2.0552016985138004, "grad_norm": 0.2270859623252876, "learning_rate": 1.7471143756558238e-05, "loss": 0.3595, "step": 1453 }, { "epoch": 2.056617126680821, "grad_norm": 0.23364629764481326, "learning_rate": 1.744491080797482e-05, "loss": 0.3295, "step": 1454 }, { "epoch": 2.0580325548478413, "grad_norm": 0.22567027139214516, "learning_rate": 1.7418677859391395e-05, "loss": 0.3831, "step": 1455 }, { "epoch": 2.059447983014862, "grad_norm": 0.21839431235112447, "learning_rate": 1.7392444910807976e-05, "loss": 0.3556, "step": 1456 }, { "epoch": 2.0608634111818827, "grad_norm": 0.22426774563431578, "learning_rate": 1.7366211962224556e-05, "loss": 0.3484, "step": 1457 }, { "epoch": 2.062278839348903, "grad_norm": 0.2233566251331184, "learning_rate": 1.7339979013641133e-05, "loss": 0.3736, "step": 1458 }, { "epoch": 2.0636942675159236, "grad_norm": 0.19773703937830703, "learning_rate": 1.7313746065057713e-05, "loss": 0.3069, "step": 1459 }, { "epoch": 2.0651096956829442, "grad_norm": 0.23091198480431166, "learning_rate": 1.7287513116474293e-05, "loss": 0.38, "step": 1460 }, { "epoch": 2.0665251238499645, "grad_norm": 0.21888763266422998, "learning_rate": 1.7261280167890874e-05, "loss": 0.3814, "step": 1461 }, { "epoch": 2.067940552016985, "grad_norm": 0.2118377785821229, "learning_rate": 1.723504721930745e-05, "loss": 0.3675, "step": 1462 }, { "epoch": 2.069355980184006, "grad_norm": 0.20564189159820773, "learning_rate": 1.720881427072403e-05, "loss": 0.3421, "step": 1463 }, { "epoch": 2.070771408351026, "grad_norm": 0.2169277793958252, "learning_rate": 1.7182581322140608e-05, "loss": 0.3781, "step": 1464 }, { "epoch": 2.0721868365180467, "grad_norm": 0.20777975555156794, "learning_rate": 1.7156348373557188e-05, "loss": 0.3421, "step": 1465 }, { "epoch": 2.0736022646850674, "grad_norm": 0.23434517928918236, "learning_rate": 1.7130115424973768e-05, "loss": 0.3926, "step": 1466 }, { "epoch": 2.0750176928520876, "grad_norm": 0.2237156164837064, "learning_rate": 1.710388247639035e-05, "loss": 0.3952, "step": 1467 }, { "epoch": 2.0764331210191083, "grad_norm": 0.20852775525749587, "learning_rate": 1.7077649527806925e-05, "loss": 0.3471, "step": 1468 }, { "epoch": 2.077848549186129, "grad_norm": 0.21165243525823502, "learning_rate": 1.7051416579223506e-05, "loss": 0.3677, "step": 1469 }, { "epoch": 2.079263977353149, "grad_norm": 1.1822920981729839, "learning_rate": 1.7025183630640086e-05, "loss": 0.4679, "step": 1470 }, { "epoch": 2.08067940552017, "grad_norm": 0.2479312294370373, "learning_rate": 1.6998950682056663e-05, "loss": 0.3967, "step": 1471 }, { "epoch": 2.0820948336871905, "grad_norm": 0.25982595855025337, "learning_rate": 1.6972717733473243e-05, "loss": 0.3706, "step": 1472 }, { "epoch": 2.0835102618542107, "grad_norm": 0.21858931940249068, "learning_rate": 1.694648478488982e-05, "loss": 0.3613, "step": 1473 }, { "epoch": 2.0849256900212314, "grad_norm": 0.22549270661935017, "learning_rate": 1.6920251836306404e-05, "loss": 0.3655, "step": 1474 }, { "epoch": 2.086341118188252, "grad_norm": 0.2121034477698962, "learning_rate": 1.689401888772298e-05, "loss": 0.3416, "step": 1475 }, { "epoch": 2.0877565463552723, "grad_norm": 0.2034045361151294, "learning_rate": 1.686778593913956e-05, "loss": 0.3352, "step": 1476 }, { "epoch": 2.089171974522293, "grad_norm": 0.21723699645973873, "learning_rate": 1.6841552990556138e-05, "loss": 0.3432, "step": 1477 }, { "epoch": 2.0905874026893136, "grad_norm": 0.23426304130031295, "learning_rate": 1.6815320041972718e-05, "loss": 0.3608, "step": 1478 }, { "epoch": 2.092002830856334, "grad_norm": 0.20725051222778673, "learning_rate": 1.67890870933893e-05, "loss": 0.347, "step": 1479 }, { "epoch": 2.0934182590233545, "grad_norm": 0.23093322390994894, "learning_rate": 1.6762854144805875e-05, "loss": 0.4253, "step": 1480 }, { "epoch": 2.094833687190375, "grad_norm": 0.22746696959851062, "learning_rate": 1.6736621196222456e-05, "loss": 0.4098, "step": 1481 }, { "epoch": 2.0962491153573954, "grad_norm": 0.223823439118962, "learning_rate": 1.6710388247639036e-05, "loss": 0.3938, "step": 1482 }, { "epoch": 2.097664543524416, "grad_norm": 0.21937232678259852, "learning_rate": 1.6684155299055616e-05, "loss": 0.3975, "step": 1483 }, { "epoch": 2.099079971691437, "grad_norm": 0.21873413992633153, "learning_rate": 1.6657922350472193e-05, "loss": 0.3965, "step": 1484 }, { "epoch": 2.100495399858457, "grad_norm": 0.22559007031341452, "learning_rate": 1.6631689401888773e-05, "loss": 0.3667, "step": 1485 }, { "epoch": 2.1019108280254777, "grad_norm": 0.18962001559298866, "learning_rate": 1.660545645330535e-05, "loss": 0.2863, "step": 1486 }, { "epoch": 2.1033262561924984, "grad_norm": 0.2256149343155719, "learning_rate": 1.657922350472193e-05, "loss": 0.3789, "step": 1487 }, { "epoch": 2.1047416843595186, "grad_norm": 0.24432634254170046, "learning_rate": 1.655299055613851e-05, "loss": 0.4192, "step": 1488 }, { "epoch": 2.1061571125265393, "grad_norm": 0.20331866520876218, "learning_rate": 1.652675760755509e-05, "loss": 0.3157, "step": 1489 }, { "epoch": 2.10757254069356, "grad_norm": 0.2218135446867955, "learning_rate": 1.6500524658971668e-05, "loss": 0.4136, "step": 1490 }, { "epoch": 2.10898796886058, "grad_norm": 0.20788675568869244, "learning_rate": 1.647429171038825e-05, "loss": 0.3828, "step": 1491 }, { "epoch": 2.110403397027601, "grad_norm": 0.2106124475328733, "learning_rate": 1.644805876180483e-05, "loss": 0.385, "step": 1492 }, { "epoch": 2.1118188251946215, "grad_norm": 0.22319116024778723, "learning_rate": 1.6421825813221406e-05, "loss": 0.3942, "step": 1493 }, { "epoch": 2.1132342533616417, "grad_norm": 0.2045669215686749, "learning_rate": 1.6395592864637986e-05, "loss": 0.3559, "step": 1494 }, { "epoch": 2.1146496815286624, "grad_norm": 0.20167320873533356, "learning_rate": 1.6369359916054566e-05, "loss": 0.3492, "step": 1495 }, { "epoch": 2.116065109695683, "grad_norm": 0.563539207714512, "learning_rate": 1.6343126967471146e-05, "loss": 0.3628, "step": 1496 }, { "epoch": 2.1174805378627033, "grad_norm": 0.21048626662633552, "learning_rate": 1.6316894018887723e-05, "loss": 0.3927, "step": 1497 }, { "epoch": 2.118895966029724, "grad_norm": 0.21525926446490412, "learning_rate": 1.6290661070304304e-05, "loss": 0.3693, "step": 1498 }, { "epoch": 2.1203113941967446, "grad_norm": 0.2278308053047193, "learning_rate": 1.626442812172088e-05, "loss": 0.4186, "step": 1499 }, { "epoch": 2.121726822363765, "grad_norm": 0.20026421206561354, "learning_rate": 1.623819517313746e-05, "loss": 0.3624, "step": 1500 }, { "epoch": 2.1231422505307855, "grad_norm": 0.21648706832363435, "learning_rate": 1.621196222455404e-05, "loss": 0.3781, "step": 1501 }, { "epoch": 2.124557678697806, "grad_norm": 0.19515781163709509, "learning_rate": 1.618572927597062e-05, "loss": 0.336, "step": 1502 }, { "epoch": 2.1259731068648264, "grad_norm": 0.24081153314019926, "learning_rate": 1.6159496327387198e-05, "loss": 0.4276, "step": 1503 }, { "epoch": 2.127388535031847, "grad_norm": 0.2156910140075714, "learning_rate": 1.613326337880378e-05, "loss": 0.3482, "step": 1504 }, { "epoch": 2.1288039631988678, "grad_norm": 0.20631513186015674, "learning_rate": 1.610703043022036e-05, "loss": 0.3488, "step": 1505 }, { "epoch": 2.130219391365888, "grad_norm": 0.2413240933427328, "learning_rate": 1.6080797481636936e-05, "loss": 0.4308, "step": 1506 }, { "epoch": 2.1316348195329087, "grad_norm": 0.19686510455803716, "learning_rate": 1.6054564533053516e-05, "loss": 0.3236, "step": 1507 }, { "epoch": 2.1330502476999293, "grad_norm": 0.21937196597825326, "learning_rate": 1.6028331584470096e-05, "loss": 0.3522, "step": 1508 }, { "epoch": 2.1344656758669496, "grad_norm": 0.2004049675013169, "learning_rate": 1.6002098635886677e-05, "loss": 0.3719, "step": 1509 }, { "epoch": 2.1358811040339702, "grad_norm": 0.21926244774280124, "learning_rate": 1.5975865687303253e-05, "loss": 0.4035, "step": 1510 }, { "epoch": 2.137296532200991, "grad_norm": 0.20670203803808587, "learning_rate": 1.5949632738719834e-05, "loss": 0.3807, "step": 1511 }, { "epoch": 2.138711960368011, "grad_norm": 0.2090052811208187, "learning_rate": 1.592339979013641e-05, "loss": 0.331, "step": 1512 }, { "epoch": 2.140127388535032, "grad_norm": 0.2218332758157514, "learning_rate": 1.589716684155299e-05, "loss": 0.3351, "step": 1513 }, { "epoch": 2.1415428167020525, "grad_norm": 0.20286719665033282, "learning_rate": 1.5870933892969568e-05, "loss": 0.3508, "step": 1514 }, { "epoch": 2.1429582448690727, "grad_norm": 0.2157182790914069, "learning_rate": 1.584470094438615e-05, "loss": 0.4096, "step": 1515 }, { "epoch": 2.1443736730360934, "grad_norm": 0.23149191498509575, "learning_rate": 1.581846799580273e-05, "loss": 0.3795, "step": 1516 }, { "epoch": 2.145789101203114, "grad_norm": 0.2179465608537574, "learning_rate": 1.579223504721931e-05, "loss": 0.3652, "step": 1517 }, { "epoch": 2.1472045293701343, "grad_norm": 0.23025993168424932, "learning_rate": 1.576600209863589e-05, "loss": 0.4007, "step": 1518 }, { "epoch": 2.148619957537155, "grad_norm": 0.22000849317187712, "learning_rate": 1.5739769150052466e-05, "loss": 0.3652, "step": 1519 }, { "epoch": 2.1500353857041756, "grad_norm": 0.24530162425275553, "learning_rate": 1.5713536201469046e-05, "loss": 0.3768, "step": 1520 }, { "epoch": 2.151450813871196, "grad_norm": 0.2085214000510065, "learning_rate": 1.5687303252885623e-05, "loss": 0.3877, "step": 1521 }, { "epoch": 2.1528662420382165, "grad_norm": 0.23192817616235128, "learning_rate": 1.5661070304302207e-05, "loss": 0.3871, "step": 1522 }, { "epoch": 2.154281670205237, "grad_norm": 0.26726756540877716, "learning_rate": 1.5634837355718784e-05, "loss": 0.3914, "step": 1523 }, { "epoch": 2.1556970983722574, "grad_norm": 0.2286927722327317, "learning_rate": 1.5608604407135364e-05, "loss": 0.3612, "step": 1524 }, { "epoch": 2.157112526539278, "grad_norm": 0.20785526014793734, "learning_rate": 1.558237145855194e-05, "loss": 0.3915, "step": 1525 }, { "epoch": 2.1585279547062988, "grad_norm": 0.22692806461653298, "learning_rate": 1.555613850996852e-05, "loss": 0.3322, "step": 1526 }, { "epoch": 2.159943382873319, "grad_norm": 0.21694441189573116, "learning_rate": 1.55299055613851e-05, "loss": 0.3668, "step": 1527 }, { "epoch": 2.1613588110403397, "grad_norm": 0.2178270572564254, "learning_rate": 1.5503672612801678e-05, "loss": 0.3506, "step": 1528 }, { "epoch": 2.1627742392073603, "grad_norm": 0.19684157868907207, "learning_rate": 1.547743966421826e-05, "loss": 0.3425, "step": 1529 }, { "epoch": 2.1641896673743806, "grad_norm": 0.2185293631247771, "learning_rate": 1.545120671563484e-05, "loss": 0.373, "step": 1530 }, { "epoch": 2.1656050955414012, "grad_norm": 0.24937859931810394, "learning_rate": 1.542497376705142e-05, "loss": 0.4453, "step": 1531 }, { "epoch": 2.167020523708422, "grad_norm": 0.20250770374257762, "learning_rate": 1.5398740818467996e-05, "loss": 0.3603, "step": 1532 }, { "epoch": 2.168435951875442, "grad_norm": 0.2229953291778268, "learning_rate": 1.5372507869884576e-05, "loss": 0.3546, "step": 1533 }, { "epoch": 2.169851380042463, "grad_norm": 0.22147353462647917, "learning_rate": 1.5346274921301153e-05, "loss": 0.4118, "step": 1534 }, { "epoch": 2.1712668082094835, "grad_norm": 0.21956103582223985, "learning_rate": 1.5320041972717733e-05, "loss": 0.3779, "step": 1535 }, { "epoch": 2.1726822363765037, "grad_norm": 0.22705905713793162, "learning_rate": 1.5293809024134314e-05, "loss": 0.3873, "step": 1536 }, { "epoch": 2.1740976645435244, "grad_norm": 0.20964973432427997, "learning_rate": 1.5267576075550894e-05, "loss": 0.3526, "step": 1537 }, { "epoch": 2.175513092710545, "grad_norm": 0.21249751481689355, "learning_rate": 1.5241343126967473e-05, "loss": 0.3628, "step": 1538 }, { "epoch": 2.1769285208775653, "grad_norm": 0.23622172484407175, "learning_rate": 1.5215110178384051e-05, "loss": 0.3835, "step": 1539 }, { "epoch": 2.178343949044586, "grad_norm": 0.23211487603406503, "learning_rate": 1.518887722980063e-05, "loss": 0.3913, "step": 1540 }, { "epoch": 2.1797593772116066, "grad_norm": 0.20060668542207974, "learning_rate": 1.5162644281217208e-05, "loss": 0.3122, "step": 1541 }, { "epoch": 2.181174805378627, "grad_norm": 0.22790100973419974, "learning_rate": 1.5136411332633787e-05, "loss": 0.3871, "step": 1542 }, { "epoch": 2.1825902335456475, "grad_norm": 0.20199828675191117, "learning_rate": 1.5110178384050369e-05, "loss": 0.3644, "step": 1543 }, { "epoch": 2.184005661712668, "grad_norm": 0.2189262776259267, "learning_rate": 1.5083945435466948e-05, "loss": 0.4043, "step": 1544 }, { "epoch": 2.1854210898796884, "grad_norm": 0.22346416372410774, "learning_rate": 1.5057712486883526e-05, "loss": 0.3789, "step": 1545 }, { "epoch": 2.186836518046709, "grad_norm": 0.20526825039948007, "learning_rate": 1.5031479538300106e-05, "loss": 0.3364, "step": 1546 }, { "epoch": 2.1882519462137298, "grad_norm": 0.2078165417942967, "learning_rate": 1.5005246589716685e-05, "loss": 0.334, "step": 1547 }, { "epoch": 2.18966737438075, "grad_norm": 0.21094891720826406, "learning_rate": 1.4979013641133264e-05, "loss": 0.373, "step": 1548 }, { "epoch": 2.1910828025477707, "grad_norm": 0.22362768217164414, "learning_rate": 1.4952780692549842e-05, "loss": 0.4026, "step": 1549 }, { "epoch": 2.1924982307147913, "grad_norm": 0.203456004686505, "learning_rate": 1.4926547743966424e-05, "loss": 0.3633, "step": 1550 }, { "epoch": 2.1939136588818116, "grad_norm": 0.23121854991500074, "learning_rate": 1.4900314795383003e-05, "loss": 0.3626, "step": 1551 }, { "epoch": 2.1953290870488322, "grad_norm": 0.2138959069426217, "learning_rate": 1.4874081846799581e-05, "loss": 0.3576, "step": 1552 }, { "epoch": 2.196744515215853, "grad_norm": 0.24699037645814892, "learning_rate": 1.484784889821616e-05, "loss": 0.3856, "step": 1553 }, { "epoch": 2.198159943382873, "grad_norm": 0.19854082616956784, "learning_rate": 1.4821615949632739e-05, "loss": 0.3418, "step": 1554 }, { "epoch": 2.199575371549894, "grad_norm": 0.21266454735849155, "learning_rate": 1.4795383001049317e-05, "loss": 0.3587, "step": 1555 }, { "epoch": 2.2009907997169145, "grad_norm": 0.21339582475218097, "learning_rate": 1.4769150052465897e-05, "loss": 0.3761, "step": 1556 }, { "epoch": 2.2024062278839347, "grad_norm": 0.20939402295306628, "learning_rate": 1.4742917103882478e-05, "loss": 0.3781, "step": 1557 }, { "epoch": 2.2038216560509554, "grad_norm": 0.2322645157397984, "learning_rate": 1.4716684155299056e-05, "loss": 0.3948, "step": 1558 }, { "epoch": 2.205237084217976, "grad_norm": 0.22585337716531292, "learning_rate": 1.4690451206715637e-05, "loss": 0.3426, "step": 1559 }, { "epoch": 2.2066525123849963, "grad_norm": 0.2148503491337666, "learning_rate": 1.4664218258132215e-05, "loss": 0.3525, "step": 1560 }, { "epoch": 2.208067940552017, "grad_norm": 0.22665694917883897, "learning_rate": 1.4637985309548794e-05, "loss": 0.3798, "step": 1561 }, { "epoch": 2.2094833687190376, "grad_norm": 0.20861364422501297, "learning_rate": 1.4611752360965372e-05, "loss": 0.3757, "step": 1562 }, { "epoch": 2.210898796886058, "grad_norm": 0.2068838790900965, "learning_rate": 1.4585519412381951e-05, "loss": 0.3345, "step": 1563 }, { "epoch": 2.2123142250530785, "grad_norm": 0.21314165921334982, "learning_rate": 1.4559286463798533e-05, "loss": 0.3743, "step": 1564 }, { "epoch": 2.213729653220099, "grad_norm": 0.20606173894877292, "learning_rate": 1.4533053515215112e-05, "loss": 0.3515, "step": 1565 }, { "epoch": 2.2151450813871194, "grad_norm": 0.21170676741754832, "learning_rate": 1.450682056663169e-05, "loss": 0.3532, "step": 1566 }, { "epoch": 2.21656050955414, "grad_norm": 0.21890755771438766, "learning_rate": 1.4480587618048269e-05, "loss": 0.3606, "step": 1567 }, { "epoch": 2.2179759377211608, "grad_norm": 0.21049879748213451, "learning_rate": 1.4454354669464847e-05, "loss": 0.3722, "step": 1568 }, { "epoch": 2.219391365888181, "grad_norm": 0.19689279358689524, "learning_rate": 1.4428121720881428e-05, "loss": 0.3332, "step": 1569 }, { "epoch": 2.2208067940552016, "grad_norm": 0.221244953805384, "learning_rate": 1.4401888772298008e-05, "loss": 0.3691, "step": 1570 }, { "epoch": 2.2222222222222223, "grad_norm": 0.21019982504697307, "learning_rate": 1.4375655823714586e-05, "loss": 0.3514, "step": 1571 }, { "epoch": 2.2236376503892425, "grad_norm": 0.22327263464343836, "learning_rate": 1.4349422875131167e-05, "loss": 0.3736, "step": 1572 }, { "epoch": 2.225053078556263, "grad_norm": 0.2337778954093035, "learning_rate": 1.4323189926547745e-05, "loss": 0.3802, "step": 1573 }, { "epoch": 2.226468506723284, "grad_norm": 0.22491847887006883, "learning_rate": 1.4296956977964324e-05, "loss": 0.3761, "step": 1574 }, { "epoch": 2.227883934890304, "grad_norm": 0.2055020512327927, "learning_rate": 1.4270724029380902e-05, "loss": 0.3534, "step": 1575 }, { "epoch": 2.229299363057325, "grad_norm": 0.22421411423810259, "learning_rate": 1.4244491080797481e-05, "loss": 0.3637, "step": 1576 }, { "epoch": 2.2307147912243455, "grad_norm": 0.19551123110438778, "learning_rate": 1.4218258132214063e-05, "loss": 0.3291, "step": 1577 }, { "epoch": 2.2321302193913657, "grad_norm": 0.2290439051675896, "learning_rate": 1.4192025183630642e-05, "loss": 0.3396, "step": 1578 }, { "epoch": 2.2335456475583864, "grad_norm": 0.20583543421027845, "learning_rate": 1.416579223504722e-05, "loss": 0.3544, "step": 1579 }, { "epoch": 2.234961075725407, "grad_norm": 0.2342349710976409, "learning_rate": 1.4139559286463799e-05, "loss": 0.4077, "step": 1580 }, { "epoch": 2.2363765038924273, "grad_norm": 0.2198729597030276, "learning_rate": 1.4113326337880377e-05, "loss": 0.3705, "step": 1581 }, { "epoch": 2.237791932059448, "grad_norm": 0.22235081348422966, "learning_rate": 1.4087093389296958e-05, "loss": 0.3551, "step": 1582 }, { "epoch": 2.2392073602264686, "grad_norm": 0.22345356674467037, "learning_rate": 1.4060860440713536e-05, "loss": 0.3462, "step": 1583 }, { "epoch": 2.240622788393489, "grad_norm": 0.20517643442250938, "learning_rate": 1.4034627492130117e-05, "loss": 0.3265, "step": 1584 }, { "epoch": 2.2420382165605095, "grad_norm": 0.21959226573979157, "learning_rate": 1.4008394543546697e-05, "loss": 0.3859, "step": 1585 }, { "epoch": 2.24345364472753, "grad_norm": 0.24355893909564938, "learning_rate": 1.3982161594963275e-05, "loss": 0.4022, "step": 1586 }, { "epoch": 2.2448690728945504, "grad_norm": 0.23377288359241896, "learning_rate": 1.3955928646379854e-05, "loss": 0.4207, "step": 1587 }, { "epoch": 2.246284501061571, "grad_norm": 0.2199085047090607, "learning_rate": 1.3929695697796433e-05, "loss": 0.379, "step": 1588 }, { "epoch": 2.2476999292285917, "grad_norm": 0.22062658898222495, "learning_rate": 1.3903462749213011e-05, "loss": 0.36, "step": 1589 }, { "epoch": 2.249115357395612, "grad_norm": 0.2491554525684039, "learning_rate": 1.387722980062959e-05, "loss": 0.4043, "step": 1590 }, { "epoch": 2.2505307855626326, "grad_norm": 0.2213422309822871, "learning_rate": 1.3850996852046172e-05, "loss": 0.3459, "step": 1591 }, { "epoch": 2.2519462137296533, "grad_norm": 0.20169981779020701, "learning_rate": 1.382476390346275e-05, "loss": 0.3562, "step": 1592 }, { "epoch": 2.2533616418966735, "grad_norm": 0.21076797503294742, "learning_rate": 1.3798530954879329e-05, "loss": 0.3392, "step": 1593 }, { "epoch": 2.254777070063694, "grad_norm": 0.21221372296515487, "learning_rate": 1.3772298006295908e-05, "loss": 0.3353, "step": 1594 }, { "epoch": 2.256192498230715, "grad_norm": 0.2413103693540377, "learning_rate": 1.3746065057712488e-05, "loss": 0.3827, "step": 1595 }, { "epoch": 2.2576079263977356, "grad_norm": 0.20595506827868224, "learning_rate": 1.3719832109129066e-05, "loss": 0.3802, "step": 1596 }, { "epoch": 2.259023354564756, "grad_norm": 0.21085923154346145, "learning_rate": 1.3693599160545645e-05, "loss": 0.4063, "step": 1597 }, { "epoch": 2.2604387827317765, "grad_norm": 0.22335746982937194, "learning_rate": 1.3667366211962227e-05, "loss": 0.3637, "step": 1598 }, { "epoch": 2.2618542108987967, "grad_norm": 0.22208789549987143, "learning_rate": 1.3641133263378806e-05, "loss": 0.3656, "step": 1599 }, { "epoch": 2.2632696390658174, "grad_norm": 0.21297418200769241, "learning_rate": 1.3614900314795384e-05, "loss": 0.3791, "step": 1600 }, { "epoch": 2.264685067232838, "grad_norm": 0.19540588053519106, "learning_rate": 1.3588667366211963e-05, "loss": 0.3324, "step": 1601 }, { "epoch": 2.2661004953998587, "grad_norm": 0.20759627984515497, "learning_rate": 1.3562434417628541e-05, "loss": 0.347, "step": 1602 }, { "epoch": 2.267515923566879, "grad_norm": 0.2245939350902509, "learning_rate": 1.353620146904512e-05, "loss": 0.3697, "step": 1603 }, { "epoch": 2.2689313517338996, "grad_norm": 0.22209454778408863, "learning_rate": 1.3509968520461699e-05, "loss": 0.356, "step": 1604 }, { "epoch": 2.27034677990092, "grad_norm": 0.2059874222511523, "learning_rate": 1.348373557187828e-05, "loss": 0.3988, "step": 1605 }, { "epoch": 2.2717622080679405, "grad_norm": 0.199393331852189, "learning_rate": 1.3457502623294859e-05, "loss": 0.3431, "step": 1606 }, { "epoch": 2.273177636234961, "grad_norm": 0.2262724322623601, "learning_rate": 1.343126967471144e-05, "loss": 0.3459, "step": 1607 }, { "epoch": 2.274593064401982, "grad_norm": 0.22062582548159224, "learning_rate": 1.3405036726128018e-05, "loss": 0.3832, "step": 1608 }, { "epoch": 2.276008492569002, "grad_norm": 0.20922151718664564, "learning_rate": 1.3378803777544597e-05, "loss": 0.3678, "step": 1609 }, { "epoch": 2.2774239207360227, "grad_norm": 0.23346011193736022, "learning_rate": 1.3352570828961175e-05, "loss": 0.4017, "step": 1610 }, { "epoch": 2.278839348903043, "grad_norm": 0.20660194929082695, "learning_rate": 1.3326337880377754e-05, "loss": 0.3547, "step": 1611 }, { "epoch": 2.2802547770700636, "grad_norm": 0.19175980626077407, "learning_rate": 1.3300104931794336e-05, "loss": 0.3275, "step": 1612 }, { "epoch": 2.2816702052370843, "grad_norm": 0.22569077925192324, "learning_rate": 1.3273871983210914e-05, "loss": 0.3949, "step": 1613 }, { "epoch": 2.283085633404105, "grad_norm": 0.22179564083723302, "learning_rate": 1.3247639034627493e-05, "loss": 0.3783, "step": 1614 }, { "epoch": 2.284501061571125, "grad_norm": 0.2225183306260963, "learning_rate": 1.3221406086044072e-05, "loss": 0.4094, "step": 1615 }, { "epoch": 2.285916489738146, "grad_norm": 0.20342682421399838, "learning_rate": 1.319517313746065e-05, "loss": 0.378, "step": 1616 }, { "epoch": 2.287331917905166, "grad_norm": 0.21405156680424392, "learning_rate": 1.3168940188877229e-05, "loss": 0.334, "step": 1617 }, { "epoch": 2.2887473460721868, "grad_norm": 0.20941510324984786, "learning_rate": 1.3142707240293809e-05, "loss": 0.3768, "step": 1618 }, { "epoch": 2.2901627742392074, "grad_norm": 0.19213037411084616, "learning_rate": 1.311647429171039e-05, "loss": 0.337, "step": 1619 }, { "epoch": 2.291578202406228, "grad_norm": 0.2160944787649923, "learning_rate": 1.309024134312697e-05, "loss": 0.3582, "step": 1620 }, { "epoch": 2.2929936305732483, "grad_norm": 0.20595871082883357, "learning_rate": 1.3064008394543548e-05, "loss": 0.3367, "step": 1621 }, { "epoch": 2.294409058740269, "grad_norm": 0.20042645092539838, "learning_rate": 1.3037775445960127e-05, "loss": 0.3576, "step": 1622 }, { "epoch": 2.2958244869072892, "grad_norm": 0.2036670157662137, "learning_rate": 1.3011542497376705e-05, "loss": 0.3471, "step": 1623 }, { "epoch": 2.29723991507431, "grad_norm": 0.19284023614390752, "learning_rate": 1.2985309548793284e-05, "loss": 0.2943, "step": 1624 }, { "epoch": 2.2986553432413306, "grad_norm": 0.20293208468285898, "learning_rate": 1.2959076600209863e-05, "loss": 0.368, "step": 1625 }, { "epoch": 2.3000707714083513, "grad_norm": 0.3737401960483002, "learning_rate": 1.2932843651626445e-05, "loss": 0.4036, "step": 1626 }, { "epoch": 2.3014861995753715, "grad_norm": 0.19354635460810501, "learning_rate": 1.2906610703043023e-05, "loss": 0.355, "step": 1627 }, { "epoch": 2.302901627742392, "grad_norm": 0.2094545599338365, "learning_rate": 1.2880377754459602e-05, "loss": 0.3914, "step": 1628 }, { "epoch": 2.3043170559094124, "grad_norm": 2.1242623654331476, "learning_rate": 1.285414480587618e-05, "loss": 0.4319, "step": 1629 }, { "epoch": 2.305732484076433, "grad_norm": 0.22344074159287974, "learning_rate": 1.282791185729276e-05, "loss": 0.3775, "step": 1630 }, { "epoch": 2.3071479122434537, "grad_norm": 0.23710780067237636, "learning_rate": 1.280167890870934e-05, "loss": 0.3786, "step": 1631 }, { "epoch": 2.3085633404104744, "grad_norm": 0.1940770930013411, "learning_rate": 1.2775445960125918e-05, "loss": 0.3128, "step": 1632 }, { "epoch": 2.3099787685774946, "grad_norm": 0.21297071997985506, "learning_rate": 1.27492130115425e-05, "loss": 0.3421, "step": 1633 }, { "epoch": 2.3113941967445153, "grad_norm": 0.20386998779645024, "learning_rate": 1.2722980062959078e-05, "loss": 0.3531, "step": 1634 }, { "epoch": 2.3128096249115355, "grad_norm": 0.21035453901853524, "learning_rate": 1.2696747114375657e-05, "loss": 0.3247, "step": 1635 }, { "epoch": 2.314225053078556, "grad_norm": 0.20189874562181195, "learning_rate": 1.2670514165792236e-05, "loss": 0.3615, "step": 1636 }, { "epoch": 2.315640481245577, "grad_norm": 0.2194447285407467, "learning_rate": 1.2644281217208814e-05, "loss": 0.3011, "step": 1637 }, { "epoch": 2.3170559094125975, "grad_norm": 0.222431639550439, "learning_rate": 1.2618048268625393e-05, "loss": 0.3835, "step": 1638 }, { "epoch": 2.3184713375796178, "grad_norm": 0.21738559273360789, "learning_rate": 1.2591815320041971e-05, "loss": 0.3132, "step": 1639 }, { "epoch": 2.3198867657466384, "grad_norm": 0.20223700517296947, "learning_rate": 1.2565582371458553e-05, "loss": 0.3552, "step": 1640 }, { "epoch": 2.3213021939136587, "grad_norm": 0.22347595598346545, "learning_rate": 1.2539349422875132e-05, "loss": 0.3623, "step": 1641 }, { "epoch": 2.3227176220806793, "grad_norm": 0.22595395152838466, "learning_rate": 1.251311647429171e-05, "loss": 0.3811, "step": 1642 }, { "epoch": 2.3241330502477, "grad_norm": 0.2218415824217748, "learning_rate": 1.248688352570829e-05, "loss": 0.3521, "step": 1643 }, { "epoch": 2.3255484784147207, "grad_norm": 0.2280833710565805, "learning_rate": 1.246065057712487e-05, "loss": 0.4319, "step": 1644 }, { "epoch": 2.326963906581741, "grad_norm": 0.23328148577348015, "learning_rate": 1.243441762854145e-05, "loss": 0.3749, "step": 1645 }, { "epoch": 2.3283793347487616, "grad_norm": 0.22987437540955832, "learning_rate": 1.2408184679958028e-05, "loss": 0.4365, "step": 1646 }, { "epoch": 2.329794762915782, "grad_norm": 0.20803738520290288, "learning_rate": 1.2381951731374607e-05, "loss": 0.35, "step": 1647 }, { "epoch": 2.3312101910828025, "grad_norm": 0.21298538936199324, "learning_rate": 1.2355718782791187e-05, "loss": 0.333, "step": 1648 }, { "epoch": 2.332625619249823, "grad_norm": 0.25212607611317467, "learning_rate": 1.2329485834207766e-05, "loss": 0.3826, "step": 1649 }, { "epoch": 2.334041047416844, "grad_norm": 0.20052310592857706, "learning_rate": 1.2303252885624344e-05, "loss": 0.3371, "step": 1650 }, { "epoch": 2.335456475583864, "grad_norm": 0.21673405119635897, "learning_rate": 1.2277019937040923e-05, "loss": 0.3482, "step": 1651 }, { "epoch": 2.3368719037508847, "grad_norm": 0.2138680795145363, "learning_rate": 1.2250786988457503e-05, "loss": 0.3636, "step": 1652 }, { "epoch": 2.338287331917905, "grad_norm": 0.20865904901786897, "learning_rate": 1.2224554039874082e-05, "loss": 0.3158, "step": 1653 }, { "epoch": 2.3397027600849256, "grad_norm": 0.2096807726320946, "learning_rate": 1.2198321091290662e-05, "loss": 0.3882, "step": 1654 }, { "epoch": 2.3411181882519463, "grad_norm": 0.19808067950244185, "learning_rate": 1.217208814270724e-05, "loss": 0.3268, "step": 1655 }, { "epoch": 2.342533616418967, "grad_norm": 0.22133221526818114, "learning_rate": 1.2145855194123821e-05, "loss": 0.3975, "step": 1656 }, { "epoch": 2.343949044585987, "grad_norm": 0.6558735068112217, "learning_rate": 1.21196222455404e-05, "loss": 0.4503, "step": 1657 }, { "epoch": 2.345364472753008, "grad_norm": 0.19935870158714386, "learning_rate": 1.2093389296956978e-05, "loss": 0.3395, "step": 1658 }, { "epoch": 2.346779900920028, "grad_norm": 0.2068664621726759, "learning_rate": 1.2067156348373558e-05, "loss": 0.3665, "step": 1659 }, { "epoch": 2.3481953290870488, "grad_norm": 0.20479958605036505, "learning_rate": 1.2040923399790137e-05, "loss": 0.3681, "step": 1660 }, { "epoch": 2.3496107572540694, "grad_norm": 0.2065543693099381, "learning_rate": 1.2014690451206716e-05, "loss": 0.3404, "step": 1661 }, { "epoch": 2.35102618542109, "grad_norm": 0.21210242048432407, "learning_rate": 1.1988457502623296e-05, "loss": 0.3415, "step": 1662 }, { "epoch": 2.3524416135881103, "grad_norm": 0.7109075104703904, "learning_rate": 1.1962224554039874e-05, "loss": 0.4539, "step": 1663 }, { "epoch": 2.353857041755131, "grad_norm": 0.22108341730671974, "learning_rate": 1.1935991605456453e-05, "loss": 0.3713, "step": 1664 }, { "epoch": 2.3552724699221512, "grad_norm": 0.21406850500866562, "learning_rate": 1.1909758656873033e-05, "loss": 0.3303, "step": 1665 }, { "epoch": 2.356687898089172, "grad_norm": 0.22414262449386538, "learning_rate": 1.1883525708289612e-05, "loss": 0.3564, "step": 1666 }, { "epoch": 2.3581033262561926, "grad_norm": 0.2211106716490476, "learning_rate": 1.1857292759706192e-05, "loss": 0.3534, "step": 1667 }, { "epoch": 2.3595187544232132, "grad_norm": 0.2113618894415322, "learning_rate": 1.183105981112277e-05, "loss": 0.3416, "step": 1668 }, { "epoch": 2.3609341825902335, "grad_norm": 0.21893020046407727, "learning_rate": 1.1804826862539351e-05, "loss": 0.3801, "step": 1669 }, { "epoch": 2.362349610757254, "grad_norm": 0.23267909424111793, "learning_rate": 1.177859391395593e-05, "loss": 0.3942, "step": 1670 }, { "epoch": 2.3637650389242744, "grad_norm": 0.22233234372987792, "learning_rate": 1.1752360965372508e-05, "loss": 0.3664, "step": 1671 }, { "epoch": 2.365180467091295, "grad_norm": 0.20200623342526824, "learning_rate": 1.1726128016789089e-05, "loss": 0.308, "step": 1672 }, { "epoch": 2.3665958952583157, "grad_norm": 0.21336860651200418, "learning_rate": 1.1699895068205667e-05, "loss": 0.343, "step": 1673 }, { "epoch": 2.3680113234253364, "grad_norm": 0.22413229666446247, "learning_rate": 1.1673662119622246e-05, "loss": 0.3956, "step": 1674 }, { "epoch": 2.3694267515923566, "grad_norm": 0.21771474988307613, "learning_rate": 1.1647429171038824e-05, "loss": 0.3652, "step": 1675 }, { "epoch": 2.3708421797593773, "grad_norm": 0.2141067143591576, "learning_rate": 1.1621196222455405e-05, "loss": 0.3405, "step": 1676 }, { "epoch": 2.3722576079263975, "grad_norm": 0.23877770632497497, "learning_rate": 1.1594963273871983e-05, "loss": 0.4238, "step": 1677 }, { "epoch": 2.373673036093418, "grad_norm": 0.22336963831812956, "learning_rate": 1.1568730325288562e-05, "loss": 0.4112, "step": 1678 }, { "epoch": 2.375088464260439, "grad_norm": 0.20451117319496986, "learning_rate": 1.1542497376705142e-05, "loss": 0.3647, "step": 1679 }, { "epoch": 2.3765038924274595, "grad_norm": 0.20754787330919905, "learning_rate": 1.1516264428121722e-05, "loss": 0.3514, "step": 1680 }, { "epoch": 2.3779193205944797, "grad_norm": 0.2161438866209889, "learning_rate": 1.1490031479538301e-05, "loss": 0.3972, "step": 1681 }, { "epoch": 2.3793347487615004, "grad_norm": 0.18714775741431736, "learning_rate": 1.146379853095488e-05, "loss": 0.3199, "step": 1682 }, { "epoch": 2.3807501769285206, "grad_norm": 0.1939274793117906, "learning_rate": 1.143756558237146e-05, "loss": 0.3408, "step": 1683 }, { "epoch": 2.3821656050955413, "grad_norm": 0.2088588496866585, "learning_rate": 1.1411332633788038e-05, "loss": 0.3597, "step": 1684 }, { "epoch": 2.383581033262562, "grad_norm": 0.19274140483291136, "learning_rate": 1.1385099685204617e-05, "loss": 0.286, "step": 1685 }, { "epoch": 2.3849964614295827, "grad_norm": 0.19465489196136257, "learning_rate": 1.1358866736621197e-05, "loss": 0.3407, "step": 1686 }, { "epoch": 2.386411889596603, "grad_norm": 0.19358748526673272, "learning_rate": 1.1332633788037776e-05, "loss": 0.3281, "step": 1687 }, { "epoch": 2.3878273177636236, "grad_norm": 0.22265113555366345, "learning_rate": 1.1306400839454354e-05, "loss": 0.4056, "step": 1688 }, { "epoch": 2.389242745930644, "grad_norm": 0.20308464928274864, "learning_rate": 1.1280167890870933e-05, "loss": 0.3679, "step": 1689 }, { "epoch": 2.3906581740976645, "grad_norm": 0.20427851173344422, "learning_rate": 1.1253934942287513e-05, "loss": 0.3826, "step": 1690 }, { "epoch": 2.392073602264685, "grad_norm": 0.20932500576312751, "learning_rate": 1.1227701993704094e-05, "loss": 0.3877, "step": 1691 }, { "epoch": 2.393489030431706, "grad_norm": 0.2100695235002279, "learning_rate": 1.1201469045120672e-05, "loss": 0.3874, "step": 1692 }, { "epoch": 2.394904458598726, "grad_norm": 0.19773857226605634, "learning_rate": 1.1175236096537252e-05, "loss": 0.3473, "step": 1693 }, { "epoch": 2.3963198867657467, "grad_norm": 0.21235501984385707, "learning_rate": 1.1149003147953831e-05, "loss": 0.3579, "step": 1694 }, { "epoch": 2.397735314932767, "grad_norm": 0.2235670395689437, "learning_rate": 1.112277019937041e-05, "loss": 0.3766, "step": 1695 }, { "epoch": 2.3991507430997876, "grad_norm": 0.21562860359515665, "learning_rate": 1.1096537250786988e-05, "loss": 0.3839, "step": 1696 }, { "epoch": 2.4005661712668083, "grad_norm": 0.20556522359943483, "learning_rate": 1.1070304302203569e-05, "loss": 0.3474, "step": 1697 }, { "epoch": 2.401981599433829, "grad_norm": 0.21632102454011515, "learning_rate": 1.1044071353620147e-05, "loss": 0.3759, "step": 1698 }, { "epoch": 2.403397027600849, "grad_norm": 0.22107312323409514, "learning_rate": 1.1017838405036726e-05, "loss": 0.3968, "step": 1699 }, { "epoch": 2.40481245576787, "grad_norm": 0.20713933717369834, "learning_rate": 1.0991605456453306e-05, "loss": 0.3699, "step": 1700 }, { "epoch": 2.40622788393489, "grad_norm": 0.2117918962164715, "learning_rate": 1.0965372507869885e-05, "loss": 0.3564, "step": 1701 }, { "epoch": 2.4076433121019107, "grad_norm": 0.19884508709349702, "learning_rate": 1.0939139559286463e-05, "loss": 0.3396, "step": 1702 }, { "epoch": 2.4090587402689314, "grad_norm": 0.20445555808171254, "learning_rate": 1.0912906610703043e-05, "loss": 0.3331, "step": 1703 }, { "epoch": 2.410474168435952, "grad_norm": 0.20438554619989838, "learning_rate": 1.0886673662119624e-05, "loss": 0.3561, "step": 1704 }, { "epoch": 2.4118895966029723, "grad_norm": 0.21961427283137555, "learning_rate": 1.0860440713536202e-05, "loss": 0.4196, "step": 1705 }, { "epoch": 2.413305024769993, "grad_norm": 0.21068372988168393, "learning_rate": 1.0834207764952781e-05, "loss": 0.4156, "step": 1706 }, { "epoch": 2.414720452937013, "grad_norm": 0.20171643481038465, "learning_rate": 1.0807974816369361e-05, "loss": 0.3708, "step": 1707 }, { "epoch": 2.416135881104034, "grad_norm": 0.21885197735233844, "learning_rate": 1.078174186778594e-05, "loss": 0.3993, "step": 1708 }, { "epoch": 2.4175513092710545, "grad_norm": 0.21675789921133995, "learning_rate": 1.0755508919202518e-05, "loss": 0.4063, "step": 1709 }, { "epoch": 2.418966737438075, "grad_norm": 3.0522202221772288, "learning_rate": 1.0729275970619099e-05, "loss": 0.4202, "step": 1710 }, { "epoch": 2.4203821656050954, "grad_norm": 0.2140766759354188, "learning_rate": 1.0703043022035677e-05, "loss": 0.4007, "step": 1711 }, { "epoch": 2.421797593772116, "grad_norm": 0.2084942514923414, "learning_rate": 1.0676810073452256e-05, "loss": 0.3396, "step": 1712 }, { "epoch": 2.4232130219391363, "grad_norm": 0.2149211224899716, "learning_rate": 1.0650577124868834e-05, "loss": 0.3422, "step": 1713 }, { "epoch": 2.424628450106157, "grad_norm": 0.2106202261748985, "learning_rate": 1.0624344176285415e-05, "loss": 0.3994, "step": 1714 }, { "epoch": 2.4260438782731777, "grad_norm": 0.22730032299583158, "learning_rate": 1.0598111227701995e-05, "loss": 0.3648, "step": 1715 }, { "epoch": 2.4274593064401984, "grad_norm": 0.22643649889098394, "learning_rate": 1.0571878279118574e-05, "loss": 0.3391, "step": 1716 }, { "epoch": 2.4288747346072186, "grad_norm": 0.23342791214204056, "learning_rate": 1.0545645330535154e-05, "loss": 0.3687, "step": 1717 }, { "epoch": 2.4302901627742393, "grad_norm": 0.20689885848622794, "learning_rate": 1.0519412381951733e-05, "loss": 0.3757, "step": 1718 }, { "epoch": 2.4317055909412595, "grad_norm": 0.210687923926889, "learning_rate": 1.0493179433368311e-05, "loss": 0.3501, "step": 1719 }, { "epoch": 2.43312101910828, "grad_norm": 0.22986089011271102, "learning_rate": 1.046694648478489e-05, "loss": 0.3836, "step": 1720 }, { "epoch": 2.434536447275301, "grad_norm": 0.22744764307215082, "learning_rate": 1.044071353620147e-05, "loss": 0.3898, "step": 1721 }, { "epoch": 2.4359518754423215, "grad_norm": 0.21194994175423806, "learning_rate": 1.0414480587618049e-05, "loss": 0.334, "step": 1722 }, { "epoch": 2.4373673036093417, "grad_norm": 0.22018978375373083, "learning_rate": 1.0388247639034627e-05, "loss": 0.4329, "step": 1723 }, { "epoch": 2.4387827317763624, "grad_norm": 0.19353091165168484, "learning_rate": 1.0362014690451207e-05, "loss": 0.3603, "step": 1724 }, { "epoch": 2.4401981599433826, "grad_norm": 0.21586391650702627, "learning_rate": 1.0335781741867786e-05, "loss": 0.4187, "step": 1725 }, { "epoch": 2.4416135881104033, "grad_norm": 0.22306618672285558, "learning_rate": 1.0309548793284365e-05, "loss": 0.4009, "step": 1726 }, { "epoch": 2.443029016277424, "grad_norm": 0.2102750167761404, "learning_rate": 1.0283315844700945e-05, "loss": 0.3828, "step": 1727 }, { "epoch": 2.4444444444444446, "grad_norm": 0.18881815647673936, "learning_rate": 1.0257082896117525e-05, "loss": 0.3189, "step": 1728 }, { "epoch": 2.445859872611465, "grad_norm": 0.20862013453932002, "learning_rate": 1.0230849947534104e-05, "loss": 0.3527, "step": 1729 }, { "epoch": 2.4472753007784855, "grad_norm": 0.21210986825118294, "learning_rate": 1.0204616998950682e-05, "loss": 0.3517, "step": 1730 }, { "epoch": 2.4486907289455058, "grad_norm": 0.21429771552588056, "learning_rate": 1.0178384050367263e-05, "loss": 0.3887, "step": 1731 }, { "epoch": 2.4501061571125264, "grad_norm": 0.21001121174052864, "learning_rate": 1.0152151101783841e-05, "loss": 0.3587, "step": 1732 }, { "epoch": 2.451521585279547, "grad_norm": 0.22002696555298376, "learning_rate": 1.012591815320042e-05, "loss": 0.3975, "step": 1733 }, { "epoch": 2.452937013446568, "grad_norm": 0.22051929953678207, "learning_rate": 1.0099685204616998e-05, "loss": 0.3839, "step": 1734 }, { "epoch": 2.454352441613588, "grad_norm": 0.20142600399295918, "learning_rate": 1.0073452256033579e-05, "loss": 0.338, "step": 1735 }, { "epoch": 2.4557678697806087, "grad_norm": 0.2174560492510006, "learning_rate": 1.0047219307450157e-05, "loss": 0.4005, "step": 1736 }, { "epoch": 2.457183297947629, "grad_norm": 0.22565669021953805, "learning_rate": 1.0020986358866736e-05, "loss": 0.3642, "step": 1737 }, { "epoch": 2.4585987261146496, "grad_norm": 0.222555475085274, "learning_rate": 9.994753410283316e-06, "loss": 0.3832, "step": 1738 }, { "epoch": 2.4600141542816703, "grad_norm": 0.20558462731843344, "learning_rate": 9.968520461699895e-06, "loss": 0.3346, "step": 1739 }, { "epoch": 2.461429582448691, "grad_norm": 0.2176380490810765, "learning_rate": 9.942287513116475e-06, "loss": 0.4208, "step": 1740 }, { "epoch": 2.462845010615711, "grad_norm": 0.19524678242958587, "learning_rate": 9.916054564533055e-06, "loss": 0.3165, "step": 1741 }, { "epoch": 2.464260438782732, "grad_norm": 0.19314190115388655, "learning_rate": 9.889821615949634e-06, "loss": 0.3311, "step": 1742 }, { "epoch": 2.4656758669497525, "grad_norm": 0.20321068255373417, "learning_rate": 9.863588667366213e-06, "loss": 0.3792, "step": 1743 }, { "epoch": 2.4670912951167727, "grad_norm": 0.21499480640234953, "learning_rate": 9.837355718782791e-06, "loss": 0.3547, "step": 1744 }, { "epoch": 2.4685067232837934, "grad_norm": 0.20409540141795263, "learning_rate": 9.811122770199371e-06, "loss": 0.3324, "step": 1745 }, { "epoch": 2.469922151450814, "grad_norm": 0.2233290959062587, "learning_rate": 9.78488982161595e-06, "loss": 0.3764, "step": 1746 }, { "epoch": 2.4713375796178343, "grad_norm": 0.20236083142308806, "learning_rate": 9.758656873032529e-06, "loss": 0.3677, "step": 1747 }, { "epoch": 2.472753007784855, "grad_norm": 0.20234208948922255, "learning_rate": 9.732423924449109e-06, "loss": 0.3755, "step": 1748 }, { "epoch": 2.4741684359518756, "grad_norm": 0.20256254108287847, "learning_rate": 9.706190975865687e-06, "loss": 0.3473, "step": 1749 }, { "epoch": 2.475583864118896, "grad_norm": 0.2188760149731594, "learning_rate": 9.679958027282266e-06, "loss": 0.3543, "step": 1750 }, { "epoch": 2.4769992922859165, "grad_norm": 0.2872207946643241, "learning_rate": 9.653725078698846e-06, "loss": 0.3412, "step": 1751 }, { "epoch": 2.478414720452937, "grad_norm": 0.19773186928805295, "learning_rate": 9.627492130115427e-06, "loss": 0.3311, "step": 1752 }, { "epoch": 2.4798301486199574, "grad_norm": 0.2109865330811176, "learning_rate": 9.601259181532005e-06, "loss": 0.364, "step": 1753 }, { "epoch": 2.481245576786978, "grad_norm": 0.20385787880377218, "learning_rate": 9.575026232948584e-06, "loss": 0.3547, "step": 1754 }, { "epoch": 2.4826610049539988, "grad_norm": 0.45454058344335446, "learning_rate": 9.548793284365164e-06, "loss": 0.3761, "step": 1755 }, { "epoch": 2.484076433121019, "grad_norm": 0.20069144673754255, "learning_rate": 9.522560335781743e-06, "loss": 0.3777, "step": 1756 }, { "epoch": 2.4854918612880397, "grad_norm": 0.20144636318040068, "learning_rate": 9.496327387198321e-06, "loss": 0.362, "step": 1757 }, { "epoch": 2.4869072894550603, "grad_norm": 0.21818576258038522, "learning_rate": 9.4700944386149e-06, "loss": 0.4137, "step": 1758 }, { "epoch": 2.4883227176220806, "grad_norm": 0.23390860882528833, "learning_rate": 9.44386149003148e-06, "loss": 0.3677, "step": 1759 }, { "epoch": 2.4897381457891012, "grad_norm": 0.21303608042166824, "learning_rate": 9.417628541448059e-06, "loss": 0.3805, "step": 1760 }, { "epoch": 2.491153573956122, "grad_norm": 0.21648637026495848, "learning_rate": 9.391395592864637e-06, "loss": 0.3957, "step": 1761 }, { "epoch": 2.492569002123142, "grad_norm": 0.21470889424752063, "learning_rate": 9.365162644281218e-06, "loss": 0.3958, "step": 1762 }, { "epoch": 2.493984430290163, "grad_norm": 0.21551369414955032, "learning_rate": 9.338929695697796e-06, "loss": 0.4112, "step": 1763 }, { "epoch": 2.4953998584571835, "grad_norm": 0.20447833108692218, "learning_rate": 9.312696747114377e-06, "loss": 0.3472, "step": 1764 }, { "epoch": 2.4968152866242037, "grad_norm": 0.20586409741257014, "learning_rate": 9.286463798530955e-06, "loss": 0.3504, "step": 1765 }, { "epoch": 2.4982307147912244, "grad_norm": 0.21323194372082052, "learning_rate": 9.260230849947535e-06, "loss": 0.3675, "step": 1766 }, { "epoch": 2.499646142958245, "grad_norm": 0.2070312822291339, "learning_rate": 9.233997901364114e-06, "loss": 0.3825, "step": 1767 }, { "epoch": 2.5010615711252653, "grad_norm": 0.2151895724043526, "learning_rate": 9.207764952780693e-06, "loss": 0.352, "step": 1768 }, { "epoch": 2.502476999292286, "grad_norm": 0.22400699091609746, "learning_rate": 9.181532004197273e-06, "loss": 0.4242, "step": 1769 }, { "epoch": 2.5038924274593066, "grad_norm": 0.20545598835648918, "learning_rate": 9.155299055613851e-06, "loss": 0.3579, "step": 1770 }, { "epoch": 2.505307855626327, "grad_norm": 0.20089201585791153, "learning_rate": 9.12906610703043e-06, "loss": 0.3429, "step": 1771 }, { "epoch": 2.5067232837933475, "grad_norm": 0.19904600554390434, "learning_rate": 9.10283315844701e-06, "loss": 0.3482, "step": 1772 }, { "epoch": 2.5081387119603678, "grad_norm": 0.20136909906575673, "learning_rate": 9.076600209863589e-06, "loss": 0.3749, "step": 1773 }, { "epoch": 2.5095541401273884, "grad_norm": 0.2039054505197896, "learning_rate": 9.050367261280168e-06, "loss": 0.3596, "step": 1774 }, { "epoch": 2.510969568294409, "grad_norm": 0.20871768692305176, "learning_rate": 9.024134312696748e-06, "loss": 0.3513, "step": 1775 }, { "epoch": 2.5123849964614298, "grad_norm": 0.1922506056528137, "learning_rate": 8.997901364113328e-06, "loss": 0.3388, "step": 1776 }, { "epoch": 2.51380042462845, "grad_norm": 4.0876903662085144, "learning_rate": 8.971668415529907e-06, "loss": 0.3932, "step": 1777 }, { "epoch": 2.5152158527954707, "grad_norm": 0.22063988754657995, "learning_rate": 8.945435466946485e-06, "loss": 0.3819, "step": 1778 }, { "epoch": 2.516631280962491, "grad_norm": 0.2182736323794303, "learning_rate": 8.919202518363066e-06, "loss": 0.3879, "step": 1779 }, { "epoch": 2.5180467091295116, "grad_norm": 0.20526743907609737, "learning_rate": 8.892969569779644e-06, "loss": 0.3797, "step": 1780 }, { "epoch": 2.5194621372965322, "grad_norm": 0.20693527123319747, "learning_rate": 8.866736621196223e-06, "loss": 0.3765, "step": 1781 }, { "epoch": 2.520877565463553, "grad_norm": 0.21418203449223625, "learning_rate": 8.840503672612801e-06, "loss": 0.3959, "step": 1782 }, { "epoch": 2.522292993630573, "grad_norm": 0.199865548007296, "learning_rate": 8.814270724029382e-06, "loss": 0.3331, "step": 1783 }, { "epoch": 2.523708421797594, "grad_norm": 0.19768360191047332, "learning_rate": 8.78803777544596e-06, "loss": 0.3387, "step": 1784 }, { "epoch": 2.525123849964614, "grad_norm": 0.220405625409773, "learning_rate": 8.761804826862539e-06, "loss": 0.3986, "step": 1785 }, { "epoch": 2.5265392781316347, "grad_norm": 0.23298794241115453, "learning_rate": 8.735571878279119e-06, "loss": 0.3726, "step": 1786 }, { "epoch": 2.5279547062986554, "grad_norm": 0.20845940872746585, "learning_rate": 8.709338929695698e-06, "loss": 0.3686, "step": 1787 }, { "epoch": 2.529370134465676, "grad_norm": 0.20739265555839045, "learning_rate": 8.683105981112278e-06, "loss": 0.341, "step": 1788 }, { "epoch": 2.5307855626326963, "grad_norm": 0.23387833738932326, "learning_rate": 8.656873032528857e-06, "loss": 0.4018, "step": 1789 }, { "epoch": 2.532200990799717, "grad_norm": 0.1991531574710069, "learning_rate": 8.630640083945437e-06, "loss": 0.3642, "step": 1790 }, { "epoch": 2.533616418966737, "grad_norm": 0.21978858324149222, "learning_rate": 8.604407135362015e-06, "loss": 0.3793, "step": 1791 }, { "epoch": 2.535031847133758, "grad_norm": 0.19771421985469945, "learning_rate": 8.578174186778594e-06, "loss": 0.3492, "step": 1792 }, { "epoch": 2.5364472753007785, "grad_norm": 0.21186116442046274, "learning_rate": 8.551941238195174e-06, "loss": 0.384, "step": 1793 }, { "epoch": 2.537862703467799, "grad_norm": 0.2159890538489202, "learning_rate": 8.525708289611753e-06, "loss": 0.4236, "step": 1794 }, { "epoch": 2.5392781316348194, "grad_norm": 0.2133088128432072, "learning_rate": 8.499475341028331e-06, "loss": 0.3928, "step": 1795 }, { "epoch": 2.54069355980184, "grad_norm": 0.20400675417562208, "learning_rate": 8.47324239244491e-06, "loss": 0.397, "step": 1796 }, { "epoch": 2.5421089879688603, "grad_norm": 0.20894650188281783, "learning_rate": 8.44700944386149e-06, "loss": 0.3664, "step": 1797 }, { "epoch": 2.543524416135881, "grad_norm": 0.20857055765655338, "learning_rate": 8.420776495278069e-06, "loss": 0.3877, "step": 1798 }, { "epoch": 2.5449398443029017, "grad_norm": 0.23029162763104252, "learning_rate": 8.39454354669465e-06, "loss": 0.4219, "step": 1799 }, { "epoch": 2.5463552724699223, "grad_norm": 0.21279556644149866, "learning_rate": 8.368310598111228e-06, "loss": 0.3522, "step": 1800 }, { "epoch": 2.5477707006369426, "grad_norm": 0.19956081610754542, "learning_rate": 8.342077649527808e-06, "loss": 0.3447, "step": 1801 }, { "epoch": 2.5491861288039632, "grad_norm": 0.21153748180846707, "learning_rate": 8.315844700944387e-06, "loss": 0.3659, "step": 1802 }, { "epoch": 2.5506015569709835, "grad_norm": 0.2064243717597128, "learning_rate": 8.289611752360965e-06, "loss": 0.3213, "step": 1803 }, { "epoch": 2.552016985138004, "grad_norm": 0.210979948172296, "learning_rate": 8.263378803777546e-06, "loss": 0.3737, "step": 1804 }, { "epoch": 2.553432413305025, "grad_norm": 0.20281319829998443, "learning_rate": 8.237145855194124e-06, "loss": 0.3581, "step": 1805 }, { "epoch": 2.5548478414720455, "grad_norm": 0.21576724260988553, "learning_rate": 8.210912906610703e-06, "loss": 0.3667, "step": 1806 }, { "epoch": 2.5562632696390657, "grad_norm": 0.19132759016872786, "learning_rate": 8.184679958027283e-06, "loss": 0.3485, "step": 1807 }, { "epoch": 2.5576786978060864, "grad_norm": 0.1974784858322557, "learning_rate": 8.158447009443862e-06, "loss": 0.3328, "step": 1808 }, { "epoch": 2.5590941259731066, "grad_norm": 0.20640998486691128, "learning_rate": 8.13221406086044e-06, "loss": 0.3642, "step": 1809 }, { "epoch": 2.5605095541401273, "grad_norm": 0.2039268310105597, "learning_rate": 8.10598111227702e-06, "loss": 0.3631, "step": 1810 }, { "epoch": 2.561924982307148, "grad_norm": 0.19215765539590668, "learning_rate": 8.079748163693599e-06, "loss": 0.3382, "step": 1811 }, { "epoch": 2.5633404104741686, "grad_norm": 0.26656039783281915, "learning_rate": 8.05351521511018e-06, "loss": 0.3783, "step": 1812 }, { "epoch": 2.564755838641189, "grad_norm": 0.21493592090680055, "learning_rate": 8.027282266526758e-06, "loss": 0.3503, "step": 1813 }, { "epoch": 2.5661712668082095, "grad_norm": 0.21040119739922322, "learning_rate": 8.001049317943338e-06, "loss": 0.347, "step": 1814 }, { "epoch": 2.5675866949752297, "grad_norm": 0.21060367789203188, "learning_rate": 7.974816369359917e-06, "loss": 0.4028, "step": 1815 }, { "epoch": 2.5690021231422504, "grad_norm": 0.20945122187457563, "learning_rate": 7.948583420776495e-06, "loss": 0.3285, "step": 1816 }, { "epoch": 2.570417551309271, "grad_norm": 0.2087227917081225, "learning_rate": 7.922350472193076e-06, "loss": 0.354, "step": 1817 }, { "epoch": 2.5718329794762917, "grad_norm": 0.2095089045850525, "learning_rate": 7.896117523609654e-06, "loss": 0.3623, "step": 1818 }, { "epoch": 2.573248407643312, "grad_norm": 0.20543371792968124, "learning_rate": 7.869884575026233e-06, "loss": 0.3671, "step": 1819 }, { "epoch": 2.5746638358103326, "grad_norm": 0.19734594910360248, "learning_rate": 7.843651626442811e-06, "loss": 0.3265, "step": 1820 }, { "epoch": 2.576079263977353, "grad_norm": 0.2074718265878983, "learning_rate": 7.817418677859392e-06, "loss": 0.363, "step": 1821 }, { "epoch": 2.5774946921443735, "grad_norm": 0.1994386092622095, "learning_rate": 7.79118572927597e-06, "loss": 0.378, "step": 1822 }, { "epoch": 2.578910120311394, "grad_norm": 0.2108485509228124, "learning_rate": 7.76495278069255e-06, "loss": 0.4183, "step": 1823 }, { "epoch": 2.580325548478415, "grad_norm": 0.2013484746339703, "learning_rate": 7.73871983210913e-06, "loss": 0.3305, "step": 1824 }, { "epoch": 2.581740976645435, "grad_norm": 0.2054314513051612, "learning_rate": 7.71248688352571e-06, "loss": 0.3454, "step": 1825 }, { "epoch": 2.583156404812456, "grad_norm": 0.20955117943330293, "learning_rate": 7.686253934942288e-06, "loss": 0.344, "step": 1826 }, { "epoch": 2.5845718329794765, "grad_norm": 0.1973210553919038, "learning_rate": 7.660020986358867e-06, "loss": 0.3204, "step": 1827 }, { "epoch": 2.5859872611464967, "grad_norm": 0.2171924649126436, "learning_rate": 7.633788037775447e-06, "loss": 0.3737, "step": 1828 }, { "epoch": 2.5874026893135174, "grad_norm": 0.21712977233946879, "learning_rate": 7.607555089192026e-06, "loss": 0.3788, "step": 1829 }, { "epoch": 2.588818117480538, "grad_norm": 0.2078317502026498, "learning_rate": 7.581322140608604e-06, "loss": 0.4006, "step": 1830 }, { "epoch": 2.5902335456475583, "grad_norm": 0.19809729361386832, "learning_rate": 7.5550891920251845e-06, "loss": 0.3704, "step": 1831 }, { "epoch": 2.591648973814579, "grad_norm": 0.1968896281185804, "learning_rate": 7.528856243441763e-06, "loss": 0.3082, "step": 1832 }, { "epoch": 2.5930644019815996, "grad_norm": 0.20981367025717132, "learning_rate": 7.5026232948583425e-06, "loss": 0.3608, "step": 1833 }, { "epoch": 2.59447983014862, "grad_norm": 0.20797161093836125, "learning_rate": 7.476390346274921e-06, "loss": 0.3362, "step": 1834 }, { "epoch": 2.5958952583156405, "grad_norm": 0.21261247348837317, "learning_rate": 7.450157397691501e-06, "loss": 0.3814, "step": 1835 }, { "epoch": 2.597310686482661, "grad_norm": 0.19882968356020506, "learning_rate": 7.42392444910808e-06, "loss": 0.343, "step": 1836 }, { "epoch": 2.5987261146496814, "grad_norm": 0.2103031596403086, "learning_rate": 7.3976915005246586e-06, "loss": 0.3829, "step": 1837 }, { "epoch": 2.600141542816702, "grad_norm": 0.19150021569844553, "learning_rate": 7.371458551941239e-06, "loss": 0.3375, "step": 1838 }, { "epoch": 2.6015569709837227, "grad_norm": 0.2011174569283403, "learning_rate": 7.345225603357818e-06, "loss": 0.3181, "step": 1839 }, { "epoch": 2.602972399150743, "grad_norm": 0.21977968424911656, "learning_rate": 7.318992654774397e-06, "loss": 0.3879, "step": 1840 }, { "epoch": 2.6043878273177636, "grad_norm": 0.2020175254696623, "learning_rate": 7.2927597061909755e-06, "loss": 0.3404, "step": 1841 }, { "epoch": 2.6058032554847843, "grad_norm": 0.19567152319659617, "learning_rate": 7.266526757607556e-06, "loss": 0.3191, "step": 1842 }, { "epoch": 2.6072186836518045, "grad_norm": 0.18597323556834847, "learning_rate": 7.240293809024134e-06, "loss": 0.2806, "step": 1843 }, { "epoch": 2.608634111818825, "grad_norm": 0.21321812440621554, "learning_rate": 7.214060860440714e-06, "loss": 0.4131, "step": 1844 }, { "epoch": 2.610049539985846, "grad_norm": 0.21088310897402524, "learning_rate": 7.187827911857293e-06, "loss": 0.3515, "step": 1845 }, { "epoch": 2.611464968152866, "grad_norm": 0.20528663135935263, "learning_rate": 7.161594963273873e-06, "loss": 0.3613, "step": 1846 }, { "epoch": 2.6128803963198868, "grad_norm": 0.20929437136431123, "learning_rate": 7.135362014690451e-06, "loss": 0.357, "step": 1847 }, { "epoch": 2.6142958244869074, "grad_norm": 0.22328273062400936, "learning_rate": 7.1091290661070315e-06, "loss": 0.4107, "step": 1848 }, { "epoch": 2.6157112526539277, "grad_norm": 0.2037256309767581, "learning_rate": 7.08289611752361e-06, "loss": 0.3576, "step": 1849 }, { "epoch": 2.6171266808209483, "grad_norm": 0.2243164183301271, "learning_rate": 7.056663168940189e-06, "loss": 0.4082, "step": 1850 }, { "epoch": 2.618542108987969, "grad_norm": 0.2186639214858174, "learning_rate": 7.030430220356768e-06, "loss": 0.3933, "step": 1851 }, { "epoch": 2.6199575371549892, "grad_norm": 0.2159361736800281, "learning_rate": 7.0041972717733484e-06, "loss": 0.3633, "step": 1852 }, { "epoch": 2.62137296532201, "grad_norm": 1.6506203295835573, "learning_rate": 6.977964323189927e-06, "loss": 0.4659, "step": 1853 }, { "epoch": 2.6227883934890306, "grad_norm": 0.19835782691780177, "learning_rate": 6.951731374606506e-06, "loss": 0.3411, "step": 1854 }, { "epoch": 2.624203821656051, "grad_norm": 0.2218844143299372, "learning_rate": 6.925498426023086e-06, "loss": 0.415, "step": 1855 }, { "epoch": 2.6256192498230715, "grad_norm": 0.21714669812635473, "learning_rate": 6.8992654774396645e-06, "loss": 0.3731, "step": 1856 }, { "epoch": 2.627034677990092, "grad_norm": 0.22171926703943656, "learning_rate": 6.873032528856244e-06, "loss": 0.373, "step": 1857 }, { "epoch": 2.6284501061571124, "grad_norm": 0.2099085807003911, "learning_rate": 6.8467995802728225e-06, "loss": 0.3741, "step": 1858 }, { "epoch": 2.629865534324133, "grad_norm": 0.19496236614733134, "learning_rate": 6.820566631689403e-06, "loss": 0.3509, "step": 1859 }, { "epoch": 2.6312809624911537, "grad_norm": 0.1939749683125769, "learning_rate": 6.794333683105981e-06, "loss": 0.3255, "step": 1860 }, { "epoch": 2.632696390658174, "grad_norm": 0.21205398195214104, "learning_rate": 6.76810073452256e-06, "loss": 0.3659, "step": 1861 }, { "epoch": 2.6341118188251946, "grad_norm": 0.2032255971712124, "learning_rate": 6.74186778593914e-06, "loss": 0.3816, "step": 1862 }, { "epoch": 2.6355272469922153, "grad_norm": 0.2101835507755894, "learning_rate": 6.71563483735572e-06, "loss": 0.3715, "step": 1863 }, { "epoch": 2.6369426751592355, "grad_norm": 0.20345481767709614, "learning_rate": 6.689401888772298e-06, "loss": 0.3676, "step": 1864 }, { "epoch": 2.638358103326256, "grad_norm": 0.20906313390295878, "learning_rate": 6.663168940188877e-06, "loss": 0.3802, "step": 1865 }, { "epoch": 2.639773531493277, "grad_norm": 0.18629132905947415, "learning_rate": 6.636935991605457e-06, "loss": 0.3209, "step": 1866 }, { "epoch": 2.641188959660297, "grad_norm": 0.2152572803841356, "learning_rate": 6.610703043022036e-06, "loss": 0.3399, "step": 1867 }, { "epoch": 2.6426043878273178, "grad_norm": 0.19539056047254483, "learning_rate": 6.584470094438614e-06, "loss": 0.3122, "step": 1868 }, { "epoch": 2.6440198159943384, "grad_norm": 0.20202618202655612, "learning_rate": 6.558237145855195e-06, "loss": 0.3608, "step": 1869 }, { "epoch": 2.6454352441613587, "grad_norm": 0.21368434378231554, "learning_rate": 6.532004197271774e-06, "loss": 0.3408, "step": 1870 }, { "epoch": 2.6468506723283793, "grad_norm": 0.2031122859886185, "learning_rate": 6.505771248688353e-06, "loss": 0.3833, "step": 1871 }, { "epoch": 2.6482661004954, "grad_norm": 0.20061405029326895, "learning_rate": 6.479538300104931e-06, "loss": 0.3759, "step": 1872 }, { "epoch": 2.6496815286624202, "grad_norm": 0.20098303425728567, "learning_rate": 6.4533053515215116e-06, "loss": 0.3467, "step": 1873 }, { "epoch": 2.651096956829441, "grad_norm": 0.20278629578617524, "learning_rate": 6.42707240293809e-06, "loss": 0.3171, "step": 1874 }, { "epoch": 2.6525123849964616, "grad_norm": 0.20242009886724874, "learning_rate": 6.40083945435467e-06, "loss": 0.366, "step": 1875 }, { "epoch": 2.653927813163482, "grad_norm": 0.20238146901560336, "learning_rate": 6.37460650577125e-06, "loss": 0.3786, "step": 1876 }, { "epoch": 2.6553432413305025, "grad_norm": 0.18806823919909846, "learning_rate": 6.3483735571878285e-06, "loss": 0.3383, "step": 1877 }, { "epoch": 2.656758669497523, "grad_norm": 0.20708956711483964, "learning_rate": 6.322140608604407e-06, "loss": 0.391, "step": 1878 }, { "epoch": 2.6581740976645434, "grad_norm": 0.20770833502603264, "learning_rate": 6.295907660020986e-06, "loss": 0.3146, "step": 1879 }, { "epoch": 2.659589525831564, "grad_norm": 0.21822289469082637, "learning_rate": 6.269674711437566e-06, "loss": 0.3934, "step": 1880 }, { "epoch": 2.6610049539985847, "grad_norm": 0.2189330109283495, "learning_rate": 6.243441762854145e-06, "loss": 0.3635, "step": 1881 }, { "epoch": 2.662420382165605, "grad_norm": 0.2025224096693157, "learning_rate": 6.217208814270725e-06, "loss": 0.3622, "step": 1882 }, { "epoch": 2.6638358103326256, "grad_norm": 0.20391333536797676, "learning_rate": 6.190975865687303e-06, "loss": 0.3605, "step": 1883 }, { "epoch": 2.6652512384996463, "grad_norm": 0.22300673855216288, "learning_rate": 6.164742917103883e-06, "loss": 0.4051, "step": 1884 }, { "epoch": 2.6666666666666665, "grad_norm": 0.1949488754489201, "learning_rate": 6.1385099685204614e-06, "loss": 0.3318, "step": 1885 }, { "epoch": 2.668082094833687, "grad_norm": 0.20257754211433252, "learning_rate": 6.112277019937041e-06, "loss": 0.3481, "step": 1886 }, { "epoch": 2.669497523000708, "grad_norm": 0.21284586193369576, "learning_rate": 6.08604407135362e-06, "loss": 0.3693, "step": 1887 }, { "epoch": 2.670912951167728, "grad_norm": 0.20975573812946513, "learning_rate": 6.0598111227702e-06, "loss": 0.3945, "step": 1888 }, { "epoch": 2.6723283793347488, "grad_norm": 0.1962050177392725, "learning_rate": 6.033578174186779e-06, "loss": 0.3527, "step": 1889 }, { "epoch": 2.6737438075017694, "grad_norm": 0.20700678452506255, "learning_rate": 6.007345225603358e-06, "loss": 0.3703, "step": 1890 }, { "epoch": 2.6751592356687897, "grad_norm": 0.21574145342448686, "learning_rate": 5.981112277019937e-06, "loss": 0.3899, "step": 1891 }, { "epoch": 2.6765746638358103, "grad_norm": 0.21797501994034846, "learning_rate": 5.954879328436517e-06, "loss": 0.3867, "step": 1892 }, { "epoch": 2.677990092002831, "grad_norm": 0.20848947685547806, "learning_rate": 5.928646379853096e-06, "loss": 0.388, "step": 1893 }, { "epoch": 2.6794055201698512, "grad_norm": 0.19866839410791579, "learning_rate": 5.9024134312696755e-06, "loss": 0.3655, "step": 1894 }, { "epoch": 2.680820948336872, "grad_norm": 0.1981411367729309, "learning_rate": 5.876180482686254e-06, "loss": 0.337, "step": 1895 }, { "epoch": 2.6822363765038926, "grad_norm": 0.195120217473148, "learning_rate": 5.8499475341028336e-06, "loss": 0.3337, "step": 1896 }, { "epoch": 2.683651804670913, "grad_norm": 0.19846949669318745, "learning_rate": 5.823714585519412e-06, "loss": 0.339, "step": 1897 }, { "epoch": 2.6850672328379335, "grad_norm": 0.5980099051970161, "learning_rate": 5.797481636935992e-06, "loss": 0.3733, "step": 1898 }, { "epoch": 2.686482661004954, "grad_norm": 0.21711699289634304, "learning_rate": 5.771248688352571e-06, "loss": 0.383, "step": 1899 }, { "epoch": 2.6878980891719744, "grad_norm": 0.2093369096634055, "learning_rate": 5.7450157397691505e-06, "loss": 0.363, "step": 1900 }, { "epoch": 2.689313517338995, "grad_norm": 0.20338112097088418, "learning_rate": 5.71878279118573e-06, "loss": 0.3513, "step": 1901 }, { "epoch": 2.6907289455060157, "grad_norm": 0.19880924179095463, "learning_rate": 5.6925498426023085e-06, "loss": 0.3974, "step": 1902 }, { "epoch": 2.692144373673036, "grad_norm": 0.21822057055933752, "learning_rate": 5.666316894018888e-06, "loss": 0.4239, "step": 1903 }, { "epoch": 2.6935598018400566, "grad_norm": 0.2076889320094873, "learning_rate": 5.6400839454354665e-06, "loss": 0.3578, "step": 1904 }, { "epoch": 2.6949752300070773, "grad_norm": 0.21095552299443449, "learning_rate": 5.613850996852047e-06, "loss": 0.3779, "step": 1905 }, { "epoch": 2.6963906581740975, "grad_norm": 0.20335690282279423, "learning_rate": 5.587618048268626e-06, "loss": 0.3592, "step": 1906 }, { "epoch": 2.697806086341118, "grad_norm": 0.2104837140088037, "learning_rate": 5.561385099685205e-06, "loss": 0.3676, "step": 1907 }, { "epoch": 2.699221514508139, "grad_norm": 0.21468127395284245, "learning_rate": 5.535152151101784e-06, "loss": 0.3606, "step": 1908 }, { "epoch": 2.700636942675159, "grad_norm": 0.2046692810039167, "learning_rate": 5.508919202518363e-06, "loss": 0.3678, "step": 1909 }, { "epoch": 2.7020523708421798, "grad_norm": 0.21112991054823232, "learning_rate": 5.482686253934942e-06, "loss": 0.372, "step": 1910 }, { "epoch": 2.7034677990092004, "grad_norm": 0.2141438116917265, "learning_rate": 5.456453305351522e-06, "loss": 0.3985, "step": 1911 }, { "epoch": 2.7048832271762207, "grad_norm": 0.21033732243828274, "learning_rate": 5.430220356768101e-06, "loss": 0.4051, "step": 1912 }, { "epoch": 2.7062986553432413, "grad_norm": 0.2111440991942263, "learning_rate": 5.403987408184681e-06, "loss": 0.4008, "step": 1913 }, { "epoch": 2.707714083510262, "grad_norm": 0.20268005737827657, "learning_rate": 5.377754459601259e-06, "loss": 0.3695, "step": 1914 }, { "epoch": 2.709129511677282, "grad_norm": 0.1951999738009383, "learning_rate": 5.351521511017839e-06, "loss": 0.3442, "step": 1915 }, { "epoch": 2.710544939844303, "grad_norm": 0.2001087039685667, "learning_rate": 5.325288562434417e-06, "loss": 0.3741, "step": 1916 }, { "epoch": 2.7119603680113236, "grad_norm": 0.19783070424964766, "learning_rate": 5.2990556138509975e-06, "loss": 0.3622, "step": 1917 }, { "epoch": 2.713375796178344, "grad_norm": 0.1935708112190888, "learning_rate": 5.272822665267577e-06, "loss": 0.3305, "step": 1918 }, { "epoch": 2.7147912243453645, "grad_norm": 0.19804726726173558, "learning_rate": 5.2465897166841556e-06, "loss": 0.343, "step": 1919 }, { "epoch": 2.716206652512385, "grad_norm": 0.2269934633201632, "learning_rate": 5.220356768100735e-06, "loss": 0.3996, "step": 1920 }, { "epoch": 2.7176220806794054, "grad_norm": 0.20480407012013166, "learning_rate": 5.194123819517314e-06, "loss": 0.3914, "step": 1921 }, { "epoch": 2.719037508846426, "grad_norm": 0.19401177570868808, "learning_rate": 5.167890870933893e-06, "loss": 0.3278, "step": 1922 }, { "epoch": 2.7204529370134467, "grad_norm": 0.20895425205289753, "learning_rate": 5.1416579223504725e-06, "loss": 0.3424, "step": 1923 }, { "epoch": 2.721868365180467, "grad_norm": 0.2438874893174407, "learning_rate": 5.115424973767052e-06, "loss": 0.3961, "step": 1924 }, { "epoch": 2.7232837933474876, "grad_norm": 0.18755381495409573, "learning_rate": 5.089192025183631e-06, "loss": 0.3203, "step": 1925 }, { "epoch": 2.7246992215145083, "grad_norm": 0.20504629032498756, "learning_rate": 5.06295907660021e-06, "loss": 0.3787, "step": 1926 }, { "epoch": 2.7261146496815285, "grad_norm": 0.19925744836851098, "learning_rate": 5.036726128016789e-06, "loss": 0.3593, "step": 1927 }, { "epoch": 2.727530077848549, "grad_norm": 0.2004589038992432, "learning_rate": 5.010493179433368e-06, "loss": 0.3639, "step": 1928 }, { "epoch": 2.72894550601557, "grad_norm": 0.20743800881777924, "learning_rate": 4.984260230849947e-06, "loss": 0.3755, "step": 1929 }, { "epoch": 2.73036093418259, "grad_norm": 0.19692865498896367, "learning_rate": 4.958027282266528e-06, "loss": 0.3399, "step": 1930 }, { "epoch": 2.7317763623496107, "grad_norm": 0.21841243044598094, "learning_rate": 4.931794333683106e-06, "loss": 0.3858, "step": 1931 }, { "epoch": 2.7331917905166314, "grad_norm": 0.21136619232889323, "learning_rate": 4.905561385099686e-06, "loss": 0.3921, "step": 1932 }, { "epoch": 2.7346072186836516, "grad_norm": 0.19484573735389385, "learning_rate": 4.879328436516264e-06, "loss": 0.3706, "step": 1933 }, { "epoch": 2.7360226468506723, "grad_norm": 0.2076830427543638, "learning_rate": 4.853095487932844e-06, "loss": 0.3456, "step": 1934 }, { "epoch": 2.737438075017693, "grad_norm": 0.20390841761357054, "learning_rate": 4.826862539349423e-06, "loss": 0.3526, "step": 1935 }, { "epoch": 2.738853503184713, "grad_norm": 0.21702033210302102, "learning_rate": 4.800629590766003e-06, "loss": 0.4627, "step": 1936 }, { "epoch": 2.740268931351734, "grad_norm": 0.20061521845716984, "learning_rate": 4.774396642182582e-06, "loss": 0.3781, "step": 1937 }, { "epoch": 2.7416843595187546, "grad_norm": 0.2560094803022627, "learning_rate": 4.748163693599161e-06, "loss": 0.4285, "step": 1938 }, { "epoch": 2.743099787685775, "grad_norm": 0.1932520148658916, "learning_rate": 4.72193074501574e-06, "loss": 0.3394, "step": 1939 }, { "epoch": 2.7445152158527955, "grad_norm": 0.20433241819246098, "learning_rate": 4.695697796432319e-06, "loss": 0.377, "step": 1940 }, { "epoch": 2.745930644019816, "grad_norm": 0.195922105002889, "learning_rate": 4.669464847848898e-06, "loss": 0.3262, "step": 1941 }, { "epoch": 2.7473460721868364, "grad_norm": 0.1993898781097721, "learning_rate": 4.6432318992654776e-06, "loss": 0.3687, "step": 1942 }, { "epoch": 2.748761500353857, "grad_norm": 0.20177198309564906, "learning_rate": 4.616998950682057e-06, "loss": 0.3741, "step": 1943 }, { "epoch": 2.7501769285208777, "grad_norm": 0.19082999793806443, "learning_rate": 4.5907660020986364e-06, "loss": 0.3534, "step": 1944 }, { "epoch": 2.7515923566878984, "grad_norm": 0.1956209768009436, "learning_rate": 4.564533053515215e-06, "loss": 0.3439, "step": 1945 }, { "epoch": 2.7530077848549186, "grad_norm": 0.20345812447505207, "learning_rate": 4.5383001049317945e-06, "loss": 0.4009, "step": 1946 }, { "epoch": 2.7544232130219393, "grad_norm": 0.1995521377352416, "learning_rate": 4.512067156348374e-06, "loss": 0.3623, "step": 1947 }, { "epoch": 2.7558386411889595, "grad_norm": 0.2170207918509561, "learning_rate": 4.485834207764953e-06, "loss": 0.3573, "step": 1948 }, { "epoch": 2.75725406935598, "grad_norm": 0.20261497371921042, "learning_rate": 4.459601259181533e-06, "loss": 0.3715, "step": 1949 }, { "epoch": 2.758669497523001, "grad_norm": 0.20015382284953034, "learning_rate": 4.433368310598111e-06, "loss": 0.327, "step": 1950 }, { "epoch": 2.7600849256900215, "grad_norm": 0.19249028445721247, "learning_rate": 4.407135362014691e-06, "loss": 0.3339, "step": 1951 }, { "epoch": 2.7615003538570417, "grad_norm": 0.19662036433893573, "learning_rate": 4.380902413431269e-06, "loss": 0.3566, "step": 1952 }, { "epoch": 2.7629157820240624, "grad_norm": 0.2018722434714739, "learning_rate": 4.354669464847849e-06, "loss": 0.3797, "step": 1953 }, { "epoch": 2.7643312101910826, "grad_norm": 0.19701375920428837, "learning_rate": 4.328436516264428e-06, "loss": 0.3813, "step": 1954 }, { "epoch": 2.7657466383581033, "grad_norm": 0.18850648320520433, "learning_rate": 4.302203567681008e-06, "loss": 0.3312, "step": 1955 }, { "epoch": 2.767162066525124, "grad_norm": 0.20634358143885162, "learning_rate": 4.275970619097587e-06, "loss": 0.3556, "step": 1956 }, { "epoch": 2.7685774946921446, "grad_norm": 0.20531390532997795, "learning_rate": 4.249737670514166e-06, "loss": 0.3771, "step": 1957 }, { "epoch": 2.769992922859165, "grad_norm": 0.19949037117645418, "learning_rate": 4.223504721930745e-06, "loss": 0.3612, "step": 1958 }, { "epoch": 2.7714083510261855, "grad_norm": 0.20340538899445754, "learning_rate": 4.197271773347325e-06, "loss": 0.3953, "step": 1959 }, { "epoch": 2.7728237791932058, "grad_norm": 0.20229010157197905, "learning_rate": 4.171038824763904e-06, "loss": 0.3768, "step": 1960 }, { "epoch": 2.7742392073602264, "grad_norm": 0.21675643705520573, "learning_rate": 4.144805876180483e-06, "loss": 0.3985, "step": 1961 }, { "epoch": 2.775654635527247, "grad_norm": 0.19955801340341722, "learning_rate": 4.118572927597062e-06, "loss": 0.3615, "step": 1962 }, { "epoch": 2.777070063694268, "grad_norm": 0.18933990643743973, "learning_rate": 4.0923399790136415e-06, "loss": 0.3394, "step": 1963 }, { "epoch": 2.778485491861288, "grad_norm": 0.19445316739805585, "learning_rate": 4.06610703043022e-06, "loss": 0.3661, "step": 1964 }, { "epoch": 2.7799009200283087, "grad_norm": 0.2110069160547559, "learning_rate": 4.0398740818467995e-06, "loss": 0.3558, "step": 1965 }, { "epoch": 2.781316348195329, "grad_norm": 0.2054544611098325, "learning_rate": 4.013641133263379e-06, "loss": 0.3855, "step": 1966 }, { "epoch": 2.7827317763623496, "grad_norm": 0.18777084253880907, "learning_rate": 3.9874081846799584e-06, "loss": 0.3239, "step": 1967 }, { "epoch": 2.7841472045293703, "grad_norm": 0.21906578216105788, "learning_rate": 3.961175236096538e-06, "loss": 0.4039, "step": 1968 }, { "epoch": 2.785562632696391, "grad_norm": 0.22430861257203039, "learning_rate": 3.9349422875131165e-06, "loss": 0.3919, "step": 1969 }, { "epoch": 2.786978060863411, "grad_norm": 0.2188750732836388, "learning_rate": 3.908709338929696e-06, "loss": 0.408, "step": 1970 }, { "epoch": 2.788393489030432, "grad_norm": 0.20746037337483966, "learning_rate": 3.882476390346275e-06, "loss": 0.3902, "step": 1971 }, { "epoch": 2.789808917197452, "grad_norm": 8.05758732383772, "learning_rate": 3.856243441762855e-06, "loss": 0.8494, "step": 1972 }, { "epoch": 2.7912243453644727, "grad_norm": 0.20122927201049465, "learning_rate": 3.830010493179433e-06, "loss": 0.3641, "step": 1973 }, { "epoch": 2.7926397735314934, "grad_norm": 0.2066591638924186, "learning_rate": 3.803777544596013e-06, "loss": 0.3927, "step": 1974 }, { "epoch": 2.794055201698514, "grad_norm": 0.2027018850431106, "learning_rate": 3.7775445960125922e-06, "loss": 0.3506, "step": 1975 }, { "epoch": 2.7954706298655343, "grad_norm": 0.21715542211666516, "learning_rate": 3.7513116474291713e-06, "loss": 0.3832, "step": 1976 }, { "epoch": 2.796886058032555, "grad_norm": 0.21860625760823418, "learning_rate": 3.7250786988457507e-06, "loss": 0.4152, "step": 1977 }, { "epoch": 2.798301486199575, "grad_norm": 0.18673496863635236, "learning_rate": 3.6988457502623293e-06, "loss": 0.2992, "step": 1978 }, { "epoch": 2.799716914366596, "grad_norm": 0.19095726885505263, "learning_rate": 3.672612801678909e-06, "loss": 0.3625, "step": 1979 }, { "epoch": 2.8011323425336165, "grad_norm": 1.321622899703763, "learning_rate": 3.6463798530954877e-06, "loss": 0.4226, "step": 1980 }, { "epoch": 2.802547770700637, "grad_norm": 5.785359000134272, "learning_rate": 3.620146904512067e-06, "loss": 0.3849, "step": 1981 }, { "epoch": 2.8039631988676574, "grad_norm": 0.21740520427626114, "learning_rate": 3.5939139559286466e-06, "loss": 0.3797, "step": 1982 }, { "epoch": 2.805378627034678, "grad_norm": 0.20837516610784643, "learning_rate": 3.5676810073452256e-06, "loss": 0.386, "step": 1983 }, { "epoch": 2.8067940552016983, "grad_norm": 0.19932361211467717, "learning_rate": 3.541448058761805e-06, "loss": 0.3831, "step": 1984 }, { "epoch": 2.808209483368719, "grad_norm": 0.19986911852545236, "learning_rate": 3.515215110178384e-06, "loss": 0.3374, "step": 1985 }, { "epoch": 2.8096249115357397, "grad_norm": 0.19704540448560762, "learning_rate": 3.4889821615949635e-06, "loss": 0.3939, "step": 1986 }, { "epoch": 2.8110403397027603, "grad_norm": 0.20783002369532763, "learning_rate": 3.462749213011543e-06, "loss": 0.3692, "step": 1987 }, { "epoch": 2.8124557678697806, "grad_norm": 0.20847088393412946, "learning_rate": 3.436516264428122e-06, "loss": 0.3922, "step": 1988 }, { "epoch": 2.8138711960368012, "grad_norm": 0.19850699462906699, "learning_rate": 3.4102833158447014e-06, "loss": 0.3682, "step": 1989 }, { "epoch": 2.8152866242038215, "grad_norm": 0.2076469565314169, "learning_rate": 3.38405036726128e-06, "loss": 0.3917, "step": 1990 }, { "epoch": 2.816702052370842, "grad_norm": 0.20081320372031727, "learning_rate": 3.35781741867786e-06, "loss": 0.3491, "step": 1991 }, { "epoch": 2.818117480537863, "grad_norm": 0.19552115470271994, "learning_rate": 3.3315844700944385e-06, "loss": 0.3604, "step": 1992 }, { "epoch": 2.8195329087048835, "grad_norm": 0.20590856907312838, "learning_rate": 3.305351521511018e-06, "loss": 0.381, "step": 1993 }, { "epoch": 2.8209483368719037, "grad_norm": 0.197122671871231, "learning_rate": 3.2791185729275973e-06, "loss": 0.3928, "step": 1994 }, { "epoch": 2.8223637650389244, "grad_norm": 0.19390053775641367, "learning_rate": 3.2528856243441763e-06, "loss": 0.3472, "step": 1995 }, { "epoch": 2.8237791932059446, "grad_norm": 0.18663152400131341, "learning_rate": 3.2266526757607558e-06, "loss": 0.3025, "step": 1996 }, { "epoch": 2.8251946213729653, "grad_norm": 0.19909280820282357, "learning_rate": 3.200419727177335e-06, "loss": 0.3353, "step": 1997 }, { "epoch": 2.826610049539986, "grad_norm": 0.19904235811084578, "learning_rate": 3.1741867785939142e-06, "loss": 0.365, "step": 1998 }, { "epoch": 2.8280254777070066, "grad_norm": 0.20649406059841743, "learning_rate": 3.147953830010493e-06, "loss": 0.3678, "step": 1999 }, { "epoch": 2.829440905874027, "grad_norm": 0.20909728780145723, "learning_rate": 3.1217208814270727e-06, "loss": 0.4091, "step": 2000 }, { "epoch": 2.8308563340410475, "grad_norm": 0.19958915848640862, "learning_rate": 3.0954879328436517e-06, "loss": 0.3632, "step": 2001 }, { "epoch": 2.8322717622080678, "grad_norm": 0.21457143457564132, "learning_rate": 3.0692549842602307e-06, "loss": 0.4022, "step": 2002 }, { "epoch": 2.8336871903750884, "grad_norm": 0.20610291251086, "learning_rate": 3.04302203567681e-06, "loss": 0.378, "step": 2003 }, { "epoch": 2.835102618542109, "grad_norm": 0.19436597825698418, "learning_rate": 3.0167890870933896e-06, "loss": 0.3432, "step": 2004 }, { "epoch": 2.8365180467091298, "grad_norm": 0.20684825531711648, "learning_rate": 2.9905561385099686e-06, "loss": 0.3605, "step": 2005 }, { "epoch": 2.83793347487615, "grad_norm": 0.7760432398237981, "learning_rate": 2.964323189926548e-06, "loss": 0.3842, "step": 2006 }, { "epoch": 2.8393489030431707, "grad_norm": 0.21211020483305554, "learning_rate": 2.938090241343127e-06, "loss": 0.4131, "step": 2007 }, { "epoch": 2.840764331210191, "grad_norm": 0.21399514331029734, "learning_rate": 2.911857292759706e-06, "loss": 0.4113, "step": 2008 }, { "epoch": 2.8421797593772116, "grad_norm": 0.19378601644607968, "learning_rate": 2.8856243441762855e-06, "loss": 0.347, "step": 2009 }, { "epoch": 2.8435951875442322, "grad_norm": 0.1933589337467757, "learning_rate": 2.859391395592865e-06, "loss": 0.3398, "step": 2010 }, { "epoch": 2.845010615711253, "grad_norm": 0.1926902438267549, "learning_rate": 2.833158447009444e-06, "loss": 0.3746, "step": 2011 }, { "epoch": 2.846426043878273, "grad_norm": 0.19736811412649616, "learning_rate": 2.8069254984260234e-06, "loss": 0.3603, "step": 2012 }, { "epoch": 2.847841472045294, "grad_norm": 0.19109684975374486, "learning_rate": 2.7806925498426024e-06, "loss": 0.3351, "step": 2013 }, { "epoch": 2.849256900212314, "grad_norm": 0.2216792369983889, "learning_rate": 2.7544596012591814e-06, "loss": 0.4175, "step": 2014 }, { "epoch": 2.8506723283793347, "grad_norm": 0.19625843195281362, "learning_rate": 2.728226652675761e-06, "loss": 0.3527, "step": 2015 }, { "epoch": 2.8520877565463554, "grad_norm": 0.20743441777460606, "learning_rate": 2.7019937040923403e-06, "loss": 0.3935, "step": 2016 }, { "epoch": 2.853503184713376, "grad_norm": 0.20561747739702188, "learning_rate": 2.6757607555089193e-06, "loss": 0.399, "step": 2017 }, { "epoch": 2.8549186128803963, "grad_norm": 0.1958029226290106, "learning_rate": 2.6495278069254988e-06, "loss": 0.3618, "step": 2018 }, { "epoch": 2.856334041047417, "grad_norm": 0.1802716298168241, "learning_rate": 2.6232948583420778e-06, "loss": 0.2994, "step": 2019 }, { "epoch": 2.857749469214437, "grad_norm": 0.22175156102935556, "learning_rate": 2.597061909758657e-06, "loss": 0.3841, "step": 2020 }, { "epoch": 2.859164897381458, "grad_norm": 0.208458674583299, "learning_rate": 2.5708289611752362e-06, "loss": 0.3687, "step": 2021 }, { "epoch": 2.8605803255484785, "grad_norm": 0.23313585596854977, "learning_rate": 2.5445960125918157e-06, "loss": 0.4213, "step": 2022 }, { "epoch": 2.861995753715499, "grad_norm": 0.20982546110462055, "learning_rate": 2.5183630640083947e-06, "loss": 0.3948, "step": 2023 }, { "epoch": 2.8634111818825194, "grad_norm": 0.22134081258828486, "learning_rate": 2.4921301154249737e-06, "loss": 0.432, "step": 2024 }, { "epoch": 2.86482661004954, "grad_norm": 0.20192104723864968, "learning_rate": 2.465897166841553e-06, "loss": 0.3546, "step": 2025 }, { "epoch": 2.8662420382165603, "grad_norm": 0.19324043372722652, "learning_rate": 2.439664218258132e-06, "loss": 0.3735, "step": 2026 }, { "epoch": 2.867657466383581, "grad_norm": 0.18986670112243625, "learning_rate": 2.4134312696747116e-06, "loss": 0.3659, "step": 2027 }, { "epoch": 2.8690728945506017, "grad_norm": 0.19523423146769045, "learning_rate": 2.387198321091291e-06, "loss": 0.3395, "step": 2028 }, { "epoch": 2.8704883227176223, "grad_norm": 0.19590239410254387, "learning_rate": 2.36096537250787e-06, "loss": 0.3456, "step": 2029 }, { "epoch": 2.8719037508846426, "grad_norm": 0.19940348026404991, "learning_rate": 2.334732423924449e-06, "loss": 0.3757, "step": 2030 }, { "epoch": 2.8733191790516632, "grad_norm": 0.20657571108654077, "learning_rate": 2.3084994753410285e-06, "loss": 0.3685, "step": 2031 }, { "epoch": 2.8747346072186835, "grad_norm": 0.24404659838605067, "learning_rate": 2.2822665267576075e-06, "loss": 0.3772, "step": 2032 }, { "epoch": 2.876150035385704, "grad_norm": 0.1930235454604922, "learning_rate": 2.256033578174187e-06, "loss": 0.3406, "step": 2033 }, { "epoch": 2.877565463552725, "grad_norm": 0.2113141777693592, "learning_rate": 2.2298006295907664e-06, "loss": 0.3722, "step": 2034 }, { "epoch": 2.8789808917197455, "grad_norm": 0.19355370824297508, "learning_rate": 2.2035676810073454e-06, "loss": 0.3416, "step": 2035 }, { "epoch": 2.8803963198867657, "grad_norm": 0.35498128723227274, "learning_rate": 2.1773347324239244e-06, "loss": 0.3439, "step": 2036 }, { "epoch": 2.8818117480537864, "grad_norm": 0.21079226500885176, "learning_rate": 2.151101783840504e-06, "loss": 0.3999, "step": 2037 }, { "epoch": 2.8832271762208066, "grad_norm": 0.19993629279499625, "learning_rate": 2.124868835257083e-06, "loss": 0.379, "step": 2038 }, { "epoch": 2.8846426043878273, "grad_norm": 0.20265852172396523, "learning_rate": 2.0986358866736623e-06, "loss": 0.3845, "step": 2039 }, { "epoch": 2.886058032554848, "grad_norm": 0.19966660066869182, "learning_rate": 2.0724029380902413e-06, "loss": 0.3429, "step": 2040 }, { "epoch": 2.8874734607218686, "grad_norm": 0.19461640316662585, "learning_rate": 2.0461699895068208e-06, "loss": 0.3311, "step": 2041 }, { "epoch": 2.888888888888889, "grad_norm": 0.18685404198759573, "learning_rate": 2.0199370409233998e-06, "loss": 0.3168, "step": 2042 }, { "epoch": 2.8903043170559095, "grad_norm": 0.18528748528930483, "learning_rate": 1.9937040923399792e-06, "loss": 0.3445, "step": 2043 }, { "epoch": 2.8917197452229297, "grad_norm": 0.20334724073798452, "learning_rate": 1.9674711437565582e-06, "loss": 0.387, "step": 2044 }, { "epoch": 2.8931351733899504, "grad_norm": 0.20068169776681, "learning_rate": 1.9412381951731377e-06, "loss": 0.3706, "step": 2045 }, { "epoch": 2.894550601556971, "grad_norm": 0.19848960345962377, "learning_rate": 1.9150052465897167e-06, "loss": 0.3848, "step": 2046 }, { "epoch": 2.8959660297239918, "grad_norm": 0.19394867329081025, "learning_rate": 1.8887722980062961e-06, "loss": 0.3823, "step": 2047 }, { "epoch": 2.897381457891012, "grad_norm": 0.20136704982699424, "learning_rate": 1.8625393494228753e-06, "loss": 0.3574, "step": 2048 }, { "epoch": 2.8987968860580327, "grad_norm": 0.18054669373990248, "learning_rate": 1.8363064008394546e-06, "loss": 0.3125, "step": 2049 }, { "epoch": 2.900212314225053, "grad_norm": 0.1951565386591877, "learning_rate": 1.8100734522560336e-06, "loss": 0.3765, "step": 2050 }, { "epoch": 2.9016277423920736, "grad_norm": 0.20474264180107443, "learning_rate": 1.7838405036726128e-06, "loss": 0.4115, "step": 2051 }, { "epoch": 2.903043170559094, "grad_norm": 0.21647592418475772, "learning_rate": 1.757607555089192e-06, "loss": 0.3917, "step": 2052 }, { "epoch": 2.904458598726115, "grad_norm": 0.18118807335850776, "learning_rate": 1.7313746065057715e-06, "loss": 0.2871, "step": 2053 }, { "epoch": 2.905874026893135, "grad_norm": 0.19448101636821263, "learning_rate": 1.7051416579223507e-06, "loss": 0.3613, "step": 2054 }, { "epoch": 2.907289455060156, "grad_norm": 0.18341263030502164, "learning_rate": 1.67890870933893e-06, "loss": 0.3271, "step": 2055 }, { "epoch": 2.908704883227176, "grad_norm": 0.21586315019415844, "learning_rate": 1.652675760755509e-06, "loss": 0.3914, "step": 2056 }, { "epoch": 2.9101203113941967, "grad_norm": 0.21643695397466847, "learning_rate": 1.6264428121720882e-06, "loss": 0.4036, "step": 2057 }, { "epoch": 2.9115357395612174, "grad_norm": 0.19145620703601973, "learning_rate": 1.6002098635886674e-06, "loss": 0.3502, "step": 2058 }, { "epoch": 2.912951167728238, "grad_norm": 0.19768294939507053, "learning_rate": 1.5739769150052464e-06, "loss": 0.3657, "step": 2059 }, { "epoch": 2.9143665958952583, "grad_norm": 0.18062247390735356, "learning_rate": 1.5477439664218259e-06, "loss": 0.3118, "step": 2060 }, { "epoch": 2.915782024062279, "grad_norm": 0.20220286484002623, "learning_rate": 1.521511017838405e-06, "loss": 0.3525, "step": 2061 }, { "epoch": 2.917197452229299, "grad_norm": 0.18868728285298295, "learning_rate": 1.4952780692549843e-06, "loss": 0.3444, "step": 2062 }, { "epoch": 2.91861288039632, "grad_norm": 0.2750350768980409, "learning_rate": 1.4690451206715635e-06, "loss": 0.4072, "step": 2063 }, { "epoch": 2.9200283085633405, "grad_norm": 0.1784673086428867, "learning_rate": 1.4428121720881428e-06, "loss": 0.3148, "step": 2064 }, { "epoch": 2.921443736730361, "grad_norm": 0.19792808089582006, "learning_rate": 1.416579223504722e-06, "loss": 0.3966, "step": 2065 }, { "epoch": 2.9228591648973814, "grad_norm": 0.19134347414255334, "learning_rate": 1.3903462749213012e-06, "loss": 0.3323, "step": 2066 }, { "epoch": 2.924274593064402, "grad_norm": 0.2035122251198033, "learning_rate": 1.3641133263378804e-06, "loss": 0.3651, "step": 2067 }, { "epoch": 2.9256900212314223, "grad_norm": 0.20033413515213966, "learning_rate": 1.3378803777544597e-06, "loss": 0.3983, "step": 2068 }, { "epoch": 2.927105449398443, "grad_norm": 0.18453302874158478, "learning_rate": 1.3116474291710389e-06, "loss": 0.3396, "step": 2069 }, { "epoch": 2.9285208775654636, "grad_norm": 0.1882995173465666, "learning_rate": 1.2854144805876181e-06, "loss": 0.3399, "step": 2070 }, { "epoch": 2.9299363057324843, "grad_norm": 0.20725779599771735, "learning_rate": 1.2591815320041973e-06, "loss": 0.3987, "step": 2071 }, { "epoch": 2.9313517338995045, "grad_norm": 0.1883440700700296, "learning_rate": 1.2329485834207766e-06, "loss": 0.345, "step": 2072 }, { "epoch": 2.932767162066525, "grad_norm": 0.20368211705548797, "learning_rate": 1.2067156348373558e-06, "loss": 0.3824, "step": 2073 }, { "epoch": 2.9341825902335454, "grad_norm": 0.2095686727474719, "learning_rate": 1.180482686253935e-06, "loss": 0.3781, "step": 2074 }, { "epoch": 2.935598018400566, "grad_norm": 0.18957399098358954, "learning_rate": 1.1542497376705142e-06, "loss": 0.3661, "step": 2075 }, { "epoch": 2.937013446567587, "grad_norm": 0.1941943711092127, "learning_rate": 1.1280167890870935e-06, "loss": 0.3539, "step": 2076 }, { "epoch": 2.9384288747346075, "grad_norm": 0.19737218357988998, "learning_rate": 1.1017838405036727e-06, "loss": 0.3701, "step": 2077 }, { "epoch": 2.9398443029016277, "grad_norm": 0.21773224729027893, "learning_rate": 1.075550891920252e-06, "loss": 0.4206, "step": 2078 }, { "epoch": 2.9412597310686484, "grad_norm": 4.797610101801964, "learning_rate": 1.0493179433368312e-06, "loss": 0.7369, "step": 2079 }, { "epoch": 2.9426751592356686, "grad_norm": 0.1935476966162268, "learning_rate": 1.0230849947534104e-06, "loss": 0.3838, "step": 2080 }, { "epoch": 2.9440905874026893, "grad_norm": 0.18450472118161934, "learning_rate": 9.968520461699896e-07, "loss": 0.3123, "step": 2081 }, { "epoch": 2.94550601556971, "grad_norm": 0.1900464839132083, "learning_rate": 9.706190975865688e-07, "loss": 0.3108, "step": 2082 }, { "epoch": 2.9469214437367306, "grad_norm": 0.19113932244352635, "learning_rate": 9.443861490031481e-07, "loss": 0.3205, "step": 2083 }, { "epoch": 2.948336871903751, "grad_norm": 0.1976782716367035, "learning_rate": 9.181532004197273e-07, "loss": 0.3611, "step": 2084 }, { "epoch": 2.9497523000707715, "grad_norm": 0.18628514897246456, "learning_rate": 8.919202518363064e-07, "loss": 0.3401, "step": 2085 }, { "epoch": 2.9511677282377917, "grad_norm": 0.20330699467998223, "learning_rate": 8.656873032528857e-07, "loss": 0.4181, "step": 2086 }, { "epoch": 2.9525831564048124, "grad_norm": 0.1986337270652071, "learning_rate": 8.39454354669465e-07, "loss": 0.3437, "step": 2087 }, { "epoch": 2.953998584571833, "grad_norm": 0.2936645275997326, "learning_rate": 8.132214060860441e-07, "loss": 0.3972, "step": 2088 }, { "epoch": 2.9554140127388537, "grad_norm": 0.18414900018164665, "learning_rate": 7.869884575026232e-07, "loss": 0.3218, "step": 2089 }, { "epoch": 2.956829440905874, "grad_norm": 0.18070732731376302, "learning_rate": 7.607555089192025e-07, "loss": 0.2919, "step": 2090 }, { "epoch": 2.9582448690728946, "grad_norm": 0.1838248187014801, "learning_rate": 7.345225603357818e-07, "loss": 0.3345, "step": 2091 }, { "epoch": 2.959660297239915, "grad_norm": 0.21380101541068397, "learning_rate": 7.08289611752361e-07, "loss": 0.4093, "step": 2092 }, { "epoch": 2.9610757254069355, "grad_norm": 0.20197658636303287, "learning_rate": 6.820566631689402e-07, "loss": 0.3728, "step": 2093 }, { "epoch": 2.962491153573956, "grad_norm": 0.21149661656774726, "learning_rate": 6.558237145855194e-07, "loss": 0.3927, "step": 2094 }, { "epoch": 2.963906581740977, "grad_norm": 0.1936150798022257, "learning_rate": 6.295907660020987e-07, "loss": 0.3682, "step": 2095 }, { "epoch": 2.965322009907997, "grad_norm": 0.2007856015130988, "learning_rate": 6.033578174186779e-07, "loss": 0.397, "step": 2096 }, { "epoch": 2.9667374380750178, "grad_norm": 0.19695599331631608, "learning_rate": 5.771248688352571e-07, "loss": 0.3648, "step": 2097 }, { "epoch": 2.968152866242038, "grad_norm": 0.19728088677482633, "learning_rate": 5.508919202518364e-07, "loss": 0.3797, "step": 2098 }, { "epoch": 2.9695682944090587, "grad_norm": 0.1858177347450067, "learning_rate": 5.246589716684156e-07, "loss": 0.3296, "step": 2099 }, { "epoch": 2.9709837225760793, "grad_norm": 0.21170677276318048, "learning_rate": 4.984260230849948e-07, "loss": 0.416, "step": 2100 }, { "epoch": 2.9723991507431, "grad_norm": 0.18831289221943484, "learning_rate": 4.7219307450157403e-07, "loss": 0.3297, "step": 2101 }, { "epoch": 2.9738145789101202, "grad_norm": 0.17619689265975416, "learning_rate": 4.459601259181532e-07, "loss": 0.3069, "step": 2102 }, { "epoch": 2.975230007077141, "grad_norm": 0.22540108458031174, "learning_rate": 4.197271773347325e-07, "loss": 0.3753, "step": 2103 }, { "epoch": 2.976645435244161, "grad_norm": 0.19400149626944135, "learning_rate": 3.934942287513116e-07, "loss": 0.3474, "step": 2104 }, { "epoch": 2.978060863411182, "grad_norm": 0.18962762860314583, "learning_rate": 3.672612801678909e-07, "loss": 0.333, "step": 2105 }, { "epoch": 2.9794762915782025, "grad_norm": 0.19093979940298456, "learning_rate": 3.410283315844701e-07, "loss": 0.3336, "step": 2106 }, { "epoch": 2.980891719745223, "grad_norm": 0.19735335084880196, "learning_rate": 3.1479538300104934e-07, "loss": 0.346, "step": 2107 }, { "epoch": 2.9823071479122434, "grad_norm": 0.19011476516224762, "learning_rate": 2.8856243441762856e-07, "loss": 0.3321, "step": 2108 }, { "epoch": 2.983722576079264, "grad_norm": 0.2017245142823575, "learning_rate": 2.623294858342078e-07, "loss": 0.3611, "step": 2109 }, { "epoch": 2.9851380042462843, "grad_norm": 0.1830758960092337, "learning_rate": 2.3609653725078701e-07, "loss": 0.3288, "step": 2110 }, { "epoch": 2.986553432413305, "grad_norm": 0.19067503603986885, "learning_rate": 2.0986358866736624e-07, "loss": 0.3642, "step": 2111 }, { "epoch": 2.9879688605803256, "grad_norm": 0.19156299414517605, "learning_rate": 1.8363064008394544e-07, "loss": 0.3569, "step": 2112 }, { "epoch": 2.9893842887473463, "grad_norm": 0.2055866092880421, "learning_rate": 1.5739769150052467e-07, "loss": 0.4054, "step": 2113 }, { "epoch": 2.9907997169143665, "grad_norm": 0.22750293431289148, "learning_rate": 1.311647429171039e-07, "loss": 0.4022, "step": 2114 }, { "epoch": 2.992215145081387, "grad_norm": 0.19484546995446372, "learning_rate": 1.0493179433368312e-07, "loss": 0.35, "step": 2115 }, { "epoch": 2.9936305732484074, "grad_norm": 0.20414099987110698, "learning_rate": 7.869884575026233e-08, "loss": 0.4187, "step": 2116 }, { "epoch": 2.995046001415428, "grad_norm": 0.19780068062196055, "learning_rate": 5.246589716684156e-08, "loss": 0.3583, "step": 2117 }, { "epoch": 2.9964614295824488, "grad_norm": 0.18612586736128045, "learning_rate": 2.623294858342078e-08, "loss": 0.3423, "step": 2118 }, { "epoch": 2.9964614295824488, "step": 2118, "total_flos": 2.3539982914407105e+19, "train_loss": 0.570159648968095, "train_runtime": 70788.7073, "train_samples_per_second": 0.479, "train_steps_per_second": 0.03 } ], "logging_steps": 1, "max_steps": 2118, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.3539982914407105e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }