{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9998993322371732, "eval_steps": 500, "global_step": 29800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00013422368376900105, "grad_norm": 42.767356872558594, "learning_rate": 2.2371364653243848e-07, "loss": 7.8985, "step": 2 }, { "epoch": 0.0002684473675380021, "grad_norm": 44.26445388793945, "learning_rate": 4.4742729306487696e-07, "loss": 6.497, "step": 4 }, { "epoch": 0.00040267105130700314, "grad_norm": 40.350433349609375, "learning_rate": 6.711409395973154e-07, "loss": 6.4155, "step": 6 }, { "epoch": 0.0005368947350760042, "grad_norm": 40.29481887817383, "learning_rate": 8.948545861297539e-07, "loss": 6.4425, "step": 8 }, { "epoch": 0.0006711184188450052, "grad_norm": 34.15610122680664, "learning_rate": 1.1185682326621925e-06, "loss": 6.1634, "step": 10 }, { "epoch": 0.0008053421026140063, "grad_norm": 35.55307388305664, "learning_rate": 1.3422818791946309e-06, "loss": 5.9724, "step": 12 }, { "epoch": 0.0009395657863830073, "grad_norm": 34.54287338256836, "learning_rate": 1.5659955257270695e-06, "loss": 5.5537, "step": 14 }, { "epoch": 0.0010737894701520084, "grad_norm": 31.5373592376709, "learning_rate": 1.7897091722595078e-06, "loss": 4.9947, "step": 16 }, { "epoch": 0.0012080131539210094, "grad_norm": 29.093385696411133, "learning_rate": 2.013422818791946e-06, "loss": 4.7559, "step": 18 }, { "epoch": 0.0013422368376900104, "grad_norm": 17.960554122924805, "learning_rate": 2.237136465324385e-06, "loss": 3.9112, "step": 20 }, { "epoch": 0.0014764605214590114, "grad_norm": 15.002157211303711, "learning_rate": 2.4608501118568234e-06, "loss": 3.6361, "step": 22 }, { "epoch": 0.0016106842052280126, "grad_norm": 12.236384391784668, "learning_rate": 2.6845637583892617e-06, "loss": 3.4589, "step": 24 }, { "epoch": 0.0017449078889970136, "grad_norm": 8.323893547058105, "learning_rate": 2.9082774049217e-06, "loss": 3.1198, "step": 26 }, { "epoch": 0.0018791315727660146, "grad_norm": 7.51531982421875, "learning_rate": 3.131991051454139e-06, "loss": 3.0164, "step": 28 }, { "epoch": 0.0020133552565350155, "grad_norm": 5.39280891418457, "learning_rate": 3.3557046979865773e-06, "loss": 2.8598, "step": 30 }, { "epoch": 0.0021475789403040168, "grad_norm": 4.7940239906311035, "learning_rate": 3.5794183445190157e-06, "loss": 2.7762, "step": 32 }, { "epoch": 0.0022818026240730175, "grad_norm": 4.117138385772705, "learning_rate": 3.803131991051454e-06, "loss": 2.6936, "step": 34 }, { "epoch": 0.0024160263078420187, "grad_norm": 3.4565351009368896, "learning_rate": 4.026845637583892e-06, "loss": 2.5653, "step": 36 }, { "epoch": 0.00255024999161102, "grad_norm": 5.539157390594482, "learning_rate": 4.250559284116332e-06, "loss": 2.592, "step": 38 }, { "epoch": 0.0026844736753800207, "grad_norm": 3.1694529056549072, "learning_rate": 4.47427293064877e-06, "loss": 2.5212, "step": 40 }, { "epoch": 0.002818697359149022, "grad_norm": 4.043001651763916, "learning_rate": 4.697986577181209e-06, "loss": 2.4646, "step": 42 }, { "epoch": 0.0029529210429180227, "grad_norm": 3.426333427429199, "learning_rate": 4.921700223713647e-06, "loss": 2.4251, "step": 44 }, { "epoch": 0.003087144726687024, "grad_norm": 2.9743473529815674, "learning_rate": 5.1454138702460855e-06, "loss": 2.3507, "step": 46 }, { "epoch": 0.003221368410456025, "grad_norm": 2.7875661849975586, "learning_rate": 5.3691275167785235e-06, "loss": 2.3166, "step": 48 }, { "epoch": 0.003355592094225026, "grad_norm": 5.016522407531738, "learning_rate": 5.592841163310962e-06, "loss": 2.2963, "step": 50 }, { "epoch": 0.003489815777994027, "grad_norm": 3.7552225589752197, "learning_rate": 5.8165548098434e-06, "loss": 2.2321, "step": 52 }, { "epoch": 0.003624039461763028, "grad_norm": 2.941863775253296, "learning_rate": 6.04026845637584e-06, "loss": 2.2089, "step": 54 }, { "epoch": 0.003758263145532029, "grad_norm": 2.9160988330841064, "learning_rate": 6.263982102908278e-06, "loss": 2.1362, "step": 56 }, { "epoch": 0.0038924868293010303, "grad_norm": 2.726508378982544, "learning_rate": 6.487695749440716e-06, "loss": 2.149, "step": 58 }, { "epoch": 0.004026710513070031, "grad_norm": 2.86901593208313, "learning_rate": 6.7114093959731546e-06, "loss": 2.0731, "step": 60 }, { "epoch": 0.004160934196839032, "grad_norm": 2.479349374771118, "learning_rate": 6.935123042505594e-06, "loss": 2.0715, "step": 62 }, { "epoch": 0.0042951578806080335, "grad_norm": 3.0036346912384033, "learning_rate": 7.158836689038031e-06, "loss": 2.0446, "step": 64 }, { "epoch": 0.004429381564377034, "grad_norm": 2.9378981590270996, "learning_rate": 7.382550335570471e-06, "loss": 1.9705, "step": 66 }, { "epoch": 0.004563605248146035, "grad_norm": 2.7089834213256836, "learning_rate": 7.606263982102908e-06, "loss": 1.9511, "step": 68 }, { "epoch": 0.004697828931915037, "grad_norm": 2.9577481746673584, "learning_rate": 7.829977628635348e-06, "loss": 1.916, "step": 70 }, { "epoch": 0.0048320526156840375, "grad_norm": 3.2720839977264404, "learning_rate": 8.053691275167785e-06, "loss": 1.838, "step": 72 }, { "epoch": 0.004966276299453038, "grad_norm": 3.753499984741211, "learning_rate": 8.277404921700224e-06, "loss": 1.7891, "step": 74 }, { "epoch": 0.00510049998322204, "grad_norm": 3.103024482727051, "learning_rate": 8.501118568232664e-06, "loss": 1.7397, "step": 76 }, { "epoch": 0.005234723666991041, "grad_norm": 3.117305040359497, "learning_rate": 8.724832214765101e-06, "loss": 1.7192, "step": 78 }, { "epoch": 0.0053689473507600415, "grad_norm": 3.049574375152588, "learning_rate": 8.94854586129754e-06, "loss": 1.649, "step": 80 }, { "epoch": 0.005503171034529042, "grad_norm": 2.904085159301758, "learning_rate": 9.172259507829977e-06, "loss": 1.5617, "step": 82 }, { "epoch": 0.005637394718298044, "grad_norm": 2.859712839126587, "learning_rate": 9.395973154362418e-06, "loss": 1.5642, "step": 84 }, { "epoch": 0.005771618402067045, "grad_norm": 2.6117401123046875, "learning_rate": 9.619686800894855e-06, "loss": 1.529, "step": 86 }, { "epoch": 0.005905842085836045, "grad_norm": 2.8871121406555176, "learning_rate": 9.843400447427293e-06, "loss": 1.4863, "step": 88 }, { "epoch": 0.006040065769605047, "grad_norm": 2.6381306648254395, "learning_rate": 1.006711409395973e-05, "loss": 1.4858, "step": 90 }, { "epoch": 0.006174289453374048, "grad_norm": 2.521726131439209, "learning_rate": 1.0290827740492171e-05, "loss": 1.443, "step": 92 }, { "epoch": 0.006308513137143049, "grad_norm": 2.3711681365966797, "learning_rate": 1.051454138702461e-05, "loss": 1.4161, "step": 94 }, { "epoch": 0.00644273682091205, "grad_norm": 2.5755703449249268, "learning_rate": 1.0738255033557047e-05, "loss": 1.461, "step": 96 }, { "epoch": 0.006576960504681051, "grad_norm": 2.566465139389038, "learning_rate": 1.0961968680089486e-05, "loss": 1.3901, "step": 98 }, { "epoch": 0.006711184188450052, "grad_norm": 2.506310224533081, "learning_rate": 1.1185682326621925e-05, "loss": 1.3596, "step": 100 }, { "epoch": 0.0068454078722190535, "grad_norm": 2.2585983276367188, "learning_rate": 1.1409395973154363e-05, "loss": 1.3593, "step": 102 }, { "epoch": 0.006979631555988054, "grad_norm": 2.6175777912139893, "learning_rate": 1.16331096196868e-05, "loss": 1.3012, "step": 104 }, { "epoch": 0.007113855239757055, "grad_norm": 2.2929115295410156, "learning_rate": 1.185682326621924e-05, "loss": 1.3368, "step": 106 }, { "epoch": 0.007248078923526056, "grad_norm": 3.431950092315674, "learning_rate": 1.208053691275168e-05, "loss": 1.2798, "step": 108 }, { "epoch": 0.007382302607295057, "grad_norm": 2.7987656593322754, "learning_rate": 1.2304250559284117e-05, "loss": 1.2695, "step": 110 }, { "epoch": 0.007516526291064058, "grad_norm": 2.331099033355713, "learning_rate": 1.2527964205816556e-05, "loss": 1.2599, "step": 112 }, { "epoch": 0.007650749974833059, "grad_norm": 2.243494749069214, "learning_rate": 1.2751677852348994e-05, "loss": 1.3625, "step": 114 }, { "epoch": 0.007784973658602061, "grad_norm": 2.421570301055908, "learning_rate": 1.2975391498881432e-05, "loss": 1.3264, "step": 116 }, { "epoch": 0.007919197342371061, "grad_norm": 2.063924789428711, "learning_rate": 1.319910514541387e-05, "loss": 1.2795, "step": 118 }, { "epoch": 0.008053421026140062, "grad_norm": 2.33018159866333, "learning_rate": 1.3422818791946309e-05, "loss": 1.2582, "step": 120 }, { "epoch": 0.008187644709909063, "grad_norm": 2.2538163661956787, "learning_rate": 1.3646532438478748e-05, "loss": 1.2508, "step": 122 }, { "epoch": 0.008321868393678064, "grad_norm": 2.2358736991882324, "learning_rate": 1.3870246085011188e-05, "loss": 1.1971, "step": 124 }, { "epoch": 0.008456092077447066, "grad_norm": 2.3848955631256104, "learning_rate": 1.4093959731543624e-05, "loss": 1.2728, "step": 126 }, { "epoch": 0.008590315761216067, "grad_norm": 2.260687828063965, "learning_rate": 1.4317673378076063e-05, "loss": 1.2465, "step": 128 }, { "epoch": 0.008724539444985068, "grad_norm": 2.33004093170166, "learning_rate": 1.4541387024608501e-05, "loss": 1.1897, "step": 130 }, { "epoch": 0.008858763128754069, "grad_norm": 2.219352960586548, "learning_rate": 1.4765100671140942e-05, "loss": 1.2068, "step": 132 }, { "epoch": 0.00899298681252307, "grad_norm": 2.1695876121520996, "learning_rate": 1.4988814317673377e-05, "loss": 1.1861, "step": 134 }, { "epoch": 0.00912721049629207, "grad_norm": 2.1830992698669434, "learning_rate": 1.5212527964205816e-05, "loss": 1.2202, "step": 136 }, { "epoch": 0.009261434180061073, "grad_norm": 2.1508290767669678, "learning_rate": 1.5436241610738255e-05, "loss": 1.152, "step": 138 }, { "epoch": 0.009395657863830073, "grad_norm": 2.224714517593384, "learning_rate": 1.5659955257270695e-05, "loss": 1.174, "step": 140 }, { "epoch": 0.009529881547599074, "grad_norm": 2.093791961669922, "learning_rate": 1.5883668903803133e-05, "loss": 1.1588, "step": 142 }, { "epoch": 0.009664105231368075, "grad_norm": 1.9894508123397827, "learning_rate": 1.610738255033557e-05, "loss": 1.1574, "step": 144 }, { "epoch": 0.009798328915137076, "grad_norm": 2.047668933868408, "learning_rate": 1.633109619686801e-05, "loss": 1.1913, "step": 146 }, { "epoch": 0.009932552598906077, "grad_norm": 2.0811944007873535, "learning_rate": 1.6554809843400447e-05, "loss": 1.1582, "step": 148 }, { "epoch": 0.010066776282675077, "grad_norm": 2.1576690673828125, "learning_rate": 1.6778523489932888e-05, "loss": 1.1453, "step": 150 }, { "epoch": 0.01020099996644408, "grad_norm": 2.1650378704071045, "learning_rate": 1.7002237136465328e-05, "loss": 1.1698, "step": 152 }, { "epoch": 0.01033522365021308, "grad_norm": 2.089860439300537, "learning_rate": 1.7225950782997762e-05, "loss": 1.111, "step": 154 }, { "epoch": 0.010469447333982081, "grad_norm": 2.0436043739318848, "learning_rate": 1.7449664429530202e-05, "loss": 1.1292, "step": 156 }, { "epoch": 0.010603671017751082, "grad_norm": 2.8769757747650146, "learning_rate": 1.767337807606264e-05, "loss": 1.1203, "step": 158 }, { "epoch": 0.010737894701520083, "grad_norm": 2.236624240875244, "learning_rate": 1.789709172259508e-05, "loss": 1.1019, "step": 160 }, { "epoch": 0.010872118385289084, "grad_norm": 2.022063970565796, "learning_rate": 1.8120805369127517e-05, "loss": 1.1133, "step": 162 }, { "epoch": 0.011006342069058084, "grad_norm": 1.8289309740066528, "learning_rate": 1.8344519015659954e-05, "loss": 1.1115, "step": 164 }, { "epoch": 0.011140565752827087, "grad_norm": 2.115225076675415, "learning_rate": 1.8568232662192395e-05, "loss": 1.1549, "step": 166 }, { "epoch": 0.011274789436596088, "grad_norm": 2.106482982635498, "learning_rate": 1.8791946308724835e-05, "loss": 1.1644, "step": 168 }, { "epoch": 0.011409013120365089, "grad_norm": 2.041001319885254, "learning_rate": 1.9015659955257272e-05, "loss": 1.0633, "step": 170 }, { "epoch": 0.01154323680413409, "grad_norm": 2.135566473007202, "learning_rate": 1.923937360178971e-05, "loss": 1.141, "step": 172 }, { "epoch": 0.01167746048790309, "grad_norm": 2.074963092803955, "learning_rate": 1.946308724832215e-05, "loss": 1.1373, "step": 174 }, { "epoch": 0.01181168417167209, "grad_norm": 5.725609302520752, "learning_rate": 1.9686800894854587e-05, "loss": 1.033, "step": 176 }, { "epoch": 0.011945907855441093, "grad_norm": 2.2156283855438232, "learning_rate": 1.9910514541387027e-05, "loss": 1.0946, "step": 178 }, { "epoch": 0.012080131539210094, "grad_norm": 1.979102373123169, "learning_rate": 2.013422818791946e-05, "loss": 1.1088, "step": 180 }, { "epoch": 0.012214355222979095, "grad_norm": 1.9270720481872559, "learning_rate": 2.03579418344519e-05, "loss": 1.116, "step": 182 }, { "epoch": 0.012348578906748096, "grad_norm": 2.0316152572631836, "learning_rate": 2.0581655480984342e-05, "loss": 1.0833, "step": 184 }, { "epoch": 0.012482802590517096, "grad_norm": 2.2409253120422363, "learning_rate": 2.080536912751678e-05, "loss": 1.1118, "step": 186 }, { "epoch": 0.012617026274286097, "grad_norm": 2.080007791519165, "learning_rate": 2.102908277404922e-05, "loss": 1.1066, "step": 188 }, { "epoch": 0.012751249958055098, "grad_norm": 1.9611796140670776, "learning_rate": 2.1252796420581657e-05, "loss": 1.1154, "step": 190 }, { "epoch": 0.0128854736418241, "grad_norm": 1.939950704574585, "learning_rate": 2.1476510067114094e-05, "loss": 1.0574, "step": 192 }, { "epoch": 0.013019697325593101, "grad_norm": 2.051105499267578, "learning_rate": 2.1700223713646534e-05, "loss": 1.1122, "step": 194 }, { "epoch": 0.013153921009362102, "grad_norm": 2.430405855178833, "learning_rate": 2.192393736017897e-05, "loss": 1.1435, "step": 196 }, { "epoch": 0.013288144693131103, "grad_norm": 2.2696692943573, "learning_rate": 2.2147651006711412e-05, "loss": 1.126, "step": 198 }, { "epoch": 0.013422368376900104, "grad_norm": 1.884081482887268, "learning_rate": 2.237136465324385e-05, "loss": 1.0558, "step": 200 }, { "epoch": 0.013556592060669104, "grad_norm": 1.9265978336334229, "learning_rate": 2.2595078299776286e-05, "loss": 1.1367, "step": 202 }, { "epoch": 0.013690815744438107, "grad_norm": 2.098081111907959, "learning_rate": 2.2818791946308727e-05, "loss": 1.0358, "step": 204 }, { "epoch": 0.013825039428207108, "grad_norm": 2.078112840652466, "learning_rate": 2.3042505592841164e-05, "loss": 1.0662, "step": 206 }, { "epoch": 0.013959263111976108, "grad_norm": 1.9842445850372314, "learning_rate": 2.32662192393736e-05, "loss": 1.0274, "step": 208 }, { "epoch": 0.01409348679574511, "grad_norm": 1.8199636936187744, "learning_rate": 2.348993288590604e-05, "loss": 0.9903, "step": 210 }, { "epoch": 0.01422771047951411, "grad_norm": 1.8890475034713745, "learning_rate": 2.371364653243848e-05, "loss": 1.0526, "step": 212 }, { "epoch": 0.01436193416328311, "grad_norm": 1.9008369445800781, "learning_rate": 2.393736017897092e-05, "loss": 1.0305, "step": 214 }, { "epoch": 0.014496157847052112, "grad_norm": 1.9247174263000488, "learning_rate": 2.416107382550336e-05, "loss": 1.0382, "step": 216 }, { "epoch": 0.014630381530821114, "grad_norm": 1.8395044803619385, "learning_rate": 2.4384787472035793e-05, "loss": 0.9981, "step": 218 }, { "epoch": 0.014764605214590115, "grad_norm": 1.9873464107513428, "learning_rate": 2.4608501118568234e-05, "loss": 1.0409, "step": 220 }, { "epoch": 0.014898828898359116, "grad_norm": 2.246591806411743, "learning_rate": 2.4832214765100674e-05, "loss": 1.0352, "step": 222 }, { "epoch": 0.015033052582128116, "grad_norm": 1.8422189950942993, "learning_rate": 2.505592841163311e-05, "loss": 1.016, "step": 224 }, { "epoch": 0.015167276265897117, "grad_norm": 2.0310373306274414, "learning_rate": 2.527964205816555e-05, "loss": 1.0845, "step": 226 }, { "epoch": 0.015301499949666118, "grad_norm": 1.8543283939361572, "learning_rate": 2.550335570469799e-05, "loss": 1.077, "step": 228 }, { "epoch": 0.01543572363343512, "grad_norm": 1.8772192001342773, "learning_rate": 2.5727069351230426e-05, "loss": 1.0509, "step": 230 }, { "epoch": 0.015569947317204121, "grad_norm": 1.918837547302246, "learning_rate": 2.5950782997762863e-05, "loss": 1.0099, "step": 232 }, { "epoch": 0.01570417100097312, "grad_norm": 2.334092140197754, "learning_rate": 2.6174496644295304e-05, "loss": 1.0434, "step": 234 }, { "epoch": 0.015838394684742123, "grad_norm": 2.450557231903076, "learning_rate": 2.639821029082774e-05, "loss": 1.0655, "step": 236 }, { "epoch": 0.015972618368511125, "grad_norm": 1.8649179935455322, "learning_rate": 2.662192393736018e-05, "loss": 1.0242, "step": 238 }, { "epoch": 0.016106842052280124, "grad_norm": 1.8493071794509888, "learning_rate": 2.6845637583892618e-05, "loss": 1.0186, "step": 240 }, { "epoch": 0.016241065736049127, "grad_norm": 1.9221071004867554, "learning_rate": 2.7069351230425055e-05, "loss": 1.0608, "step": 242 }, { "epoch": 0.016375289419818126, "grad_norm": 1.8620386123657227, "learning_rate": 2.7293064876957496e-05, "loss": 0.9906, "step": 244 }, { "epoch": 0.01650951310358713, "grad_norm": 2.163280963897705, "learning_rate": 2.7516778523489933e-05, "loss": 0.9914, "step": 246 }, { "epoch": 0.016643736787356127, "grad_norm": 9.815702438354492, "learning_rate": 2.7740492170022377e-05, "loss": 1.0617, "step": 248 }, { "epoch": 0.01677796047112513, "grad_norm": 1.879356026649475, "learning_rate": 2.796420581655481e-05, "loss": 0.9977, "step": 250 }, { "epoch": 0.016912184154894132, "grad_norm": 1.9691245555877686, "learning_rate": 2.8187919463087248e-05, "loss": 1.0689, "step": 252 }, { "epoch": 0.01704640783866313, "grad_norm": 1.7776098251342773, "learning_rate": 2.8411633109619688e-05, "loss": 0.9844, "step": 254 }, { "epoch": 0.017180631522432134, "grad_norm": 2.298013925552368, "learning_rate": 2.8635346756152125e-05, "loss": 1.0274, "step": 256 }, { "epoch": 0.017314855206201133, "grad_norm": 1.880511999130249, "learning_rate": 2.885906040268457e-05, "loss": 1.0339, "step": 258 }, { "epoch": 0.017449078889970136, "grad_norm": 2.170077323913574, "learning_rate": 2.9082774049217003e-05, "loss": 0.9712, "step": 260 }, { "epoch": 0.017583302573739135, "grad_norm": 1.9864907264709473, "learning_rate": 2.930648769574944e-05, "loss": 1.0496, "step": 262 }, { "epoch": 0.017717526257508137, "grad_norm": 1.7236459255218506, "learning_rate": 2.9530201342281884e-05, "loss": 0.9276, "step": 264 }, { "epoch": 0.01785174994127714, "grad_norm": 1.8995026350021362, "learning_rate": 2.9753914988814318e-05, "loss": 0.9976, "step": 266 }, { "epoch": 0.01798597362504614, "grad_norm": 1.8944072723388672, "learning_rate": 2.9977628635346755e-05, "loss": 0.9927, "step": 268 }, { "epoch": 0.01812019730881514, "grad_norm": 1.9850423336029053, "learning_rate": 3.02013422818792e-05, "loss": 1.0039, "step": 270 }, { "epoch": 0.01825442099258414, "grad_norm": 2.0685665607452393, "learning_rate": 3.0425055928411632e-05, "loss": 1.0208, "step": 272 }, { "epoch": 0.018388644676353143, "grad_norm": 1.7906346321105957, "learning_rate": 3.0648769574944076e-05, "loss": 1.0011, "step": 274 }, { "epoch": 0.018522868360122145, "grad_norm": 2.530409097671509, "learning_rate": 3.087248322147651e-05, "loss": 1.0667, "step": 276 }, { "epoch": 0.018657092043891144, "grad_norm": 1.7903863191604614, "learning_rate": 3.109619686800895e-05, "loss": 0.9743, "step": 278 }, { "epoch": 0.018791315727660147, "grad_norm": 2.157287120819092, "learning_rate": 3.131991051454139e-05, "loss": 1.036, "step": 280 }, { "epoch": 0.018925539411429146, "grad_norm": 1.8101394176483154, "learning_rate": 3.1543624161073825e-05, "loss": 1.0131, "step": 282 }, { "epoch": 0.01905976309519815, "grad_norm": 1.8443541526794434, "learning_rate": 3.1767337807606265e-05, "loss": 0.9968, "step": 284 }, { "epoch": 0.019193986778967147, "grad_norm": 3.132554531097412, "learning_rate": 3.1991051454138706e-05, "loss": 0.9758, "step": 286 }, { "epoch": 0.01932821046273615, "grad_norm": 1.8236656188964844, "learning_rate": 3.221476510067114e-05, "loss": 0.9747, "step": 288 }, { "epoch": 0.019462434146505152, "grad_norm": 2.611768960952759, "learning_rate": 3.243847874720358e-05, "loss": 0.9804, "step": 290 }, { "epoch": 0.01959665783027415, "grad_norm": 19.7513427734375, "learning_rate": 3.266219239373602e-05, "loss": 1.0925, "step": 292 }, { "epoch": 0.019730881514043154, "grad_norm": 2.0081746578216553, "learning_rate": 3.288590604026846e-05, "loss": 1.0048, "step": 294 }, { "epoch": 0.019865105197812153, "grad_norm": 2.01193904876709, "learning_rate": 3.3109619686800894e-05, "loss": 0.9301, "step": 296 }, { "epoch": 0.019999328881581156, "grad_norm": 1.8901691436767578, "learning_rate": 3.3333333333333335e-05, "loss": 0.9385, "step": 298 }, { "epoch": 0.020133552565350155, "grad_norm": 2.4309654235839844, "learning_rate": 3.3557046979865775e-05, "loss": 1.0148, "step": 300 }, { "epoch": 0.020267776249119157, "grad_norm": 1.8135578632354736, "learning_rate": 3.378076062639821e-05, "loss": 0.9746, "step": 302 }, { "epoch": 0.02040199993288816, "grad_norm": 2.1611545085906982, "learning_rate": 3.4004474272930656e-05, "loss": 0.9685, "step": 304 }, { "epoch": 0.02053622361665716, "grad_norm": 1.7547329664230347, "learning_rate": 3.422818791946309e-05, "loss": 1.0273, "step": 306 }, { "epoch": 0.02067044730042616, "grad_norm": 1.6789021492004395, "learning_rate": 3.4451901565995524e-05, "loss": 0.962, "step": 308 }, { "epoch": 0.02080467098419516, "grad_norm": 1.7096121311187744, "learning_rate": 3.4675615212527964e-05, "loss": 1.0389, "step": 310 }, { "epoch": 0.020938894667964163, "grad_norm": 2.264564037322998, "learning_rate": 3.4899328859060405e-05, "loss": 1.0188, "step": 312 }, { "epoch": 0.021073118351733162, "grad_norm": 2.4550116062164307, "learning_rate": 3.512304250559284e-05, "loss": 0.9969, "step": 314 }, { "epoch": 0.021207342035502164, "grad_norm": 1.9647586345672607, "learning_rate": 3.534675615212528e-05, "loss": 0.934, "step": 316 }, { "epoch": 0.021341565719271167, "grad_norm": 1.7615418434143066, "learning_rate": 3.557046979865772e-05, "loss": 0.9689, "step": 318 }, { "epoch": 0.021475789403040166, "grad_norm": 1.8934327363967896, "learning_rate": 3.579418344519016e-05, "loss": 1.0117, "step": 320 }, { "epoch": 0.02161001308680917, "grad_norm": 1.7119038105010986, "learning_rate": 3.6017897091722594e-05, "loss": 0.9061, "step": 322 }, { "epoch": 0.021744236770578167, "grad_norm": 1.767134666442871, "learning_rate": 3.6241610738255034e-05, "loss": 0.9391, "step": 324 }, { "epoch": 0.02187846045434717, "grad_norm": 2.058410406112671, "learning_rate": 3.6465324384787475e-05, "loss": 0.9918, "step": 326 }, { "epoch": 0.02201268413811617, "grad_norm": 1.6547894477844238, "learning_rate": 3.668903803131991e-05, "loss": 1.0361, "step": 328 }, { "epoch": 0.02214690782188517, "grad_norm": 1.8306688070297241, "learning_rate": 3.6912751677852356e-05, "loss": 0.9505, "step": 330 }, { "epoch": 0.022281131505654174, "grad_norm": 1.8601231575012207, "learning_rate": 3.713646532438479e-05, "loss": 0.9661, "step": 332 }, { "epoch": 0.022415355189423173, "grad_norm": 1.877942442893982, "learning_rate": 3.736017897091722e-05, "loss": 0.9401, "step": 334 }, { "epoch": 0.022549578873192175, "grad_norm": 1.8389990329742432, "learning_rate": 3.758389261744967e-05, "loss": 0.9276, "step": 336 }, { "epoch": 0.022683802556961175, "grad_norm": 1.6361632347106934, "learning_rate": 3.7807606263982104e-05, "loss": 0.9313, "step": 338 }, { "epoch": 0.022818026240730177, "grad_norm": 1.736832857131958, "learning_rate": 3.8031319910514545e-05, "loss": 1.0363, "step": 340 }, { "epoch": 0.02295224992449918, "grad_norm": 1.808683156967163, "learning_rate": 3.8255033557046985e-05, "loss": 0.9628, "step": 342 }, { "epoch": 0.02308647360826818, "grad_norm": 1.7691764831542969, "learning_rate": 3.847874720357942e-05, "loss": 0.9395, "step": 344 }, { "epoch": 0.02322069729203718, "grad_norm": 1.6702338457107544, "learning_rate": 3.870246085011186e-05, "loss": 0.9588, "step": 346 }, { "epoch": 0.02335492097580618, "grad_norm": 1.7142435312271118, "learning_rate": 3.89261744966443e-05, "loss": 0.903, "step": 348 }, { "epoch": 0.023489144659575183, "grad_norm": 1.6802270412445068, "learning_rate": 3.914988814317674e-05, "loss": 0.9571, "step": 350 }, { "epoch": 0.02362336834334418, "grad_norm": 1.9458545446395874, "learning_rate": 3.9373601789709174e-05, "loss": 0.9587, "step": 352 }, { "epoch": 0.023757592027113184, "grad_norm": 1.8980857133865356, "learning_rate": 3.959731543624161e-05, "loss": 0.9452, "step": 354 }, { "epoch": 0.023891815710882187, "grad_norm": 1.8086082935333252, "learning_rate": 3.9821029082774055e-05, "loss": 0.8951, "step": 356 }, { "epoch": 0.024026039394651186, "grad_norm": 1.7780733108520508, "learning_rate": 4.004474272930649e-05, "loss": 0.8955, "step": 358 }, { "epoch": 0.02416026307842019, "grad_norm": 2.0118095874786377, "learning_rate": 4.026845637583892e-05, "loss": 0.9621, "step": 360 }, { "epoch": 0.024294486762189187, "grad_norm": 1.6194727420806885, "learning_rate": 4.049217002237137e-05, "loss": 0.8891, "step": 362 }, { "epoch": 0.02442871044595819, "grad_norm": 1.8351916074752808, "learning_rate": 4.07158836689038e-05, "loss": 0.9886, "step": 364 }, { "epoch": 0.02456293412972719, "grad_norm": 1.8265271186828613, "learning_rate": 4.0939597315436244e-05, "loss": 1.0141, "step": 366 }, { "epoch": 0.02469715781349619, "grad_norm": 1.84307861328125, "learning_rate": 4.1163310961968684e-05, "loss": 0.987, "step": 368 }, { "epoch": 0.024831381497265194, "grad_norm": 2.075024127960205, "learning_rate": 4.138702460850112e-05, "loss": 0.9706, "step": 370 }, { "epoch": 0.024965605181034193, "grad_norm": 1.7007570266723633, "learning_rate": 4.161073825503356e-05, "loss": 0.9534, "step": 372 }, { "epoch": 0.025099828864803195, "grad_norm": 1.9438354969024658, "learning_rate": 4.1834451901566e-05, "loss": 0.9997, "step": 374 }, { "epoch": 0.025234052548572194, "grad_norm": 1.5149259567260742, "learning_rate": 4.205816554809844e-05, "loss": 0.9208, "step": 376 }, { "epoch": 0.025368276232341197, "grad_norm": 1.6345700025558472, "learning_rate": 4.228187919463087e-05, "loss": 0.9345, "step": 378 }, { "epoch": 0.025502499916110196, "grad_norm": 2.3215291500091553, "learning_rate": 4.2505592841163314e-05, "loss": 0.9775, "step": 380 }, { "epoch": 0.0256367235998792, "grad_norm": 2.234036684036255, "learning_rate": 4.2729306487695754e-05, "loss": 0.9709, "step": 382 }, { "epoch": 0.0257709472836482, "grad_norm": 1.8222061395645142, "learning_rate": 4.295302013422819e-05, "loss": 0.9838, "step": 384 }, { "epoch": 0.0259051709674172, "grad_norm": 1.5880529880523682, "learning_rate": 4.317673378076063e-05, "loss": 0.941, "step": 386 }, { "epoch": 0.026039394651186203, "grad_norm": 1.7884408235549927, "learning_rate": 4.340044742729307e-05, "loss": 1.025, "step": 388 }, { "epoch": 0.0261736183349552, "grad_norm": 1.5966376066207886, "learning_rate": 4.36241610738255e-05, "loss": 0.9362, "step": 390 }, { "epoch": 0.026307842018724204, "grad_norm": 1.804855227470398, "learning_rate": 4.384787472035794e-05, "loss": 0.9968, "step": 392 }, { "epoch": 0.026442065702493203, "grad_norm": 1.6474556922912598, "learning_rate": 4.4071588366890384e-05, "loss": 0.9353, "step": 394 }, { "epoch": 0.026576289386262206, "grad_norm": 2.1431021690368652, "learning_rate": 4.4295302013422824e-05, "loss": 0.9527, "step": 396 }, { "epoch": 0.026710513070031208, "grad_norm": 2.439883232116699, "learning_rate": 4.451901565995526e-05, "loss": 0.9578, "step": 398 }, { "epoch": 0.026844736753800207, "grad_norm": 2.5266048908233643, "learning_rate": 4.47427293064877e-05, "loss": 0.9169, "step": 400 }, { "epoch": 0.02697896043756921, "grad_norm": 7.857470512390137, "learning_rate": 4.496644295302014e-05, "loss": 0.9875, "step": 402 }, { "epoch": 0.02711318412133821, "grad_norm": 1.719964861869812, "learning_rate": 4.519015659955257e-05, "loss": 0.9324, "step": 404 }, { "epoch": 0.02724740780510721, "grad_norm": 1.7738561630249023, "learning_rate": 4.541387024608501e-05, "loss": 0.9648, "step": 406 }, { "epoch": 0.027381631488876214, "grad_norm": 4.7296953201293945, "learning_rate": 4.5637583892617453e-05, "loss": 0.9465, "step": 408 }, { "epoch": 0.027515855172645213, "grad_norm": 1.8566882610321045, "learning_rate": 4.586129753914989e-05, "loss": 0.9163, "step": 410 }, { "epoch": 0.027650078856414215, "grad_norm": 1.6954864263534546, "learning_rate": 4.608501118568233e-05, "loss": 0.9287, "step": 412 }, { "epoch": 0.027784302540183214, "grad_norm": 1.734711766242981, "learning_rate": 4.630872483221477e-05, "loss": 0.9313, "step": 414 }, { "epoch": 0.027918526223952217, "grad_norm": 1.7496058940887451, "learning_rate": 4.65324384787472e-05, "loss": 0.926, "step": 416 }, { "epoch": 0.028052749907721216, "grad_norm": 1.6309984922409058, "learning_rate": 4.675615212527964e-05, "loss": 0.8993, "step": 418 }, { "epoch": 0.02818697359149022, "grad_norm": 3.022975444793701, "learning_rate": 4.697986577181208e-05, "loss": 0.8894, "step": 420 }, { "epoch": 0.02832119727525922, "grad_norm": 1.860723614692688, "learning_rate": 4.720357941834452e-05, "loss": 0.9366, "step": 422 }, { "epoch": 0.02845542095902822, "grad_norm": 1.670973777770996, "learning_rate": 4.742729306487696e-05, "loss": 0.9438, "step": 424 }, { "epoch": 0.028589644642797223, "grad_norm": 1.8910622596740723, "learning_rate": 4.76510067114094e-05, "loss": 0.9366, "step": 426 }, { "epoch": 0.02872386832656622, "grad_norm": 1.6946468353271484, "learning_rate": 4.787472035794184e-05, "loss": 0.9455, "step": 428 }, { "epoch": 0.028858092010335224, "grad_norm": 1.9580165147781372, "learning_rate": 4.809843400447427e-05, "loss": 0.9467, "step": 430 }, { "epoch": 0.028992315694104223, "grad_norm": 1.765116810798645, "learning_rate": 4.832214765100672e-05, "loss": 0.9342, "step": 432 }, { "epoch": 0.029126539377873226, "grad_norm": 2.1234090328216553, "learning_rate": 4.854586129753915e-05, "loss": 0.9607, "step": 434 }, { "epoch": 0.029260763061642228, "grad_norm": 2.8784382343292236, "learning_rate": 4.8769574944071586e-05, "loss": 0.9569, "step": 436 }, { "epoch": 0.029394986745411227, "grad_norm": 2.9103879928588867, "learning_rate": 4.8993288590604034e-05, "loss": 0.8813, "step": 438 }, { "epoch": 0.02952921042918023, "grad_norm": 1.9417873620986938, "learning_rate": 4.921700223713647e-05, "loss": 0.9726, "step": 440 }, { "epoch": 0.02966343411294923, "grad_norm": 1.5041873455047607, "learning_rate": 4.944071588366891e-05, "loss": 0.9298, "step": 442 }, { "epoch": 0.02979765779671823, "grad_norm": 2.0654478073120117, "learning_rate": 4.966442953020135e-05, "loss": 1.0488, "step": 444 }, { "epoch": 0.02993188148048723, "grad_norm": 1.5653998851776123, "learning_rate": 4.988814317673378e-05, "loss": 0.9493, "step": 446 }, { "epoch": 0.030066105164256233, "grad_norm": 2.1645994186401367, "learning_rate": 5.011185682326622e-05, "loss": 0.9166, "step": 448 }, { "epoch": 0.030200328848025235, "grad_norm": 1.866364598274231, "learning_rate": 5.033557046979866e-05, "loss": 0.9697, "step": 450 }, { "epoch": 0.030334552531794234, "grad_norm": 1.411466360092163, "learning_rate": 5.05592841163311e-05, "loss": 0.92, "step": 452 }, { "epoch": 0.030468776215563237, "grad_norm": 1.7987719774246216, "learning_rate": 5.078299776286354e-05, "loss": 0.995, "step": 454 }, { "epoch": 0.030602999899332236, "grad_norm": 2.065213680267334, "learning_rate": 5.100671140939598e-05, "loss": 0.9905, "step": 456 }, { "epoch": 0.03073722358310124, "grad_norm": 3.715909004211426, "learning_rate": 5.123042505592841e-05, "loss": 0.9419, "step": 458 }, { "epoch": 0.03087144726687024, "grad_norm": 1.5822199583053589, "learning_rate": 5.145413870246085e-05, "loss": 0.9306, "step": 460 }, { "epoch": 0.03100567095063924, "grad_norm": 1.9820096492767334, "learning_rate": 5.167785234899329e-05, "loss": 1.0166, "step": 462 }, { "epoch": 0.031139894634408243, "grad_norm": 7.59510612487793, "learning_rate": 5.1901565995525726e-05, "loss": 0.9553, "step": 464 }, { "epoch": 0.031274118318177245, "grad_norm": 1.6523329019546509, "learning_rate": 5.212527964205817e-05, "loss": 0.9754, "step": 466 }, { "epoch": 0.03140834200194624, "grad_norm": 1.549682378768921, "learning_rate": 5.234899328859061e-05, "loss": 0.8897, "step": 468 }, { "epoch": 0.03154256568571524, "grad_norm": 2.2443296909332275, "learning_rate": 5.257270693512305e-05, "loss": 0.9679, "step": 470 }, { "epoch": 0.031676789369484246, "grad_norm": 3.2198832035064697, "learning_rate": 5.279642058165548e-05, "loss": 0.9503, "step": 472 }, { "epoch": 0.03181101305325325, "grad_norm": 1.7302112579345703, "learning_rate": 5.302013422818792e-05, "loss": 0.9569, "step": 474 }, { "epoch": 0.03194523673702225, "grad_norm": 1.6648476123809814, "learning_rate": 5.324384787472036e-05, "loss": 0.9226, "step": 476 }, { "epoch": 0.032079460420791246, "grad_norm": 1.5022692680358887, "learning_rate": 5.3467561521252796e-05, "loss": 0.8694, "step": 478 }, { "epoch": 0.03221368410456025, "grad_norm": 1.6610078811645508, "learning_rate": 5.3691275167785237e-05, "loss": 0.9539, "step": 480 }, { "epoch": 0.03234790778832925, "grad_norm": 1.7546595335006714, "learning_rate": 5.391498881431768e-05, "loss": 0.8589, "step": 482 }, { "epoch": 0.032482131472098254, "grad_norm": 3.204794406890869, "learning_rate": 5.413870246085011e-05, "loss": 0.885, "step": 484 }, { "epoch": 0.03261635515586725, "grad_norm": 1.8887150287628174, "learning_rate": 5.436241610738255e-05, "loss": 0.9423, "step": 486 }, { "epoch": 0.03275057883963625, "grad_norm": 1.6688555479049683, "learning_rate": 5.458612975391499e-05, "loss": 0.9337, "step": 488 }, { "epoch": 0.032884802523405254, "grad_norm": 1.4868736267089844, "learning_rate": 5.4809843400447426e-05, "loss": 0.8962, "step": 490 }, { "epoch": 0.03301902620717426, "grad_norm": 2.3861632347106934, "learning_rate": 5.5033557046979866e-05, "loss": 0.9746, "step": 492 }, { "epoch": 0.03315324989094326, "grad_norm": 1.8033841848373413, "learning_rate": 5.5257270693512306e-05, "loss": 0.9252, "step": 494 }, { "epoch": 0.033287473574712255, "grad_norm": 1.5896399021148682, "learning_rate": 5.5480984340044754e-05, "loss": 0.8963, "step": 496 }, { "epoch": 0.03342169725848126, "grad_norm": 1.4402724504470825, "learning_rate": 5.570469798657718e-05, "loss": 0.9249, "step": 498 }, { "epoch": 0.03355592094225026, "grad_norm": 1.7836982011795044, "learning_rate": 5.592841163310962e-05, "loss": 0.8972, "step": 500 }, { "epoch": 0.03369014462601926, "grad_norm": 1.5181845426559448, "learning_rate": 5.615212527964206e-05, "loss": 0.8425, "step": 502 }, { "epoch": 0.033824368309788265, "grad_norm": 1.7089219093322754, "learning_rate": 5.6375838926174495e-05, "loss": 0.9004, "step": 504 }, { "epoch": 0.03395859199355726, "grad_norm": 2.010953903198242, "learning_rate": 5.6599552572706936e-05, "loss": 0.9736, "step": 506 }, { "epoch": 0.03409281567732626, "grad_norm": 2.0464272499084473, "learning_rate": 5.6823266219239376e-05, "loss": 0.9807, "step": 508 }, { "epoch": 0.034227039361095266, "grad_norm": 1.6039035320281982, "learning_rate": 5.704697986577181e-05, "loss": 0.8955, "step": 510 }, { "epoch": 0.03436126304486427, "grad_norm": 5.811178207397461, "learning_rate": 5.727069351230425e-05, "loss": 0.9033, "step": 512 }, { "epoch": 0.03449548672863327, "grad_norm": 3.845050096511841, "learning_rate": 5.749440715883669e-05, "loss": 0.9547, "step": 514 }, { "epoch": 0.034629710412402266, "grad_norm": 1.4917224645614624, "learning_rate": 5.771812080536914e-05, "loss": 0.9389, "step": 516 }, { "epoch": 0.03476393409617127, "grad_norm": 1.537062644958496, "learning_rate": 5.7941834451901565e-05, "loss": 0.8869, "step": 518 }, { "epoch": 0.03489815777994027, "grad_norm": 1.544792890548706, "learning_rate": 5.8165548098434006e-05, "loss": 0.9185, "step": 520 }, { "epoch": 0.035032381463709274, "grad_norm": 1.5967966318130493, "learning_rate": 5.838926174496645e-05, "loss": 0.9002, "step": 522 }, { "epoch": 0.03516660514747827, "grad_norm": 1.4114470481872559, "learning_rate": 5.861297539149888e-05, "loss": 0.8574, "step": 524 }, { "epoch": 0.03530082883124727, "grad_norm": 1.5922441482543945, "learning_rate": 5.883668903803132e-05, "loss": 0.9034, "step": 526 }, { "epoch": 0.035435052515016274, "grad_norm": 2.062743663787842, "learning_rate": 5.906040268456377e-05, "loss": 0.9114, "step": 528 }, { "epoch": 0.03556927619878528, "grad_norm": 1.4497499465942383, "learning_rate": 5.9284116331096195e-05, "loss": 0.8768, "step": 530 }, { "epoch": 0.03570349988255428, "grad_norm": 2.300522565841675, "learning_rate": 5.9507829977628635e-05, "loss": 0.9472, "step": 532 }, { "epoch": 0.035837723566323275, "grad_norm": 2.232165575027466, "learning_rate": 5.973154362416108e-05, "loss": 0.8641, "step": 534 }, { "epoch": 0.03597194725009228, "grad_norm": 4.513120174407959, "learning_rate": 5.995525727069351e-05, "loss": 0.8452, "step": 536 }, { "epoch": 0.03610617093386128, "grad_norm": 1.7872861623764038, "learning_rate": 6.017897091722595e-05, "loss": 1.0023, "step": 538 }, { "epoch": 0.03624039461763028, "grad_norm": 3.58272647857666, "learning_rate": 6.04026845637584e-05, "loss": 0.9371, "step": 540 }, { "epoch": 0.036374618301399285, "grad_norm": 1.8548665046691895, "learning_rate": 6.062639821029084e-05, "loss": 0.9036, "step": 542 }, { "epoch": 0.03650884198516828, "grad_norm": 1.9858945608139038, "learning_rate": 6.0850111856823265e-05, "loss": 0.9069, "step": 544 }, { "epoch": 0.03664306566893728, "grad_norm": 1.9066975116729736, "learning_rate": 6.107382550335571e-05, "loss": 0.9916, "step": 546 }, { "epoch": 0.036777289352706286, "grad_norm": 1.6789088249206543, "learning_rate": 6.129753914988815e-05, "loss": 0.9138, "step": 548 }, { "epoch": 0.03691151303647529, "grad_norm": 2.7989633083343506, "learning_rate": 6.152125279642058e-05, "loss": 0.895, "step": 550 }, { "epoch": 0.03704573672024429, "grad_norm": 2.9937233924865723, "learning_rate": 6.174496644295302e-05, "loss": 0.9045, "step": 552 }, { "epoch": 0.037179960404013286, "grad_norm": 2.912078380584717, "learning_rate": 6.196868008948546e-05, "loss": 0.9123, "step": 554 }, { "epoch": 0.03731418408778229, "grad_norm": 2.75984525680542, "learning_rate": 6.21923937360179e-05, "loss": 0.9155, "step": 556 }, { "epoch": 0.03744840777155129, "grad_norm": 2.853611707687378, "learning_rate": 6.241610738255034e-05, "loss": 0.9572, "step": 558 }, { "epoch": 0.037582631455320294, "grad_norm": 2.731768846511841, "learning_rate": 6.263982102908278e-05, "loss": 0.8789, "step": 560 }, { "epoch": 0.03771685513908929, "grad_norm": 1.396867036819458, "learning_rate": 6.286353467561522e-05, "loss": 0.9149, "step": 562 }, { "epoch": 0.03785107882285829, "grad_norm": 3.020078659057617, "learning_rate": 6.308724832214765e-05, "loss": 0.9413, "step": 564 }, { "epoch": 0.037985302506627294, "grad_norm": 9.90017318725586, "learning_rate": 6.331096196868009e-05, "loss": 0.8593, "step": 566 }, { "epoch": 0.0381195261903963, "grad_norm": 3.2886898517608643, "learning_rate": 6.353467561521253e-05, "loss": 0.9688, "step": 568 }, { "epoch": 0.0382537498741653, "grad_norm": 2.8568029403686523, "learning_rate": 6.375838926174497e-05, "loss": 0.8769, "step": 570 }, { "epoch": 0.038387973557934295, "grad_norm": 1.5326567888259888, "learning_rate": 6.398210290827741e-05, "loss": 0.8189, "step": 572 }, { "epoch": 0.0385221972417033, "grad_norm": 9.17284107208252, "learning_rate": 6.420581655480985e-05, "loss": 0.9056, "step": 574 }, { "epoch": 0.0386564209254723, "grad_norm": 1.7160284519195557, "learning_rate": 6.442953020134228e-05, "loss": 0.8903, "step": 576 }, { "epoch": 0.0387906446092413, "grad_norm": 1.760678768157959, "learning_rate": 6.465324384787472e-05, "loss": 0.9099, "step": 578 }, { "epoch": 0.038924868293010305, "grad_norm": 3.0596044063568115, "learning_rate": 6.487695749440716e-05, "loss": 0.8456, "step": 580 }, { "epoch": 0.0390590919767793, "grad_norm": 1.4947329759597778, "learning_rate": 6.51006711409396e-05, "loss": 0.9065, "step": 582 }, { "epoch": 0.0391933156605483, "grad_norm": 2.0749616622924805, "learning_rate": 6.532438478747204e-05, "loss": 0.9234, "step": 584 }, { "epoch": 0.039327539344317305, "grad_norm": 1.5716087818145752, "learning_rate": 6.554809843400448e-05, "loss": 0.9152, "step": 586 }, { "epoch": 0.03946176302808631, "grad_norm": 2.856884717941284, "learning_rate": 6.577181208053692e-05, "loss": 0.8517, "step": 588 }, { "epoch": 0.039595986711855304, "grad_norm": 1.5679426193237305, "learning_rate": 6.599552572706935e-05, "loss": 0.8785, "step": 590 }, { "epoch": 0.039730210395624306, "grad_norm": 1.7654825448989868, "learning_rate": 6.621923937360179e-05, "loss": 0.8845, "step": 592 }, { "epoch": 0.03986443407939331, "grad_norm": 2.7273428440093994, "learning_rate": 6.644295302013423e-05, "loss": 0.9262, "step": 594 }, { "epoch": 0.03999865776316231, "grad_norm": 1.99534273147583, "learning_rate": 6.666666666666667e-05, "loss": 0.8871, "step": 596 }, { "epoch": 0.040132881446931314, "grad_norm": 2.444655656814575, "learning_rate": 6.689038031319911e-05, "loss": 0.9251, "step": 598 }, { "epoch": 0.04026710513070031, "grad_norm": 3.578488826751709, "learning_rate": 6.711409395973155e-05, "loss": 0.8834, "step": 600 }, { "epoch": 0.04040132881446931, "grad_norm": 6.7625837326049805, "learning_rate": 6.733780760626398e-05, "loss": 0.9143, "step": 602 }, { "epoch": 0.040535552498238314, "grad_norm": 1.28508722782135, "learning_rate": 6.756152125279642e-05, "loss": 0.8798, "step": 604 }, { "epoch": 0.04066977618200732, "grad_norm": 1.5711547136306763, "learning_rate": 6.778523489932886e-05, "loss": 0.9331, "step": 606 }, { "epoch": 0.04080399986577632, "grad_norm": 1.9061214923858643, "learning_rate": 6.800894854586131e-05, "loss": 0.918, "step": 608 }, { "epoch": 0.040938223549545315, "grad_norm": 1.4584358930587769, "learning_rate": 6.823266219239374e-05, "loss": 0.8318, "step": 610 }, { "epoch": 0.04107244723331432, "grad_norm": 1.6867270469665527, "learning_rate": 6.845637583892618e-05, "loss": 0.9338, "step": 612 }, { "epoch": 0.04120667091708332, "grad_norm": 3.235154151916504, "learning_rate": 6.868008948545862e-05, "loss": 0.9091, "step": 614 }, { "epoch": 0.04134089460085232, "grad_norm": 1.7500040531158447, "learning_rate": 6.890380313199105e-05, "loss": 0.9071, "step": 616 }, { "epoch": 0.041475118284621325, "grad_norm": 1.5308209657669067, "learning_rate": 6.912751677852349e-05, "loss": 0.8818, "step": 618 }, { "epoch": 0.04160934196839032, "grad_norm": 1.5142631530761719, "learning_rate": 6.935123042505593e-05, "loss": 0.98, "step": 620 }, { "epoch": 0.04174356565215932, "grad_norm": 1.6299635171890259, "learning_rate": 6.957494407158837e-05, "loss": 0.8267, "step": 622 }, { "epoch": 0.041877789335928325, "grad_norm": 1.7747793197631836, "learning_rate": 6.979865771812081e-05, "loss": 0.8581, "step": 624 }, { "epoch": 0.04201201301969733, "grad_norm": 1.6121944189071655, "learning_rate": 7.002237136465325e-05, "loss": 0.8964, "step": 626 }, { "epoch": 0.042146236703466324, "grad_norm": 1.6563115119934082, "learning_rate": 7.024608501118568e-05, "loss": 0.9141, "step": 628 }, { "epoch": 0.042280460387235326, "grad_norm": 1.6622352600097656, "learning_rate": 7.046979865771812e-05, "loss": 0.8935, "step": 630 }, { "epoch": 0.04241468407100433, "grad_norm": 1.409108281135559, "learning_rate": 7.069351230425056e-05, "loss": 0.9248, "step": 632 }, { "epoch": 0.04254890775477333, "grad_norm": 1.6550589799880981, "learning_rate": 7.091722595078301e-05, "loss": 0.997, "step": 634 }, { "epoch": 0.042683131438542334, "grad_norm": 1.4970465898513794, "learning_rate": 7.114093959731544e-05, "loss": 0.844, "step": 636 }, { "epoch": 0.04281735512231133, "grad_norm": 2.5428895950317383, "learning_rate": 7.136465324384788e-05, "loss": 0.8903, "step": 638 }, { "epoch": 0.04295157880608033, "grad_norm": 1.9868848323822021, "learning_rate": 7.158836689038032e-05, "loss": 0.8593, "step": 640 }, { "epoch": 0.043085802489849334, "grad_norm": 1.4770828485488892, "learning_rate": 7.181208053691275e-05, "loss": 0.8941, "step": 642 }, { "epoch": 0.04322002617361834, "grad_norm": 1.4597129821777344, "learning_rate": 7.203579418344519e-05, "loss": 0.8786, "step": 644 }, { "epoch": 0.04335424985738734, "grad_norm": 2.5897531509399414, "learning_rate": 7.225950782997764e-05, "loss": 0.9089, "step": 646 }, { "epoch": 0.043488473541156335, "grad_norm": 3.4969544410705566, "learning_rate": 7.248322147651007e-05, "loss": 0.8873, "step": 648 }, { "epoch": 0.04362269722492534, "grad_norm": 2.3789241313934326, "learning_rate": 7.270693512304251e-05, "loss": 0.8342, "step": 650 }, { "epoch": 0.04375692090869434, "grad_norm": 2.3509631156921387, "learning_rate": 7.293064876957495e-05, "loss": 0.9211, "step": 652 }, { "epoch": 0.04389114459246334, "grad_norm": 2.083207845687866, "learning_rate": 7.315436241610739e-05, "loss": 0.9198, "step": 654 }, { "epoch": 0.04402536827623234, "grad_norm": 3.4255850315093994, "learning_rate": 7.337807606263982e-05, "loss": 0.9134, "step": 656 }, { "epoch": 0.04415959196000134, "grad_norm": 3.6358864307403564, "learning_rate": 7.360178970917227e-05, "loss": 0.8847, "step": 658 }, { "epoch": 0.04429381564377034, "grad_norm": 2.201115131378174, "learning_rate": 7.382550335570471e-05, "loss": 0.9123, "step": 660 }, { "epoch": 0.044428039327539345, "grad_norm": 1.5794076919555664, "learning_rate": 7.404921700223714e-05, "loss": 0.8586, "step": 662 }, { "epoch": 0.04456226301130835, "grad_norm": 4.102825164794922, "learning_rate": 7.427293064876958e-05, "loss": 0.9089, "step": 664 }, { "epoch": 0.04469648669507734, "grad_norm": 1.4240411520004272, "learning_rate": 7.449664429530202e-05, "loss": 0.87, "step": 666 }, { "epoch": 0.044830710378846346, "grad_norm": 1.6875033378601074, "learning_rate": 7.472035794183445e-05, "loss": 0.8644, "step": 668 }, { "epoch": 0.04496493406261535, "grad_norm": 1.5658398866653442, "learning_rate": 7.494407158836689e-05, "loss": 0.9216, "step": 670 }, { "epoch": 0.04509915774638435, "grad_norm": 1.396859884262085, "learning_rate": 7.516778523489934e-05, "loss": 0.8531, "step": 672 }, { "epoch": 0.045233381430153353, "grad_norm": 1.5729717016220093, "learning_rate": 7.539149888143177e-05, "loss": 0.9405, "step": 674 }, { "epoch": 0.04536760511392235, "grad_norm": 2.3833537101745605, "learning_rate": 7.561521252796421e-05, "loss": 0.9004, "step": 676 }, { "epoch": 0.04550182879769135, "grad_norm": 2.166736364364624, "learning_rate": 7.583892617449665e-05, "loss": 0.9107, "step": 678 }, { "epoch": 0.045636052481460354, "grad_norm": 1.4044255018234253, "learning_rate": 7.606263982102909e-05, "loss": 0.8947, "step": 680 }, { "epoch": 0.04577027616522936, "grad_norm": 1.9272996187210083, "learning_rate": 7.628635346756152e-05, "loss": 0.9089, "step": 682 }, { "epoch": 0.04590449984899836, "grad_norm": 3.3325276374816895, "learning_rate": 7.651006711409397e-05, "loss": 0.8839, "step": 684 }, { "epoch": 0.046038723532767355, "grad_norm": 2.3696067333221436, "learning_rate": 7.673378076062641e-05, "loss": 0.8734, "step": 686 }, { "epoch": 0.04617294721653636, "grad_norm": 1.3236571550369263, "learning_rate": 7.695749440715884e-05, "loss": 0.8133, "step": 688 }, { "epoch": 0.04630717090030536, "grad_norm": 1.7845690250396729, "learning_rate": 7.718120805369128e-05, "loss": 0.8593, "step": 690 }, { "epoch": 0.04644139458407436, "grad_norm": 1.9854339361190796, "learning_rate": 7.740492170022372e-05, "loss": 0.8604, "step": 692 }, { "epoch": 0.04657561826784336, "grad_norm": 2.004243850708008, "learning_rate": 7.762863534675615e-05, "loss": 0.9429, "step": 694 }, { "epoch": 0.04670984195161236, "grad_norm": 1.6457959413528442, "learning_rate": 7.78523489932886e-05, "loss": 0.9183, "step": 696 }, { "epoch": 0.04684406563538136, "grad_norm": 1.4421991109848022, "learning_rate": 7.807606263982104e-05, "loss": 0.8882, "step": 698 }, { "epoch": 0.046978289319150365, "grad_norm": 1.4433611631393433, "learning_rate": 7.829977628635348e-05, "loss": 0.8521, "step": 700 }, { "epoch": 0.04711251300291937, "grad_norm": 1.7360423803329468, "learning_rate": 7.852348993288591e-05, "loss": 0.8972, "step": 702 }, { "epoch": 0.04724673668668836, "grad_norm": 2.0711636543273926, "learning_rate": 7.874720357941835e-05, "loss": 0.896, "step": 704 }, { "epoch": 0.047380960370457366, "grad_norm": 1.478452444076538, "learning_rate": 7.897091722595079e-05, "loss": 0.8851, "step": 706 }, { "epoch": 0.04751518405422637, "grad_norm": 1.7454901933670044, "learning_rate": 7.919463087248322e-05, "loss": 0.9579, "step": 708 }, { "epoch": 0.04764940773799537, "grad_norm": 2.5465352535247803, "learning_rate": 7.941834451901567e-05, "loss": 0.8532, "step": 710 }, { "epoch": 0.04778363142176437, "grad_norm": 1.8807421922683716, "learning_rate": 7.964205816554811e-05, "loss": 0.8988, "step": 712 }, { "epoch": 0.04791785510553337, "grad_norm": 1.62044095993042, "learning_rate": 7.986577181208054e-05, "loss": 0.8757, "step": 714 }, { "epoch": 0.04805207878930237, "grad_norm": 1.3398281335830688, "learning_rate": 8.008948545861298e-05, "loss": 0.838, "step": 716 }, { "epoch": 0.048186302473071374, "grad_norm": 1.6650171279907227, "learning_rate": 8.031319910514542e-05, "loss": 0.9152, "step": 718 }, { "epoch": 0.04832052615684038, "grad_norm": 1.874007225036621, "learning_rate": 8.053691275167784e-05, "loss": 0.9129, "step": 720 }, { "epoch": 0.04845474984060937, "grad_norm": 1.5779155492782593, "learning_rate": 8.07606263982103e-05, "loss": 0.8926, "step": 722 }, { "epoch": 0.048588973524378375, "grad_norm": 1.4982414245605469, "learning_rate": 8.098434004474274e-05, "loss": 0.9389, "step": 724 }, { "epoch": 0.04872319720814738, "grad_norm": 2.013594388961792, "learning_rate": 8.120805369127518e-05, "loss": 0.8542, "step": 726 }, { "epoch": 0.04885742089191638, "grad_norm": 1.6982237100601196, "learning_rate": 8.14317673378076e-05, "loss": 0.9011, "step": 728 }, { "epoch": 0.04899164457568538, "grad_norm": 3.0389931201934814, "learning_rate": 8.165548098434005e-05, "loss": 0.8657, "step": 730 }, { "epoch": 0.04912586825945438, "grad_norm": 2.0701937675476074, "learning_rate": 8.187919463087249e-05, "loss": 0.8703, "step": 732 }, { "epoch": 0.04926009194322338, "grad_norm": 1.7981239557266235, "learning_rate": 8.210290827740493e-05, "loss": 0.9151, "step": 734 }, { "epoch": 0.04939431562699238, "grad_norm": 2.2304553985595703, "learning_rate": 8.232662192393737e-05, "loss": 0.8721, "step": 736 }, { "epoch": 0.049528539310761385, "grad_norm": 2.8598451614379883, "learning_rate": 8.255033557046981e-05, "loss": 0.925, "step": 738 }, { "epoch": 0.04966276299453039, "grad_norm": 1.5226657390594482, "learning_rate": 8.277404921700224e-05, "loss": 0.8633, "step": 740 }, { "epoch": 0.04979698667829938, "grad_norm": 1.9706357717514038, "learning_rate": 8.299776286353468e-05, "loss": 0.8499, "step": 742 }, { "epoch": 0.049931210362068386, "grad_norm": 1.6518230438232422, "learning_rate": 8.322147651006712e-05, "loss": 0.7827, "step": 744 }, { "epoch": 0.05006543404583739, "grad_norm": 1.5772185325622559, "learning_rate": 8.344519015659956e-05, "loss": 0.946, "step": 746 }, { "epoch": 0.05019965772960639, "grad_norm": 2.105870246887207, "learning_rate": 8.3668903803132e-05, "loss": 0.849, "step": 748 }, { "epoch": 0.05033388141337539, "grad_norm": 1.616303563117981, "learning_rate": 8.389261744966444e-05, "loss": 0.8467, "step": 750 }, { "epoch": 0.05046810509714439, "grad_norm": 2.292529582977295, "learning_rate": 8.411633109619688e-05, "loss": 0.9027, "step": 752 }, { "epoch": 0.05060232878091339, "grad_norm": 1.5480836629867554, "learning_rate": 8.43400447427293e-05, "loss": 0.8651, "step": 754 }, { "epoch": 0.050736552464682394, "grad_norm": 1.5585349798202515, "learning_rate": 8.456375838926175e-05, "loss": 0.8554, "step": 756 }, { "epoch": 0.050870776148451397, "grad_norm": 1.6302483081817627, "learning_rate": 8.478747203579419e-05, "loss": 0.8242, "step": 758 }, { "epoch": 0.05100499983222039, "grad_norm": 1.8211963176727295, "learning_rate": 8.501118568232663e-05, "loss": 0.9038, "step": 760 }, { "epoch": 0.051139223515989395, "grad_norm": 1.9375659227371216, "learning_rate": 8.523489932885907e-05, "loss": 0.8742, "step": 762 }, { "epoch": 0.0512734471997584, "grad_norm": 1.5411709547042847, "learning_rate": 8.545861297539151e-05, "loss": 0.8768, "step": 764 }, { "epoch": 0.0514076708835274, "grad_norm": 1.6576482057571411, "learning_rate": 8.568232662192394e-05, "loss": 0.8962, "step": 766 }, { "epoch": 0.0515418945672964, "grad_norm": 1.7040925025939941, "learning_rate": 8.590604026845638e-05, "loss": 0.8104, "step": 768 }, { "epoch": 0.0516761182510654, "grad_norm": 1.7019544839859009, "learning_rate": 8.612975391498882e-05, "loss": 0.885, "step": 770 }, { "epoch": 0.0518103419348344, "grad_norm": 1.4850620031356812, "learning_rate": 8.635346756152126e-05, "loss": 0.9345, "step": 772 }, { "epoch": 0.0519445656186034, "grad_norm": 2.2853121757507324, "learning_rate": 8.65771812080537e-05, "loss": 0.9004, "step": 774 }, { "epoch": 0.052078789302372405, "grad_norm": 1.718302845954895, "learning_rate": 8.680089485458614e-05, "loss": 0.8885, "step": 776 }, { "epoch": 0.05221301298614141, "grad_norm": 2.4001340866088867, "learning_rate": 8.702460850111858e-05, "loss": 0.883, "step": 778 }, { "epoch": 0.0523472366699104, "grad_norm": 1.4307639598846436, "learning_rate": 8.7248322147651e-05, "loss": 0.8925, "step": 780 }, { "epoch": 0.052481460353679406, "grad_norm": 1.3757809400558472, "learning_rate": 8.747203579418345e-05, "loss": 0.8874, "step": 782 }, { "epoch": 0.05261568403744841, "grad_norm": 1.4396053552627563, "learning_rate": 8.769574944071589e-05, "loss": 0.9032, "step": 784 }, { "epoch": 0.05274990772121741, "grad_norm": 1.6641589403152466, "learning_rate": 8.791946308724833e-05, "loss": 0.9667, "step": 786 }, { "epoch": 0.052884131404986406, "grad_norm": 1.66237211227417, "learning_rate": 8.814317673378077e-05, "loss": 0.9228, "step": 788 }, { "epoch": 0.05301835508875541, "grad_norm": 1.4634531736373901, "learning_rate": 8.836689038031321e-05, "loss": 0.9049, "step": 790 }, { "epoch": 0.05315257877252441, "grad_norm": 2.0408074855804443, "learning_rate": 8.859060402684565e-05, "loss": 0.9163, "step": 792 }, { "epoch": 0.053286802456293414, "grad_norm": 1.4953192472457886, "learning_rate": 8.881431767337808e-05, "loss": 0.922, "step": 794 }, { "epoch": 0.053421026140062416, "grad_norm": 2.698669672012329, "learning_rate": 8.903803131991052e-05, "loss": 0.8641, "step": 796 }, { "epoch": 0.05355524982383141, "grad_norm": 1.5639328956604004, "learning_rate": 8.926174496644296e-05, "loss": 0.8754, "step": 798 }, { "epoch": 0.053689473507600415, "grad_norm": 1.4404573440551758, "learning_rate": 8.94854586129754e-05, "loss": 0.8646, "step": 800 }, { "epoch": 0.05382369719136942, "grad_norm": 2.0549895763397217, "learning_rate": 8.970917225950784e-05, "loss": 0.8184, "step": 802 }, { "epoch": 0.05395792087513842, "grad_norm": 2.4482076168060303, "learning_rate": 8.993288590604028e-05, "loss": 0.8865, "step": 804 }, { "epoch": 0.05409214455890742, "grad_norm": 2.617295980453491, "learning_rate": 9.01565995525727e-05, "loss": 0.9126, "step": 806 }, { "epoch": 0.05422636824267642, "grad_norm": 1.6565312147140503, "learning_rate": 9.038031319910515e-05, "loss": 0.8668, "step": 808 }, { "epoch": 0.05436059192644542, "grad_norm": 1.6836875677108765, "learning_rate": 9.060402684563759e-05, "loss": 0.8412, "step": 810 }, { "epoch": 0.05449481561021442, "grad_norm": 3.3620588779449463, "learning_rate": 9.082774049217003e-05, "loss": 0.8967, "step": 812 }, { "epoch": 0.054629039293983425, "grad_norm": 1.7039563655853271, "learning_rate": 9.105145413870247e-05, "loss": 0.8374, "step": 814 }, { "epoch": 0.05476326297775243, "grad_norm": 1.5863765478134155, "learning_rate": 9.127516778523491e-05, "loss": 0.821, "step": 816 }, { "epoch": 0.05489748666152142, "grad_norm": 1.436874270439148, "learning_rate": 9.149888143176735e-05, "loss": 0.9001, "step": 818 }, { "epoch": 0.055031710345290426, "grad_norm": 1.9585341215133667, "learning_rate": 9.172259507829977e-05, "loss": 0.8977, "step": 820 }, { "epoch": 0.05516593402905943, "grad_norm": 1.5336610078811646, "learning_rate": 9.194630872483221e-05, "loss": 0.8992, "step": 822 }, { "epoch": 0.05530015771282843, "grad_norm": 2.7829861640930176, "learning_rate": 9.217002237136466e-05, "loss": 0.922, "step": 824 }, { "epoch": 0.055434381396597426, "grad_norm": 1.4213629961013794, "learning_rate": 9.23937360178971e-05, "loss": 0.9008, "step": 826 }, { "epoch": 0.05556860508036643, "grad_norm": 1.4592857360839844, "learning_rate": 9.261744966442954e-05, "loss": 0.884, "step": 828 }, { "epoch": 0.05570282876413543, "grad_norm": 1.7585809230804443, "learning_rate": 9.284116331096198e-05, "loss": 0.8807, "step": 830 }, { "epoch": 0.055837052447904434, "grad_norm": 1.8087443113327026, "learning_rate": 9.30648769574944e-05, "loss": 0.867, "step": 832 }, { "epoch": 0.055971276131673436, "grad_norm": 2.120878219604492, "learning_rate": 9.328859060402684e-05, "loss": 0.8508, "step": 834 }, { "epoch": 0.05610549981544243, "grad_norm": 3.5708112716674805, "learning_rate": 9.351230425055928e-05, "loss": 0.9109, "step": 836 }, { "epoch": 0.056239723499211434, "grad_norm": 1.4622738361358643, "learning_rate": 9.373601789709174e-05, "loss": 0.8472, "step": 838 }, { "epoch": 0.05637394718298044, "grad_norm": 1.4327478408813477, "learning_rate": 9.395973154362417e-05, "loss": 0.8546, "step": 840 }, { "epoch": 0.05650817086674944, "grad_norm": 1.8220202922821045, "learning_rate": 9.41834451901566e-05, "loss": 0.9217, "step": 842 }, { "epoch": 0.05664239455051844, "grad_norm": 3.180845260620117, "learning_rate": 9.440715883668905e-05, "loss": 0.9166, "step": 844 }, { "epoch": 0.05677661823428744, "grad_norm": 1.602735161781311, "learning_rate": 9.463087248322147e-05, "loss": 0.9221, "step": 846 }, { "epoch": 0.05691084191805644, "grad_norm": 1.3909780979156494, "learning_rate": 9.485458612975391e-05, "loss": 0.8352, "step": 848 }, { "epoch": 0.05704506560182544, "grad_norm": 1.4142225980758667, "learning_rate": 9.507829977628635e-05, "loss": 0.8591, "step": 850 }, { "epoch": 0.057179289285594445, "grad_norm": 3.468468427658081, "learning_rate": 9.53020134228188e-05, "loss": 0.8948, "step": 852 }, { "epoch": 0.05731351296936345, "grad_norm": 1.839051604270935, "learning_rate": 9.552572706935124e-05, "loss": 0.8648, "step": 854 }, { "epoch": 0.05744773665313244, "grad_norm": 1.7877297401428223, "learning_rate": 9.574944071588368e-05, "loss": 0.8961, "step": 856 }, { "epoch": 0.057581960336901446, "grad_norm": 1.9843209981918335, "learning_rate": 9.59731543624161e-05, "loss": 0.8792, "step": 858 }, { "epoch": 0.05771618402067045, "grad_norm": 1.921899437904358, "learning_rate": 9.619686800894854e-05, "loss": 0.8588, "step": 860 }, { "epoch": 0.05785040770443945, "grad_norm": 2.093482255935669, "learning_rate": 9.642058165548098e-05, "loss": 0.9604, "step": 862 }, { "epoch": 0.057984631388208446, "grad_norm": 1.8832120895385742, "learning_rate": 9.664429530201344e-05, "loss": 0.908, "step": 864 }, { "epoch": 0.05811885507197745, "grad_norm": 1.2885879278182983, "learning_rate": 9.686800894854587e-05, "loss": 0.8453, "step": 866 }, { "epoch": 0.05825307875574645, "grad_norm": 1.905735969543457, "learning_rate": 9.70917225950783e-05, "loss": 0.9275, "step": 868 }, { "epoch": 0.058387302439515454, "grad_norm": 2.0157928466796875, "learning_rate": 9.731543624161075e-05, "loss": 0.8819, "step": 870 }, { "epoch": 0.058521526123284456, "grad_norm": 1.3370037078857422, "learning_rate": 9.753914988814317e-05, "loss": 0.8153, "step": 872 }, { "epoch": 0.05865574980705345, "grad_norm": 1.4842098951339722, "learning_rate": 9.776286353467561e-05, "loss": 0.8796, "step": 874 }, { "epoch": 0.058789973490822454, "grad_norm": 1.4972621202468872, "learning_rate": 9.798657718120807e-05, "loss": 0.8647, "step": 876 }, { "epoch": 0.05892419717459146, "grad_norm": 1.740038275718689, "learning_rate": 9.82102908277405e-05, "loss": 0.8925, "step": 878 }, { "epoch": 0.05905842085836046, "grad_norm": 1.52963387966156, "learning_rate": 9.843400447427293e-05, "loss": 0.9363, "step": 880 }, { "epoch": 0.05919264454212946, "grad_norm": 2.290513038635254, "learning_rate": 9.865771812080538e-05, "loss": 0.8879, "step": 882 }, { "epoch": 0.05932686822589846, "grad_norm": 1.6631953716278076, "learning_rate": 9.888143176733782e-05, "loss": 0.8762, "step": 884 }, { "epoch": 0.05946109190966746, "grad_norm": 1.472398042678833, "learning_rate": 9.910514541387024e-05, "loss": 0.8953, "step": 886 }, { "epoch": 0.05959531559343646, "grad_norm": 1.6663011312484741, "learning_rate": 9.93288590604027e-05, "loss": 0.8678, "step": 888 }, { "epoch": 0.059729539277205465, "grad_norm": 2.3357253074645996, "learning_rate": 9.955257270693514e-05, "loss": 0.898, "step": 890 }, { "epoch": 0.05986376296097446, "grad_norm": 1.9164422750473022, "learning_rate": 9.977628635346756e-05, "loss": 0.8616, "step": 892 }, { "epoch": 0.05999798664474346, "grad_norm": 1.9118332862854004, "learning_rate": 0.0001, "loss": 0.9123, "step": 894 }, { "epoch": 0.060132210328512466, "grad_norm": 1.4247066974639893, "learning_rate": 9.999999881879917e-05, "loss": 0.8485, "step": 896 }, { "epoch": 0.06026643401228147, "grad_norm": 1.6532388925552368, "learning_rate": 9.99999952751967e-05, "loss": 0.8958, "step": 898 }, { "epoch": 0.06040065769605047, "grad_norm": 1.411152958869934, "learning_rate": 9.999998936919278e-05, "loss": 0.8478, "step": 900 }, { "epoch": 0.060534881379819466, "grad_norm": 1.296415090560913, "learning_rate": 9.999998110078769e-05, "loss": 0.8662, "step": 902 }, { "epoch": 0.06066910506358847, "grad_norm": 1.436594009399414, "learning_rate": 9.99999704699818e-05, "loss": 0.8919, "step": 904 }, { "epoch": 0.06080332874735747, "grad_norm": 1.9760971069335938, "learning_rate": 9.999995747677564e-05, "loss": 0.801, "step": 906 }, { "epoch": 0.060937552431126474, "grad_norm": 2.1906111240386963, "learning_rate": 9.99999421211698e-05, "loss": 0.85, "step": 908 }, { "epoch": 0.061071776114895476, "grad_norm": 1.5004881620407104, "learning_rate": 9.999992440316502e-05, "loss": 0.8712, "step": 910 }, { "epoch": 0.06120599979866447, "grad_norm": 1.4594885110855103, "learning_rate": 9.999990432276214e-05, "loss": 0.8529, "step": 912 }, { "epoch": 0.061340223482433474, "grad_norm": 1.8031158447265625, "learning_rate": 9.999988187996208e-05, "loss": 0.885, "step": 914 }, { "epoch": 0.06147444716620248, "grad_norm": 1.571834921836853, "learning_rate": 9.999985707476594e-05, "loss": 0.8803, "step": 916 }, { "epoch": 0.06160867084997148, "grad_norm": 1.6122395992279053, "learning_rate": 9.999982990717487e-05, "loss": 0.882, "step": 918 }, { "epoch": 0.06174289453374048, "grad_norm": 2.581127643585205, "learning_rate": 9.999980037719016e-05, "loss": 0.8602, "step": 920 }, { "epoch": 0.06187711821750948, "grad_norm": 1.770944595336914, "learning_rate": 9.99997684848132e-05, "loss": 0.9097, "step": 922 }, { "epoch": 0.06201134190127848, "grad_norm": 1.7628514766693115, "learning_rate": 9.99997342300455e-05, "loss": 0.8812, "step": 924 }, { "epoch": 0.06214556558504748, "grad_norm": 1.4451676607131958, "learning_rate": 9.999969761288868e-05, "loss": 0.8952, "step": 926 }, { "epoch": 0.062279789268816485, "grad_norm": 2.1536195278167725, "learning_rate": 9.999965863334445e-05, "loss": 0.8244, "step": 928 }, { "epoch": 0.06241401295258548, "grad_norm": 1.5444164276123047, "learning_rate": 9.99996172914147e-05, "loss": 0.9155, "step": 930 }, { "epoch": 0.06254823663635449, "grad_norm": 1.5740525722503662, "learning_rate": 9.999957358710132e-05, "loss": 0.8623, "step": 932 }, { "epoch": 0.06268246032012349, "grad_norm": 1.4494963884353638, "learning_rate": 9.999952752040643e-05, "loss": 0.9155, "step": 934 }, { "epoch": 0.06281668400389248, "grad_norm": 1.6744060516357422, "learning_rate": 9.999947909133219e-05, "loss": 0.8704, "step": 936 }, { "epoch": 0.06295090768766148, "grad_norm": 1.5943561792373657, "learning_rate": 9.999942829988086e-05, "loss": 0.8005, "step": 938 }, { "epoch": 0.06308513137143049, "grad_norm": 1.6639518737792969, "learning_rate": 9.999937514605486e-05, "loss": 0.834, "step": 940 }, { "epoch": 0.06321935505519949, "grad_norm": 1.3946268558502197, "learning_rate": 9.999931962985674e-05, "loss": 0.8614, "step": 942 }, { "epoch": 0.06335357873896849, "grad_norm": 1.7072865962982178, "learning_rate": 9.999926175128905e-05, "loss": 0.8979, "step": 944 }, { "epoch": 0.0634878024227375, "grad_norm": 2.105038642883301, "learning_rate": 9.999920151035458e-05, "loss": 0.9333, "step": 946 }, { "epoch": 0.0636220261065065, "grad_norm": 1.498768925666809, "learning_rate": 9.999913890705616e-05, "loss": 0.8761, "step": 948 }, { "epoch": 0.0637562497902755, "grad_norm": 1.3514175415039062, "learning_rate": 9.999907394139674e-05, "loss": 0.8673, "step": 950 }, { "epoch": 0.0638904734740445, "grad_norm": 1.5056959390640259, "learning_rate": 9.99990066133794e-05, "loss": 0.89, "step": 952 }, { "epoch": 0.06402469715781349, "grad_norm": 1.3694546222686768, "learning_rate": 9.99989369230073e-05, "loss": 0.8391, "step": 954 }, { "epoch": 0.06415892084158249, "grad_norm": 1.7158812284469604, "learning_rate": 9.999886487028376e-05, "loss": 0.9251, "step": 956 }, { "epoch": 0.0642931445253515, "grad_norm": 1.927911400794983, "learning_rate": 9.999879045521218e-05, "loss": 0.9081, "step": 958 }, { "epoch": 0.0644273682091205, "grad_norm": 1.925579309463501, "learning_rate": 9.999871367779606e-05, "loss": 0.854, "step": 960 }, { "epoch": 0.0645615918928895, "grad_norm": 1.2850353717803955, "learning_rate": 9.999863453803904e-05, "loss": 0.8755, "step": 962 }, { "epoch": 0.0646958155766585, "grad_norm": 1.4252276420593262, "learning_rate": 9.999855303594485e-05, "loss": 0.8555, "step": 964 }, { "epoch": 0.0648300392604275, "grad_norm": 3.0897445678710938, "learning_rate": 9.999846917151737e-05, "loss": 0.8602, "step": 966 }, { "epoch": 0.06496426294419651, "grad_norm": 1.5266180038452148, "learning_rate": 9.999838294476051e-05, "loss": 0.8598, "step": 968 }, { "epoch": 0.06509848662796551, "grad_norm": 2.7690606117248535, "learning_rate": 9.99982943556784e-05, "loss": 0.885, "step": 970 }, { "epoch": 0.0652327103117345, "grad_norm": 1.3687183856964111, "learning_rate": 9.999820340427517e-05, "loss": 0.8545, "step": 972 }, { "epoch": 0.0653669339955035, "grad_norm": 1.5934265851974487, "learning_rate": 9.999811009055518e-05, "loss": 0.8277, "step": 974 }, { "epoch": 0.0655011576792725, "grad_norm": 2.1981096267700195, "learning_rate": 9.999801441452278e-05, "loss": 0.7975, "step": 976 }, { "epoch": 0.0656353813630415, "grad_norm": 1.4350436925888062, "learning_rate": 9.999791637618252e-05, "loss": 0.8324, "step": 978 }, { "epoch": 0.06576960504681051, "grad_norm": 2.4169490337371826, "learning_rate": 9.999781597553903e-05, "loss": 0.8526, "step": 980 }, { "epoch": 0.06590382873057951, "grad_norm": 1.2815511226654053, "learning_rate": 9.999771321259705e-05, "loss": 0.8884, "step": 982 }, { "epoch": 0.06603805241434851, "grad_norm": 1.4573643207550049, "learning_rate": 9.999760808736145e-05, "loss": 0.8733, "step": 984 }, { "epoch": 0.06617227609811752, "grad_norm": 1.8162658214569092, "learning_rate": 9.999750059983716e-05, "loss": 0.8241, "step": 986 }, { "epoch": 0.06630649978188652, "grad_norm": 1.3659766912460327, "learning_rate": 9.999739075002931e-05, "loss": 0.895, "step": 988 }, { "epoch": 0.06644072346565552, "grad_norm": 1.7499064207077026, "learning_rate": 9.999727853794305e-05, "loss": 0.928, "step": 990 }, { "epoch": 0.06657494714942451, "grad_norm": 1.3851163387298584, "learning_rate": 9.999716396358369e-05, "loss": 0.8947, "step": 992 }, { "epoch": 0.06670917083319351, "grad_norm": 2.0054306983947754, "learning_rate": 9.999704702695664e-05, "loss": 0.884, "step": 994 }, { "epoch": 0.06684339451696251, "grad_norm": 1.4638731479644775, "learning_rate": 9.999692772806746e-05, "loss": 0.9536, "step": 996 }, { "epoch": 0.06697761820073152, "grad_norm": 1.4908126592636108, "learning_rate": 9.999680606692174e-05, "loss": 0.8523, "step": 998 }, { "epoch": 0.06711184188450052, "grad_norm": 1.5436162948608398, "learning_rate": 9.999668204352526e-05, "loss": 0.8326, "step": 1000 }, { "epoch": 0.06724606556826952, "grad_norm": 1.7729190587997437, "learning_rate": 9.999655565788385e-05, "loss": 0.859, "step": 1002 }, { "epoch": 0.06738028925203852, "grad_norm": 2.2067604064941406, "learning_rate": 9.99964269100035e-05, "loss": 0.793, "step": 1004 }, { "epoch": 0.06751451293580753, "grad_norm": 1.7928110361099243, "learning_rate": 9.999629579989032e-05, "loss": 0.8179, "step": 1006 }, { "epoch": 0.06764873661957653, "grad_norm": 1.3039089441299438, "learning_rate": 9.999616232755045e-05, "loss": 0.8545, "step": 1008 }, { "epoch": 0.06778296030334552, "grad_norm": 1.576532244682312, "learning_rate": 9.999602649299022e-05, "loss": 0.842, "step": 1010 }, { "epoch": 0.06791718398711452, "grad_norm": 1.6710187196731567, "learning_rate": 9.999588829621606e-05, "loss": 0.911, "step": 1012 }, { "epoch": 0.06805140767088352, "grad_norm": 1.279616117477417, "learning_rate": 9.99957477372345e-05, "loss": 0.8623, "step": 1014 }, { "epoch": 0.06818563135465253, "grad_norm": 1.345814824104309, "learning_rate": 9.999560481605217e-05, "loss": 0.8028, "step": 1016 }, { "epoch": 0.06831985503842153, "grad_norm": 1.5044865608215332, "learning_rate": 9.999545953267582e-05, "loss": 0.8385, "step": 1018 }, { "epoch": 0.06845407872219053, "grad_norm": 1.6629892587661743, "learning_rate": 9.999531188711232e-05, "loss": 0.8897, "step": 1020 }, { "epoch": 0.06858830240595953, "grad_norm": 1.2309874296188354, "learning_rate": 9.999516187936864e-05, "loss": 0.8042, "step": 1022 }, { "epoch": 0.06872252608972854, "grad_norm": 1.2960044145584106, "learning_rate": 9.999500950945188e-05, "loss": 0.8592, "step": 1024 }, { "epoch": 0.06885674977349754, "grad_norm": 3.2066714763641357, "learning_rate": 9.999485477736923e-05, "loss": 0.8411, "step": 1026 }, { "epoch": 0.06899097345726654, "grad_norm": 1.5629687309265137, "learning_rate": 9.999469768312799e-05, "loss": 0.9141, "step": 1028 }, { "epoch": 0.06912519714103553, "grad_norm": 1.96613609790802, "learning_rate": 9.99945382267356e-05, "loss": 0.9002, "step": 1030 }, { "epoch": 0.06925942082480453, "grad_norm": 1.3429502248764038, "learning_rate": 9.999437640819959e-05, "loss": 0.9162, "step": 1032 }, { "epoch": 0.06939364450857353, "grad_norm": 1.2782080173492432, "learning_rate": 9.999421222752763e-05, "loss": 0.8635, "step": 1034 }, { "epoch": 0.06952786819234254, "grad_norm": 1.330491065979004, "learning_rate": 9.999404568472742e-05, "loss": 0.884, "step": 1036 }, { "epoch": 0.06966209187611154, "grad_norm": 1.3575077056884766, "learning_rate": 9.999387677980687e-05, "loss": 0.8559, "step": 1038 }, { "epoch": 0.06979631555988054, "grad_norm": 1.6015293598175049, "learning_rate": 9.999370551277395e-05, "loss": 0.8575, "step": 1040 }, { "epoch": 0.06993053924364954, "grad_norm": 1.6036041975021362, "learning_rate": 9.999353188363676e-05, "loss": 0.8912, "step": 1042 }, { "epoch": 0.07006476292741855, "grad_norm": 1.4285792112350464, "learning_rate": 9.999335589240348e-05, "loss": 0.8507, "step": 1044 }, { "epoch": 0.07019898661118755, "grad_norm": 1.7658170461654663, "learning_rate": 9.999317753908246e-05, "loss": 0.869, "step": 1046 }, { "epoch": 0.07033321029495654, "grad_norm": 1.4544686079025269, "learning_rate": 9.999299682368211e-05, "loss": 0.8702, "step": 1048 }, { "epoch": 0.07046743397872554, "grad_norm": 1.428667426109314, "learning_rate": 9.999281374621095e-05, "loss": 0.8041, "step": 1050 }, { "epoch": 0.07060165766249454, "grad_norm": 1.5885188579559326, "learning_rate": 9.999262830667766e-05, "loss": 0.8531, "step": 1052 }, { "epoch": 0.07073588134626355, "grad_norm": 6.514537334442139, "learning_rate": 9.999244050509098e-05, "loss": 0.869, "step": 1054 }, { "epoch": 0.07087010503003255, "grad_norm": 1.5298740863800049, "learning_rate": 9.999225034145979e-05, "loss": 0.8104, "step": 1056 }, { "epoch": 0.07100432871380155, "grad_norm": 1.3994238376617432, "learning_rate": 9.999205781579309e-05, "loss": 0.8596, "step": 1058 }, { "epoch": 0.07113855239757055, "grad_norm": 1.8185546398162842, "learning_rate": 9.999186292809995e-05, "loss": 0.8485, "step": 1060 }, { "epoch": 0.07127277608133956, "grad_norm": 1.4394915103912354, "learning_rate": 9.99916656783896e-05, "loss": 0.8019, "step": 1062 }, { "epoch": 0.07140699976510856, "grad_norm": 1.567618727684021, "learning_rate": 9.999146606667135e-05, "loss": 0.8233, "step": 1064 }, { "epoch": 0.07154122344887756, "grad_norm": 1.8687235116958618, "learning_rate": 9.999126409295463e-05, "loss": 0.8539, "step": 1066 }, { "epoch": 0.07167544713264655, "grad_norm": 1.3390531539916992, "learning_rate": 9.999105975724898e-05, "loss": 0.8283, "step": 1068 }, { "epoch": 0.07180967081641555, "grad_norm": 2.7048981189727783, "learning_rate": 9.999085305956406e-05, "loss": 0.819, "step": 1070 }, { "epoch": 0.07194389450018455, "grad_norm": 1.721372365951538, "learning_rate": 9.999064399990964e-05, "loss": 0.8514, "step": 1072 }, { "epoch": 0.07207811818395356, "grad_norm": 1.280475378036499, "learning_rate": 9.999043257829561e-05, "loss": 0.7827, "step": 1074 }, { "epoch": 0.07221234186772256, "grad_norm": 3.6173923015594482, "learning_rate": 9.999021879473192e-05, "loss": 0.7947, "step": 1076 }, { "epoch": 0.07234656555149156, "grad_norm": 2.0217697620391846, "learning_rate": 9.99900026492287e-05, "loss": 0.8223, "step": 1078 }, { "epoch": 0.07248078923526056, "grad_norm": 1.4395509958267212, "learning_rate": 9.998978414179617e-05, "loss": 0.8473, "step": 1080 }, { "epoch": 0.07261501291902957, "grad_norm": 1.426866054534912, "learning_rate": 9.998956327244462e-05, "loss": 0.9186, "step": 1082 }, { "epoch": 0.07274923660279857, "grad_norm": 1.339833378791809, "learning_rate": 9.998934004118452e-05, "loss": 0.8104, "step": 1084 }, { "epoch": 0.07288346028656756, "grad_norm": 1.3739291429519653, "learning_rate": 9.99891144480264e-05, "loss": 0.808, "step": 1086 }, { "epoch": 0.07301768397033656, "grad_norm": 1.93855881690979, "learning_rate": 9.99888864929809e-05, "loss": 0.8228, "step": 1088 }, { "epoch": 0.07315190765410556, "grad_norm": 1.507676362991333, "learning_rate": 9.998865617605883e-05, "loss": 0.8457, "step": 1090 }, { "epoch": 0.07328613133787457, "grad_norm": 1.841133952140808, "learning_rate": 9.998842349727107e-05, "loss": 0.8103, "step": 1092 }, { "epoch": 0.07342035502164357, "grad_norm": 1.2533656358718872, "learning_rate": 9.998818845662859e-05, "loss": 0.8232, "step": 1094 }, { "epoch": 0.07355457870541257, "grad_norm": 1.4022369384765625, "learning_rate": 9.998795105414248e-05, "loss": 0.8436, "step": 1096 }, { "epoch": 0.07368880238918157, "grad_norm": 1.5202876329421997, "learning_rate": 9.998771128982399e-05, "loss": 0.8699, "step": 1098 }, { "epoch": 0.07382302607295058, "grad_norm": 1.4362784624099731, "learning_rate": 9.998746916368444e-05, "loss": 0.8737, "step": 1100 }, { "epoch": 0.07395724975671958, "grad_norm": 1.4843257665634155, "learning_rate": 9.998722467573528e-05, "loss": 0.8605, "step": 1102 }, { "epoch": 0.07409147344048858, "grad_norm": 1.275189757347107, "learning_rate": 9.998697782598804e-05, "loss": 0.7974, "step": 1104 }, { "epoch": 0.07422569712425757, "grad_norm": 1.3372191190719604, "learning_rate": 9.998672861445439e-05, "loss": 0.9027, "step": 1106 }, { "epoch": 0.07435992080802657, "grad_norm": 1.7303627729415894, "learning_rate": 9.998647704114612e-05, "loss": 0.8333, "step": 1108 }, { "epoch": 0.07449414449179557, "grad_norm": 1.6017955541610718, "learning_rate": 9.998622310607508e-05, "loss": 0.8252, "step": 1110 }, { "epoch": 0.07462836817556458, "grad_norm": 1.5187143087387085, "learning_rate": 9.998596680925331e-05, "loss": 0.868, "step": 1112 }, { "epoch": 0.07476259185933358, "grad_norm": 1.3681526184082031, "learning_rate": 9.99857081506929e-05, "loss": 0.8218, "step": 1114 }, { "epoch": 0.07489681554310258, "grad_norm": 1.9026960134506226, "learning_rate": 9.998544713040608e-05, "loss": 0.7745, "step": 1116 }, { "epoch": 0.07503103922687158, "grad_norm": 1.8776766061782837, "learning_rate": 9.998518374840515e-05, "loss": 0.8585, "step": 1118 }, { "epoch": 0.07516526291064059, "grad_norm": 2.2423903942108154, "learning_rate": 9.998491800470259e-05, "loss": 0.8332, "step": 1120 }, { "epoch": 0.07529948659440959, "grad_norm": 1.6247942447662354, "learning_rate": 9.998464989931097e-05, "loss": 0.8407, "step": 1122 }, { "epoch": 0.07543371027817858, "grad_norm": 1.3288408517837524, "learning_rate": 9.998437943224292e-05, "loss": 0.865, "step": 1124 }, { "epoch": 0.07556793396194758, "grad_norm": 1.5527228116989136, "learning_rate": 9.998410660351121e-05, "loss": 0.8223, "step": 1126 }, { "epoch": 0.07570215764571658, "grad_norm": 1.4501368999481201, "learning_rate": 9.998383141312877e-05, "loss": 0.8355, "step": 1128 }, { "epoch": 0.07583638132948559, "grad_norm": 1.5103774070739746, "learning_rate": 9.99835538611086e-05, "loss": 0.8287, "step": 1130 }, { "epoch": 0.07597060501325459, "grad_norm": 1.4005253314971924, "learning_rate": 9.998327394746378e-05, "loss": 0.8736, "step": 1132 }, { "epoch": 0.07610482869702359, "grad_norm": 2.2452094554901123, "learning_rate": 9.998299167220755e-05, "loss": 0.8813, "step": 1134 }, { "epoch": 0.0762390523807926, "grad_norm": 1.320500373840332, "learning_rate": 9.998270703535326e-05, "loss": 0.7934, "step": 1136 }, { "epoch": 0.0763732760645616, "grad_norm": 1.582572340965271, "learning_rate": 9.998242003691434e-05, "loss": 0.9164, "step": 1138 }, { "epoch": 0.0765074997483306, "grad_norm": 1.6067897081375122, "learning_rate": 9.998213067690436e-05, "loss": 0.8988, "step": 1140 }, { "epoch": 0.07664172343209959, "grad_norm": 1.3594622611999512, "learning_rate": 9.998183895533701e-05, "loss": 0.8876, "step": 1142 }, { "epoch": 0.07677594711586859, "grad_norm": 1.7617809772491455, "learning_rate": 9.998154487222602e-05, "loss": 0.8808, "step": 1144 }, { "epoch": 0.07691017079963759, "grad_norm": 1.5601156949996948, "learning_rate": 9.998124842758535e-05, "loss": 0.8613, "step": 1146 }, { "epoch": 0.0770443944834066, "grad_norm": 1.5522006750106812, "learning_rate": 9.998094962142897e-05, "loss": 0.8229, "step": 1148 }, { "epoch": 0.0771786181671756, "grad_norm": 1.2793421745300293, "learning_rate": 9.9980648453771e-05, "loss": 0.8181, "step": 1150 }, { "epoch": 0.0773128418509446, "grad_norm": 1.309046745300293, "learning_rate": 9.998034492462567e-05, "loss": 0.8801, "step": 1152 }, { "epoch": 0.0774470655347136, "grad_norm": 1.305838704109192, "learning_rate": 9.998003903400732e-05, "loss": 0.8232, "step": 1154 }, { "epoch": 0.0775812892184826, "grad_norm": 1.352232575416565, "learning_rate": 9.997973078193041e-05, "loss": 0.9048, "step": 1156 }, { "epoch": 0.07771551290225161, "grad_norm": 1.3795783519744873, "learning_rate": 9.99794201684095e-05, "loss": 0.8422, "step": 1158 }, { "epoch": 0.07784973658602061, "grad_norm": 1.46577787399292, "learning_rate": 9.997910719345928e-05, "loss": 0.8454, "step": 1160 }, { "epoch": 0.0779839602697896, "grad_norm": 1.3010011911392212, "learning_rate": 9.997879185709453e-05, "loss": 0.8205, "step": 1162 }, { "epoch": 0.0781181839535586, "grad_norm": 1.2190300226211548, "learning_rate": 9.997847415933012e-05, "loss": 0.7828, "step": 1164 }, { "epoch": 0.0782524076373276, "grad_norm": 1.3660764694213867, "learning_rate": 9.997815410018111e-05, "loss": 0.9083, "step": 1166 }, { "epoch": 0.0783866313210966, "grad_norm": 1.1576415300369263, "learning_rate": 9.997783167966258e-05, "loss": 0.8664, "step": 1168 }, { "epoch": 0.07852085500486561, "grad_norm": 3.223151445388794, "learning_rate": 9.997750689778978e-05, "loss": 0.8677, "step": 1170 }, { "epoch": 0.07865507868863461, "grad_norm": 1.2480010986328125, "learning_rate": 9.997717975457807e-05, "loss": 0.8692, "step": 1172 }, { "epoch": 0.07878930237240361, "grad_norm": 1.337230920791626, "learning_rate": 9.997685025004288e-05, "loss": 0.84, "step": 1174 }, { "epoch": 0.07892352605617262, "grad_norm": 1.5925257205963135, "learning_rate": 9.997651838419979e-05, "loss": 0.828, "step": 1176 }, { "epoch": 0.07905774973994162, "grad_norm": 1.4853055477142334, "learning_rate": 9.997618415706448e-05, "loss": 0.8157, "step": 1178 }, { "epoch": 0.07919197342371061, "grad_norm": 1.3261574506759644, "learning_rate": 9.997584756865274e-05, "loss": 0.8429, "step": 1180 }, { "epoch": 0.07932619710747961, "grad_norm": 1.4375333786010742, "learning_rate": 9.997550861898049e-05, "loss": 0.9062, "step": 1182 }, { "epoch": 0.07946042079124861, "grad_norm": 1.9652042388916016, "learning_rate": 9.997516730806372e-05, "loss": 0.8545, "step": 1184 }, { "epoch": 0.07959464447501761, "grad_norm": 1.3153257369995117, "learning_rate": 9.997482363591857e-05, "loss": 0.8545, "step": 1186 }, { "epoch": 0.07972886815878662, "grad_norm": 1.7656164169311523, "learning_rate": 9.997447760256126e-05, "loss": 0.8648, "step": 1188 }, { "epoch": 0.07986309184255562, "grad_norm": 1.261059284210205, "learning_rate": 9.997412920800817e-05, "loss": 0.7956, "step": 1190 }, { "epoch": 0.07999731552632462, "grad_norm": 1.4606881141662598, "learning_rate": 9.997377845227576e-05, "loss": 0.8576, "step": 1192 }, { "epoch": 0.08013153921009362, "grad_norm": 1.3973236083984375, "learning_rate": 9.997342533538056e-05, "loss": 0.8701, "step": 1194 }, { "epoch": 0.08026576289386263, "grad_norm": 1.4433778524398804, "learning_rate": 9.99730698573393e-05, "loss": 0.8336, "step": 1196 }, { "epoch": 0.08039998657763163, "grad_norm": 1.434862732887268, "learning_rate": 9.997271201816873e-05, "loss": 0.825, "step": 1198 }, { "epoch": 0.08053421026140062, "grad_norm": 1.2260946035385132, "learning_rate": 9.99723518178858e-05, "loss": 0.7844, "step": 1200 }, { "epoch": 0.08066843394516962, "grad_norm": 1.6296569108963013, "learning_rate": 9.997198925650753e-05, "loss": 0.8702, "step": 1202 }, { "epoch": 0.08080265762893862, "grad_norm": 1.4024724960327148, "learning_rate": 9.9971624334051e-05, "loss": 0.8359, "step": 1204 }, { "epoch": 0.08093688131270763, "grad_norm": 2.7965261936187744, "learning_rate": 9.997125705053352e-05, "loss": 0.853, "step": 1206 }, { "epoch": 0.08107110499647663, "grad_norm": 1.1983357667922974, "learning_rate": 9.997088740597237e-05, "loss": 0.835, "step": 1208 }, { "epoch": 0.08120532868024563, "grad_norm": 1.1802141666412354, "learning_rate": 9.997051540038508e-05, "loss": 0.8028, "step": 1210 }, { "epoch": 0.08133955236401463, "grad_norm": 5.62499475479126, "learning_rate": 9.997014103378921e-05, "loss": 0.8374, "step": 1212 }, { "epoch": 0.08147377604778364, "grad_norm": 1.3405348062515259, "learning_rate": 9.996976430620241e-05, "loss": 0.8063, "step": 1214 }, { "epoch": 0.08160799973155264, "grad_norm": 1.4804438352584839, "learning_rate": 9.996938521764254e-05, "loss": 0.8195, "step": 1216 }, { "epoch": 0.08174222341532163, "grad_norm": 1.225172758102417, "learning_rate": 9.996900376812746e-05, "loss": 0.8781, "step": 1218 }, { "epoch": 0.08187644709909063, "grad_norm": 1.286395788192749, "learning_rate": 9.996861995767522e-05, "loss": 0.8469, "step": 1220 }, { "epoch": 0.08201067078285963, "grad_norm": 1.6874159574508667, "learning_rate": 9.996823378630393e-05, "loss": 0.8572, "step": 1222 }, { "epoch": 0.08214489446662863, "grad_norm": 1.407341718673706, "learning_rate": 9.996784525403186e-05, "loss": 0.7872, "step": 1224 }, { "epoch": 0.08227911815039764, "grad_norm": 1.9326817989349365, "learning_rate": 9.996745436087736e-05, "loss": 0.8664, "step": 1226 }, { "epoch": 0.08241334183416664, "grad_norm": 1.7948651313781738, "learning_rate": 9.99670611068589e-05, "loss": 0.8661, "step": 1228 }, { "epoch": 0.08254756551793564, "grad_norm": 1.3383947610855103, "learning_rate": 9.996666549199505e-05, "loss": 0.8789, "step": 1230 }, { "epoch": 0.08268178920170464, "grad_norm": 1.5959744453430176, "learning_rate": 9.996626751630453e-05, "loss": 0.8675, "step": 1232 }, { "epoch": 0.08281601288547365, "grad_norm": 1.5789822340011597, "learning_rate": 9.996586717980611e-05, "loss": 0.8179, "step": 1234 }, { "epoch": 0.08295023656924265, "grad_norm": 1.6039397716522217, "learning_rate": 9.996546448251871e-05, "loss": 0.8497, "step": 1236 }, { "epoch": 0.08308446025301164, "grad_norm": 1.5541913509368896, "learning_rate": 9.996505942446139e-05, "loss": 0.8189, "step": 1238 }, { "epoch": 0.08321868393678064, "grad_norm": 1.6467313766479492, "learning_rate": 9.996465200565324e-05, "loss": 0.908, "step": 1240 }, { "epoch": 0.08335290762054964, "grad_norm": 1.3486804962158203, "learning_rate": 9.996424222611356e-05, "loss": 0.7749, "step": 1242 }, { "epoch": 0.08348713130431865, "grad_norm": 1.6886804103851318, "learning_rate": 9.996383008586165e-05, "loss": 0.8229, "step": 1244 }, { "epoch": 0.08362135498808765, "grad_norm": 1.831268310546875, "learning_rate": 9.996341558491706e-05, "loss": 0.9068, "step": 1246 }, { "epoch": 0.08375557867185665, "grad_norm": 1.4950191974639893, "learning_rate": 9.996299872329931e-05, "loss": 0.8092, "step": 1248 }, { "epoch": 0.08388980235562565, "grad_norm": 1.272981882095337, "learning_rate": 9.996257950102811e-05, "loss": 0.7781, "step": 1250 }, { "epoch": 0.08402402603939466, "grad_norm": 1.3095083236694336, "learning_rate": 9.996215791812328e-05, "loss": 0.8071, "step": 1252 }, { "epoch": 0.08415824972316366, "grad_norm": 1.740781545639038, "learning_rate": 9.996173397460475e-05, "loss": 0.8787, "step": 1254 }, { "epoch": 0.08429247340693265, "grad_norm": 2.427889347076416, "learning_rate": 9.996130767049252e-05, "loss": 0.7607, "step": 1256 }, { "epoch": 0.08442669709070165, "grad_norm": 1.7244151830673218, "learning_rate": 9.996087900580675e-05, "loss": 0.8009, "step": 1258 }, { "epoch": 0.08456092077447065, "grad_norm": 2.351905107498169, "learning_rate": 9.996044798056769e-05, "loss": 0.8508, "step": 1260 }, { "epoch": 0.08469514445823965, "grad_norm": 1.3471221923828125, "learning_rate": 9.996001459479572e-05, "loss": 0.8101, "step": 1262 }, { "epoch": 0.08482936814200866, "grad_norm": 1.319439172744751, "learning_rate": 9.995957884851129e-05, "loss": 0.827, "step": 1264 }, { "epoch": 0.08496359182577766, "grad_norm": 1.4922178983688354, "learning_rate": 9.995914074173501e-05, "loss": 0.7704, "step": 1266 }, { "epoch": 0.08509781550954666, "grad_norm": 2.1654322147369385, "learning_rate": 9.995870027448756e-05, "loss": 0.8398, "step": 1268 }, { "epoch": 0.08523203919331566, "grad_norm": 1.5658771991729736, "learning_rate": 9.995825744678976e-05, "loss": 0.8643, "step": 1270 }, { "epoch": 0.08536626287708467, "grad_norm": 1.5930389165878296, "learning_rate": 9.995781225866254e-05, "loss": 0.8004, "step": 1272 }, { "epoch": 0.08550048656085366, "grad_norm": 1.2441797256469727, "learning_rate": 9.995736471012693e-05, "loss": 0.8926, "step": 1274 }, { "epoch": 0.08563471024462266, "grad_norm": 1.4061142206192017, "learning_rate": 9.995691480120408e-05, "loss": 0.7998, "step": 1276 }, { "epoch": 0.08576893392839166, "grad_norm": 1.9022417068481445, "learning_rate": 9.995646253191522e-05, "loss": 0.8846, "step": 1278 }, { "epoch": 0.08590315761216066, "grad_norm": 1.862823486328125, "learning_rate": 9.995600790228176e-05, "loss": 0.7921, "step": 1280 }, { "epoch": 0.08603738129592967, "grad_norm": 1.2491618394851685, "learning_rate": 9.995555091232516e-05, "loss": 0.7846, "step": 1282 }, { "epoch": 0.08617160497969867, "grad_norm": 1.292881727218628, "learning_rate": 9.995509156206701e-05, "loss": 0.8009, "step": 1284 }, { "epoch": 0.08630582866346767, "grad_norm": 2.969083309173584, "learning_rate": 9.995462985152902e-05, "loss": 0.8369, "step": 1286 }, { "epoch": 0.08644005234723667, "grad_norm": 1.9158148765563965, "learning_rate": 9.995416578073299e-05, "loss": 0.7613, "step": 1288 }, { "epoch": 0.08657427603100568, "grad_norm": 1.4865663051605225, "learning_rate": 9.995369934970085e-05, "loss": 0.8059, "step": 1290 }, { "epoch": 0.08670849971477468, "grad_norm": 1.2080954313278198, "learning_rate": 9.995323055845466e-05, "loss": 0.7653, "step": 1292 }, { "epoch": 0.08684272339854367, "grad_norm": 1.413210391998291, "learning_rate": 9.995275940701657e-05, "loss": 0.8324, "step": 1294 }, { "epoch": 0.08697694708231267, "grad_norm": 1.22736394405365, "learning_rate": 9.995228589540881e-05, "loss": 0.7899, "step": 1296 }, { "epoch": 0.08711117076608167, "grad_norm": 1.8747268915176392, "learning_rate": 9.995181002365376e-05, "loss": 0.8409, "step": 1298 }, { "epoch": 0.08724539444985067, "grad_norm": 1.453848123550415, "learning_rate": 9.995133179177391e-05, "loss": 0.7499, "step": 1300 }, { "epoch": 0.08737961813361968, "grad_norm": 1.2807241678237915, "learning_rate": 9.995085119979189e-05, "loss": 0.7725, "step": 1302 }, { "epoch": 0.08751384181738868, "grad_norm": 1.6054444313049316, "learning_rate": 9.995036824773034e-05, "loss": 0.7963, "step": 1304 }, { "epoch": 0.08764806550115768, "grad_norm": 1.579632043838501, "learning_rate": 9.994988293561213e-05, "loss": 0.8583, "step": 1306 }, { "epoch": 0.08778228918492668, "grad_norm": 1.4864579439163208, "learning_rate": 9.994939526346016e-05, "loss": 0.8355, "step": 1308 }, { "epoch": 0.08791651286869569, "grad_norm": 1.6583774089813232, "learning_rate": 9.99489052312975e-05, "loss": 0.7561, "step": 1310 }, { "epoch": 0.08805073655246468, "grad_norm": 2.079428195953369, "learning_rate": 9.99484128391473e-05, "loss": 0.7822, "step": 1312 }, { "epoch": 0.08818496023623368, "grad_norm": 1.3199563026428223, "learning_rate": 9.994791808703279e-05, "loss": 0.7904, "step": 1314 }, { "epoch": 0.08831918392000268, "grad_norm": 2.335811138153076, "learning_rate": 9.994742097497737e-05, "loss": 0.9113, "step": 1316 }, { "epoch": 0.08845340760377168, "grad_norm": 1.2894052267074585, "learning_rate": 9.994692150300453e-05, "loss": 0.8324, "step": 1318 }, { "epoch": 0.08858763128754069, "grad_norm": 1.2020514011383057, "learning_rate": 9.994641967113787e-05, "loss": 0.7979, "step": 1320 }, { "epoch": 0.08872185497130969, "grad_norm": 1.4542573690414429, "learning_rate": 9.994591547940109e-05, "loss": 0.7408, "step": 1322 }, { "epoch": 0.08885607865507869, "grad_norm": 1.7090483903884888, "learning_rate": 9.994540892781802e-05, "loss": 0.8605, "step": 1324 }, { "epoch": 0.0889903023388477, "grad_norm": 1.3629621267318726, "learning_rate": 9.994490001641258e-05, "loss": 0.8248, "step": 1326 }, { "epoch": 0.0891245260226167, "grad_norm": 1.5533065795898438, "learning_rate": 9.994438874520885e-05, "loss": 0.801, "step": 1328 }, { "epoch": 0.0892587497063857, "grad_norm": 1.7974445819854736, "learning_rate": 9.994387511423096e-05, "loss": 0.8156, "step": 1330 }, { "epoch": 0.08939297339015469, "grad_norm": 1.3404579162597656, "learning_rate": 9.994335912350317e-05, "loss": 0.8292, "step": 1332 }, { "epoch": 0.08952719707392369, "grad_norm": 1.7092863321304321, "learning_rate": 9.994284077304987e-05, "loss": 0.8455, "step": 1334 }, { "epoch": 0.08966142075769269, "grad_norm": 1.2564599514007568, "learning_rate": 9.994232006289554e-05, "loss": 0.8207, "step": 1336 }, { "epoch": 0.0897956444414617, "grad_norm": 8.875020980834961, "learning_rate": 9.994179699306483e-05, "loss": 0.7694, "step": 1338 }, { "epoch": 0.0899298681252307, "grad_norm": 1.4143531322479248, "learning_rate": 9.99412715635824e-05, "loss": 0.9383, "step": 1340 }, { "epoch": 0.0900640918089997, "grad_norm": 1.8751009702682495, "learning_rate": 9.994074377447309e-05, "loss": 0.8285, "step": 1342 }, { "epoch": 0.0901983154927687, "grad_norm": 1.322713851928711, "learning_rate": 9.994021362576184e-05, "loss": 0.7885, "step": 1344 }, { "epoch": 0.0903325391765377, "grad_norm": 1.5340343713760376, "learning_rate": 9.99396811174737e-05, "loss": 0.8407, "step": 1346 }, { "epoch": 0.09046676286030671, "grad_norm": 1.544365406036377, "learning_rate": 9.993914624963383e-05, "loss": 0.7966, "step": 1348 }, { "epoch": 0.0906009865440757, "grad_norm": 1.9333579540252686, "learning_rate": 9.99386090222675e-05, "loss": 0.7956, "step": 1350 }, { "epoch": 0.0907352102278447, "grad_norm": 1.710904836654663, "learning_rate": 9.993806943540009e-05, "loss": 0.8167, "step": 1352 }, { "epoch": 0.0908694339116137, "grad_norm": 1.344147801399231, "learning_rate": 9.993752748905712e-05, "loss": 0.8044, "step": 1354 }, { "epoch": 0.0910036575953827, "grad_norm": 1.5693535804748535, "learning_rate": 9.993698318326416e-05, "loss": 0.8063, "step": 1356 }, { "epoch": 0.0911378812791517, "grad_norm": 1.494345784187317, "learning_rate": 9.993643651804694e-05, "loss": 0.8564, "step": 1358 }, { "epoch": 0.09127210496292071, "grad_norm": 1.3548997640609741, "learning_rate": 9.99358874934313e-05, "loss": 0.8763, "step": 1360 }, { "epoch": 0.09140632864668971, "grad_norm": 1.4821975231170654, "learning_rate": 9.993533610944315e-05, "loss": 0.8288, "step": 1362 }, { "epoch": 0.09154055233045871, "grad_norm": 1.3371686935424805, "learning_rate": 9.993478236610858e-05, "loss": 0.8288, "step": 1364 }, { "epoch": 0.09167477601422772, "grad_norm": 1.7221264839172363, "learning_rate": 9.993422626345373e-05, "loss": 0.8626, "step": 1366 }, { "epoch": 0.09180899969799672, "grad_norm": 1.32262122631073, "learning_rate": 9.993366780150488e-05, "loss": 0.8873, "step": 1368 }, { "epoch": 0.0919432233817657, "grad_norm": 1.3162240982055664, "learning_rate": 9.993310698028842e-05, "loss": 0.8074, "step": 1370 }, { "epoch": 0.09207744706553471, "grad_norm": 1.3181543350219727, "learning_rate": 9.993254379983084e-05, "loss": 0.8264, "step": 1372 }, { "epoch": 0.09221167074930371, "grad_norm": 1.367629885673523, "learning_rate": 9.993197826015874e-05, "loss": 0.8927, "step": 1374 }, { "epoch": 0.09234589443307271, "grad_norm": 1.7153240442276, "learning_rate": 9.993141036129887e-05, "loss": 0.7883, "step": 1376 }, { "epoch": 0.09248011811684172, "grad_norm": 1.9687062501907349, "learning_rate": 9.993084010327804e-05, "loss": 0.8382, "step": 1378 }, { "epoch": 0.09261434180061072, "grad_norm": 1.3036432266235352, "learning_rate": 9.993026748612322e-05, "loss": 0.8184, "step": 1380 }, { "epoch": 0.09274856548437972, "grad_norm": 1.311419129371643, "learning_rate": 9.992969250986142e-05, "loss": 0.863, "step": 1382 }, { "epoch": 0.09288278916814872, "grad_norm": 1.3586655855178833, "learning_rate": 9.992911517451985e-05, "loss": 0.8568, "step": 1384 }, { "epoch": 0.09301701285191773, "grad_norm": 1.15739107131958, "learning_rate": 9.992853548012576e-05, "loss": 0.8365, "step": 1386 }, { "epoch": 0.09315123653568672, "grad_norm": 1.3566397428512573, "learning_rate": 9.992795342670656e-05, "loss": 0.7763, "step": 1388 }, { "epoch": 0.09328546021945572, "grad_norm": 1.740161418914795, "learning_rate": 9.992736901428971e-05, "loss": 0.7798, "step": 1390 }, { "epoch": 0.09341968390322472, "grad_norm": 6.3728485107421875, "learning_rate": 9.992678224290288e-05, "loss": 0.7612, "step": 1392 }, { "epoch": 0.09355390758699372, "grad_norm": 1.6933761835098267, "learning_rate": 9.992619311257376e-05, "loss": 0.7963, "step": 1394 }, { "epoch": 0.09368813127076273, "grad_norm": 1.3581308126449585, "learning_rate": 9.992560162333019e-05, "loss": 0.8041, "step": 1396 }, { "epoch": 0.09382235495453173, "grad_norm": 1.2820748090744019, "learning_rate": 9.992500777520011e-05, "loss": 0.8395, "step": 1398 }, { "epoch": 0.09395657863830073, "grad_norm": 1.3277995586395264, "learning_rate": 9.99244115682116e-05, "loss": 0.8477, "step": 1400 }, { "epoch": 0.09409080232206973, "grad_norm": 1.349748134613037, "learning_rate": 9.992381300239281e-05, "loss": 0.8246, "step": 1402 }, { "epoch": 0.09422502600583874, "grad_norm": 1.4513624906539917, "learning_rate": 9.992321207777202e-05, "loss": 0.8087, "step": 1404 }, { "epoch": 0.09435924968960774, "grad_norm": 1.3624792098999023, "learning_rate": 9.992260879437763e-05, "loss": 0.7429, "step": 1406 }, { "epoch": 0.09449347337337673, "grad_norm": 1.508120059967041, "learning_rate": 9.992200315223815e-05, "loss": 0.8151, "step": 1408 }, { "epoch": 0.09462769705714573, "grad_norm": 1.3839117288589478, "learning_rate": 9.992139515138219e-05, "loss": 0.839, "step": 1410 }, { "epoch": 0.09476192074091473, "grad_norm": 1.907886028289795, "learning_rate": 9.992078479183847e-05, "loss": 0.8181, "step": 1412 }, { "epoch": 0.09489614442468373, "grad_norm": 1.682834506034851, "learning_rate": 9.992017207363584e-05, "loss": 0.8222, "step": 1414 }, { "epoch": 0.09503036810845274, "grad_norm": 1.2854427099227905, "learning_rate": 9.991955699680322e-05, "loss": 0.7843, "step": 1416 }, { "epoch": 0.09516459179222174, "grad_norm": 1.4885661602020264, "learning_rate": 9.991893956136973e-05, "loss": 0.8888, "step": 1418 }, { "epoch": 0.09529881547599074, "grad_norm": 1.8255189657211304, "learning_rate": 9.991831976736447e-05, "loss": 0.8754, "step": 1420 }, { "epoch": 0.09543303915975974, "grad_norm": 1.947938084602356, "learning_rate": 9.99176976148168e-05, "loss": 0.8436, "step": 1422 }, { "epoch": 0.09556726284352875, "grad_norm": 1.3944926261901855, "learning_rate": 9.991707310375604e-05, "loss": 0.7863, "step": 1424 }, { "epoch": 0.09570148652729774, "grad_norm": 2.084347724914551, "learning_rate": 9.991644623421176e-05, "loss": 0.7776, "step": 1426 }, { "epoch": 0.09583571021106674, "grad_norm": 1.5374326705932617, "learning_rate": 9.991581700621355e-05, "loss": 0.82, "step": 1428 }, { "epoch": 0.09596993389483574, "grad_norm": 1.429213523864746, "learning_rate": 9.991518541979113e-05, "loss": 0.8114, "step": 1430 }, { "epoch": 0.09610415757860474, "grad_norm": 1.3940026760101318, "learning_rate": 9.991455147497435e-05, "loss": 0.8303, "step": 1432 }, { "epoch": 0.09623838126237375, "grad_norm": 1.309457540512085, "learning_rate": 9.991391517179318e-05, "loss": 0.8646, "step": 1434 }, { "epoch": 0.09637260494614275, "grad_norm": 1.4260421991348267, "learning_rate": 9.991327651027765e-05, "loss": 0.9089, "step": 1436 }, { "epoch": 0.09650682862991175, "grad_norm": 1.6267179250717163, "learning_rate": 9.991263549045797e-05, "loss": 0.7937, "step": 1438 }, { "epoch": 0.09664105231368075, "grad_norm": 1.6733784675598145, "learning_rate": 9.991199211236442e-05, "loss": 0.8057, "step": 1440 }, { "epoch": 0.09677527599744976, "grad_norm": 1.1236393451690674, "learning_rate": 9.991134637602737e-05, "loss": 0.7769, "step": 1442 }, { "epoch": 0.09690949968121874, "grad_norm": 1.145785927772522, "learning_rate": 9.991069828147737e-05, "loss": 0.7233, "step": 1444 }, { "epoch": 0.09704372336498775, "grad_norm": 1.2535909414291382, "learning_rate": 9.9910047828745e-05, "loss": 0.8332, "step": 1446 }, { "epoch": 0.09717794704875675, "grad_norm": 1.3010847568511963, "learning_rate": 9.990939501786103e-05, "loss": 0.761, "step": 1448 }, { "epoch": 0.09731217073252575, "grad_norm": 2.345953941345215, "learning_rate": 9.990873984885629e-05, "loss": 0.8331, "step": 1450 }, { "epoch": 0.09744639441629475, "grad_norm": 1.5649274587631226, "learning_rate": 9.990808232176172e-05, "loss": 0.8558, "step": 1452 }, { "epoch": 0.09758061810006376, "grad_norm": 1.2294329404830933, "learning_rate": 9.99074224366084e-05, "loss": 0.7881, "step": 1454 }, { "epoch": 0.09771484178383276, "grad_norm": 1.3104562759399414, "learning_rate": 9.990676019342752e-05, "loss": 0.798, "step": 1456 }, { "epoch": 0.09784906546760176, "grad_norm": 1.2666356563568115, "learning_rate": 9.990609559225036e-05, "loss": 0.8305, "step": 1458 }, { "epoch": 0.09798328915137076, "grad_norm": 1.2787644863128662, "learning_rate": 9.990542863310831e-05, "loss": 0.7108, "step": 1460 }, { "epoch": 0.09811751283513977, "grad_norm": 1.3989942073822021, "learning_rate": 9.990475931603289e-05, "loss": 0.838, "step": 1462 }, { "epoch": 0.09825173651890876, "grad_norm": 1.323578953742981, "learning_rate": 9.990408764105575e-05, "loss": 0.81, "step": 1464 }, { "epoch": 0.09838596020267776, "grad_norm": 1.2336007356643677, "learning_rate": 9.990341360820856e-05, "loss": 0.8711, "step": 1466 }, { "epoch": 0.09852018388644676, "grad_norm": 1.4644780158996582, "learning_rate": 9.990273721752324e-05, "loss": 0.8542, "step": 1468 }, { "epoch": 0.09865440757021576, "grad_norm": 1.164896011352539, "learning_rate": 9.99020584690317e-05, "loss": 0.7895, "step": 1470 }, { "epoch": 0.09878863125398477, "grad_norm": 2.3305513858795166, "learning_rate": 9.990137736276604e-05, "loss": 0.796, "step": 1472 }, { "epoch": 0.09892285493775377, "grad_norm": 1.4515163898468018, "learning_rate": 9.990069389875843e-05, "loss": 0.7958, "step": 1474 }, { "epoch": 0.09905707862152277, "grad_norm": 1.335404634475708, "learning_rate": 9.990000807704114e-05, "loss": 0.782, "step": 1476 }, { "epoch": 0.09919130230529177, "grad_norm": 1.3235502243041992, "learning_rate": 9.98993198976466e-05, "loss": 0.8255, "step": 1478 }, { "epoch": 0.09932552598906078, "grad_norm": 2.2413065433502197, "learning_rate": 9.989862936060731e-05, "loss": 0.8303, "step": 1480 }, { "epoch": 0.09945974967282976, "grad_norm": 5.080524921417236, "learning_rate": 9.989793646595591e-05, "loss": 0.7998, "step": 1482 }, { "epoch": 0.09959397335659877, "grad_norm": 1.2764524221420288, "learning_rate": 9.989724121372514e-05, "loss": 0.7724, "step": 1484 }, { "epoch": 0.09972819704036777, "grad_norm": 1.6263426542282104, "learning_rate": 9.989654360394782e-05, "loss": 0.8485, "step": 1486 }, { "epoch": 0.09986242072413677, "grad_norm": 1.8585193157196045, "learning_rate": 9.989584363665696e-05, "loss": 0.8305, "step": 1488 }, { "epoch": 0.09999664440790577, "grad_norm": 3.082002878189087, "learning_rate": 9.989514131188559e-05, "loss": 0.787, "step": 1490 }, { "epoch": 0.10013086809167478, "grad_norm": 1.289965271949768, "learning_rate": 9.989443662966691e-05, "loss": 0.7783, "step": 1492 }, { "epoch": 0.10026509177544378, "grad_norm": 1.5314586162567139, "learning_rate": 9.989372959003421e-05, "loss": 0.8543, "step": 1494 }, { "epoch": 0.10039931545921278, "grad_norm": 3.5703372955322266, "learning_rate": 9.98930201930209e-05, "loss": 0.8007, "step": 1496 }, { "epoch": 0.10053353914298178, "grad_norm": 1.9307204484939575, "learning_rate": 9.989230843866049e-05, "loss": 0.824, "step": 1498 }, { "epoch": 0.10066776282675079, "grad_norm": 2.095611810684204, "learning_rate": 9.989159432698663e-05, "loss": 0.7722, "step": 1500 }, { "epoch": 0.10080198651051978, "grad_norm": 1.2669951915740967, "learning_rate": 9.989087785803303e-05, "loss": 0.8146, "step": 1502 }, { "epoch": 0.10093621019428878, "grad_norm": 1.3628166913986206, "learning_rate": 9.989015903183357e-05, "loss": 0.7819, "step": 1504 }, { "epoch": 0.10107043387805778, "grad_norm": 1.2423489093780518, "learning_rate": 9.98894378484222e-05, "loss": 0.8651, "step": 1506 }, { "epoch": 0.10120465756182678, "grad_norm": 1.501184105873108, "learning_rate": 9.988871430783298e-05, "loss": 0.7944, "step": 1508 }, { "epoch": 0.10133888124559579, "grad_norm": 1.2144012451171875, "learning_rate": 9.988798841010012e-05, "loss": 0.8069, "step": 1510 }, { "epoch": 0.10147310492936479, "grad_norm": 2.4115240573883057, "learning_rate": 9.98872601552579e-05, "loss": 0.7664, "step": 1512 }, { "epoch": 0.10160732861313379, "grad_norm": 2.0335206985473633, "learning_rate": 9.988652954334076e-05, "loss": 0.8062, "step": 1514 }, { "epoch": 0.10174155229690279, "grad_norm": 1.4311667680740356, "learning_rate": 9.988579657438317e-05, "loss": 0.8448, "step": 1516 }, { "epoch": 0.1018757759806718, "grad_norm": 3.5951297283172607, "learning_rate": 9.988506124841981e-05, "loss": 0.7939, "step": 1518 }, { "epoch": 0.10200999966444078, "grad_norm": 4.994314670562744, "learning_rate": 9.98843235654854e-05, "loss": 0.7847, "step": 1520 }, { "epoch": 0.10214422334820979, "grad_norm": 4.0133256912231445, "learning_rate": 9.988358352561478e-05, "loss": 0.7028, "step": 1522 }, { "epoch": 0.10227844703197879, "grad_norm": 1.512060284614563, "learning_rate": 9.988284112884294e-05, "loss": 0.7783, "step": 1524 }, { "epoch": 0.10241267071574779, "grad_norm": 1.9130018949508667, "learning_rate": 9.988209637520494e-05, "loss": 0.8547, "step": 1526 }, { "epoch": 0.1025468943995168, "grad_norm": 1.427221417427063, "learning_rate": 9.988134926473598e-05, "loss": 0.8566, "step": 1528 }, { "epoch": 0.1026811180832858, "grad_norm": 1.2133654356002808, "learning_rate": 9.988059979747135e-05, "loss": 0.7705, "step": 1530 }, { "epoch": 0.1028153417670548, "grad_norm": 2.223356246948242, "learning_rate": 9.987984797344648e-05, "loss": 0.8361, "step": 1532 }, { "epoch": 0.1029495654508238, "grad_norm": 1.1635984182357788, "learning_rate": 9.987909379269686e-05, "loss": 0.8032, "step": 1534 }, { "epoch": 0.1030837891345928, "grad_norm": 1.4475617408752441, "learning_rate": 9.987833725525815e-05, "loss": 0.8738, "step": 1536 }, { "epoch": 0.1032180128183618, "grad_norm": 1.554733157157898, "learning_rate": 9.987757836116608e-05, "loss": 0.7886, "step": 1538 }, { "epoch": 0.1033522365021308, "grad_norm": 1.3741666078567505, "learning_rate": 9.987681711045652e-05, "loss": 0.8057, "step": 1540 }, { "epoch": 0.1034864601858998, "grad_norm": 1.1904531717300415, "learning_rate": 9.987605350316542e-05, "loss": 0.8379, "step": 1542 }, { "epoch": 0.1036206838696688, "grad_norm": 1.6170859336853027, "learning_rate": 9.987528753932888e-05, "loss": 0.8036, "step": 1544 }, { "epoch": 0.1037549075534378, "grad_norm": 1.4818545579910278, "learning_rate": 9.987451921898307e-05, "loss": 0.7636, "step": 1546 }, { "epoch": 0.1038891312372068, "grad_norm": 1.4961395263671875, "learning_rate": 9.987374854216431e-05, "loss": 0.8496, "step": 1548 }, { "epoch": 0.10402335492097581, "grad_norm": 1.1231153011322021, "learning_rate": 9.9872975508909e-05, "loss": 0.7967, "step": 1550 }, { "epoch": 0.10415757860474481, "grad_norm": 1.0882415771484375, "learning_rate": 9.987220011925367e-05, "loss": 0.6914, "step": 1552 }, { "epoch": 0.10429180228851381, "grad_norm": 1.2804229259490967, "learning_rate": 9.987142237323495e-05, "loss": 0.7904, "step": 1554 }, { "epoch": 0.10442602597228282, "grad_norm": 1.3405996561050415, "learning_rate": 9.98706422708896e-05, "loss": 0.8317, "step": 1556 }, { "epoch": 0.1045602496560518, "grad_norm": 1.7399771213531494, "learning_rate": 9.986985981225445e-05, "loss": 0.7981, "step": 1558 }, { "epoch": 0.1046944733398208, "grad_norm": 1.1339941024780273, "learning_rate": 9.98690749973665e-05, "loss": 0.771, "step": 1560 }, { "epoch": 0.10482869702358981, "grad_norm": 1.4163483381271362, "learning_rate": 9.986828782626282e-05, "loss": 0.7263, "step": 1562 }, { "epoch": 0.10496292070735881, "grad_norm": 1.2849148511886597, "learning_rate": 9.986749829898061e-05, "loss": 0.7603, "step": 1564 }, { "epoch": 0.10509714439112781, "grad_norm": 1.2235387563705444, "learning_rate": 9.986670641555715e-05, "loss": 0.8495, "step": 1566 }, { "epoch": 0.10523136807489682, "grad_norm": 1.3458445072174072, "learning_rate": 9.986591217602988e-05, "loss": 0.7594, "step": 1568 }, { "epoch": 0.10536559175866582, "grad_norm": 1.1595379114151, "learning_rate": 9.986511558043631e-05, "loss": 0.7621, "step": 1570 }, { "epoch": 0.10549981544243482, "grad_norm": 1.22896146774292, "learning_rate": 9.98643166288141e-05, "loss": 0.7538, "step": 1572 }, { "epoch": 0.10563403912620382, "grad_norm": 1.3033969402313232, "learning_rate": 9.986351532120097e-05, "loss": 0.8397, "step": 1574 }, { "epoch": 0.10576826280997281, "grad_norm": 1.3369084596633911, "learning_rate": 9.98627116576348e-05, "loss": 0.7519, "step": 1576 }, { "epoch": 0.10590248649374182, "grad_norm": 1.3159362077713013, "learning_rate": 9.986190563815355e-05, "loss": 0.7714, "step": 1578 }, { "epoch": 0.10603671017751082, "grad_norm": 1.2473654747009277, "learning_rate": 9.986109726279531e-05, "loss": 0.7457, "step": 1580 }, { "epoch": 0.10617093386127982, "grad_norm": 1.3414995670318604, "learning_rate": 9.986028653159826e-05, "loss": 0.8176, "step": 1582 }, { "epoch": 0.10630515754504882, "grad_norm": 3.310617446899414, "learning_rate": 9.985947344460074e-05, "loss": 0.8616, "step": 1584 }, { "epoch": 0.10643938122881783, "grad_norm": 1.5509215593338013, "learning_rate": 9.985865800184113e-05, "loss": 0.8301, "step": 1586 }, { "epoch": 0.10657360491258683, "grad_norm": 1.2767397165298462, "learning_rate": 9.985784020335798e-05, "loss": 0.8217, "step": 1588 }, { "epoch": 0.10670782859635583, "grad_norm": 1.5381598472595215, "learning_rate": 9.985702004918992e-05, "loss": 0.83, "step": 1590 }, { "epoch": 0.10684205228012483, "grad_norm": 6.233184337615967, "learning_rate": 9.98561975393757e-05, "loss": 0.8475, "step": 1592 }, { "epoch": 0.10697627596389384, "grad_norm": 2.839657783508301, "learning_rate": 9.985537267395418e-05, "loss": 0.7681, "step": 1594 }, { "epoch": 0.10711049964766282, "grad_norm": 1.9623489379882812, "learning_rate": 9.985454545296434e-05, "loss": 0.8005, "step": 1596 }, { "epoch": 0.10724472333143183, "grad_norm": 1.5133533477783203, "learning_rate": 9.985371587644526e-05, "loss": 0.868, "step": 1598 }, { "epoch": 0.10737894701520083, "grad_norm": 1.517056941986084, "learning_rate": 9.985288394443615e-05, "loss": 0.803, "step": 1600 }, { "epoch": 0.10751317069896983, "grad_norm": 1.431883454322815, "learning_rate": 9.98520496569763e-05, "loss": 0.7725, "step": 1602 }, { "epoch": 0.10764739438273883, "grad_norm": 2.0910532474517822, "learning_rate": 9.985121301410511e-05, "loss": 0.7608, "step": 1604 }, { "epoch": 0.10778161806650784, "grad_norm": 1.8099147081375122, "learning_rate": 9.985037401586217e-05, "loss": 0.9247, "step": 1606 }, { "epoch": 0.10791584175027684, "grad_norm": 1.9033899307250977, "learning_rate": 9.984953266228707e-05, "loss": 0.821, "step": 1608 }, { "epoch": 0.10805006543404584, "grad_norm": 1.296483039855957, "learning_rate": 9.984868895341957e-05, "loss": 0.7958, "step": 1610 }, { "epoch": 0.10818428911781484, "grad_norm": 1.6510436534881592, "learning_rate": 9.984784288929953e-05, "loss": 0.769, "step": 1612 }, { "epoch": 0.10831851280158383, "grad_norm": 1.635801911354065, "learning_rate": 9.984699446996697e-05, "loss": 0.8483, "step": 1614 }, { "epoch": 0.10845273648535284, "grad_norm": 1.598320484161377, "learning_rate": 9.984614369546191e-05, "loss": 0.8505, "step": 1616 }, { "epoch": 0.10858696016912184, "grad_norm": 1.5453705787658691, "learning_rate": 9.984529056582459e-05, "loss": 0.8062, "step": 1618 }, { "epoch": 0.10872118385289084, "grad_norm": 1.5651856660842896, "learning_rate": 9.984443508109531e-05, "loss": 0.7983, "step": 1620 }, { "epoch": 0.10885540753665984, "grad_norm": 1.8090884685516357, "learning_rate": 9.984357724131448e-05, "loss": 0.8308, "step": 1622 }, { "epoch": 0.10898963122042885, "grad_norm": 1.8616065979003906, "learning_rate": 9.984271704652263e-05, "loss": 0.7526, "step": 1624 }, { "epoch": 0.10912385490419785, "grad_norm": 1.312053918838501, "learning_rate": 9.984185449676044e-05, "loss": 0.7644, "step": 1626 }, { "epoch": 0.10925807858796685, "grad_norm": 1.403348445892334, "learning_rate": 9.984098959206863e-05, "loss": 0.7711, "step": 1628 }, { "epoch": 0.10939230227173585, "grad_norm": 1.3508669137954712, "learning_rate": 9.984012233248805e-05, "loss": 0.7732, "step": 1630 }, { "epoch": 0.10952652595550486, "grad_norm": 1.1792352199554443, "learning_rate": 9.98392527180597e-05, "loss": 0.7759, "step": 1632 }, { "epoch": 0.10966074963927384, "grad_norm": 1.2566429376602173, "learning_rate": 9.983838074882467e-05, "loss": 0.8051, "step": 1634 }, { "epoch": 0.10979497332304285, "grad_norm": 1.272409200668335, "learning_rate": 9.983750642482414e-05, "loss": 0.7872, "step": 1636 }, { "epoch": 0.10992919700681185, "grad_norm": 1.3588621616363525, "learning_rate": 9.983662974609945e-05, "loss": 0.8099, "step": 1638 }, { "epoch": 0.11006342069058085, "grad_norm": 1.3502743244171143, "learning_rate": 9.9835750712692e-05, "loss": 0.7576, "step": 1640 }, { "epoch": 0.11019764437434985, "grad_norm": 1.369414210319519, "learning_rate": 9.983486932464332e-05, "loss": 0.7747, "step": 1642 }, { "epoch": 0.11033186805811886, "grad_norm": 1.192598581314087, "learning_rate": 9.983398558199506e-05, "loss": 0.7855, "step": 1644 }, { "epoch": 0.11046609174188786, "grad_norm": 1.53266441822052, "learning_rate": 9.983309948478898e-05, "loss": 0.8317, "step": 1646 }, { "epoch": 0.11060031542565686, "grad_norm": 1.2673954963684082, "learning_rate": 9.983221103306695e-05, "loss": 0.7258, "step": 1648 }, { "epoch": 0.11073453910942586, "grad_norm": 1.2609288692474365, "learning_rate": 9.983132022687093e-05, "loss": 0.7564, "step": 1650 }, { "epoch": 0.11086876279319485, "grad_norm": 1.4011356830596924, "learning_rate": 9.983042706624302e-05, "loss": 0.8574, "step": 1652 }, { "epoch": 0.11100298647696386, "grad_norm": 3.235201835632324, "learning_rate": 9.982953155122542e-05, "loss": 0.7768, "step": 1654 }, { "epoch": 0.11113721016073286, "grad_norm": 1.3096420764923096, "learning_rate": 9.982863368186044e-05, "loss": 0.7925, "step": 1656 }, { "epoch": 0.11127143384450186, "grad_norm": 1.4204448461532593, "learning_rate": 9.98277334581905e-05, "loss": 0.7952, "step": 1658 }, { "epoch": 0.11140565752827086, "grad_norm": 1.3249876499176025, "learning_rate": 9.982683088025813e-05, "loss": 0.7743, "step": 1660 }, { "epoch": 0.11153988121203987, "grad_norm": 1.5150166749954224, "learning_rate": 9.982592594810599e-05, "loss": 0.8212, "step": 1662 }, { "epoch": 0.11167410489580887, "grad_norm": 1.3663066625595093, "learning_rate": 9.982501866177682e-05, "loss": 0.8411, "step": 1664 }, { "epoch": 0.11180832857957787, "grad_norm": 1.3154817819595337, "learning_rate": 9.98241090213135e-05, "loss": 0.7834, "step": 1666 }, { "epoch": 0.11194255226334687, "grad_norm": 1.3979624509811401, "learning_rate": 9.982319702675901e-05, "loss": 0.8053, "step": 1668 }, { "epoch": 0.11207677594711588, "grad_norm": 1.3888903856277466, "learning_rate": 9.982228267815643e-05, "loss": 0.7757, "step": 1670 }, { "epoch": 0.11221099963088486, "grad_norm": 1.4778214693069458, "learning_rate": 9.982136597554896e-05, "loss": 0.8161, "step": 1672 }, { "epoch": 0.11234522331465387, "grad_norm": 1.4513918161392212, "learning_rate": 9.982044691897991e-05, "loss": 0.7305, "step": 1674 }, { "epoch": 0.11247944699842287, "grad_norm": 1.3353772163391113, "learning_rate": 9.981952550849273e-05, "loss": 0.7575, "step": 1676 }, { "epoch": 0.11261367068219187, "grad_norm": 1.2371762990951538, "learning_rate": 9.981860174413092e-05, "loss": 0.819, "step": 1678 }, { "epoch": 0.11274789436596087, "grad_norm": 3.9118690490722656, "learning_rate": 9.981767562593815e-05, "loss": 0.6956, "step": 1680 }, { "epoch": 0.11288211804972988, "grad_norm": 1.2148282527923584, "learning_rate": 9.981674715395816e-05, "loss": 0.7897, "step": 1682 }, { "epoch": 0.11301634173349888, "grad_norm": 1.2970951795578003, "learning_rate": 9.981581632823485e-05, "loss": 0.8284, "step": 1684 }, { "epoch": 0.11315056541726788, "grad_norm": 1.3304328918457031, "learning_rate": 9.981488314881215e-05, "loss": 0.8311, "step": 1686 }, { "epoch": 0.11328478910103688, "grad_norm": 1.2172995805740356, "learning_rate": 9.981394761573419e-05, "loss": 0.7544, "step": 1688 }, { "epoch": 0.11341901278480587, "grad_norm": 1.2176858186721802, "learning_rate": 9.981300972904515e-05, "loss": 0.8118, "step": 1690 }, { "epoch": 0.11355323646857488, "grad_norm": 1.348036527633667, "learning_rate": 9.981206948878937e-05, "loss": 0.7689, "step": 1692 }, { "epoch": 0.11368746015234388, "grad_norm": 1.1857105493545532, "learning_rate": 9.981112689501126e-05, "loss": 0.8124, "step": 1694 }, { "epoch": 0.11382168383611288, "grad_norm": 1.2678489685058594, "learning_rate": 9.981018194775533e-05, "loss": 0.8076, "step": 1696 }, { "epoch": 0.11395590751988188, "grad_norm": 1.3750667572021484, "learning_rate": 9.980923464706627e-05, "loss": 0.8819, "step": 1698 }, { "epoch": 0.11409013120365089, "grad_norm": 1.183547854423523, "learning_rate": 9.980828499298882e-05, "loss": 0.8644, "step": 1700 }, { "epoch": 0.11422435488741989, "grad_norm": 1.2928284406661987, "learning_rate": 9.980733298556783e-05, "loss": 0.8045, "step": 1702 }, { "epoch": 0.11435857857118889, "grad_norm": 1.5439093112945557, "learning_rate": 9.980637862484832e-05, "loss": 0.8166, "step": 1704 }, { "epoch": 0.11449280225495789, "grad_norm": 1.4209173917770386, "learning_rate": 9.980542191087535e-05, "loss": 0.8048, "step": 1706 }, { "epoch": 0.1146270259387269, "grad_norm": 2.5467655658721924, "learning_rate": 9.980446284369413e-05, "loss": 0.7793, "step": 1708 }, { "epoch": 0.11476124962249588, "grad_norm": 1.5948156118392944, "learning_rate": 9.980350142334998e-05, "loss": 0.797, "step": 1710 }, { "epoch": 0.11489547330626489, "grad_norm": 1.2480472326278687, "learning_rate": 9.980253764988832e-05, "loss": 0.8435, "step": 1712 }, { "epoch": 0.11502969699003389, "grad_norm": 1.4319919347763062, "learning_rate": 9.980157152335467e-05, "loss": 0.8062, "step": 1714 }, { "epoch": 0.11516392067380289, "grad_norm": 1.3824477195739746, "learning_rate": 9.980060304379472e-05, "loss": 0.7959, "step": 1716 }, { "epoch": 0.1152981443575719, "grad_norm": 1.350187063217163, "learning_rate": 9.979963221125421e-05, "loss": 0.73, "step": 1718 }, { "epoch": 0.1154323680413409, "grad_norm": 1.605605125427246, "learning_rate": 9.9798659025779e-05, "loss": 0.7791, "step": 1720 }, { "epoch": 0.1155665917251099, "grad_norm": 1.3328030109405518, "learning_rate": 9.979768348741507e-05, "loss": 0.7717, "step": 1722 }, { "epoch": 0.1157008154088789, "grad_norm": 1.287878155708313, "learning_rate": 9.979670559620851e-05, "loss": 0.7704, "step": 1724 }, { "epoch": 0.1158350390926479, "grad_norm": 1.2989848852157593, "learning_rate": 9.979572535220555e-05, "loss": 0.7726, "step": 1726 }, { "epoch": 0.11596926277641689, "grad_norm": 1.447620153427124, "learning_rate": 9.979474275545248e-05, "loss": 0.7897, "step": 1728 }, { "epoch": 0.1161034864601859, "grad_norm": 1.369988203048706, "learning_rate": 9.979375780599573e-05, "loss": 0.7513, "step": 1730 }, { "epoch": 0.1162377101439549, "grad_norm": 1.8917629718780518, "learning_rate": 9.979277050388183e-05, "loss": 0.7262, "step": 1732 }, { "epoch": 0.1163719338277239, "grad_norm": 1.0357433557510376, "learning_rate": 9.979178084915745e-05, "loss": 0.7571, "step": 1734 }, { "epoch": 0.1165061575114929, "grad_norm": 1.284207820892334, "learning_rate": 9.979078884186933e-05, "loss": 0.7646, "step": 1736 }, { "epoch": 0.1166403811952619, "grad_norm": 1.2012271881103516, "learning_rate": 9.978979448206434e-05, "loss": 0.8242, "step": 1738 }, { "epoch": 0.11677460487903091, "grad_norm": 1.6155526638031006, "learning_rate": 9.978879776978949e-05, "loss": 0.8832, "step": 1740 }, { "epoch": 0.11690882856279991, "grad_norm": 1.3117786645889282, "learning_rate": 9.978779870509182e-05, "loss": 0.7964, "step": 1742 }, { "epoch": 0.11704305224656891, "grad_norm": 1.2522165775299072, "learning_rate": 9.978679728801859e-05, "loss": 0.8288, "step": 1744 }, { "epoch": 0.1171772759303379, "grad_norm": 1.4759392738342285, "learning_rate": 9.978579351861707e-05, "loss": 0.815, "step": 1746 }, { "epoch": 0.1173114996141069, "grad_norm": 1.2629553079605103, "learning_rate": 9.978478739693473e-05, "loss": 0.7777, "step": 1748 }, { "epoch": 0.1174457232978759, "grad_norm": 1.3460296392440796, "learning_rate": 9.978377892301906e-05, "loss": 0.7317, "step": 1750 }, { "epoch": 0.11757994698164491, "grad_norm": 1.3826546669006348, "learning_rate": 9.978276809691776e-05, "loss": 0.822, "step": 1752 }, { "epoch": 0.11771417066541391, "grad_norm": 1.285803198814392, "learning_rate": 9.978175491867854e-05, "loss": 0.7972, "step": 1754 }, { "epoch": 0.11784839434918291, "grad_norm": 1.267152190208435, "learning_rate": 9.978073938834929e-05, "loss": 0.7369, "step": 1756 }, { "epoch": 0.11798261803295192, "grad_norm": 1.4843337535858154, "learning_rate": 9.977972150597799e-05, "loss": 0.8302, "step": 1758 }, { "epoch": 0.11811684171672092, "grad_norm": 1.3544261455535889, "learning_rate": 9.977870127161275e-05, "loss": 0.7733, "step": 1760 }, { "epoch": 0.11825106540048992, "grad_norm": 1.9223833084106445, "learning_rate": 9.977767868530176e-05, "loss": 0.8213, "step": 1762 }, { "epoch": 0.11838528908425892, "grad_norm": 1.3144159317016602, "learning_rate": 9.977665374709333e-05, "loss": 0.8232, "step": 1764 }, { "epoch": 0.11851951276802791, "grad_norm": 1.1810052394866943, "learning_rate": 9.977562645703589e-05, "loss": 0.7614, "step": 1766 }, { "epoch": 0.11865373645179692, "grad_norm": 1.575556993484497, "learning_rate": 9.977459681517798e-05, "loss": 0.8387, "step": 1768 }, { "epoch": 0.11878796013556592, "grad_norm": 1.113950252532959, "learning_rate": 9.977356482156825e-05, "loss": 0.7608, "step": 1770 }, { "epoch": 0.11892218381933492, "grad_norm": 1.2446372509002686, "learning_rate": 9.977253047625546e-05, "loss": 0.7678, "step": 1772 }, { "epoch": 0.11905640750310392, "grad_norm": 1.3604047298431396, "learning_rate": 9.977149377928847e-05, "loss": 0.7857, "step": 1774 }, { "epoch": 0.11919063118687293, "grad_norm": 1.0783966779708862, "learning_rate": 9.977045473071627e-05, "loss": 0.7241, "step": 1776 }, { "epoch": 0.11932485487064193, "grad_norm": 1.2782478332519531, "learning_rate": 9.976941333058796e-05, "loss": 0.8046, "step": 1778 }, { "epoch": 0.11945907855441093, "grad_norm": 1.35762619972229, "learning_rate": 9.976836957895275e-05, "loss": 0.8035, "step": 1780 }, { "epoch": 0.11959330223817993, "grad_norm": 1.316946268081665, "learning_rate": 9.976732347585993e-05, "loss": 0.8468, "step": 1782 }, { "epoch": 0.11972752592194892, "grad_norm": 1.3204083442687988, "learning_rate": 9.976627502135894e-05, "loss": 0.8752, "step": 1784 }, { "epoch": 0.11986174960571792, "grad_norm": 1.148432970046997, "learning_rate": 9.976522421549932e-05, "loss": 0.6699, "step": 1786 }, { "epoch": 0.11999597328948693, "grad_norm": 1.3192583322525024, "learning_rate": 9.97641710583307e-05, "loss": 0.7993, "step": 1788 }, { "epoch": 0.12013019697325593, "grad_norm": 1.2164489030838013, "learning_rate": 9.976311554990287e-05, "loss": 0.7252, "step": 1790 }, { "epoch": 0.12026442065702493, "grad_norm": 1.1925063133239746, "learning_rate": 9.976205769026568e-05, "loss": 0.7562, "step": 1792 }, { "epoch": 0.12039864434079393, "grad_norm": 1.2228387594223022, "learning_rate": 9.976099747946912e-05, "loss": 0.7322, "step": 1794 }, { "epoch": 0.12053286802456294, "grad_norm": 1.2622096538543701, "learning_rate": 9.975993491756328e-05, "loss": 0.8457, "step": 1796 }, { "epoch": 0.12066709170833194, "grad_norm": 1.1262799501419067, "learning_rate": 9.975887000459835e-05, "loss": 0.8479, "step": 1798 }, { "epoch": 0.12080131539210094, "grad_norm": 34.44312286376953, "learning_rate": 9.975780274062468e-05, "loss": 0.765, "step": 1800 }, { "epoch": 0.12093553907586994, "grad_norm": 1.2758334875106812, "learning_rate": 9.975673312569267e-05, "loss": 0.7277, "step": 1802 }, { "epoch": 0.12106976275963893, "grad_norm": 1.367239236831665, "learning_rate": 9.975566115985284e-05, "loss": 0.7898, "step": 1804 }, { "epoch": 0.12120398644340794, "grad_norm": 1.4623262882232666, "learning_rate": 9.975458684315588e-05, "loss": 0.7649, "step": 1806 }, { "epoch": 0.12133821012717694, "grad_norm": 2.027376651763916, "learning_rate": 9.975351017565253e-05, "loss": 0.7479, "step": 1808 }, { "epoch": 0.12147243381094594, "grad_norm": 1.2474169731140137, "learning_rate": 9.975243115739366e-05, "loss": 0.7568, "step": 1810 }, { "epoch": 0.12160665749471494, "grad_norm": 1.9691351652145386, "learning_rate": 9.975134978843026e-05, "loss": 0.8102, "step": 1812 }, { "epoch": 0.12174088117848395, "grad_norm": 1.175947666168213, "learning_rate": 9.97502660688134e-05, "loss": 0.7581, "step": 1814 }, { "epoch": 0.12187510486225295, "grad_norm": 1.529521107673645, "learning_rate": 9.97491799985943e-05, "loss": 0.8436, "step": 1816 }, { "epoch": 0.12200932854602195, "grad_norm": 1.532962441444397, "learning_rate": 9.974809157782427e-05, "loss": 0.7984, "step": 1818 }, { "epoch": 0.12214355222979095, "grad_norm": 1.0973433256149292, "learning_rate": 9.974700080655475e-05, "loss": 0.7525, "step": 1820 }, { "epoch": 0.12227777591355994, "grad_norm": 1.2945022583007812, "learning_rate": 9.974590768483725e-05, "loss": 0.767, "step": 1822 }, { "epoch": 0.12241199959732894, "grad_norm": 1.1806293725967407, "learning_rate": 9.974481221272345e-05, "loss": 0.7899, "step": 1824 }, { "epoch": 0.12254622328109795, "grad_norm": 1.2637357711791992, "learning_rate": 9.974371439026508e-05, "loss": 0.7686, "step": 1826 }, { "epoch": 0.12268044696486695, "grad_norm": 1.3107693195343018, "learning_rate": 9.974261421751403e-05, "loss": 0.7571, "step": 1828 }, { "epoch": 0.12281467064863595, "grad_norm": 1.2245479822158813, "learning_rate": 9.974151169452226e-05, "loss": 0.8113, "step": 1830 }, { "epoch": 0.12294889433240495, "grad_norm": 1.2956514358520508, "learning_rate": 9.974040682134189e-05, "loss": 0.8173, "step": 1832 }, { "epoch": 0.12308311801617396, "grad_norm": 1.2990756034851074, "learning_rate": 9.97392995980251e-05, "loss": 0.8011, "step": 1834 }, { "epoch": 0.12321734169994296, "grad_norm": 1.2273094654083252, "learning_rate": 9.973819002462421e-05, "loss": 0.8189, "step": 1836 }, { "epoch": 0.12335156538371196, "grad_norm": 1.1717866659164429, "learning_rate": 9.973707810119165e-05, "loss": 0.759, "step": 1838 }, { "epoch": 0.12348578906748096, "grad_norm": 1.330411434173584, "learning_rate": 9.973596382777995e-05, "loss": 0.8096, "step": 1840 }, { "epoch": 0.12362001275124995, "grad_norm": 1.3452109098434448, "learning_rate": 9.973484720444178e-05, "loss": 0.7882, "step": 1842 }, { "epoch": 0.12375423643501895, "grad_norm": 1.2101787328720093, "learning_rate": 9.973372823122985e-05, "loss": 0.8313, "step": 1844 }, { "epoch": 0.12388846011878796, "grad_norm": 1.41432523727417, "learning_rate": 9.973260690819708e-05, "loss": 0.7717, "step": 1846 }, { "epoch": 0.12402268380255696, "grad_norm": 1.2841819524765015, "learning_rate": 9.973148323539641e-05, "loss": 0.7722, "step": 1848 }, { "epoch": 0.12415690748632596, "grad_norm": 1.3827474117279053, "learning_rate": 9.973035721288096e-05, "loss": 0.7811, "step": 1850 }, { "epoch": 0.12429113117009497, "grad_norm": 1.5363903045654297, "learning_rate": 9.972922884070392e-05, "loss": 0.752, "step": 1852 }, { "epoch": 0.12442535485386397, "grad_norm": 1.2475894689559937, "learning_rate": 9.972809811891861e-05, "loss": 0.7954, "step": 1854 }, { "epoch": 0.12455957853763297, "grad_norm": 1.1693847179412842, "learning_rate": 9.972696504757846e-05, "loss": 0.8019, "step": 1856 }, { "epoch": 0.12469380222140197, "grad_norm": 1.9490900039672852, "learning_rate": 9.972582962673698e-05, "loss": 0.7501, "step": 1858 }, { "epoch": 0.12482802590517096, "grad_norm": 2.058365821838379, "learning_rate": 9.972469185644783e-05, "loss": 0.7717, "step": 1860 }, { "epoch": 0.12496224958893996, "grad_norm": 2.113640069961548, "learning_rate": 9.972355173676478e-05, "loss": 0.8192, "step": 1862 }, { "epoch": 0.12509647327270898, "grad_norm": 1.2641154527664185, "learning_rate": 9.972240926774168e-05, "loss": 0.7613, "step": 1864 }, { "epoch": 0.12523069695647798, "grad_norm": 1.1489163637161255, "learning_rate": 9.972126444943252e-05, "loss": 0.7567, "step": 1866 }, { "epoch": 0.12536492064024699, "grad_norm": 1.123280644416809, "learning_rate": 9.97201172818914e-05, "loss": 0.7765, "step": 1868 }, { "epoch": 0.12549914432401596, "grad_norm": 1.388689637184143, "learning_rate": 9.97189677651725e-05, "loss": 0.8062, "step": 1870 }, { "epoch": 0.12563336800778496, "grad_norm": 1.3234788179397583, "learning_rate": 9.971781589933012e-05, "loss": 0.7392, "step": 1872 }, { "epoch": 0.12576759169155396, "grad_norm": 1.1948199272155762, "learning_rate": 9.971666168441872e-05, "loss": 0.8034, "step": 1874 }, { "epoch": 0.12590181537532297, "grad_norm": 2.0882232189178467, "learning_rate": 9.971550512049281e-05, "loss": 0.7683, "step": 1876 }, { "epoch": 0.12603603905909197, "grad_norm": 1.1946005821228027, "learning_rate": 9.971434620760707e-05, "loss": 0.7405, "step": 1878 }, { "epoch": 0.12617026274286097, "grad_norm": 1.5380767583847046, "learning_rate": 9.97131849458162e-05, "loss": 0.7712, "step": 1880 }, { "epoch": 0.12630448642662997, "grad_norm": 1.0007857084274292, "learning_rate": 9.971202133517512e-05, "loss": 0.7146, "step": 1882 }, { "epoch": 0.12643871011039898, "grad_norm": 1.260170578956604, "learning_rate": 9.971085537573879e-05, "loss": 0.8351, "step": 1884 }, { "epoch": 0.12657293379416798, "grad_norm": 1.1478792428970337, "learning_rate": 9.970968706756227e-05, "loss": 0.7442, "step": 1886 }, { "epoch": 0.12670715747793698, "grad_norm": 1.2533572912216187, "learning_rate": 9.970851641070081e-05, "loss": 0.8, "step": 1888 }, { "epoch": 0.12684138116170598, "grad_norm": 1.5237817764282227, "learning_rate": 9.970734340520969e-05, "loss": 0.7082, "step": 1890 }, { "epoch": 0.126975604845475, "grad_norm": 1.4419610500335693, "learning_rate": 9.970616805114434e-05, "loss": 0.801, "step": 1892 }, { "epoch": 0.127109828529244, "grad_norm": 1.7004404067993164, "learning_rate": 9.970499034856029e-05, "loss": 0.8262, "step": 1894 }, { "epoch": 0.127244052213013, "grad_norm": 1.3557778596878052, "learning_rate": 9.970381029751319e-05, "loss": 0.8065, "step": 1896 }, { "epoch": 0.127378275896782, "grad_norm": 1.2651209831237793, "learning_rate": 9.970262789805878e-05, "loss": 0.7901, "step": 1898 }, { "epoch": 0.127512499580551, "grad_norm": 1.9335846900939941, "learning_rate": 9.970144315025296e-05, "loss": 0.8032, "step": 1900 }, { "epoch": 0.12764672326432, "grad_norm": 3.0118868350982666, "learning_rate": 9.970025605415166e-05, "loss": 0.7242, "step": 1902 }, { "epoch": 0.127780946948089, "grad_norm": 1.3224263191223145, "learning_rate": 9.9699066609811e-05, "loss": 0.8135, "step": 1904 }, { "epoch": 0.127915170631858, "grad_norm": 1.142478108406067, "learning_rate": 9.969787481728718e-05, "loss": 0.7557, "step": 1906 }, { "epoch": 0.12804939431562698, "grad_norm": 1.0734022855758667, "learning_rate": 9.969668067663652e-05, "loss": 0.7483, "step": 1908 }, { "epoch": 0.12818361799939598, "grad_norm": 1.2380741834640503, "learning_rate": 9.969548418791539e-05, "loss": 0.7703, "step": 1910 }, { "epoch": 0.12831784168316498, "grad_norm": 1.385408639907837, "learning_rate": 9.969428535118036e-05, "loss": 0.7688, "step": 1912 }, { "epoch": 0.128452065366934, "grad_norm": 1.2676340341567993, "learning_rate": 9.969308416648807e-05, "loss": 0.7792, "step": 1914 }, { "epoch": 0.128586289050703, "grad_norm": 1.098507046699524, "learning_rate": 9.969188063389528e-05, "loss": 0.8204, "step": 1916 }, { "epoch": 0.128720512734472, "grad_norm": 1.145999550819397, "learning_rate": 9.969067475345884e-05, "loss": 0.7539, "step": 1918 }, { "epoch": 0.128854736418241, "grad_norm": 1.6709622144699097, "learning_rate": 9.968946652523572e-05, "loss": 0.7803, "step": 1920 }, { "epoch": 0.12898896010201, "grad_norm": 1.552955985069275, "learning_rate": 9.968825594928302e-05, "loss": 0.8612, "step": 1922 }, { "epoch": 0.129123183785779, "grad_norm": 1.707472324371338, "learning_rate": 9.968704302565794e-05, "loss": 0.78, "step": 1924 }, { "epoch": 0.129257407469548, "grad_norm": 1.2199907302856445, "learning_rate": 9.968582775441778e-05, "loss": 0.7672, "step": 1926 }, { "epoch": 0.129391631153317, "grad_norm": 1.230023741722107, "learning_rate": 9.968461013561995e-05, "loss": 0.7606, "step": 1928 }, { "epoch": 0.129525854837086, "grad_norm": 1.389344334602356, "learning_rate": 9.968339016932202e-05, "loss": 0.7466, "step": 1930 }, { "epoch": 0.129660078520855, "grad_norm": 2.679879665374756, "learning_rate": 9.968216785558158e-05, "loss": 0.7859, "step": 1932 }, { "epoch": 0.129794302204624, "grad_norm": 1.282226324081421, "learning_rate": 9.968094319445642e-05, "loss": 0.7685, "step": 1934 }, { "epoch": 0.12992852588839301, "grad_norm": 1.1735801696777344, "learning_rate": 9.967971618600437e-05, "loss": 0.7154, "step": 1936 }, { "epoch": 0.13006274957216202, "grad_norm": 1.2819246053695679, "learning_rate": 9.967848683028343e-05, "loss": 0.8303, "step": 1938 }, { "epoch": 0.13019697325593102, "grad_norm": 1.1663737297058105, "learning_rate": 9.967725512735169e-05, "loss": 0.6946, "step": 1940 }, { "epoch": 0.13033119693970002, "grad_norm": 1.334855556488037, "learning_rate": 9.96760210772673e-05, "loss": 0.7953, "step": 1942 }, { "epoch": 0.130465420623469, "grad_norm": 1.795897126197815, "learning_rate": 9.967478468008861e-05, "loss": 0.8466, "step": 1944 }, { "epoch": 0.130599644307238, "grad_norm": 1.366719126701355, "learning_rate": 9.967354593587403e-05, "loss": 0.8178, "step": 1946 }, { "epoch": 0.130733867991007, "grad_norm": 1.5758782625198364, "learning_rate": 9.96723048446821e-05, "loss": 0.7448, "step": 1948 }, { "epoch": 0.130868091674776, "grad_norm": 6.708095073699951, "learning_rate": 9.967106140657143e-05, "loss": 0.8304, "step": 1950 }, { "epoch": 0.131002315358545, "grad_norm": 1.4056740999221802, "learning_rate": 9.966981562160077e-05, "loss": 0.8205, "step": 1952 }, { "epoch": 0.131136539042314, "grad_norm": 1.2714340686798096, "learning_rate": 9.9668567489829e-05, "loss": 0.8092, "step": 1954 }, { "epoch": 0.131270762726083, "grad_norm": 1.224878191947937, "learning_rate": 9.966731701131509e-05, "loss": 0.7486, "step": 1956 }, { "epoch": 0.13140498640985201, "grad_norm": 2.1951212882995605, "learning_rate": 9.966606418611811e-05, "loss": 0.7046, "step": 1958 }, { "epoch": 0.13153921009362102, "grad_norm": 1.5360804796218872, "learning_rate": 9.966480901429727e-05, "loss": 0.7998, "step": 1960 }, { "epoch": 0.13167343377739002, "grad_norm": 1.2967774868011475, "learning_rate": 9.966355149591187e-05, "loss": 0.8602, "step": 1962 }, { "epoch": 0.13180765746115902, "grad_norm": 1.301439642906189, "learning_rate": 9.96622916310213e-05, "loss": 0.7479, "step": 1964 }, { "epoch": 0.13194188114492802, "grad_norm": 1.2981282472610474, "learning_rate": 9.966102941968512e-05, "loss": 0.749, "step": 1966 }, { "epoch": 0.13207610482869703, "grad_norm": 1.6208810806274414, "learning_rate": 9.965976486196295e-05, "loss": 0.724, "step": 1968 }, { "epoch": 0.13221032851246603, "grad_norm": 1.2995665073394775, "learning_rate": 9.965849795791455e-05, "loss": 0.7397, "step": 1970 }, { "epoch": 0.13234455219623503, "grad_norm": 1.1837626695632935, "learning_rate": 9.965722870759977e-05, "loss": 0.7759, "step": 1972 }, { "epoch": 0.13247877588000403, "grad_norm": 1.2880407571792603, "learning_rate": 9.965595711107858e-05, "loss": 0.8343, "step": 1974 }, { "epoch": 0.13261299956377304, "grad_norm": 1.2756503820419312, "learning_rate": 9.965468316841106e-05, "loss": 0.7975, "step": 1976 }, { "epoch": 0.13274722324754204, "grad_norm": 1.225896954536438, "learning_rate": 9.96534068796574e-05, "loss": 0.7585, "step": 1978 }, { "epoch": 0.13288144693131104, "grad_norm": 2.2436439990997314, "learning_rate": 9.965212824487791e-05, "loss": 0.7737, "step": 1980 }, { "epoch": 0.13301567061508002, "grad_norm": 3.3554556369781494, "learning_rate": 9.965084726413298e-05, "loss": 0.7495, "step": 1982 }, { "epoch": 0.13314989429884902, "grad_norm": 1.4639302492141724, "learning_rate": 9.964956393748317e-05, "loss": 0.7274, "step": 1984 }, { "epoch": 0.13328411798261802, "grad_norm": 1.5518916845321655, "learning_rate": 9.964827826498909e-05, "loss": 0.7824, "step": 1986 }, { "epoch": 0.13341834166638702, "grad_norm": 1.7486345767974854, "learning_rate": 9.964699024671148e-05, "loss": 0.8203, "step": 1988 }, { "epoch": 0.13355256535015603, "grad_norm": 1.6883330345153809, "learning_rate": 9.964569988271122e-05, "loss": 0.7276, "step": 1990 }, { "epoch": 0.13368678903392503, "grad_norm": 1.7586199045181274, "learning_rate": 9.964440717304926e-05, "loss": 0.748, "step": 1992 }, { "epoch": 0.13382101271769403, "grad_norm": 1.22627854347229, "learning_rate": 9.964311211778667e-05, "loss": 0.7706, "step": 1994 }, { "epoch": 0.13395523640146303, "grad_norm": 1.7036383152008057, "learning_rate": 9.964181471698469e-05, "loss": 0.7716, "step": 1996 }, { "epoch": 0.13408946008523204, "grad_norm": 1.3230777978897095, "learning_rate": 9.964051497070455e-05, "loss": 0.7399, "step": 1998 }, { "epoch": 0.13422368376900104, "grad_norm": 1.5753309726715088, "learning_rate": 9.963921287900769e-05, "loss": 0.8049, "step": 2000 }, { "epoch": 0.13435790745277004, "grad_norm": 1.1064035892486572, "learning_rate": 9.963790844195563e-05, "loss": 0.7231, "step": 2002 }, { "epoch": 0.13449213113653904, "grad_norm": 1.4460866451263428, "learning_rate": 9.963660165961002e-05, "loss": 0.7572, "step": 2004 }, { "epoch": 0.13462635482030805, "grad_norm": 1.347019910812378, "learning_rate": 9.963529253203259e-05, "loss": 0.7903, "step": 2006 }, { "epoch": 0.13476057850407705, "grad_norm": 2.20070481300354, "learning_rate": 9.963398105928519e-05, "loss": 0.7821, "step": 2008 }, { "epoch": 0.13489480218784605, "grad_norm": 1.2994256019592285, "learning_rate": 9.963266724142976e-05, "loss": 0.7841, "step": 2010 }, { "epoch": 0.13502902587161505, "grad_norm": 2.113593339920044, "learning_rate": 9.963135107852844e-05, "loss": 0.7278, "step": 2012 }, { "epoch": 0.13516324955538406, "grad_norm": 1.2047135829925537, "learning_rate": 9.963003257064336e-05, "loss": 0.7626, "step": 2014 }, { "epoch": 0.13529747323915306, "grad_norm": 1.3871992826461792, "learning_rate": 9.962871171783684e-05, "loss": 0.8016, "step": 2016 }, { "epoch": 0.13543169692292206, "grad_norm": 1.2552212476730347, "learning_rate": 9.962738852017126e-05, "loss": 0.7771, "step": 2018 }, { "epoch": 0.13556592060669104, "grad_norm": 3.0351154804229736, "learning_rate": 9.962606297770917e-05, "loss": 0.8385, "step": 2020 }, { "epoch": 0.13570014429046004, "grad_norm": 1.8649026155471802, "learning_rate": 9.962473509051319e-05, "loss": 0.7947, "step": 2022 }, { "epoch": 0.13583436797422904, "grad_norm": 1.1793270111083984, "learning_rate": 9.962340485864608e-05, "loss": 0.779, "step": 2024 }, { "epoch": 0.13596859165799804, "grad_norm": 3.530226230621338, "learning_rate": 9.962207228217066e-05, "loss": 0.7629, "step": 2026 }, { "epoch": 0.13610281534176705, "grad_norm": 1.3800299167633057, "learning_rate": 9.962073736114989e-05, "loss": 0.7415, "step": 2028 }, { "epoch": 0.13623703902553605, "grad_norm": 1.15921950340271, "learning_rate": 9.961940009564688e-05, "loss": 0.7576, "step": 2030 }, { "epoch": 0.13637126270930505, "grad_norm": 1.1775150299072266, "learning_rate": 9.961806048572477e-05, "loss": 0.7293, "step": 2032 }, { "epoch": 0.13650548639307405, "grad_norm": 1.3775979280471802, "learning_rate": 9.961671853144687e-05, "loss": 0.7293, "step": 2034 }, { "epoch": 0.13663971007684306, "grad_norm": 1.3519424200057983, "learning_rate": 9.96153742328766e-05, "loss": 0.8251, "step": 2036 }, { "epoch": 0.13677393376061206, "grad_norm": 1.2024866342544556, "learning_rate": 9.961402759007742e-05, "loss": 0.7602, "step": 2038 }, { "epoch": 0.13690815744438106, "grad_norm": 1.2619240283966064, "learning_rate": 9.961267860311305e-05, "loss": 0.7521, "step": 2040 }, { "epoch": 0.13704238112815006, "grad_norm": 1.3698983192443848, "learning_rate": 9.961132727204716e-05, "loss": 0.8188, "step": 2042 }, { "epoch": 0.13717660481191907, "grad_norm": 1.5280874967575073, "learning_rate": 9.96099735969436e-05, "loss": 0.8327, "step": 2044 }, { "epoch": 0.13731082849568807, "grad_norm": 1.436668038368225, "learning_rate": 9.960861757786634e-05, "loss": 0.7705, "step": 2046 }, { "epoch": 0.13744505217945707, "grad_norm": 1.3540871143341064, "learning_rate": 9.960725921487947e-05, "loss": 0.8292, "step": 2048 }, { "epoch": 0.13757927586322607, "grad_norm": 1.2535234689712524, "learning_rate": 9.960589850804713e-05, "loss": 0.8032, "step": 2050 }, { "epoch": 0.13771349954699508, "grad_norm": 1.1608707904815674, "learning_rate": 9.960453545743365e-05, "loss": 0.7469, "step": 2052 }, { "epoch": 0.13784772323076408, "grad_norm": 1.1416162252426147, "learning_rate": 9.96031700631034e-05, "loss": 0.7772, "step": 2054 }, { "epoch": 0.13798194691453308, "grad_norm": 1.2960506677627563, "learning_rate": 9.96018023251209e-05, "loss": 0.7593, "step": 2056 }, { "epoch": 0.13811617059830206, "grad_norm": 1.4882822036743164, "learning_rate": 9.96004322435508e-05, "loss": 0.7428, "step": 2058 }, { "epoch": 0.13825039428207106, "grad_norm": 1.3273555040359497, "learning_rate": 9.959905981845781e-05, "loss": 0.7656, "step": 2060 }, { "epoch": 0.13838461796584006, "grad_norm": 1.2009254693984985, "learning_rate": 9.959768504990675e-05, "loss": 0.7915, "step": 2062 }, { "epoch": 0.13851884164960906, "grad_norm": 1.2466890811920166, "learning_rate": 9.959630793796262e-05, "loss": 0.7866, "step": 2064 }, { "epoch": 0.13865306533337807, "grad_norm": 1.2222994565963745, "learning_rate": 9.959492848269047e-05, "loss": 0.7919, "step": 2066 }, { "epoch": 0.13878728901714707, "grad_norm": 1.122150182723999, "learning_rate": 9.959354668415546e-05, "loss": 0.7601, "step": 2068 }, { "epoch": 0.13892151270091607, "grad_norm": 1.397875428199768, "learning_rate": 9.95921625424229e-05, "loss": 0.7757, "step": 2070 }, { "epoch": 0.13905573638468507, "grad_norm": 1.2442747354507446, "learning_rate": 9.959077605755818e-05, "loss": 0.7413, "step": 2072 }, { "epoch": 0.13918996006845408, "grad_norm": 1.3520575761795044, "learning_rate": 9.95893872296268e-05, "loss": 0.7089, "step": 2074 }, { "epoch": 0.13932418375222308, "grad_norm": 1.3335973024368286, "learning_rate": 9.958799605869438e-05, "loss": 0.7301, "step": 2076 }, { "epoch": 0.13945840743599208, "grad_norm": 1.2340714931488037, "learning_rate": 9.958660254482667e-05, "loss": 0.7388, "step": 2078 }, { "epoch": 0.13959263111976108, "grad_norm": 1.120605707168579, "learning_rate": 9.95852066880895e-05, "loss": 0.8445, "step": 2080 }, { "epoch": 0.1397268548035301, "grad_norm": 1.170307993888855, "learning_rate": 9.95838084885488e-05, "loss": 0.7694, "step": 2082 }, { "epoch": 0.1398610784872991, "grad_norm": 1.257315754890442, "learning_rate": 9.958240794627067e-05, "loss": 0.7403, "step": 2084 }, { "epoch": 0.1399953021710681, "grad_norm": 1.2532832622528076, "learning_rate": 9.958100506132127e-05, "loss": 0.8383, "step": 2086 }, { "epoch": 0.1401295258548371, "grad_norm": 1.278404951095581, "learning_rate": 9.957959983376686e-05, "loss": 0.7382, "step": 2088 }, { "epoch": 0.1402637495386061, "grad_norm": 1.2620716094970703, "learning_rate": 9.957819226367385e-05, "loss": 0.7767, "step": 2090 }, { "epoch": 0.1403979732223751, "grad_norm": 1.158979058265686, "learning_rate": 9.957678235110877e-05, "loss": 0.7545, "step": 2092 }, { "epoch": 0.1405321969061441, "grad_norm": 1.125051498413086, "learning_rate": 9.957537009613819e-05, "loss": 0.7712, "step": 2094 }, { "epoch": 0.14066642058991308, "grad_norm": 1.1082146167755127, "learning_rate": 9.957395549882887e-05, "loss": 0.811, "step": 2096 }, { "epoch": 0.14080064427368208, "grad_norm": 1.2813564538955688, "learning_rate": 9.957253855924761e-05, "loss": 0.7631, "step": 2098 }, { "epoch": 0.14093486795745108, "grad_norm": 1.2051936388015747, "learning_rate": 9.957111927746143e-05, "loss": 0.7933, "step": 2100 }, { "epoch": 0.14106909164122008, "grad_norm": 1.1182860136032104, "learning_rate": 9.956969765353731e-05, "loss": 0.7487, "step": 2102 }, { "epoch": 0.1412033153249891, "grad_norm": 1.590226650238037, "learning_rate": 9.956827368754246e-05, "loss": 0.8099, "step": 2104 }, { "epoch": 0.1413375390087581, "grad_norm": 1.163851261138916, "learning_rate": 9.956684737954414e-05, "loss": 0.7571, "step": 2106 }, { "epoch": 0.1414717626925271, "grad_norm": 1.29503333568573, "learning_rate": 9.956541872960976e-05, "loss": 0.7603, "step": 2108 }, { "epoch": 0.1416059863762961, "grad_norm": 1.2573994398117065, "learning_rate": 9.956398773780682e-05, "loss": 0.7937, "step": 2110 }, { "epoch": 0.1417402100600651, "grad_norm": 1.3426892757415771, "learning_rate": 9.95625544042029e-05, "loss": 0.7615, "step": 2112 }, { "epoch": 0.1418744337438341, "grad_norm": 1.237913727760315, "learning_rate": 9.956111872886576e-05, "loss": 0.7442, "step": 2114 }, { "epoch": 0.1420086574276031, "grad_norm": 1.1821072101593018, "learning_rate": 9.95596807118632e-05, "loss": 0.6547, "step": 2116 }, { "epoch": 0.1421428811113721, "grad_norm": 1.3989217281341553, "learning_rate": 9.955824035326321e-05, "loss": 0.738, "step": 2118 }, { "epoch": 0.1422771047951411, "grad_norm": 1.27780282497406, "learning_rate": 9.95567976531338e-05, "loss": 0.8016, "step": 2120 }, { "epoch": 0.1424113284789101, "grad_norm": 1.3274186849594116, "learning_rate": 9.955535261154316e-05, "loss": 0.7972, "step": 2122 }, { "epoch": 0.1425455521626791, "grad_norm": 1.2454166412353516, "learning_rate": 9.955390522855954e-05, "loss": 0.7838, "step": 2124 }, { "epoch": 0.14267977584644811, "grad_norm": 1.2846773862838745, "learning_rate": 9.955245550425135e-05, "loss": 0.7931, "step": 2126 }, { "epoch": 0.14281399953021712, "grad_norm": 1.2487084865570068, "learning_rate": 9.955100343868709e-05, "loss": 0.7326, "step": 2128 }, { "epoch": 0.14294822321398612, "grad_norm": 1.3661062717437744, "learning_rate": 9.954954903193533e-05, "loss": 0.7615, "step": 2130 }, { "epoch": 0.14308244689775512, "grad_norm": 1.4582746028900146, "learning_rate": 9.954809228406483e-05, "loss": 0.7534, "step": 2132 }, { "epoch": 0.1432166705815241, "grad_norm": 1.4687801599502563, "learning_rate": 9.954663319514439e-05, "loss": 0.8336, "step": 2134 }, { "epoch": 0.1433508942652931, "grad_norm": 1.1722335815429688, "learning_rate": 9.954517176524298e-05, "loss": 0.7461, "step": 2136 }, { "epoch": 0.1434851179490621, "grad_norm": 1.2503559589385986, "learning_rate": 9.954370799442961e-05, "loss": 0.8353, "step": 2138 }, { "epoch": 0.1436193416328311, "grad_norm": 1.614410400390625, "learning_rate": 9.954224188277347e-05, "loss": 0.6964, "step": 2140 }, { "epoch": 0.1437535653166001, "grad_norm": 1.238930344581604, "learning_rate": 9.954077343034383e-05, "loss": 0.7151, "step": 2142 }, { "epoch": 0.1438877890003691, "grad_norm": 1.317571997642517, "learning_rate": 9.953930263721003e-05, "loss": 0.8119, "step": 2144 }, { "epoch": 0.1440220126841381, "grad_norm": 1.0393397808074951, "learning_rate": 9.953782950344164e-05, "loss": 0.7762, "step": 2146 }, { "epoch": 0.14415623636790711, "grad_norm": 1.179162621498108, "learning_rate": 9.95363540291082e-05, "loss": 0.8064, "step": 2148 }, { "epoch": 0.14429046005167612, "grad_norm": 1.8216201066970825, "learning_rate": 9.953487621427942e-05, "loss": 0.7538, "step": 2150 }, { "epoch": 0.14442468373544512, "grad_norm": 1.174189805984497, "learning_rate": 9.953339605902517e-05, "loss": 0.8036, "step": 2152 }, { "epoch": 0.14455890741921412, "grad_norm": 1.2796666622161865, "learning_rate": 9.953191356341535e-05, "loss": 0.7963, "step": 2154 }, { "epoch": 0.14469313110298312, "grad_norm": 1.1608229875564575, "learning_rate": 9.953042872752003e-05, "loss": 0.7536, "step": 2156 }, { "epoch": 0.14482735478675213, "grad_norm": 1.2113354206085205, "learning_rate": 9.952894155140932e-05, "loss": 0.7396, "step": 2158 }, { "epoch": 0.14496157847052113, "grad_norm": 1.3290759325027466, "learning_rate": 9.952745203515354e-05, "loss": 0.7895, "step": 2160 }, { "epoch": 0.14509580215429013, "grad_norm": 1.383651614189148, "learning_rate": 9.952596017882304e-05, "loss": 0.7644, "step": 2162 }, { "epoch": 0.14523002583805913, "grad_norm": 1.2829383611679077, "learning_rate": 9.95244659824883e-05, "loss": 0.7665, "step": 2164 }, { "epoch": 0.14536424952182814, "grad_norm": 1.3651251792907715, "learning_rate": 9.952296944621994e-05, "loss": 0.7422, "step": 2166 }, { "epoch": 0.14549847320559714, "grad_norm": 1.5312341451644897, "learning_rate": 9.952147057008864e-05, "loss": 0.7046, "step": 2168 }, { "epoch": 0.14563269688936614, "grad_norm": 1.7080609798431396, "learning_rate": 9.951996935416525e-05, "loss": 0.7642, "step": 2170 }, { "epoch": 0.14576692057313512, "grad_norm": 1.1974818706512451, "learning_rate": 9.951846579852069e-05, "loss": 0.7425, "step": 2172 }, { "epoch": 0.14590114425690412, "grad_norm": 1.213015079498291, "learning_rate": 9.951695990322598e-05, "loss": 0.8013, "step": 2174 }, { "epoch": 0.14603536794067312, "grad_norm": 1.1024574041366577, "learning_rate": 9.95154516683523e-05, "loss": 0.7126, "step": 2176 }, { "epoch": 0.14616959162444212, "grad_norm": 1.175889492034912, "learning_rate": 9.951394109397088e-05, "loss": 0.7567, "step": 2178 }, { "epoch": 0.14630381530821113, "grad_norm": 3.2251601219177246, "learning_rate": 9.951242818015312e-05, "loss": 0.7992, "step": 2180 }, { "epoch": 0.14643803899198013, "grad_norm": 1.39309561252594, "learning_rate": 9.951091292697048e-05, "loss": 0.756, "step": 2182 }, { "epoch": 0.14657226267574913, "grad_norm": 1.3651174306869507, "learning_rate": 9.950939533449458e-05, "loss": 0.8032, "step": 2184 }, { "epoch": 0.14670648635951813, "grad_norm": 1.6942890882492065, "learning_rate": 9.950787540279711e-05, "loss": 0.7455, "step": 2186 }, { "epoch": 0.14684071004328714, "grad_norm": 1.5253287553787231, "learning_rate": 9.950635313194986e-05, "loss": 0.7913, "step": 2188 }, { "epoch": 0.14697493372705614, "grad_norm": 1.0081595182418823, "learning_rate": 9.950482852202478e-05, "loss": 0.71, "step": 2190 }, { "epoch": 0.14710915741082514, "grad_norm": 1.2482656240463257, "learning_rate": 9.95033015730939e-05, "loss": 0.8345, "step": 2192 }, { "epoch": 0.14724338109459414, "grad_norm": 1.436821460723877, "learning_rate": 9.950177228522937e-05, "loss": 0.783, "step": 2194 }, { "epoch": 0.14737760477836315, "grad_norm": 1.3209800720214844, "learning_rate": 9.950024065850343e-05, "loss": 0.708, "step": 2196 }, { "epoch": 0.14751182846213215, "grad_norm": 1.070220947265625, "learning_rate": 9.949870669298846e-05, "loss": 0.7609, "step": 2198 }, { "epoch": 0.14764605214590115, "grad_norm": 1.7431410551071167, "learning_rate": 9.949717038875695e-05, "loss": 0.7533, "step": 2200 }, { "epoch": 0.14778027582967015, "grad_norm": 1.4989782571792603, "learning_rate": 9.949563174588146e-05, "loss": 0.7089, "step": 2202 }, { "epoch": 0.14791449951343916, "grad_norm": 1.8121305704116821, "learning_rate": 9.949409076443468e-05, "loss": 0.7706, "step": 2204 }, { "epoch": 0.14804872319720816, "grad_norm": 1.2245267629623413, "learning_rate": 9.949254744448946e-05, "loss": 0.8267, "step": 2206 }, { "epoch": 0.14818294688097716, "grad_norm": 2.5223753452301025, "learning_rate": 9.94910017861187e-05, "loss": 0.7736, "step": 2208 }, { "epoch": 0.14831717056474614, "grad_norm": 1.1899269819259644, "learning_rate": 9.948945378939542e-05, "loss": 0.7145, "step": 2210 }, { "epoch": 0.14845139424851514, "grad_norm": 1.1237188577651978, "learning_rate": 9.948790345439276e-05, "loss": 0.8445, "step": 2212 }, { "epoch": 0.14858561793228414, "grad_norm": 1.4916682243347168, "learning_rate": 9.948635078118398e-05, "loss": 0.8115, "step": 2214 }, { "epoch": 0.14871984161605314, "grad_norm": 1.5387001037597656, "learning_rate": 9.948479576984242e-05, "loss": 0.7568, "step": 2216 }, { "epoch": 0.14885406529982215, "grad_norm": 1.1785844564437866, "learning_rate": 9.948323842044159e-05, "loss": 0.7708, "step": 2218 }, { "epoch": 0.14898828898359115, "grad_norm": 1.6287435293197632, "learning_rate": 9.948167873305503e-05, "loss": 0.7399, "step": 2220 }, { "epoch": 0.14912251266736015, "grad_norm": 1.3606981039047241, "learning_rate": 9.948011670775647e-05, "loss": 0.8457, "step": 2222 }, { "epoch": 0.14925673635112915, "grad_norm": 1.2144297361373901, "learning_rate": 9.94785523446197e-05, "loss": 0.7864, "step": 2224 }, { "epoch": 0.14939096003489816, "grad_norm": 1.1448192596435547, "learning_rate": 9.947698564371859e-05, "loss": 0.751, "step": 2226 }, { "epoch": 0.14952518371866716, "grad_norm": 1.1770398616790771, "learning_rate": 9.947541660512722e-05, "loss": 0.7395, "step": 2228 }, { "epoch": 0.14965940740243616, "grad_norm": 1.417425274848938, "learning_rate": 9.947384522891972e-05, "loss": 0.803, "step": 2230 }, { "epoch": 0.14979363108620516, "grad_norm": 1.1650129556655884, "learning_rate": 9.947227151517032e-05, "loss": 0.7982, "step": 2232 }, { "epoch": 0.14992785476997417, "grad_norm": 1.4188095331192017, "learning_rate": 9.947069546395336e-05, "loss": 0.7795, "step": 2234 }, { "epoch": 0.15006207845374317, "grad_norm": 1.1999107599258423, "learning_rate": 9.946911707534331e-05, "loss": 0.7865, "step": 2236 }, { "epoch": 0.15019630213751217, "grad_norm": 1.2209099531173706, "learning_rate": 9.946753634941477e-05, "loss": 0.817, "step": 2238 }, { "epoch": 0.15033052582128117, "grad_norm": 1.3145641088485718, "learning_rate": 9.94659532862424e-05, "loss": 0.7132, "step": 2240 }, { "epoch": 0.15046474950505018, "grad_norm": 1.1500871181488037, "learning_rate": 9.9464367885901e-05, "loss": 0.7218, "step": 2242 }, { "epoch": 0.15059897318881918, "grad_norm": 1.2443076372146606, "learning_rate": 9.946278014846551e-05, "loss": 0.8128, "step": 2244 }, { "epoch": 0.15073319687258815, "grad_norm": 1.2562689781188965, "learning_rate": 9.946119007401091e-05, "loss": 0.7358, "step": 2246 }, { "epoch": 0.15086742055635716, "grad_norm": 2.3361754417419434, "learning_rate": 9.945959766261235e-05, "loss": 0.7404, "step": 2248 }, { "epoch": 0.15100164424012616, "grad_norm": 1.2442772388458252, "learning_rate": 9.945800291434504e-05, "loss": 0.7834, "step": 2250 }, { "epoch": 0.15113586792389516, "grad_norm": 1.1339282989501953, "learning_rate": 9.945640582928437e-05, "loss": 0.6752, "step": 2252 }, { "epoch": 0.15127009160766416, "grad_norm": 1.4449889659881592, "learning_rate": 9.945480640750577e-05, "loss": 0.7898, "step": 2254 }, { "epoch": 0.15140431529143317, "grad_norm": 1.2559311389923096, "learning_rate": 9.945320464908481e-05, "loss": 0.7696, "step": 2256 }, { "epoch": 0.15153853897520217, "grad_norm": 1.77239191532135, "learning_rate": 9.945160055409717e-05, "loss": 0.8126, "step": 2258 }, { "epoch": 0.15167276265897117, "grad_norm": 1.2011572122573853, "learning_rate": 9.944999412261866e-05, "loss": 0.7859, "step": 2260 }, { "epoch": 0.15180698634274017, "grad_norm": 1.8600788116455078, "learning_rate": 9.944838535472515e-05, "loss": 0.7774, "step": 2262 }, { "epoch": 0.15194121002650918, "grad_norm": 1.2087199687957764, "learning_rate": 9.944677425049268e-05, "loss": 0.7841, "step": 2264 }, { "epoch": 0.15207543371027818, "grad_norm": 1.272459864616394, "learning_rate": 9.944516080999735e-05, "loss": 0.7719, "step": 2266 }, { "epoch": 0.15220965739404718, "grad_norm": 1.721838116645813, "learning_rate": 9.944354503331541e-05, "loss": 0.7264, "step": 2268 }, { "epoch": 0.15234388107781618, "grad_norm": 1.2229008674621582, "learning_rate": 9.94419269205232e-05, "loss": 0.7513, "step": 2270 }, { "epoch": 0.1524781047615852, "grad_norm": 1.1715528964996338, "learning_rate": 9.944030647169715e-05, "loss": 0.7562, "step": 2272 }, { "epoch": 0.1526123284453542, "grad_norm": 1.1329717636108398, "learning_rate": 9.943868368691386e-05, "loss": 0.702, "step": 2274 }, { "epoch": 0.1527465521291232, "grad_norm": 1.1848491430282593, "learning_rate": 9.943705856624996e-05, "loss": 0.7484, "step": 2276 }, { "epoch": 0.1528807758128922, "grad_norm": 1.3173327445983887, "learning_rate": 9.943543110978227e-05, "loss": 0.7915, "step": 2278 }, { "epoch": 0.1530149994966612, "grad_norm": 1.2425185441970825, "learning_rate": 9.943380131758768e-05, "loss": 0.7463, "step": 2280 }, { "epoch": 0.1531492231804302, "grad_norm": 1.7477085590362549, "learning_rate": 9.943216918974317e-05, "loss": 0.7949, "step": 2282 }, { "epoch": 0.15328344686419917, "grad_norm": 1.1816725730895996, "learning_rate": 9.943053472632587e-05, "loss": 0.7867, "step": 2284 }, { "epoch": 0.15341767054796818, "grad_norm": 1.3466622829437256, "learning_rate": 9.942889792741302e-05, "loss": 0.7513, "step": 2286 }, { "epoch": 0.15355189423173718, "grad_norm": 1.5214899778366089, "learning_rate": 9.942725879308192e-05, "loss": 0.7092, "step": 2288 }, { "epoch": 0.15368611791550618, "grad_norm": 1.2291330099105835, "learning_rate": 9.942561732341005e-05, "loss": 0.7821, "step": 2290 }, { "epoch": 0.15382034159927518, "grad_norm": 1.209675908088684, "learning_rate": 9.942397351847494e-05, "loss": 0.6945, "step": 2292 }, { "epoch": 0.1539545652830442, "grad_norm": 1.1875338554382324, "learning_rate": 9.942232737835428e-05, "loss": 0.8003, "step": 2294 }, { "epoch": 0.1540887889668132, "grad_norm": 1.1507275104522705, "learning_rate": 9.942067890312584e-05, "loss": 0.8093, "step": 2296 }, { "epoch": 0.1542230126505822, "grad_norm": 1.1249173879623413, "learning_rate": 9.94190280928675e-05, "loss": 0.7864, "step": 2298 }, { "epoch": 0.1543572363343512, "grad_norm": 1.3070746660232544, "learning_rate": 9.941737494765725e-05, "loss": 0.6958, "step": 2300 }, { "epoch": 0.1544914600181202, "grad_norm": 1.1317988634109497, "learning_rate": 9.941571946757321e-05, "loss": 0.7691, "step": 2302 }, { "epoch": 0.1546256837018892, "grad_norm": 1.2521330118179321, "learning_rate": 9.941406165269362e-05, "loss": 0.7484, "step": 2304 }, { "epoch": 0.1547599073856582, "grad_norm": 1.2915416955947876, "learning_rate": 9.941240150309676e-05, "loss": 0.829, "step": 2306 }, { "epoch": 0.1548941310694272, "grad_norm": 0.9967454671859741, "learning_rate": 9.94107390188611e-05, "loss": 0.6904, "step": 2308 }, { "epoch": 0.1550283547531962, "grad_norm": 1.4336018562316895, "learning_rate": 9.94090742000652e-05, "loss": 0.7557, "step": 2310 }, { "epoch": 0.1551625784369652, "grad_norm": 1.0555182695388794, "learning_rate": 9.940740704678768e-05, "loss": 0.7726, "step": 2312 }, { "epoch": 0.1552968021207342, "grad_norm": 1.298625111579895, "learning_rate": 9.940573755910735e-05, "loss": 0.7815, "step": 2314 }, { "epoch": 0.15543102580450321, "grad_norm": 1.1737966537475586, "learning_rate": 9.940406573710306e-05, "loss": 0.7837, "step": 2316 }, { "epoch": 0.15556524948827222, "grad_norm": 2.8313827514648438, "learning_rate": 9.940239158085382e-05, "loss": 0.8697, "step": 2318 }, { "epoch": 0.15569947317204122, "grad_norm": 1.422286033630371, "learning_rate": 9.940071509043872e-05, "loss": 0.7542, "step": 2320 }, { "epoch": 0.1558336968558102, "grad_norm": 1.3354989290237427, "learning_rate": 9.939903626593698e-05, "loss": 0.8018, "step": 2322 }, { "epoch": 0.1559679205395792, "grad_norm": 1.183380126953125, "learning_rate": 9.939735510742792e-05, "loss": 0.777, "step": 2324 }, { "epoch": 0.1561021442233482, "grad_norm": 1.3062756061553955, "learning_rate": 9.939567161499095e-05, "loss": 0.7899, "step": 2326 }, { "epoch": 0.1562363679071172, "grad_norm": 1.3748116493225098, "learning_rate": 9.939398578870563e-05, "loss": 0.7707, "step": 2328 }, { "epoch": 0.1563705915908862, "grad_norm": 1.2097219228744507, "learning_rate": 9.939229762865164e-05, "loss": 0.7341, "step": 2330 }, { "epoch": 0.1565048152746552, "grad_norm": 1.2697770595550537, "learning_rate": 9.939060713490868e-05, "loss": 0.6468, "step": 2332 }, { "epoch": 0.1566390389584242, "grad_norm": 1.3282665014266968, "learning_rate": 9.938891430755666e-05, "loss": 0.7525, "step": 2334 }, { "epoch": 0.1567732626421932, "grad_norm": 1.0913504362106323, "learning_rate": 9.938721914667557e-05, "loss": 0.7398, "step": 2336 }, { "epoch": 0.15690748632596221, "grad_norm": 1.1177167892456055, "learning_rate": 9.938552165234548e-05, "loss": 0.7344, "step": 2338 }, { "epoch": 0.15704171000973122, "grad_norm": 1.3176368474960327, "learning_rate": 9.93838218246466e-05, "loss": 0.7472, "step": 2340 }, { "epoch": 0.15717593369350022, "grad_norm": 1.211304783821106, "learning_rate": 9.938211966365926e-05, "loss": 0.7897, "step": 2342 }, { "epoch": 0.15731015737726922, "grad_norm": 1.1732478141784668, "learning_rate": 9.938041516946389e-05, "loss": 0.7204, "step": 2344 }, { "epoch": 0.15744438106103822, "grad_norm": 1.1366077661514282, "learning_rate": 9.937870834214097e-05, "loss": 0.6967, "step": 2346 }, { "epoch": 0.15757860474480723, "grad_norm": 5.17569637298584, "learning_rate": 9.93769991817712e-05, "loss": 0.7289, "step": 2348 }, { "epoch": 0.15771282842857623, "grad_norm": 1.351745843887329, "learning_rate": 9.93752876884353e-05, "loss": 0.7601, "step": 2350 }, { "epoch": 0.15784705211234523, "grad_norm": 1.280282974243164, "learning_rate": 9.937357386221416e-05, "loss": 0.7786, "step": 2352 }, { "epoch": 0.15798127579611423, "grad_norm": 1.163209080696106, "learning_rate": 9.937185770318874e-05, "loss": 0.748, "step": 2354 }, { "epoch": 0.15811549947988324, "grad_norm": 1.4269531965255737, "learning_rate": 9.937013921144014e-05, "loss": 0.7422, "step": 2356 }, { "epoch": 0.15824972316365224, "grad_norm": 1.350040078163147, "learning_rate": 9.936841838704953e-05, "loss": 0.7644, "step": 2358 }, { "epoch": 0.15838394684742121, "grad_norm": 1.2202732563018799, "learning_rate": 9.936669523009823e-05, "loss": 0.7566, "step": 2360 }, { "epoch": 0.15851817053119022, "grad_norm": 1.2352739572525024, "learning_rate": 9.936496974066767e-05, "loss": 0.7225, "step": 2362 }, { "epoch": 0.15865239421495922, "grad_norm": 1.2050681114196777, "learning_rate": 9.936324191883935e-05, "loss": 0.7364, "step": 2364 }, { "epoch": 0.15878661789872822, "grad_norm": 1.1253108978271484, "learning_rate": 9.936151176469493e-05, "loss": 0.7951, "step": 2366 }, { "epoch": 0.15892084158249722, "grad_norm": 1.2299094200134277, "learning_rate": 9.935977927831612e-05, "loss": 0.7207, "step": 2368 }, { "epoch": 0.15905506526626623, "grad_norm": 1.3539718389511108, "learning_rate": 9.935804445978483e-05, "loss": 0.8654, "step": 2370 }, { "epoch": 0.15918928895003523, "grad_norm": 1.3129518032073975, "learning_rate": 9.935630730918297e-05, "loss": 0.7026, "step": 2372 }, { "epoch": 0.15932351263380423, "grad_norm": 1.2837015390396118, "learning_rate": 9.935456782659267e-05, "loss": 0.8147, "step": 2374 }, { "epoch": 0.15945773631757323, "grad_norm": 1.6276519298553467, "learning_rate": 9.935282601209608e-05, "loss": 0.7906, "step": 2376 }, { "epoch": 0.15959196000134224, "grad_norm": 1.1305961608886719, "learning_rate": 9.93510818657755e-05, "loss": 0.7622, "step": 2378 }, { "epoch": 0.15972618368511124, "grad_norm": 1.7124518156051636, "learning_rate": 9.934933538771336e-05, "loss": 0.7523, "step": 2380 }, { "epoch": 0.15986040736888024, "grad_norm": 1.5002055168151855, "learning_rate": 9.934758657799218e-05, "loss": 0.8082, "step": 2382 }, { "epoch": 0.15999463105264924, "grad_norm": 1.216407060623169, "learning_rate": 9.934583543669453e-05, "loss": 0.7598, "step": 2384 }, { "epoch": 0.16012885473641825, "grad_norm": 1.3850566148757935, "learning_rate": 9.934408196390322e-05, "loss": 0.7452, "step": 2386 }, { "epoch": 0.16026307842018725, "grad_norm": 1.1947745084762573, "learning_rate": 9.934232615970107e-05, "loss": 0.7384, "step": 2388 }, { "epoch": 0.16039730210395625, "grad_norm": 1.2187963724136353, "learning_rate": 9.934056802417101e-05, "loss": 0.7995, "step": 2390 }, { "epoch": 0.16053152578772525, "grad_norm": 1.179424524307251, "learning_rate": 9.933880755739616e-05, "loss": 0.8486, "step": 2392 }, { "epoch": 0.16066574947149426, "grad_norm": 1.32096266746521, "learning_rate": 9.933704475945966e-05, "loss": 0.7991, "step": 2394 }, { "epoch": 0.16079997315526326, "grad_norm": 1.4193588495254517, "learning_rate": 9.933527963044483e-05, "loss": 0.7433, "step": 2396 }, { "epoch": 0.16093419683903223, "grad_norm": 1.4676076173782349, "learning_rate": 9.933351217043504e-05, "loss": 0.7406, "step": 2398 }, { "epoch": 0.16106842052280124, "grad_norm": 1.3093665838241577, "learning_rate": 9.933174237951381e-05, "loss": 0.7232, "step": 2400 }, { "epoch": 0.16120264420657024, "grad_norm": 1.3477256298065186, "learning_rate": 9.932997025776475e-05, "loss": 0.7955, "step": 2402 }, { "epoch": 0.16133686789033924, "grad_norm": 1.4739902019500732, "learning_rate": 9.932819580527162e-05, "loss": 0.7765, "step": 2404 }, { "epoch": 0.16147109157410824, "grad_norm": 0.9575392007827759, "learning_rate": 9.932641902211821e-05, "loss": 0.7153, "step": 2406 }, { "epoch": 0.16160531525787725, "grad_norm": 1.135473370552063, "learning_rate": 9.932463990838851e-05, "loss": 0.7088, "step": 2408 }, { "epoch": 0.16173953894164625, "grad_norm": 1.268178939819336, "learning_rate": 9.932285846416658e-05, "loss": 0.7715, "step": 2410 }, { "epoch": 0.16187376262541525, "grad_norm": 1.320458173751831, "learning_rate": 9.932107468953656e-05, "loss": 0.7896, "step": 2412 }, { "epoch": 0.16200798630918425, "grad_norm": 1.2818015813827515, "learning_rate": 9.931928858458276e-05, "loss": 0.7705, "step": 2414 }, { "epoch": 0.16214220999295326, "grad_norm": 1.1835918426513672, "learning_rate": 9.931750014938954e-05, "loss": 0.7115, "step": 2416 }, { "epoch": 0.16227643367672226, "grad_norm": 1.1886675357818604, "learning_rate": 9.931570938404144e-05, "loss": 0.7545, "step": 2418 }, { "epoch": 0.16241065736049126, "grad_norm": 1.196901559829712, "learning_rate": 9.931391628862304e-05, "loss": 0.7196, "step": 2420 }, { "epoch": 0.16254488104426026, "grad_norm": 1.235708475112915, "learning_rate": 9.931212086321905e-05, "loss": 0.8074, "step": 2422 }, { "epoch": 0.16267910472802927, "grad_norm": 1.2353520393371582, "learning_rate": 9.931032310791432e-05, "loss": 0.7673, "step": 2424 }, { "epoch": 0.16281332841179827, "grad_norm": 1.1860045194625854, "learning_rate": 9.930852302279381e-05, "loss": 0.7585, "step": 2426 }, { "epoch": 0.16294755209556727, "grad_norm": 1.1550395488739014, "learning_rate": 9.930672060794253e-05, "loss": 0.7325, "step": 2428 }, { "epoch": 0.16308177577933627, "grad_norm": 1.867236614227295, "learning_rate": 9.930491586344565e-05, "loss": 0.7485, "step": 2430 }, { "epoch": 0.16321599946310528, "grad_norm": 1.3778027296066284, "learning_rate": 9.930310878938846e-05, "loss": 0.7444, "step": 2432 }, { "epoch": 0.16335022314687428, "grad_norm": 1.1479414701461792, "learning_rate": 9.930129938585633e-05, "loss": 0.781, "step": 2434 }, { "epoch": 0.16348444683064325, "grad_norm": 1.8815853595733643, "learning_rate": 9.929948765293477e-05, "loss": 0.6867, "step": 2436 }, { "epoch": 0.16361867051441226, "grad_norm": 1.6327052116394043, "learning_rate": 9.929767359070934e-05, "loss": 0.6813, "step": 2438 }, { "epoch": 0.16375289419818126, "grad_norm": 1.1803863048553467, "learning_rate": 9.929585719926577e-05, "loss": 0.7515, "step": 2440 }, { "epoch": 0.16388711788195026, "grad_norm": 1.522848129272461, "learning_rate": 9.92940384786899e-05, "loss": 0.7298, "step": 2442 }, { "epoch": 0.16402134156571926, "grad_norm": 1.3202654123306274, "learning_rate": 9.929221742906763e-05, "loss": 0.8428, "step": 2444 }, { "epoch": 0.16415556524948827, "grad_norm": 1.2601298093795776, "learning_rate": 9.929039405048501e-05, "loss": 0.7443, "step": 2446 }, { "epoch": 0.16428978893325727, "grad_norm": 1.2158600091934204, "learning_rate": 9.928856834302823e-05, "loss": 0.7908, "step": 2448 }, { "epoch": 0.16442401261702627, "grad_norm": 1.2021377086639404, "learning_rate": 9.928674030678348e-05, "loss": 0.7545, "step": 2450 }, { "epoch": 0.16455823630079527, "grad_norm": 1.3095088005065918, "learning_rate": 9.928490994183719e-05, "loss": 0.7767, "step": 2452 }, { "epoch": 0.16469245998456428, "grad_norm": 1.6881237030029297, "learning_rate": 9.928307724827581e-05, "loss": 0.739, "step": 2454 }, { "epoch": 0.16482668366833328, "grad_norm": 1.1968653202056885, "learning_rate": 9.928124222618594e-05, "loss": 0.6853, "step": 2456 }, { "epoch": 0.16496090735210228, "grad_norm": 2.4138987064361572, "learning_rate": 9.92794048756543e-05, "loss": 0.7295, "step": 2458 }, { "epoch": 0.16509513103587128, "grad_norm": 1.1000226736068726, "learning_rate": 9.927756519676765e-05, "loss": 0.7345, "step": 2460 }, { "epoch": 0.1652293547196403, "grad_norm": 1.4773651361465454, "learning_rate": 9.927572318961299e-05, "loss": 0.7501, "step": 2462 }, { "epoch": 0.1653635784034093, "grad_norm": 1.304039478302002, "learning_rate": 9.927387885427726e-05, "loss": 0.6835, "step": 2464 }, { "epoch": 0.1654978020871783, "grad_norm": 1.3185865879058838, "learning_rate": 9.927203219084766e-05, "loss": 0.7706, "step": 2466 }, { "epoch": 0.1656320257709473, "grad_norm": 1.181311845779419, "learning_rate": 9.927018319941145e-05, "loss": 0.7814, "step": 2468 }, { "epoch": 0.1657662494547163, "grad_norm": 2.8532204627990723, "learning_rate": 9.926833188005595e-05, "loss": 0.7775, "step": 2470 }, { "epoch": 0.1659004731384853, "grad_norm": 1.2215230464935303, "learning_rate": 9.926647823286865e-05, "loss": 0.7126, "step": 2472 }, { "epoch": 0.16603469682225427, "grad_norm": 1.2203389406204224, "learning_rate": 9.926462225793714e-05, "loss": 0.7673, "step": 2474 }, { "epoch": 0.16616892050602328, "grad_norm": 1.7585606575012207, "learning_rate": 9.926276395534911e-05, "loss": 0.7177, "step": 2476 }, { "epoch": 0.16630314418979228, "grad_norm": 1.2644014358520508, "learning_rate": 9.926090332519234e-05, "loss": 0.8059, "step": 2478 }, { "epoch": 0.16643736787356128, "grad_norm": 1.0948673486709595, "learning_rate": 9.925904036755475e-05, "loss": 0.8598, "step": 2480 }, { "epoch": 0.16657159155733028, "grad_norm": 1.1442246437072754, "learning_rate": 9.92571750825244e-05, "loss": 0.7273, "step": 2482 }, { "epoch": 0.1667058152410993, "grad_norm": 1.0907037258148193, "learning_rate": 9.925530747018935e-05, "loss": 0.7578, "step": 2484 }, { "epoch": 0.1668400389248683, "grad_norm": 1.1021397113800049, "learning_rate": 9.92534375306379e-05, "loss": 0.6874, "step": 2486 }, { "epoch": 0.1669742626086373, "grad_norm": 1.3376507759094238, "learning_rate": 9.925156526395836e-05, "loss": 0.7817, "step": 2488 }, { "epoch": 0.1671084862924063, "grad_norm": 1.2467944622039795, "learning_rate": 9.924969067023922e-05, "loss": 0.7608, "step": 2490 }, { "epoch": 0.1672427099761753, "grad_norm": 1.2369590997695923, "learning_rate": 9.924781374956905e-05, "loss": 0.7391, "step": 2492 }, { "epoch": 0.1673769336599443, "grad_norm": 1.2221477031707764, "learning_rate": 9.924593450203652e-05, "loss": 0.7096, "step": 2494 }, { "epoch": 0.1675111573437133, "grad_norm": 1.2566503286361694, "learning_rate": 9.924405292773041e-05, "loss": 0.7927, "step": 2496 }, { "epoch": 0.1676453810274823, "grad_norm": 1.7431105375289917, "learning_rate": 9.924216902673964e-05, "loss": 0.7695, "step": 2498 }, { "epoch": 0.1677796047112513, "grad_norm": 1.2656195163726807, "learning_rate": 9.924028279915323e-05, "loss": 0.7307, "step": 2500 }, { "epoch": 0.1679138283950203, "grad_norm": 1.2934205532073975, "learning_rate": 9.923839424506026e-05, "loss": 0.7277, "step": 2502 }, { "epoch": 0.1680480520787893, "grad_norm": 1.0522221326828003, "learning_rate": 9.923650336454999e-05, "loss": 0.6516, "step": 2504 }, { "epoch": 0.16818227576255831, "grad_norm": 1.1378988027572632, "learning_rate": 9.923461015771177e-05, "loss": 0.749, "step": 2506 }, { "epoch": 0.16831649944632732, "grad_norm": 1.086676001548767, "learning_rate": 9.923271462463502e-05, "loss": 0.7114, "step": 2508 }, { "epoch": 0.16845072313009632, "grad_norm": 1.1618993282318115, "learning_rate": 9.923081676540933e-05, "loss": 0.7368, "step": 2510 }, { "epoch": 0.1685849468138653, "grad_norm": 1.5692874193191528, "learning_rate": 9.922891658012435e-05, "loss": 0.7363, "step": 2512 }, { "epoch": 0.1687191704976343, "grad_norm": 1.5985716581344604, "learning_rate": 9.922701406886987e-05, "loss": 0.7275, "step": 2514 }, { "epoch": 0.1688533941814033, "grad_norm": 1.7021191120147705, "learning_rate": 9.922510923173576e-05, "loss": 0.7594, "step": 2516 }, { "epoch": 0.1689876178651723, "grad_norm": 1.0292367935180664, "learning_rate": 9.922320206881205e-05, "loss": 0.7647, "step": 2518 }, { "epoch": 0.1691218415489413, "grad_norm": 1.0268237590789795, "learning_rate": 9.922129258018883e-05, "loss": 0.7616, "step": 2520 }, { "epoch": 0.1692560652327103, "grad_norm": 1.1957460641860962, "learning_rate": 9.921938076595632e-05, "loss": 0.7571, "step": 2522 }, { "epoch": 0.1693902889164793, "grad_norm": 1.319554328918457, "learning_rate": 9.921746662620488e-05, "loss": 0.7842, "step": 2524 }, { "epoch": 0.1695245126002483, "grad_norm": 1.523512601852417, "learning_rate": 9.92155501610249e-05, "loss": 0.7729, "step": 2526 }, { "epoch": 0.16965873628401731, "grad_norm": 1.1168030500411987, "learning_rate": 9.921363137050696e-05, "loss": 0.7467, "step": 2528 }, { "epoch": 0.16979295996778632, "grad_norm": 1.394824504852295, "learning_rate": 9.921171025474171e-05, "loss": 0.7121, "step": 2530 }, { "epoch": 0.16992718365155532, "grad_norm": 1.3063008785247803, "learning_rate": 9.920978681381991e-05, "loss": 0.7918, "step": 2532 }, { "epoch": 0.17006140733532432, "grad_norm": 1.3497661352157593, "learning_rate": 9.920786104783247e-05, "loss": 0.7768, "step": 2534 }, { "epoch": 0.17019563101909332, "grad_norm": 1.41562819480896, "learning_rate": 9.920593295687035e-05, "loss": 0.7616, "step": 2536 }, { "epoch": 0.17032985470286233, "grad_norm": 1.2921581268310547, "learning_rate": 9.920400254102466e-05, "loss": 0.7841, "step": 2538 }, { "epoch": 0.17046407838663133, "grad_norm": 1.0917983055114746, "learning_rate": 9.92020698003866e-05, "loss": 0.7187, "step": 2540 }, { "epoch": 0.17059830207040033, "grad_norm": 1.3144159317016602, "learning_rate": 9.92001347350475e-05, "loss": 0.7462, "step": 2542 }, { "epoch": 0.17073252575416933, "grad_norm": 1.0488479137420654, "learning_rate": 9.919819734509878e-05, "loss": 0.704, "step": 2544 }, { "epoch": 0.17086674943793834, "grad_norm": 1.1551920175552368, "learning_rate": 9.919625763063197e-05, "loss": 0.7127, "step": 2546 }, { "epoch": 0.1710009731217073, "grad_norm": 1.1880587339401245, "learning_rate": 9.919431559173874e-05, "loss": 0.8004, "step": 2548 }, { "epoch": 0.17113519680547631, "grad_norm": 1.3170595169067383, "learning_rate": 9.919237122851084e-05, "loss": 0.7678, "step": 2550 }, { "epoch": 0.17126942048924532, "grad_norm": 1.301263451576233, "learning_rate": 9.91904245410401e-05, "loss": 0.7392, "step": 2552 }, { "epoch": 0.17140364417301432, "grad_norm": 1.723443627357483, "learning_rate": 9.918847552941856e-05, "loss": 0.7954, "step": 2554 }, { "epoch": 0.17153786785678332, "grad_norm": 1.2289232015609741, "learning_rate": 9.918652419373827e-05, "loss": 0.78, "step": 2556 }, { "epoch": 0.17167209154055232, "grad_norm": 1.0448615550994873, "learning_rate": 9.918457053409142e-05, "loss": 0.7678, "step": 2558 }, { "epoch": 0.17180631522432133, "grad_norm": 1.0582432746887207, "learning_rate": 9.918261455057036e-05, "loss": 0.6728, "step": 2560 }, { "epoch": 0.17194053890809033, "grad_norm": 1.1932666301727295, "learning_rate": 9.918065624326745e-05, "loss": 0.7933, "step": 2562 }, { "epoch": 0.17207476259185933, "grad_norm": 1.028738021850586, "learning_rate": 9.917869561227524e-05, "loss": 0.685, "step": 2564 }, { "epoch": 0.17220898627562833, "grad_norm": 1.0490139722824097, "learning_rate": 9.917673265768639e-05, "loss": 0.687, "step": 2566 }, { "epoch": 0.17234320995939734, "grad_norm": 1.2028472423553467, "learning_rate": 9.917476737959361e-05, "loss": 0.7658, "step": 2568 }, { "epoch": 0.17247743364316634, "grad_norm": 2.2009756565093994, "learning_rate": 9.917279977808976e-05, "loss": 0.7275, "step": 2570 }, { "epoch": 0.17261165732693534, "grad_norm": 1.2236677408218384, "learning_rate": 9.917082985326782e-05, "loss": 0.7573, "step": 2572 }, { "epoch": 0.17274588101070434, "grad_norm": 1.2253838777542114, "learning_rate": 9.916885760522087e-05, "loss": 0.7352, "step": 2574 }, { "epoch": 0.17288010469447335, "grad_norm": 1.1593698263168335, "learning_rate": 9.916688303404208e-05, "loss": 0.7484, "step": 2576 }, { "epoch": 0.17301432837824235, "grad_norm": 1.2147655487060547, "learning_rate": 9.916490613982474e-05, "loss": 0.7174, "step": 2578 }, { "epoch": 0.17314855206201135, "grad_norm": 1.321434736251831, "learning_rate": 9.916292692266228e-05, "loss": 0.7979, "step": 2580 }, { "epoch": 0.17328277574578035, "grad_norm": 1.3140137195587158, "learning_rate": 9.91609453826482e-05, "loss": 0.7674, "step": 2582 }, { "epoch": 0.17341699942954936, "grad_norm": 1.2361663579940796, "learning_rate": 9.915896151987609e-05, "loss": 0.704, "step": 2584 }, { "epoch": 0.17355122311331833, "grad_norm": 1.2271445989608765, "learning_rate": 9.915697533443976e-05, "loss": 0.7282, "step": 2586 }, { "epoch": 0.17368544679708733, "grad_norm": 1.3761489391326904, "learning_rate": 9.915498682643297e-05, "loss": 0.7634, "step": 2588 }, { "epoch": 0.17381967048085634, "grad_norm": 1.174750566482544, "learning_rate": 9.915299599594973e-05, "loss": 0.7094, "step": 2590 }, { "epoch": 0.17395389416462534, "grad_norm": 1.1923400163650513, "learning_rate": 9.915100284308407e-05, "loss": 0.7698, "step": 2592 }, { "epoch": 0.17408811784839434, "grad_norm": 1.2211850881576538, "learning_rate": 9.91490073679302e-05, "loss": 0.7688, "step": 2594 }, { "epoch": 0.17422234153216334, "grad_norm": 1.0700055360794067, "learning_rate": 9.914700957058235e-05, "loss": 0.6985, "step": 2596 }, { "epoch": 0.17435656521593235, "grad_norm": 1.142652153968811, "learning_rate": 9.914500945113496e-05, "loss": 0.7337, "step": 2598 }, { "epoch": 0.17449078889970135, "grad_norm": 1.1199309825897217, "learning_rate": 9.91430070096825e-05, "loss": 0.7095, "step": 2600 }, { "epoch": 0.17462501258347035, "grad_norm": 1.0610016584396362, "learning_rate": 9.914100224631962e-05, "loss": 0.7307, "step": 2602 }, { "epoch": 0.17475923626723935, "grad_norm": 1.1697499752044678, "learning_rate": 9.9138995161141e-05, "loss": 0.7198, "step": 2604 }, { "epoch": 0.17489345995100836, "grad_norm": 1.088790774345398, "learning_rate": 9.913698575424148e-05, "loss": 0.6844, "step": 2606 }, { "epoch": 0.17502768363477736, "grad_norm": 1.5284569263458252, "learning_rate": 9.913497402571602e-05, "loss": 0.7349, "step": 2608 }, { "epoch": 0.17516190731854636, "grad_norm": 1.1123801469802856, "learning_rate": 9.913295997565966e-05, "loss": 0.6767, "step": 2610 }, { "epoch": 0.17529613100231536, "grad_norm": 1.2049285173416138, "learning_rate": 9.913094360416755e-05, "loss": 0.7509, "step": 2612 }, { "epoch": 0.17543035468608437, "grad_norm": 1.16732656955719, "learning_rate": 9.912892491133496e-05, "loss": 0.7729, "step": 2614 }, { "epoch": 0.17556457836985337, "grad_norm": 1.1883516311645508, "learning_rate": 9.912690389725727e-05, "loss": 0.7081, "step": 2616 }, { "epoch": 0.17569880205362237, "grad_norm": 1.1990382671356201, "learning_rate": 9.912488056202998e-05, "loss": 0.7833, "step": 2618 }, { "epoch": 0.17583302573739137, "grad_norm": 1.6179962158203125, "learning_rate": 9.91228549057487e-05, "loss": 0.7501, "step": 2620 }, { "epoch": 0.17596724942116038, "grad_norm": 1.2496074438095093, "learning_rate": 9.91208269285091e-05, "loss": 0.8079, "step": 2622 }, { "epoch": 0.17610147310492935, "grad_norm": 2.569533109664917, "learning_rate": 9.911879663040704e-05, "loss": 0.7947, "step": 2624 }, { "epoch": 0.17623569678869835, "grad_norm": 1.2028528451919556, "learning_rate": 9.911676401153842e-05, "loss": 0.7544, "step": 2626 }, { "epoch": 0.17636992047246736, "grad_norm": 1.208221435546875, "learning_rate": 9.911472907199928e-05, "loss": 0.7546, "step": 2628 }, { "epoch": 0.17650414415623636, "grad_norm": 1.1322695016860962, "learning_rate": 9.911269181188575e-05, "loss": 0.7467, "step": 2630 }, { "epoch": 0.17663836784000536, "grad_norm": 1.445098638534546, "learning_rate": 9.911065223129414e-05, "loss": 0.7239, "step": 2632 }, { "epoch": 0.17677259152377436, "grad_norm": 1.2913486957550049, "learning_rate": 9.910861033032079e-05, "loss": 0.7957, "step": 2634 }, { "epoch": 0.17690681520754337, "grad_norm": 1.304382085800171, "learning_rate": 9.910656610906214e-05, "loss": 0.7491, "step": 2636 }, { "epoch": 0.17704103889131237, "grad_norm": 1.1376643180847168, "learning_rate": 9.910451956761482e-05, "loss": 0.7934, "step": 2638 }, { "epoch": 0.17717526257508137, "grad_norm": 1.0980056524276733, "learning_rate": 9.910247070607552e-05, "loss": 0.7526, "step": 2640 }, { "epoch": 0.17730948625885037, "grad_norm": 1.2822057008743286, "learning_rate": 9.910041952454103e-05, "loss": 0.7221, "step": 2642 }, { "epoch": 0.17744370994261938, "grad_norm": 1.14082932472229, "learning_rate": 9.909836602310824e-05, "loss": 0.7206, "step": 2644 }, { "epoch": 0.17757793362638838, "grad_norm": 1.1832576990127563, "learning_rate": 9.909631020187424e-05, "loss": 0.7794, "step": 2646 }, { "epoch": 0.17771215731015738, "grad_norm": 1.0556637048721313, "learning_rate": 9.909425206093613e-05, "loss": 0.7347, "step": 2648 }, { "epoch": 0.17784638099392638, "grad_norm": 1.192689299583435, "learning_rate": 9.909219160039112e-05, "loss": 0.7317, "step": 2650 }, { "epoch": 0.1779806046776954, "grad_norm": 1.3593887090682983, "learning_rate": 9.90901288203366e-05, "loss": 0.7381, "step": 2652 }, { "epoch": 0.1781148283614644, "grad_norm": 1.3047209978103638, "learning_rate": 9.908806372087002e-05, "loss": 0.7594, "step": 2654 }, { "epoch": 0.1782490520452334, "grad_norm": 1.2017946243286133, "learning_rate": 9.9085996302089e-05, "loss": 0.881, "step": 2656 }, { "epoch": 0.1783832757290024, "grad_norm": 1.2056033611297607, "learning_rate": 9.908392656409113e-05, "loss": 0.6891, "step": 2658 }, { "epoch": 0.1785174994127714, "grad_norm": 1.1817090511322021, "learning_rate": 9.908185450697428e-05, "loss": 0.7388, "step": 2660 }, { "epoch": 0.17865172309654037, "grad_norm": 1.1967507600784302, "learning_rate": 9.907978013083629e-05, "loss": 0.7806, "step": 2662 }, { "epoch": 0.17878594678030937, "grad_norm": 1.319650411605835, "learning_rate": 9.907770343577522e-05, "loss": 0.7853, "step": 2664 }, { "epoch": 0.17892017046407838, "grad_norm": 2.1139121055603027, "learning_rate": 9.907562442188916e-05, "loss": 0.7972, "step": 2666 }, { "epoch": 0.17905439414784738, "grad_norm": 1.6605976819992065, "learning_rate": 9.907354308927635e-05, "loss": 0.7991, "step": 2668 }, { "epoch": 0.17918861783161638, "grad_norm": 1.182556390762329, "learning_rate": 9.907145943803513e-05, "loss": 0.7995, "step": 2670 }, { "epoch": 0.17932284151538538, "grad_norm": 1.0889657735824585, "learning_rate": 9.906937346826395e-05, "loss": 0.7726, "step": 2672 }, { "epoch": 0.1794570651991544, "grad_norm": 1.0999469757080078, "learning_rate": 9.906728518006136e-05, "loss": 0.7414, "step": 2674 }, { "epoch": 0.1795912888829234, "grad_norm": 1.4655449390411377, "learning_rate": 9.906519457352605e-05, "loss": 0.7299, "step": 2676 }, { "epoch": 0.1797255125666924, "grad_norm": 1.482847809791565, "learning_rate": 9.906310164875676e-05, "loss": 0.7254, "step": 2678 }, { "epoch": 0.1798597362504614, "grad_norm": 1.191411018371582, "learning_rate": 9.90610064058524e-05, "loss": 0.6878, "step": 2680 }, { "epoch": 0.1799939599342304, "grad_norm": 1.2766669988632202, "learning_rate": 9.905890884491195e-05, "loss": 0.7835, "step": 2682 }, { "epoch": 0.1801281836179994, "grad_norm": 1.205926537513733, "learning_rate": 9.905680896603455e-05, "loss": 0.6985, "step": 2684 }, { "epoch": 0.1802624073017684, "grad_norm": 1.0893765687942505, "learning_rate": 9.905470676931938e-05, "loss": 0.7839, "step": 2686 }, { "epoch": 0.1803966309855374, "grad_norm": 42.618675231933594, "learning_rate": 9.905260225486577e-05, "loss": 0.7174, "step": 2688 }, { "epoch": 0.1805308546693064, "grad_norm": 1.4459474086761475, "learning_rate": 9.905049542277318e-05, "loss": 0.7727, "step": 2690 }, { "epoch": 0.1806650783530754, "grad_norm": 1.1787508726119995, "learning_rate": 9.904838627314112e-05, "loss": 0.7852, "step": 2692 }, { "epoch": 0.1807993020368444, "grad_norm": 1.2559529542922974, "learning_rate": 9.904627480606926e-05, "loss": 0.78, "step": 2694 }, { "epoch": 0.18093352572061341, "grad_norm": 1.2696396112442017, "learning_rate": 9.904416102165736e-05, "loss": 0.6922, "step": 2696 }, { "epoch": 0.18106774940438242, "grad_norm": 1.4660208225250244, "learning_rate": 9.90420449200053e-05, "loss": 0.7632, "step": 2698 }, { "epoch": 0.1812019730881514, "grad_norm": 1.0987337827682495, "learning_rate": 9.903992650121306e-05, "loss": 0.6834, "step": 2700 }, { "epoch": 0.1813361967719204, "grad_norm": 1.1610790491104126, "learning_rate": 9.903780576538071e-05, "loss": 0.7703, "step": 2702 }, { "epoch": 0.1814704204556894, "grad_norm": 1.284844160079956, "learning_rate": 9.903568271260847e-05, "loss": 0.8427, "step": 2704 }, { "epoch": 0.1816046441394584, "grad_norm": 1.110213041305542, "learning_rate": 9.903355734299664e-05, "loss": 0.6959, "step": 2706 }, { "epoch": 0.1817388678232274, "grad_norm": 1.186299443244934, "learning_rate": 9.903142965664566e-05, "loss": 0.7515, "step": 2708 }, { "epoch": 0.1818730915069964, "grad_norm": 1.2254539728164673, "learning_rate": 9.902929965365603e-05, "loss": 0.7767, "step": 2710 }, { "epoch": 0.1820073151907654, "grad_norm": 1.0632877349853516, "learning_rate": 9.902716733412842e-05, "loss": 0.7166, "step": 2712 }, { "epoch": 0.1821415388745344, "grad_norm": 1.3026443719863892, "learning_rate": 9.902503269816356e-05, "loss": 0.7088, "step": 2714 }, { "epoch": 0.1822757625583034, "grad_norm": 1.2889468669891357, "learning_rate": 9.90228957458623e-05, "loss": 0.7798, "step": 2716 }, { "epoch": 0.1824099862420724, "grad_norm": 1.083335041999817, "learning_rate": 9.902075647732563e-05, "loss": 0.7136, "step": 2718 }, { "epoch": 0.18254420992584142, "grad_norm": 1.224172830581665, "learning_rate": 9.90186148926546e-05, "loss": 0.7451, "step": 2720 }, { "epoch": 0.18267843360961042, "grad_norm": 1.1782457828521729, "learning_rate": 9.90164709919504e-05, "loss": 0.8336, "step": 2722 }, { "epoch": 0.18281265729337942, "grad_norm": 1.1387137174606323, "learning_rate": 9.901432477531433e-05, "loss": 0.7322, "step": 2724 }, { "epoch": 0.18294688097714842, "grad_norm": 1.1830766201019287, "learning_rate": 9.901217624284782e-05, "loss": 0.8164, "step": 2726 }, { "epoch": 0.18308110466091743, "grad_norm": 1.3433092832565308, "learning_rate": 9.901002539465234e-05, "loss": 0.8382, "step": 2728 }, { "epoch": 0.18321532834468643, "grad_norm": 1.133381962776184, "learning_rate": 9.900787223082955e-05, "loss": 0.779, "step": 2730 }, { "epoch": 0.18334955202845543, "grad_norm": 1.3183567523956299, "learning_rate": 9.900571675148116e-05, "loss": 0.701, "step": 2732 }, { "epoch": 0.18348377571222443, "grad_norm": 1.3738702535629272, "learning_rate": 9.9003558956709e-05, "loss": 0.7616, "step": 2734 }, { "epoch": 0.18361799939599344, "grad_norm": 1.0567593574523926, "learning_rate": 9.900139884661507e-05, "loss": 0.7768, "step": 2736 }, { "epoch": 0.1837522230797624, "grad_norm": 1.253952145576477, "learning_rate": 9.899923642130139e-05, "loss": 0.7426, "step": 2738 }, { "epoch": 0.1838864467635314, "grad_norm": 1.1834818124771118, "learning_rate": 9.899707168087013e-05, "loss": 0.7258, "step": 2740 }, { "epoch": 0.18402067044730042, "grad_norm": 1.1730256080627441, "learning_rate": 9.89949046254236e-05, "loss": 0.7649, "step": 2742 }, { "epoch": 0.18415489413106942, "grad_norm": 1.2688709497451782, "learning_rate": 9.899273525506417e-05, "loss": 0.7731, "step": 2744 }, { "epoch": 0.18428911781483842, "grad_norm": 0.988409698009491, "learning_rate": 9.899056356989434e-05, "loss": 0.6975, "step": 2746 }, { "epoch": 0.18442334149860742, "grad_norm": 1.1317416429519653, "learning_rate": 9.89883895700167e-05, "loss": 0.6988, "step": 2748 }, { "epoch": 0.18455756518237643, "grad_norm": 1.400775671005249, "learning_rate": 9.8986213255534e-05, "loss": 0.8486, "step": 2750 }, { "epoch": 0.18469178886614543, "grad_norm": 2.1011314392089844, "learning_rate": 9.898403462654904e-05, "loss": 0.8074, "step": 2752 }, { "epoch": 0.18482601254991443, "grad_norm": 1.2800923585891724, "learning_rate": 9.898185368316477e-05, "loss": 0.7832, "step": 2754 }, { "epoch": 0.18496023623368343, "grad_norm": 7.7373270988464355, "learning_rate": 9.897967042548424e-05, "loss": 0.7684, "step": 2756 }, { "epoch": 0.18509445991745244, "grad_norm": 1.1353752613067627, "learning_rate": 9.897748485361059e-05, "loss": 0.7304, "step": 2758 }, { "epoch": 0.18522868360122144, "grad_norm": 2.234422445297241, "learning_rate": 9.89752969676471e-05, "loss": 0.7119, "step": 2760 }, { "epoch": 0.18536290728499044, "grad_norm": 1.1421473026275635, "learning_rate": 9.897310676769712e-05, "loss": 0.8281, "step": 2762 }, { "epoch": 0.18549713096875944, "grad_norm": 1.2406892776489258, "learning_rate": 9.897091425386415e-05, "loss": 0.7532, "step": 2764 }, { "epoch": 0.18563135465252845, "grad_norm": 1.4110692739486694, "learning_rate": 9.896871942625179e-05, "loss": 0.7723, "step": 2766 }, { "epoch": 0.18576557833629745, "grad_norm": 1.292081356048584, "learning_rate": 9.896652228496372e-05, "loss": 0.7827, "step": 2768 }, { "epoch": 0.18589980202006645, "grad_norm": 1.1562920808792114, "learning_rate": 9.896432283010376e-05, "loss": 0.7404, "step": 2770 }, { "epoch": 0.18603402570383545, "grad_norm": 1.7341324090957642, "learning_rate": 9.896212106177583e-05, "loss": 0.7552, "step": 2772 }, { "epoch": 0.18616824938760446, "grad_norm": 1.6327306032180786, "learning_rate": 9.895991698008397e-05, "loss": 0.7682, "step": 2774 }, { "epoch": 0.18630247307137343, "grad_norm": 1.3916778564453125, "learning_rate": 9.89577105851323e-05, "loss": 0.7868, "step": 2776 }, { "epoch": 0.18643669675514243, "grad_norm": 1.4612019062042236, "learning_rate": 9.895550187702506e-05, "loss": 0.7867, "step": 2778 }, { "epoch": 0.18657092043891144, "grad_norm": 1.1375740766525269, "learning_rate": 9.895329085586667e-05, "loss": 0.7608, "step": 2780 }, { "epoch": 0.18670514412268044, "grad_norm": 1.4708541631698608, "learning_rate": 9.895107752176152e-05, "loss": 0.8019, "step": 2782 }, { "epoch": 0.18683936780644944, "grad_norm": 1.0651980638504028, "learning_rate": 9.894886187481421e-05, "loss": 0.7935, "step": 2784 }, { "epoch": 0.18697359149021844, "grad_norm": 1.7364146709442139, "learning_rate": 9.894664391512943e-05, "loss": 0.7939, "step": 2786 }, { "epoch": 0.18710781517398745, "grad_norm": 1.8116856813430786, "learning_rate": 9.894442364281197e-05, "loss": 0.7369, "step": 2788 }, { "epoch": 0.18724203885775645, "grad_norm": 1.2194713354110718, "learning_rate": 9.894220105796676e-05, "loss": 0.8202, "step": 2790 }, { "epoch": 0.18737626254152545, "grad_norm": 1.329644799232483, "learning_rate": 9.893997616069878e-05, "loss": 0.7858, "step": 2792 }, { "epoch": 0.18751048622529445, "grad_norm": 1.1791480779647827, "learning_rate": 9.893774895111317e-05, "loss": 0.7183, "step": 2794 }, { "epoch": 0.18764470990906346, "grad_norm": 1.3106368780136108, "learning_rate": 9.893551942931514e-05, "loss": 0.7934, "step": 2796 }, { "epoch": 0.18777893359283246, "grad_norm": 1.014682412147522, "learning_rate": 9.893328759541003e-05, "loss": 0.794, "step": 2798 }, { "epoch": 0.18791315727660146, "grad_norm": 1.1085726022720337, "learning_rate": 9.893105344950333e-05, "loss": 0.6797, "step": 2800 }, { "epoch": 0.18804738096037046, "grad_norm": 1.0696598291397095, "learning_rate": 9.892881699170058e-05, "loss": 0.6568, "step": 2802 }, { "epoch": 0.18818160464413947, "grad_norm": 1.4964417219161987, "learning_rate": 9.892657822210742e-05, "loss": 0.7017, "step": 2804 }, { "epoch": 0.18831582832790847, "grad_norm": 1.2987319231033325, "learning_rate": 9.892433714082966e-05, "loss": 0.7513, "step": 2806 }, { "epoch": 0.18845005201167747, "grad_norm": 1.3627650737762451, "learning_rate": 9.892209374797318e-05, "loss": 0.7948, "step": 2808 }, { "epoch": 0.18858427569544647, "grad_norm": 1.2367647886276245, "learning_rate": 9.891984804364395e-05, "loss": 0.7252, "step": 2810 }, { "epoch": 0.18871849937921548, "grad_norm": 1.2189453840255737, "learning_rate": 9.891760002794812e-05, "loss": 0.7563, "step": 2812 }, { "epoch": 0.18885272306298445, "grad_norm": 1.0545687675476074, "learning_rate": 9.891534970099188e-05, "loss": 0.7729, "step": 2814 }, { "epoch": 0.18898694674675345, "grad_norm": 1.0687874555587769, "learning_rate": 9.891309706288154e-05, "loss": 0.7083, "step": 2816 }, { "epoch": 0.18912117043052246, "grad_norm": 1.707055687904358, "learning_rate": 9.891084211372356e-05, "loss": 0.8, "step": 2818 }, { "epoch": 0.18925539411429146, "grad_norm": 1.1517010927200317, "learning_rate": 9.890858485362447e-05, "loss": 0.7723, "step": 2820 }, { "epoch": 0.18938961779806046, "grad_norm": 1.7331767082214355, "learning_rate": 9.89063252826909e-05, "loss": 0.7459, "step": 2822 }, { "epoch": 0.18952384148182946, "grad_norm": 1.2639738321304321, "learning_rate": 9.890406340102964e-05, "loss": 0.7863, "step": 2824 }, { "epoch": 0.18965806516559847, "grad_norm": 1.186524748802185, "learning_rate": 9.890179920874756e-05, "loss": 0.7206, "step": 2826 }, { "epoch": 0.18979228884936747, "grad_norm": 1.2795987129211426, "learning_rate": 9.889953270595162e-05, "loss": 0.751, "step": 2828 }, { "epoch": 0.18992651253313647, "grad_norm": 1.4070851802825928, "learning_rate": 9.889726389274892e-05, "loss": 0.7062, "step": 2830 }, { "epoch": 0.19006073621690547, "grad_norm": 2.162493944168091, "learning_rate": 9.889499276924666e-05, "loss": 0.7934, "step": 2832 }, { "epoch": 0.19019495990067448, "grad_norm": 1.2679576873779297, "learning_rate": 9.889271933555213e-05, "loss": 0.793, "step": 2834 }, { "epoch": 0.19032918358444348, "grad_norm": 1.405962586402893, "learning_rate": 9.889044359177277e-05, "loss": 0.7238, "step": 2836 }, { "epoch": 0.19046340726821248, "grad_norm": 1.2353532314300537, "learning_rate": 9.888816553801608e-05, "loss": 0.7139, "step": 2838 }, { "epoch": 0.19059763095198148, "grad_norm": 1.1028170585632324, "learning_rate": 9.888588517438968e-05, "loss": 0.7244, "step": 2840 }, { "epoch": 0.1907318546357505, "grad_norm": 1.1188403367996216, "learning_rate": 9.888360250100137e-05, "loss": 0.7216, "step": 2842 }, { "epoch": 0.1908660783195195, "grad_norm": 1.157393455505371, "learning_rate": 9.888131751795895e-05, "loss": 0.7636, "step": 2844 }, { "epoch": 0.1910003020032885, "grad_norm": 1.4509886503219604, "learning_rate": 9.88790302253704e-05, "loss": 0.7643, "step": 2846 }, { "epoch": 0.1911345256870575, "grad_norm": 1.1343168020248413, "learning_rate": 9.887674062334377e-05, "loss": 0.8027, "step": 2848 }, { "epoch": 0.19126874937082647, "grad_norm": 1.3045190572738647, "learning_rate": 9.88744487119873e-05, "loss": 0.7703, "step": 2850 }, { "epoch": 0.19140297305459547, "grad_norm": 1.1510552167892456, "learning_rate": 9.88721544914092e-05, "loss": 0.7222, "step": 2852 }, { "epoch": 0.19153719673836447, "grad_norm": 1.2157753705978394, "learning_rate": 9.886985796171792e-05, "loss": 0.7933, "step": 2854 }, { "epoch": 0.19167142042213348, "grad_norm": 1.9779458045959473, "learning_rate": 9.886755912302194e-05, "loss": 0.68, "step": 2856 }, { "epoch": 0.19180564410590248, "grad_norm": 1.150827169418335, "learning_rate": 9.886525797542989e-05, "loss": 0.761, "step": 2858 }, { "epoch": 0.19193986778967148, "grad_norm": 1.3666388988494873, "learning_rate": 9.88629545190505e-05, "loss": 0.784, "step": 2860 }, { "epoch": 0.19207409147344048, "grad_norm": 1.2915219068527222, "learning_rate": 9.88606487539926e-05, "loss": 0.7966, "step": 2862 }, { "epoch": 0.1922083151572095, "grad_norm": 1.4001399278640747, "learning_rate": 9.88583406803651e-05, "loss": 0.7708, "step": 2864 }, { "epoch": 0.1923425388409785, "grad_norm": 1.7778263092041016, "learning_rate": 9.885603029827707e-05, "loss": 0.7163, "step": 2866 }, { "epoch": 0.1924767625247475, "grad_norm": 1.3420580625534058, "learning_rate": 9.885371760783772e-05, "loss": 0.7487, "step": 2868 }, { "epoch": 0.1926109862085165, "grad_norm": 1.06627357006073, "learning_rate": 9.885140260915625e-05, "loss": 0.7426, "step": 2870 }, { "epoch": 0.1927452098922855, "grad_norm": 1.1110459566116333, "learning_rate": 9.884908530234208e-05, "loss": 0.7587, "step": 2872 }, { "epoch": 0.1928794335760545, "grad_norm": 0.9999328851699829, "learning_rate": 9.884676568750469e-05, "loss": 0.7246, "step": 2874 }, { "epoch": 0.1930136572598235, "grad_norm": 1.1353189945220947, "learning_rate": 9.884444376475367e-05, "loss": 0.7546, "step": 2876 }, { "epoch": 0.1931478809435925, "grad_norm": 1.070284128189087, "learning_rate": 9.884211953419873e-05, "loss": 0.6763, "step": 2878 }, { "epoch": 0.1932821046273615, "grad_norm": 1.1456369161605835, "learning_rate": 9.883979299594969e-05, "loss": 0.7325, "step": 2880 }, { "epoch": 0.1934163283111305, "grad_norm": 1.1208562850952148, "learning_rate": 9.883746415011646e-05, "loss": 0.7168, "step": 2882 }, { "epoch": 0.1935505519948995, "grad_norm": 1.173606276512146, "learning_rate": 9.88351329968091e-05, "loss": 0.7472, "step": 2884 }, { "epoch": 0.1936847756786685, "grad_norm": 1.1535316705703735, "learning_rate": 9.883279953613771e-05, "loss": 0.7096, "step": 2886 }, { "epoch": 0.1938189993624375, "grad_norm": 1.0993741750717163, "learning_rate": 9.88304637682126e-05, "loss": 0.7317, "step": 2888 }, { "epoch": 0.1939532230462065, "grad_norm": 1.6121855974197388, "learning_rate": 9.882812569314408e-05, "loss": 0.6954, "step": 2890 }, { "epoch": 0.1940874467299755, "grad_norm": 1.5288883447647095, "learning_rate": 9.882578531104263e-05, "loss": 0.772, "step": 2892 }, { "epoch": 0.1942216704137445, "grad_norm": 0.9637719988822937, "learning_rate": 9.882344262201884e-05, "loss": 0.72, "step": 2894 }, { "epoch": 0.1943558940975135, "grad_norm": 1.0743874311447144, "learning_rate": 9.882109762618342e-05, "loss": 0.725, "step": 2896 }, { "epoch": 0.1944901177812825, "grad_norm": 1.1397347450256348, "learning_rate": 9.88187503236471e-05, "loss": 0.7995, "step": 2898 }, { "epoch": 0.1946243414650515, "grad_norm": 1.1381994485855103, "learning_rate": 9.881640071452085e-05, "loss": 0.7373, "step": 2900 }, { "epoch": 0.1947585651488205, "grad_norm": 2.827315330505371, "learning_rate": 9.881404879891565e-05, "loss": 0.7867, "step": 2902 }, { "epoch": 0.1948927888325895, "grad_norm": 1.2178807258605957, "learning_rate": 9.881169457694263e-05, "loss": 0.7632, "step": 2904 }, { "epoch": 0.1950270125163585, "grad_norm": 1.188071846961975, "learning_rate": 9.880933804871304e-05, "loss": 0.7936, "step": 2906 }, { "epoch": 0.1951612362001275, "grad_norm": 1.5488351583480835, "learning_rate": 9.88069792143382e-05, "loss": 0.7501, "step": 2908 }, { "epoch": 0.19529545988389652, "grad_norm": 1.2302963733673096, "learning_rate": 9.880461807392956e-05, "loss": 0.7269, "step": 2910 }, { "epoch": 0.19542968356766552, "grad_norm": 1.212137222290039, "learning_rate": 9.880225462759869e-05, "loss": 0.7771, "step": 2912 }, { "epoch": 0.19556390725143452, "grad_norm": 1.40646493434906, "learning_rate": 9.879988887545726e-05, "loss": 0.7424, "step": 2914 }, { "epoch": 0.19569813093520352, "grad_norm": 1.6506620645523071, "learning_rate": 9.879752081761704e-05, "loss": 0.719, "step": 2916 }, { "epoch": 0.19583235461897253, "grad_norm": 1.1950587034225464, "learning_rate": 9.87951504541899e-05, "loss": 0.7966, "step": 2918 }, { "epoch": 0.19596657830274153, "grad_norm": 1.2168084383010864, "learning_rate": 9.879277778528788e-05, "loss": 0.7682, "step": 2920 }, { "epoch": 0.19610080198651053, "grad_norm": 1.1859941482543945, "learning_rate": 9.879040281102303e-05, "loss": 0.7092, "step": 2922 }, { "epoch": 0.19623502567027953, "grad_norm": 1.2239441871643066, "learning_rate": 9.878802553150762e-05, "loss": 0.7683, "step": 2924 }, { "epoch": 0.1963692493540485, "grad_norm": 1.762573480606079, "learning_rate": 9.87856459468539e-05, "loss": 0.7422, "step": 2926 }, { "epoch": 0.1965034730378175, "grad_norm": 1.6585900783538818, "learning_rate": 9.878326405717438e-05, "loss": 0.7766, "step": 2928 }, { "epoch": 0.1966376967215865, "grad_norm": 1.1981463432312012, "learning_rate": 9.878087986258156e-05, "loss": 0.7542, "step": 2930 }, { "epoch": 0.19677192040535552, "grad_norm": 1.0024964809417725, "learning_rate": 9.877849336318807e-05, "loss": 0.7245, "step": 2932 }, { "epoch": 0.19690614408912452, "grad_norm": 1.1262283325195312, "learning_rate": 9.877610455910668e-05, "loss": 0.7791, "step": 2934 }, { "epoch": 0.19704036777289352, "grad_norm": 1.1668745279312134, "learning_rate": 9.877371345045029e-05, "loss": 0.7428, "step": 2936 }, { "epoch": 0.19717459145666252, "grad_norm": 1.3850470781326294, "learning_rate": 9.877132003733182e-05, "loss": 0.6995, "step": 2938 }, { "epoch": 0.19730881514043153, "grad_norm": 1.1988468170166016, "learning_rate": 9.876892431986442e-05, "loss": 0.7714, "step": 2940 }, { "epoch": 0.19744303882420053, "grad_norm": 1.1874380111694336, "learning_rate": 9.876652629816122e-05, "loss": 0.7281, "step": 2942 }, { "epoch": 0.19757726250796953, "grad_norm": 1.0807650089263916, "learning_rate": 9.876412597233555e-05, "loss": 0.8305, "step": 2944 }, { "epoch": 0.19771148619173853, "grad_norm": 1.122131586074829, "learning_rate": 9.876172334250082e-05, "loss": 0.7169, "step": 2946 }, { "epoch": 0.19784570987550754, "grad_norm": 1.265494704246521, "learning_rate": 9.875931840877055e-05, "loss": 0.7753, "step": 2948 }, { "epoch": 0.19797993355927654, "grad_norm": 1.19541597366333, "learning_rate": 9.875691117125837e-05, "loss": 0.7584, "step": 2950 }, { "epoch": 0.19811415724304554, "grad_norm": 1.149675965309143, "learning_rate": 9.875450163007801e-05, "loss": 0.7827, "step": 2952 }, { "epoch": 0.19824838092681454, "grad_norm": 1.1761040687561035, "learning_rate": 9.875208978534331e-05, "loss": 0.6823, "step": 2954 }, { "epoch": 0.19838260461058355, "grad_norm": 1.0206571817398071, "learning_rate": 9.874967563716826e-05, "loss": 0.6873, "step": 2956 }, { "epoch": 0.19851682829435255, "grad_norm": 1.573801040649414, "learning_rate": 9.874725918566687e-05, "loss": 0.7422, "step": 2958 }, { "epoch": 0.19865105197812155, "grad_norm": 3.296018123626709, "learning_rate": 9.874484043095336e-05, "loss": 0.7217, "step": 2960 }, { "epoch": 0.19878527566189055, "grad_norm": 2.1112844944000244, "learning_rate": 9.874241937314199e-05, "loss": 0.6715, "step": 2962 }, { "epoch": 0.19891949934565953, "grad_norm": 1.3631985187530518, "learning_rate": 9.873999601234715e-05, "loss": 0.7548, "step": 2964 }, { "epoch": 0.19905372302942853, "grad_norm": 1.7285358905792236, "learning_rate": 9.873757034868333e-05, "loss": 0.7102, "step": 2966 }, { "epoch": 0.19918794671319753, "grad_norm": 1.4606434106826782, "learning_rate": 9.873514238226515e-05, "loss": 0.7486, "step": 2968 }, { "epoch": 0.19932217039696654, "grad_norm": 1.2391377687454224, "learning_rate": 9.873271211320735e-05, "loss": 0.7302, "step": 2970 }, { "epoch": 0.19945639408073554, "grad_norm": 1.065616250038147, "learning_rate": 9.873027954162471e-05, "loss": 0.7424, "step": 2972 }, { "epoch": 0.19959061776450454, "grad_norm": 1.0928055047988892, "learning_rate": 9.87278446676322e-05, "loss": 0.7631, "step": 2974 }, { "epoch": 0.19972484144827354, "grad_norm": 1.2849832773208618, "learning_rate": 9.872540749134484e-05, "loss": 0.8011, "step": 2976 }, { "epoch": 0.19985906513204255, "grad_norm": 1.105771780014038, "learning_rate": 9.872296801287779e-05, "loss": 0.8108, "step": 2978 }, { "epoch": 0.19999328881581155, "grad_norm": 1.8333889245986938, "learning_rate": 9.872052623234632e-05, "loss": 0.7429, "step": 2980 }, { "epoch": 0.20012751249958055, "grad_norm": 1.1117559671401978, "learning_rate": 9.871808214986578e-05, "loss": 0.6958, "step": 2982 }, { "epoch": 0.20026173618334955, "grad_norm": 1.1552224159240723, "learning_rate": 9.871563576555165e-05, "loss": 0.7331, "step": 2984 }, { "epoch": 0.20039595986711856, "grad_norm": 1.2907450199127197, "learning_rate": 9.871318707951953e-05, "loss": 0.7443, "step": 2986 }, { "epoch": 0.20053018355088756, "grad_norm": 1.1583023071289062, "learning_rate": 9.871073609188513e-05, "loss": 0.8166, "step": 2988 }, { "epoch": 0.20066440723465656, "grad_norm": 1.2795884609222412, "learning_rate": 9.87082828027642e-05, "loss": 0.7008, "step": 2990 }, { "epoch": 0.20079863091842556, "grad_norm": 1.3416017293930054, "learning_rate": 9.870582721227273e-05, "loss": 0.822, "step": 2992 }, { "epoch": 0.20093285460219457, "grad_norm": 1.2741094827651978, "learning_rate": 9.870336932052667e-05, "loss": 0.7139, "step": 2994 }, { "epoch": 0.20106707828596357, "grad_norm": 1.5517258644104004, "learning_rate": 9.87009091276422e-05, "loss": 0.7441, "step": 2996 }, { "epoch": 0.20120130196973257, "grad_norm": 1.0434761047363281, "learning_rate": 9.869844663373553e-05, "loss": 0.7364, "step": 2998 }, { "epoch": 0.20133552565350157, "grad_norm": 1.3809441328048706, "learning_rate": 9.8695981838923e-05, "loss": 0.7398, "step": 3000 }, { "epoch": 0.20146974933727055, "grad_norm": 1.060219168663025, "learning_rate": 9.869351474332111e-05, "loss": 0.7235, "step": 3002 }, { "epoch": 0.20160397302103955, "grad_norm": 1.6353628635406494, "learning_rate": 9.869104534704641e-05, "loss": 0.7376, "step": 3004 }, { "epoch": 0.20173819670480855, "grad_norm": 1.215761661529541, "learning_rate": 9.868857365021553e-05, "loss": 0.8069, "step": 3006 }, { "epoch": 0.20187242038857756, "grad_norm": 1.2343037128448486, "learning_rate": 9.86860996529453e-05, "loss": 0.72, "step": 3008 }, { "epoch": 0.20200664407234656, "grad_norm": 1.1783651113510132, "learning_rate": 9.868362335535262e-05, "loss": 0.7734, "step": 3010 }, { "epoch": 0.20214086775611556, "grad_norm": 1.0304306745529175, "learning_rate": 9.868114475755445e-05, "loss": 0.6904, "step": 3012 }, { "epoch": 0.20227509143988456, "grad_norm": 1.1342130899429321, "learning_rate": 9.867866385966791e-05, "loss": 0.7364, "step": 3014 }, { "epoch": 0.20240931512365357, "grad_norm": 1.1649483442306519, "learning_rate": 9.867618066181023e-05, "loss": 0.7879, "step": 3016 }, { "epoch": 0.20254353880742257, "grad_norm": 1.2916783094406128, "learning_rate": 9.867369516409874e-05, "loss": 0.738, "step": 3018 }, { "epoch": 0.20267776249119157, "grad_norm": 1.258411169052124, "learning_rate": 9.867120736665087e-05, "loss": 0.7294, "step": 3020 }, { "epoch": 0.20281198617496057, "grad_norm": 1.307192087173462, "learning_rate": 9.866871726958415e-05, "loss": 0.8615, "step": 3022 }, { "epoch": 0.20294620985872958, "grad_norm": 1.1338318586349487, "learning_rate": 9.866622487301624e-05, "loss": 0.6777, "step": 3024 }, { "epoch": 0.20308043354249858, "grad_norm": 1.2602858543395996, "learning_rate": 9.866373017706492e-05, "loss": 0.7936, "step": 3026 }, { "epoch": 0.20321465722626758, "grad_norm": 0.9843763113021851, "learning_rate": 9.866123318184803e-05, "loss": 0.6511, "step": 3028 }, { "epoch": 0.20334888091003658, "grad_norm": 1.1347159147262573, "learning_rate": 9.865873388748354e-05, "loss": 0.7671, "step": 3030 }, { "epoch": 0.20348310459380559, "grad_norm": 1.4263908863067627, "learning_rate": 9.865623229408959e-05, "loss": 0.7339, "step": 3032 }, { "epoch": 0.2036173282775746, "grad_norm": 1.1282978057861328, "learning_rate": 9.865372840178433e-05, "loss": 0.7729, "step": 3034 }, { "epoch": 0.2037515519613436, "grad_norm": 1.1372108459472656, "learning_rate": 9.865122221068608e-05, "loss": 0.7619, "step": 3036 }, { "epoch": 0.2038857756451126, "grad_norm": 1.1885900497436523, "learning_rate": 9.864871372091324e-05, "loss": 0.7027, "step": 3038 }, { "epoch": 0.20401999932888157, "grad_norm": 1.1478683948516846, "learning_rate": 9.864620293258434e-05, "loss": 0.7334, "step": 3040 }, { "epoch": 0.20415422301265057, "grad_norm": 1.4754353761672974, "learning_rate": 9.864368984581803e-05, "loss": 0.6868, "step": 3042 }, { "epoch": 0.20428844669641957, "grad_norm": 1.1996697187423706, "learning_rate": 9.8641174460733e-05, "loss": 0.7198, "step": 3044 }, { "epoch": 0.20442267038018858, "grad_norm": 1.235029697418213, "learning_rate": 9.863865677744814e-05, "loss": 0.7288, "step": 3046 }, { "epoch": 0.20455689406395758, "grad_norm": 1.098764419555664, "learning_rate": 9.863613679608239e-05, "loss": 0.6877, "step": 3048 }, { "epoch": 0.20469111774772658, "grad_norm": 1.268288254737854, "learning_rate": 9.863361451675481e-05, "loss": 0.817, "step": 3050 }, { "epoch": 0.20482534143149558, "grad_norm": 1.1356425285339355, "learning_rate": 9.863108993958459e-05, "loss": 0.7465, "step": 3052 }, { "epoch": 0.20495956511526459, "grad_norm": 1.187319278717041, "learning_rate": 9.862856306469099e-05, "loss": 0.7622, "step": 3054 }, { "epoch": 0.2050937887990336, "grad_norm": 1.3413240909576416, "learning_rate": 9.86260338921934e-05, "loss": 0.6907, "step": 3056 }, { "epoch": 0.2052280124828026, "grad_norm": 1.3412761688232422, "learning_rate": 9.862350242221135e-05, "loss": 0.7947, "step": 3058 }, { "epoch": 0.2053622361665716, "grad_norm": 1.2627809047698975, "learning_rate": 9.862096865486441e-05, "loss": 0.7168, "step": 3060 }, { "epoch": 0.2054964598503406, "grad_norm": 1.1120837926864624, "learning_rate": 9.861843259027233e-05, "loss": 0.7121, "step": 3062 }, { "epoch": 0.2056306835341096, "grad_norm": 1.2272433042526245, "learning_rate": 9.861589422855488e-05, "loss": 0.7741, "step": 3064 }, { "epoch": 0.2057649072178786, "grad_norm": 2.011345863342285, "learning_rate": 9.861335356983206e-05, "loss": 0.7038, "step": 3066 }, { "epoch": 0.2058991309016476, "grad_norm": 1.1610370874404907, "learning_rate": 9.861081061422386e-05, "loss": 0.7797, "step": 3068 }, { "epoch": 0.2060333545854166, "grad_norm": 1.5199135541915894, "learning_rate": 9.860826536185044e-05, "loss": 0.7256, "step": 3070 }, { "epoch": 0.2061675782691856, "grad_norm": 0.9823091626167297, "learning_rate": 9.860571781283208e-05, "loss": 0.7143, "step": 3072 }, { "epoch": 0.2063018019529546, "grad_norm": 1.1092530488967896, "learning_rate": 9.860316796728912e-05, "loss": 0.7797, "step": 3074 }, { "epoch": 0.2064360256367236, "grad_norm": 1.133023977279663, "learning_rate": 9.860061582534205e-05, "loss": 0.7105, "step": 3076 }, { "epoch": 0.2065702493204926, "grad_norm": 1.3843505382537842, "learning_rate": 9.859806138711148e-05, "loss": 0.7875, "step": 3078 }, { "epoch": 0.2067044730042616, "grad_norm": 1.0685838460922241, "learning_rate": 9.859550465271804e-05, "loss": 0.6607, "step": 3080 }, { "epoch": 0.2068386966880306, "grad_norm": 1.1929254531860352, "learning_rate": 9.859294562228258e-05, "loss": 0.8085, "step": 3082 }, { "epoch": 0.2069729203717996, "grad_norm": 2.1357104778289795, "learning_rate": 9.859038429592599e-05, "loss": 0.7222, "step": 3084 }, { "epoch": 0.2071071440555686, "grad_norm": 1.1408915519714355, "learning_rate": 9.858782067376928e-05, "loss": 0.6628, "step": 3086 }, { "epoch": 0.2072413677393376, "grad_norm": 1.2059365510940552, "learning_rate": 9.85852547559336e-05, "loss": 0.8336, "step": 3088 }, { "epoch": 0.2073755914231066, "grad_norm": 1.0125360488891602, "learning_rate": 9.858268654254017e-05, "loss": 0.7075, "step": 3090 }, { "epoch": 0.2075098151068756, "grad_norm": 1.3566681146621704, "learning_rate": 9.858011603371033e-05, "loss": 0.7074, "step": 3092 }, { "epoch": 0.2076440387906446, "grad_norm": 1.080464482307434, "learning_rate": 9.857754322956554e-05, "loss": 0.696, "step": 3094 }, { "epoch": 0.2077782624744136, "grad_norm": 1.0486165285110474, "learning_rate": 9.857496813022735e-05, "loss": 0.7453, "step": 3096 }, { "epoch": 0.2079124861581826, "grad_norm": 1.0846431255340576, "learning_rate": 9.857239073581743e-05, "loss": 0.7727, "step": 3098 }, { "epoch": 0.20804670984195162, "grad_norm": 1.1815550327301025, "learning_rate": 9.856981104645757e-05, "loss": 0.7271, "step": 3100 }, { "epoch": 0.20818093352572062, "grad_norm": 1.200812578201294, "learning_rate": 9.856722906226965e-05, "loss": 0.7766, "step": 3102 }, { "epoch": 0.20831515720948962, "grad_norm": 1.3870935440063477, "learning_rate": 9.856464478337566e-05, "loss": 0.7684, "step": 3104 }, { "epoch": 0.20844938089325862, "grad_norm": 2.0267977714538574, "learning_rate": 9.85620582098977e-05, "loss": 0.727, "step": 3106 }, { "epoch": 0.20858360457702763, "grad_norm": 1.2615206241607666, "learning_rate": 9.855946934195799e-05, "loss": 0.7919, "step": 3108 }, { "epoch": 0.20871782826079663, "grad_norm": 1.1016186475753784, "learning_rate": 9.855687817967882e-05, "loss": 0.7548, "step": 3110 }, { "epoch": 0.20885205194456563, "grad_norm": 0.9793462157249451, "learning_rate": 9.855428472318267e-05, "loss": 0.7518, "step": 3112 }, { "epoch": 0.20898627562833463, "grad_norm": 1.139752745628357, "learning_rate": 9.855168897259202e-05, "loss": 0.7284, "step": 3114 }, { "epoch": 0.2091204993121036, "grad_norm": 1.2097362279891968, "learning_rate": 9.854909092802955e-05, "loss": 0.7188, "step": 3116 }, { "epoch": 0.2092547229958726, "grad_norm": 1.1110161542892456, "learning_rate": 9.854649058961799e-05, "loss": 0.7279, "step": 3118 }, { "epoch": 0.2093889466796416, "grad_norm": 1.1635180711746216, "learning_rate": 9.854388795748022e-05, "loss": 0.7384, "step": 3120 }, { "epoch": 0.20952317036341062, "grad_norm": 1.0395607948303223, "learning_rate": 9.854128303173919e-05, "loss": 0.6984, "step": 3122 }, { "epoch": 0.20965739404717962, "grad_norm": 1.2108814716339111, "learning_rate": 9.8538675812518e-05, "loss": 0.7615, "step": 3124 }, { "epoch": 0.20979161773094862, "grad_norm": 1.134364128112793, "learning_rate": 9.853606629993983e-05, "loss": 0.7576, "step": 3126 }, { "epoch": 0.20992584141471762, "grad_norm": 1.1714712381362915, "learning_rate": 9.853345449412796e-05, "loss": 0.6579, "step": 3128 }, { "epoch": 0.21006006509848663, "grad_norm": 1.1478958129882812, "learning_rate": 9.853084039520581e-05, "loss": 0.7723, "step": 3130 }, { "epoch": 0.21019428878225563, "grad_norm": 1.2991362810134888, "learning_rate": 9.852822400329688e-05, "loss": 0.7406, "step": 3132 }, { "epoch": 0.21032851246602463, "grad_norm": 1.1406856775283813, "learning_rate": 9.852560531852479e-05, "loss": 0.7334, "step": 3134 }, { "epoch": 0.21046273614979363, "grad_norm": 1.0486959218978882, "learning_rate": 9.852298434101328e-05, "loss": 0.7012, "step": 3136 }, { "epoch": 0.21059695983356264, "grad_norm": 1.4187973737716675, "learning_rate": 9.852036107088617e-05, "loss": 0.7311, "step": 3138 }, { "epoch": 0.21073118351733164, "grad_norm": 1.5801842212677002, "learning_rate": 9.851773550826742e-05, "loss": 0.7044, "step": 3140 }, { "epoch": 0.21086540720110064, "grad_norm": 1.2946140766143799, "learning_rate": 9.851510765328105e-05, "loss": 0.7428, "step": 3142 }, { "epoch": 0.21099963088486964, "grad_norm": 1.1796568632125854, "learning_rate": 9.851247750605126e-05, "loss": 0.7211, "step": 3144 }, { "epoch": 0.21113385456863865, "grad_norm": 1.0621049404144287, "learning_rate": 9.85098450667023e-05, "loss": 0.7527, "step": 3146 }, { "epoch": 0.21126807825240765, "grad_norm": 1.4693660736083984, "learning_rate": 9.850721033535854e-05, "loss": 0.759, "step": 3148 }, { "epoch": 0.21140230193617665, "grad_norm": 1.233821153640747, "learning_rate": 9.85045733121445e-05, "loss": 0.7018, "step": 3150 }, { "epoch": 0.21153652561994563, "grad_norm": 1.652405023574829, "learning_rate": 9.850193399718475e-05, "loss": 0.7492, "step": 3152 }, { "epoch": 0.21167074930371463, "grad_norm": 1.35104501247406, "learning_rate": 9.849929239060398e-05, "loss": 0.7357, "step": 3154 }, { "epoch": 0.21180497298748363, "grad_norm": 1.1674814224243164, "learning_rate": 9.849664849252701e-05, "loss": 0.7184, "step": 3156 }, { "epoch": 0.21193919667125263, "grad_norm": 1.2722527980804443, "learning_rate": 9.849400230307877e-05, "loss": 0.7412, "step": 3158 }, { "epoch": 0.21207342035502164, "grad_norm": 1.0681673288345337, "learning_rate": 9.849135382238428e-05, "loss": 0.7271, "step": 3160 }, { "epoch": 0.21220764403879064, "grad_norm": 1.156601071357727, "learning_rate": 9.848870305056867e-05, "loss": 0.7621, "step": 3162 }, { "epoch": 0.21234186772255964, "grad_norm": 1.1709725856781006, "learning_rate": 9.848604998775721e-05, "loss": 0.7921, "step": 3164 }, { "epoch": 0.21247609140632864, "grad_norm": 1.0551742315292358, "learning_rate": 9.848339463407521e-05, "loss": 0.6156, "step": 3166 }, { "epoch": 0.21261031509009765, "grad_norm": 1.2272999286651611, "learning_rate": 9.848073698964817e-05, "loss": 0.6912, "step": 3168 }, { "epoch": 0.21274453877386665, "grad_norm": 1.103464126586914, "learning_rate": 9.847807705460163e-05, "loss": 0.7523, "step": 3170 }, { "epoch": 0.21287876245763565, "grad_norm": 1.215071439743042, "learning_rate": 9.847541482906129e-05, "loss": 0.7767, "step": 3172 }, { "epoch": 0.21301298614140465, "grad_norm": 1.2491865158081055, "learning_rate": 9.84727503131529e-05, "loss": 0.7524, "step": 3174 }, { "epoch": 0.21314720982517366, "grad_norm": 1.4180876016616821, "learning_rate": 9.847008350700239e-05, "loss": 0.767, "step": 3176 }, { "epoch": 0.21328143350894266, "grad_norm": 1.2222609519958496, "learning_rate": 9.846741441073574e-05, "loss": 0.7615, "step": 3178 }, { "epoch": 0.21341565719271166, "grad_norm": 1.1580231189727783, "learning_rate": 9.846474302447907e-05, "loss": 0.7834, "step": 3180 }, { "epoch": 0.21354988087648066, "grad_norm": 1.3348591327667236, "learning_rate": 9.846206934835859e-05, "loss": 0.7133, "step": 3182 }, { "epoch": 0.21368410456024967, "grad_norm": 1.0967212915420532, "learning_rate": 9.845939338250063e-05, "loss": 0.707, "step": 3184 }, { "epoch": 0.21381832824401867, "grad_norm": 1.2718641757965088, "learning_rate": 9.845671512703163e-05, "loss": 0.7405, "step": 3186 }, { "epoch": 0.21395255192778767, "grad_norm": 1.4138965606689453, "learning_rate": 9.845403458207813e-05, "loss": 0.7712, "step": 3188 }, { "epoch": 0.21408677561155665, "grad_norm": 6.629738807678223, "learning_rate": 9.845135174776676e-05, "loss": 0.7772, "step": 3190 }, { "epoch": 0.21422099929532565, "grad_norm": 1.3174983263015747, "learning_rate": 9.844866662422432e-05, "loss": 0.7585, "step": 3192 }, { "epoch": 0.21435522297909465, "grad_norm": 1.3816879987716675, "learning_rate": 9.844597921157764e-05, "loss": 0.7098, "step": 3194 }, { "epoch": 0.21448944666286365, "grad_norm": 1.203576683998108, "learning_rate": 9.84432895099537e-05, "loss": 0.8389, "step": 3196 }, { "epoch": 0.21462367034663266, "grad_norm": 1.1303554773330688, "learning_rate": 9.844059751947959e-05, "loss": 0.7791, "step": 3198 }, { "epoch": 0.21475789403040166, "grad_norm": 1.2763458490371704, "learning_rate": 9.84379032402825e-05, "loss": 0.6964, "step": 3200 }, { "epoch": 0.21489211771417066, "grad_norm": 1.4167712926864624, "learning_rate": 9.843520667248974e-05, "loss": 0.6939, "step": 3202 }, { "epoch": 0.21502634139793966, "grad_norm": 1.0749961137771606, "learning_rate": 9.84325078162287e-05, "loss": 0.7213, "step": 3204 }, { "epoch": 0.21516056508170867, "grad_norm": 1.409028172492981, "learning_rate": 9.84298066716269e-05, "loss": 0.707, "step": 3206 }, { "epoch": 0.21529478876547767, "grad_norm": 2.0518510341644287, "learning_rate": 9.842710323881199e-05, "loss": 0.7577, "step": 3208 }, { "epoch": 0.21542901244924667, "grad_norm": 0.9805545806884766, "learning_rate": 9.842439751791169e-05, "loss": 0.7061, "step": 3210 }, { "epoch": 0.21556323613301567, "grad_norm": 1.0923629999160767, "learning_rate": 9.842168950905379e-05, "loss": 0.6747, "step": 3212 }, { "epoch": 0.21569745981678468, "grad_norm": 1.0803589820861816, "learning_rate": 9.84189792123663e-05, "loss": 0.6973, "step": 3214 }, { "epoch": 0.21583168350055368, "grad_norm": 1.091895580291748, "learning_rate": 9.841626662797725e-05, "loss": 0.7432, "step": 3216 }, { "epoch": 0.21596590718432268, "grad_norm": 1.1444942951202393, "learning_rate": 9.841355175601481e-05, "loss": 0.744, "step": 3218 }, { "epoch": 0.21610013086809168, "grad_norm": 1.2099897861480713, "learning_rate": 9.841083459660725e-05, "loss": 0.7323, "step": 3220 }, { "epoch": 0.21623435455186069, "grad_norm": 1.3110309839248657, "learning_rate": 9.840811514988294e-05, "loss": 0.7389, "step": 3222 }, { "epoch": 0.2163685782356297, "grad_norm": 1.2931196689605713, "learning_rate": 9.840539341597039e-05, "loss": 0.7125, "step": 3224 }, { "epoch": 0.2165028019193987, "grad_norm": 1.1830400228500366, "learning_rate": 9.840266939499818e-05, "loss": 0.7337, "step": 3226 }, { "epoch": 0.21663702560316767, "grad_norm": 1.1818979978561401, "learning_rate": 9.839994308709504e-05, "loss": 0.6791, "step": 3228 }, { "epoch": 0.21677124928693667, "grad_norm": 1.1020506620407104, "learning_rate": 9.839721449238974e-05, "loss": 0.6538, "step": 3230 }, { "epoch": 0.21690547297070567, "grad_norm": 1.7511346340179443, "learning_rate": 9.839448361101124e-05, "loss": 0.7263, "step": 3232 }, { "epoch": 0.21703969665447467, "grad_norm": 1.1966168880462646, "learning_rate": 9.839175044308854e-05, "loss": 0.7244, "step": 3234 }, { "epoch": 0.21717392033824368, "grad_norm": 1.5103763341903687, "learning_rate": 9.838901498875081e-05, "loss": 0.7085, "step": 3236 }, { "epoch": 0.21730814402201268, "grad_norm": 1.7880642414093018, "learning_rate": 9.838627724812725e-05, "loss": 0.7046, "step": 3238 }, { "epoch": 0.21744236770578168, "grad_norm": 1.1055731773376465, "learning_rate": 9.838353722134725e-05, "loss": 0.7563, "step": 3240 }, { "epoch": 0.21757659138955068, "grad_norm": 1.201663851737976, "learning_rate": 9.838079490854027e-05, "loss": 0.7018, "step": 3242 }, { "epoch": 0.21771081507331969, "grad_norm": 1.2205414772033691, "learning_rate": 9.837805030983585e-05, "loss": 0.7647, "step": 3244 }, { "epoch": 0.2178450387570887, "grad_norm": 1.239234447479248, "learning_rate": 9.837530342536368e-05, "loss": 0.7548, "step": 3246 }, { "epoch": 0.2179792624408577, "grad_norm": 1.5525140762329102, "learning_rate": 9.837255425525356e-05, "loss": 0.7301, "step": 3248 }, { "epoch": 0.2181134861246267, "grad_norm": 1.1218043565750122, "learning_rate": 9.836980279963537e-05, "loss": 0.7664, "step": 3250 }, { "epoch": 0.2182477098083957, "grad_norm": 1.2930965423583984, "learning_rate": 9.836704905863911e-05, "loss": 0.715, "step": 3252 }, { "epoch": 0.2183819334921647, "grad_norm": 1.274599552154541, "learning_rate": 9.836429303239491e-05, "loss": 0.7677, "step": 3254 }, { "epoch": 0.2185161571759337, "grad_norm": 1.1840789318084717, "learning_rate": 9.836153472103296e-05, "loss": 0.7674, "step": 3256 }, { "epoch": 0.2186503808597027, "grad_norm": 1.3341907262802124, "learning_rate": 9.835877412468357e-05, "loss": 0.7489, "step": 3258 }, { "epoch": 0.2187846045434717, "grad_norm": 1.1839338541030884, "learning_rate": 9.835601124347722e-05, "loss": 0.6916, "step": 3260 }, { "epoch": 0.2189188282272407, "grad_norm": 1.022170901298523, "learning_rate": 9.835324607754442e-05, "loss": 0.689, "step": 3262 }, { "epoch": 0.2190530519110097, "grad_norm": 1.1673866510391235, "learning_rate": 9.835047862701583e-05, "loss": 0.8045, "step": 3264 }, { "epoch": 0.21918727559477869, "grad_norm": 1.1706578731536865, "learning_rate": 9.834770889202219e-05, "loss": 0.7545, "step": 3266 }, { "epoch": 0.2193214992785477, "grad_norm": 1.2601052522659302, "learning_rate": 9.834493687269438e-05, "loss": 0.7145, "step": 3268 }, { "epoch": 0.2194557229623167, "grad_norm": 1.116899847984314, "learning_rate": 9.834216256916337e-05, "loss": 0.7458, "step": 3270 }, { "epoch": 0.2195899466460857, "grad_norm": 1.189274787902832, "learning_rate": 9.833938598156025e-05, "loss": 0.7316, "step": 3272 }, { "epoch": 0.2197241703298547, "grad_norm": 1.1440093517303467, "learning_rate": 9.833660711001619e-05, "loss": 0.7593, "step": 3274 }, { "epoch": 0.2198583940136237, "grad_norm": 1.5465357303619385, "learning_rate": 9.833382595466249e-05, "loss": 0.6599, "step": 3276 }, { "epoch": 0.2199926176973927, "grad_norm": 2.1189398765563965, "learning_rate": 9.833104251563056e-05, "loss": 0.6945, "step": 3278 }, { "epoch": 0.2201268413811617, "grad_norm": 1.1823686361312866, "learning_rate": 9.832825679305191e-05, "loss": 0.7261, "step": 3280 }, { "epoch": 0.2202610650649307, "grad_norm": 1.389985203742981, "learning_rate": 9.832546878705817e-05, "loss": 0.6992, "step": 3282 }, { "epoch": 0.2203952887486997, "grad_norm": 1.2685588598251343, "learning_rate": 9.832267849778106e-05, "loss": 0.74, "step": 3284 }, { "epoch": 0.2205295124324687, "grad_norm": 1.1789789199829102, "learning_rate": 9.83198859253524e-05, "loss": 0.7712, "step": 3286 }, { "epoch": 0.2206637361162377, "grad_norm": 1.079740047454834, "learning_rate": 9.831709106990414e-05, "loss": 0.7319, "step": 3288 }, { "epoch": 0.22079795980000672, "grad_norm": 1.25736665725708, "learning_rate": 9.831429393156834e-05, "loss": 0.7148, "step": 3290 }, { "epoch": 0.22093218348377572, "grad_norm": 1.6037640571594238, "learning_rate": 9.831149451047718e-05, "loss": 0.7842, "step": 3292 }, { "epoch": 0.22106640716754472, "grad_norm": 1.8668400049209595, "learning_rate": 9.83086928067629e-05, "loss": 0.7163, "step": 3294 }, { "epoch": 0.22120063085131372, "grad_norm": 1.6986749172210693, "learning_rate": 9.830588882055786e-05, "loss": 0.7139, "step": 3296 }, { "epoch": 0.22133485453508273, "grad_norm": 1.2642039060592651, "learning_rate": 9.830308255199457e-05, "loss": 0.747, "step": 3298 }, { "epoch": 0.22146907821885173, "grad_norm": 1.3078261613845825, "learning_rate": 9.830027400120561e-05, "loss": 0.7421, "step": 3300 }, { "epoch": 0.22160330190262073, "grad_norm": 1.1431974172592163, "learning_rate": 9.82974631683237e-05, "loss": 0.7438, "step": 3302 }, { "epoch": 0.2217375255863897, "grad_norm": 1.0984482765197754, "learning_rate": 9.829465005348162e-05, "loss": 0.7345, "step": 3304 }, { "epoch": 0.2218717492701587, "grad_norm": 1.1009687185287476, "learning_rate": 9.829183465681229e-05, "loss": 0.686, "step": 3306 }, { "epoch": 0.2220059729539277, "grad_norm": 1.0883309841156006, "learning_rate": 9.828901697844872e-05, "loss": 0.6682, "step": 3308 }, { "epoch": 0.2221401966376967, "grad_norm": 1.512415885925293, "learning_rate": 9.828619701852407e-05, "loss": 0.7897, "step": 3310 }, { "epoch": 0.22227442032146572, "grad_norm": 1.1906039714813232, "learning_rate": 9.828337477717157e-05, "loss": 0.6654, "step": 3312 }, { "epoch": 0.22240864400523472, "grad_norm": 1.738184928894043, "learning_rate": 9.828055025452454e-05, "loss": 0.8456, "step": 3314 }, { "epoch": 0.22254286768900372, "grad_norm": 1.3221595287322998, "learning_rate": 9.827772345071647e-05, "loss": 0.697, "step": 3316 }, { "epoch": 0.22267709137277272, "grad_norm": 1.672614336013794, "learning_rate": 9.827489436588088e-05, "loss": 0.7487, "step": 3318 }, { "epoch": 0.22281131505654173, "grad_norm": 1.4702619314193726, "learning_rate": 9.827206300015147e-05, "loss": 0.7818, "step": 3320 }, { "epoch": 0.22294553874031073, "grad_norm": 1.2230552434921265, "learning_rate": 9.8269229353662e-05, "loss": 0.7932, "step": 3322 }, { "epoch": 0.22307976242407973, "grad_norm": 1.1639833450317383, "learning_rate": 9.826639342654636e-05, "loss": 0.7041, "step": 3324 }, { "epoch": 0.22321398610784873, "grad_norm": 1.333631992340088, "learning_rate": 9.826355521893855e-05, "loss": 0.7125, "step": 3326 }, { "epoch": 0.22334820979161774, "grad_norm": 1.2048940658569336, "learning_rate": 9.826071473097265e-05, "loss": 0.7808, "step": 3328 }, { "epoch": 0.22348243347538674, "grad_norm": 1.131531834602356, "learning_rate": 9.82578719627829e-05, "loss": 0.7575, "step": 3330 }, { "epoch": 0.22361665715915574, "grad_norm": 1.1711511611938477, "learning_rate": 9.825502691450357e-05, "loss": 0.6958, "step": 3332 }, { "epoch": 0.22375088084292474, "grad_norm": 1.0046534538269043, "learning_rate": 9.825217958626913e-05, "loss": 0.7534, "step": 3334 }, { "epoch": 0.22388510452669375, "grad_norm": 1.0132005214691162, "learning_rate": 9.824932997821408e-05, "loss": 0.6879, "step": 3336 }, { "epoch": 0.22401932821046275, "grad_norm": 1.1994860172271729, "learning_rate": 9.824647809047306e-05, "loss": 0.7525, "step": 3338 }, { "epoch": 0.22415355189423175, "grad_norm": 1.171962022781372, "learning_rate": 9.824362392318082e-05, "loss": 0.6529, "step": 3340 }, { "epoch": 0.22428777557800073, "grad_norm": 1.1129786968231201, "learning_rate": 9.824076747647223e-05, "loss": 0.7097, "step": 3342 }, { "epoch": 0.22442199926176973, "grad_norm": 1.3192722797393799, "learning_rate": 9.823790875048224e-05, "loss": 0.7779, "step": 3344 }, { "epoch": 0.22455622294553873, "grad_norm": 1.1418710947036743, "learning_rate": 9.823504774534591e-05, "loss": 0.7458, "step": 3346 }, { "epoch": 0.22469044662930773, "grad_norm": 1.180791974067688, "learning_rate": 9.823218446119842e-05, "loss": 0.7012, "step": 3348 }, { "epoch": 0.22482467031307674, "grad_norm": 1.1937333345413208, "learning_rate": 9.822931889817506e-05, "loss": 0.7746, "step": 3350 }, { "epoch": 0.22495889399684574, "grad_norm": 1.2063286304473877, "learning_rate": 9.822645105641123e-05, "loss": 0.7043, "step": 3352 }, { "epoch": 0.22509311768061474, "grad_norm": 1.2942754030227661, "learning_rate": 9.822358093604242e-05, "loss": 0.7285, "step": 3354 }, { "epoch": 0.22522734136438374, "grad_norm": 1.1679829359054565, "learning_rate": 9.822070853720421e-05, "loss": 0.683, "step": 3356 }, { "epoch": 0.22536156504815275, "grad_norm": 1.1674364805221558, "learning_rate": 9.821783386003239e-05, "loss": 0.6875, "step": 3358 }, { "epoch": 0.22549578873192175, "grad_norm": 1.1562505960464478, "learning_rate": 9.821495690466272e-05, "loss": 0.7452, "step": 3360 }, { "epoch": 0.22563001241569075, "grad_norm": 1.2197908163070679, "learning_rate": 9.821207767123113e-05, "loss": 0.8058, "step": 3362 }, { "epoch": 0.22576423609945975, "grad_norm": 1.1897050142288208, "learning_rate": 9.820919615987368e-05, "loss": 0.6991, "step": 3364 }, { "epoch": 0.22589845978322876, "grad_norm": 1.2827985286712646, "learning_rate": 9.820631237072652e-05, "loss": 0.84, "step": 3366 }, { "epoch": 0.22603268346699776, "grad_norm": 1.0616706609725952, "learning_rate": 9.82034263039259e-05, "loss": 0.6679, "step": 3368 }, { "epoch": 0.22616690715076676, "grad_norm": 1.5039023160934448, "learning_rate": 9.820053795960815e-05, "loss": 0.7722, "step": 3370 }, { "epoch": 0.22630113083453576, "grad_norm": 1.2132271528244019, "learning_rate": 9.819764733790979e-05, "loss": 0.696, "step": 3372 }, { "epoch": 0.22643535451830477, "grad_norm": 1.3512502908706665, "learning_rate": 9.819475443896736e-05, "loss": 0.6855, "step": 3374 }, { "epoch": 0.22656957820207377, "grad_norm": 1.2658389806747437, "learning_rate": 9.819185926291754e-05, "loss": 0.7526, "step": 3376 }, { "epoch": 0.22670380188584277, "grad_norm": 1.086442232131958, "learning_rate": 9.818896180989716e-05, "loss": 0.7226, "step": 3378 }, { "epoch": 0.22683802556961175, "grad_norm": 1.312869668006897, "learning_rate": 9.818606208004309e-05, "loss": 0.7276, "step": 3380 }, { "epoch": 0.22697224925338075, "grad_norm": 1.1030393838882446, "learning_rate": 9.818316007349232e-05, "loss": 0.8007, "step": 3382 }, { "epoch": 0.22710647293714975, "grad_norm": 1.296844244003296, "learning_rate": 9.8180255790382e-05, "loss": 0.7222, "step": 3384 }, { "epoch": 0.22724069662091875, "grad_norm": 2.0905978679656982, "learning_rate": 9.817734923084934e-05, "loss": 0.7195, "step": 3386 }, { "epoch": 0.22737492030468776, "grad_norm": 1.3652046918869019, "learning_rate": 9.817444039503165e-05, "loss": 0.7275, "step": 3388 }, { "epoch": 0.22750914398845676, "grad_norm": 1.1128994226455688, "learning_rate": 9.817152928306638e-05, "loss": 0.7207, "step": 3390 }, { "epoch": 0.22764336767222576, "grad_norm": 1.7706009149551392, "learning_rate": 9.81686158950911e-05, "loss": 0.7097, "step": 3392 }, { "epoch": 0.22777759135599476, "grad_norm": 1.4544775485992432, "learning_rate": 9.816570023124342e-05, "loss": 0.834, "step": 3394 }, { "epoch": 0.22791181503976377, "grad_norm": 1.928645133972168, "learning_rate": 9.816278229166114e-05, "loss": 0.7695, "step": 3396 }, { "epoch": 0.22804603872353277, "grad_norm": 1.073394775390625, "learning_rate": 9.815986207648208e-05, "loss": 0.7945, "step": 3398 }, { "epoch": 0.22818026240730177, "grad_norm": 1.1026980876922607, "learning_rate": 9.815693958584424e-05, "loss": 0.7195, "step": 3400 }, { "epoch": 0.22831448609107077, "grad_norm": 1.298104166984558, "learning_rate": 9.815401481988571e-05, "loss": 0.7352, "step": 3402 }, { "epoch": 0.22844870977483978, "grad_norm": 1.1682909727096558, "learning_rate": 9.815108777874467e-05, "loss": 0.7504, "step": 3404 }, { "epoch": 0.22858293345860878, "grad_norm": 1.234245777130127, "learning_rate": 9.814815846255942e-05, "loss": 0.819, "step": 3406 }, { "epoch": 0.22871715714237778, "grad_norm": 1.2632322311401367, "learning_rate": 9.814522687146837e-05, "loss": 0.6739, "step": 3408 }, { "epoch": 0.22885138082614678, "grad_norm": 1.0744209289550781, "learning_rate": 9.814229300560999e-05, "loss": 0.7162, "step": 3410 }, { "epoch": 0.22898560450991579, "grad_norm": 1.0212805271148682, "learning_rate": 9.813935686512297e-05, "loss": 0.7079, "step": 3412 }, { "epoch": 0.2291198281936848, "grad_norm": 1.13587486743927, "learning_rate": 9.813641845014599e-05, "loss": 0.8045, "step": 3414 }, { "epoch": 0.2292540518774538, "grad_norm": 1.1257480382919312, "learning_rate": 9.813347776081789e-05, "loss": 0.6868, "step": 3416 }, { "epoch": 0.22938827556122277, "grad_norm": 1.2473704814910889, "learning_rate": 9.813053479727761e-05, "loss": 0.714, "step": 3418 }, { "epoch": 0.22952249924499177, "grad_norm": 1.0271937847137451, "learning_rate": 9.812758955966421e-05, "loss": 0.6735, "step": 3420 }, { "epoch": 0.22965672292876077, "grad_norm": 1.1803767681121826, "learning_rate": 9.812464204811686e-05, "loss": 0.7347, "step": 3422 }, { "epoch": 0.22979094661252977, "grad_norm": 1.1167150735855103, "learning_rate": 9.812169226277479e-05, "loss": 0.6919, "step": 3424 }, { "epoch": 0.22992517029629878, "grad_norm": 1.2198318243026733, "learning_rate": 9.811874020377738e-05, "loss": 0.7548, "step": 3426 }, { "epoch": 0.23005939398006778, "grad_norm": 1.0730876922607422, "learning_rate": 9.811578587126413e-05, "loss": 0.6881, "step": 3428 }, { "epoch": 0.23019361766383678, "grad_norm": 1.3007014989852905, "learning_rate": 9.81128292653746e-05, "loss": 0.6966, "step": 3430 }, { "epoch": 0.23032784134760578, "grad_norm": 1.1373518705368042, "learning_rate": 9.810987038624851e-05, "loss": 0.7638, "step": 3432 }, { "epoch": 0.23046206503137479, "grad_norm": 1.1992700099945068, "learning_rate": 9.810690923402566e-05, "loss": 0.7312, "step": 3434 }, { "epoch": 0.2305962887151438, "grad_norm": 1.0444557666778564, "learning_rate": 9.810394580884592e-05, "loss": 0.7302, "step": 3436 }, { "epoch": 0.2307305123989128, "grad_norm": 1.0445059537887573, "learning_rate": 9.810098011084935e-05, "loss": 0.7635, "step": 3438 }, { "epoch": 0.2308647360826818, "grad_norm": 1.4251956939697266, "learning_rate": 9.809801214017604e-05, "loss": 0.7156, "step": 3440 }, { "epoch": 0.2309989597664508, "grad_norm": 1.1660226583480835, "learning_rate": 9.809504189696626e-05, "loss": 0.7105, "step": 3442 }, { "epoch": 0.2311331834502198, "grad_norm": 1.0929772853851318, "learning_rate": 9.809206938136031e-05, "loss": 0.738, "step": 3444 }, { "epoch": 0.2312674071339888, "grad_norm": 1.3212532997131348, "learning_rate": 9.808909459349865e-05, "loss": 0.7501, "step": 3446 }, { "epoch": 0.2314016308177578, "grad_norm": 1.182216763496399, "learning_rate": 9.808611753352184e-05, "loss": 0.737, "step": 3448 }, { "epoch": 0.2315358545015268, "grad_norm": 1.378854513168335, "learning_rate": 9.808313820157052e-05, "loss": 0.7129, "step": 3450 }, { "epoch": 0.2316700781852958, "grad_norm": 1.1713718175888062, "learning_rate": 9.808015659778549e-05, "loss": 0.7528, "step": 3452 }, { "epoch": 0.23180430186906478, "grad_norm": 1.11350679397583, "learning_rate": 9.80771727223076e-05, "loss": 0.6979, "step": 3454 }, { "epoch": 0.23193852555283379, "grad_norm": 1.063978910446167, "learning_rate": 9.807418657527782e-05, "loss": 0.7262, "step": 3456 }, { "epoch": 0.2320727492366028, "grad_norm": 1.4309468269348145, "learning_rate": 9.807119815683728e-05, "loss": 0.6923, "step": 3458 }, { "epoch": 0.2322069729203718, "grad_norm": 1.379732370376587, "learning_rate": 9.806820746712716e-05, "loss": 0.7342, "step": 3460 }, { "epoch": 0.2323411966041408, "grad_norm": 1.1750215291976929, "learning_rate": 9.806521450628875e-05, "loss": 0.7704, "step": 3462 }, { "epoch": 0.2324754202879098, "grad_norm": 1.1312384605407715, "learning_rate": 9.806221927446347e-05, "loss": 0.7967, "step": 3464 }, { "epoch": 0.2326096439716788, "grad_norm": 1.3090623617172241, "learning_rate": 9.805922177179283e-05, "loss": 0.7725, "step": 3466 }, { "epoch": 0.2327438676554478, "grad_norm": 1.1082390546798706, "learning_rate": 9.805622199841848e-05, "loss": 0.6999, "step": 3468 }, { "epoch": 0.2328780913392168, "grad_norm": 1.0914182662963867, "learning_rate": 9.805321995448214e-05, "loss": 0.7262, "step": 3470 }, { "epoch": 0.2330123150229858, "grad_norm": 1.1261818408966064, "learning_rate": 9.805021564012564e-05, "loss": 0.7699, "step": 3472 }, { "epoch": 0.2331465387067548, "grad_norm": 1.0641417503356934, "learning_rate": 9.804720905549094e-05, "loss": 0.7033, "step": 3474 }, { "epoch": 0.2332807623905238, "grad_norm": 1.4888118505477905, "learning_rate": 9.80442002007201e-05, "loss": 0.6867, "step": 3476 }, { "epoch": 0.2334149860742928, "grad_norm": 1.3334269523620605, "learning_rate": 9.804118907595527e-05, "loss": 0.6697, "step": 3478 }, { "epoch": 0.23354920975806182, "grad_norm": 1.0359936952590942, "learning_rate": 9.803817568133872e-05, "loss": 0.7034, "step": 3480 }, { "epoch": 0.23368343344183082, "grad_norm": 1.256037712097168, "learning_rate": 9.803516001701286e-05, "loss": 0.7702, "step": 3482 }, { "epoch": 0.23381765712559982, "grad_norm": 1.2835520505905151, "learning_rate": 9.80321420831201e-05, "loss": 0.7173, "step": 3484 }, { "epoch": 0.23395188080936882, "grad_norm": 1.2579686641693115, "learning_rate": 9.80291218798031e-05, "loss": 0.7454, "step": 3486 }, { "epoch": 0.23408610449313783, "grad_norm": 1.1662805080413818, "learning_rate": 9.802609940720455e-05, "loss": 0.7573, "step": 3488 }, { "epoch": 0.23422032817690683, "grad_norm": 1.117235541343689, "learning_rate": 9.802307466546723e-05, "loss": 0.7579, "step": 3490 }, { "epoch": 0.2343545518606758, "grad_norm": 1.1694008111953735, "learning_rate": 9.802004765473407e-05, "loss": 0.7195, "step": 3492 }, { "epoch": 0.2344887755444448, "grad_norm": 1.1704126596450806, "learning_rate": 9.801701837514808e-05, "loss": 0.6869, "step": 3494 }, { "epoch": 0.2346229992282138, "grad_norm": 1.1846356391906738, "learning_rate": 9.801398682685238e-05, "loss": 0.7383, "step": 3496 }, { "epoch": 0.2347572229119828, "grad_norm": 1.1462559700012207, "learning_rate": 9.801095300999024e-05, "loss": 0.7334, "step": 3498 }, { "epoch": 0.2348914465957518, "grad_norm": 1.1507761478424072, "learning_rate": 9.800791692470497e-05, "loss": 0.66, "step": 3500 }, { "epoch": 0.23502567027952082, "grad_norm": 1.2698543071746826, "learning_rate": 9.800487857114004e-05, "loss": 0.7343, "step": 3502 }, { "epoch": 0.23515989396328982, "grad_norm": 1.1707394123077393, "learning_rate": 9.800183794943898e-05, "loss": 0.7248, "step": 3504 }, { "epoch": 0.23529411764705882, "grad_norm": 1.3977922201156616, "learning_rate": 9.799879505974548e-05, "loss": 0.7618, "step": 3506 }, { "epoch": 0.23542834133082782, "grad_norm": 1.113440752029419, "learning_rate": 9.799574990220328e-05, "loss": 0.6986, "step": 3508 }, { "epoch": 0.23556256501459683, "grad_norm": 0.9917773008346558, "learning_rate": 9.79927024769563e-05, "loss": 0.6308, "step": 3510 }, { "epoch": 0.23569678869836583, "grad_norm": 1.1032965183258057, "learning_rate": 9.798965278414849e-05, "loss": 0.7304, "step": 3512 }, { "epoch": 0.23583101238213483, "grad_norm": 1.1461377143859863, "learning_rate": 9.798660082392396e-05, "loss": 0.6918, "step": 3514 }, { "epoch": 0.23596523606590383, "grad_norm": 1.1212458610534668, "learning_rate": 9.798354659642691e-05, "loss": 0.7687, "step": 3516 }, { "epoch": 0.23609945974967284, "grad_norm": 1.6978212594985962, "learning_rate": 9.798049010180161e-05, "loss": 0.7388, "step": 3518 }, { "epoch": 0.23623368343344184, "grad_norm": 1.3478388786315918, "learning_rate": 9.797743134019253e-05, "loss": 0.7947, "step": 3520 }, { "epoch": 0.23636790711721084, "grad_norm": 1.1857460737228394, "learning_rate": 9.797437031174414e-05, "loss": 0.7414, "step": 3522 }, { "epoch": 0.23650213080097984, "grad_norm": 1.2205296754837036, "learning_rate": 9.797130701660111e-05, "loss": 0.7317, "step": 3524 }, { "epoch": 0.23663635448474885, "grad_norm": 1.3379762172698975, "learning_rate": 9.796824145490815e-05, "loss": 0.7165, "step": 3526 }, { "epoch": 0.23677057816851785, "grad_norm": 1.1166516542434692, "learning_rate": 9.79651736268101e-05, "loss": 0.7088, "step": 3528 }, { "epoch": 0.23690480185228682, "grad_norm": 1.2244601249694824, "learning_rate": 9.796210353245192e-05, "loss": 0.7266, "step": 3530 }, { "epoch": 0.23703902553605583, "grad_norm": 1.0700269937515259, "learning_rate": 9.795903117197867e-05, "loss": 0.7143, "step": 3532 }, { "epoch": 0.23717324921982483, "grad_norm": 1.1071746349334717, "learning_rate": 9.795595654553548e-05, "loss": 0.697, "step": 3534 }, { "epoch": 0.23730747290359383, "grad_norm": 1.3859741687774658, "learning_rate": 9.795287965326767e-05, "loss": 0.7405, "step": 3536 }, { "epoch": 0.23744169658736283, "grad_norm": 1.2678968906402588, "learning_rate": 9.794980049532058e-05, "loss": 0.706, "step": 3538 }, { "epoch": 0.23757592027113184, "grad_norm": 1.073473572731018, "learning_rate": 9.79467190718397e-05, "loss": 0.7804, "step": 3540 }, { "epoch": 0.23771014395490084, "grad_norm": 1.770054578781128, "learning_rate": 9.794363538297065e-05, "loss": 0.7759, "step": 3542 }, { "epoch": 0.23784436763866984, "grad_norm": 1.256622552871704, "learning_rate": 9.794054942885909e-05, "loss": 0.7375, "step": 3544 }, { "epoch": 0.23797859132243884, "grad_norm": 1.2577251195907593, "learning_rate": 9.793746120965083e-05, "loss": 0.7204, "step": 3546 }, { "epoch": 0.23811281500620785, "grad_norm": 1.1353765726089478, "learning_rate": 9.793437072549181e-05, "loss": 0.6705, "step": 3548 }, { "epoch": 0.23824703868997685, "grad_norm": 1.0381370782852173, "learning_rate": 9.793127797652801e-05, "loss": 0.6757, "step": 3550 }, { "epoch": 0.23838126237374585, "grad_norm": 1.1946134567260742, "learning_rate": 9.79281829629056e-05, "loss": 0.7714, "step": 3552 }, { "epoch": 0.23851548605751485, "grad_norm": 1.4038199186325073, "learning_rate": 9.792508568477078e-05, "loss": 0.723, "step": 3554 }, { "epoch": 0.23864970974128386, "grad_norm": 1.2247889041900635, "learning_rate": 9.792198614226992e-05, "loss": 0.7129, "step": 3556 }, { "epoch": 0.23878393342505286, "grad_norm": 1.3369977474212646, "learning_rate": 9.791888433554943e-05, "loss": 0.7371, "step": 3558 }, { "epoch": 0.23891815710882186, "grad_norm": 1.1259257793426514, "learning_rate": 9.79157802647559e-05, "loss": 0.6986, "step": 3560 }, { "epoch": 0.23905238079259086, "grad_norm": 1.241822600364685, "learning_rate": 9.791267393003596e-05, "loss": 0.7072, "step": 3562 }, { "epoch": 0.23918660447635987, "grad_norm": 1.2874438762664795, "learning_rate": 9.79095653315364e-05, "loss": 0.7947, "step": 3564 }, { "epoch": 0.23932082816012887, "grad_norm": 1.553829312324524, "learning_rate": 9.790645446940408e-05, "loss": 0.7108, "step": 3566 }, { "epoch": 0.23945505184389784, "grad_norm": 1.1112264394760132, "learning_rate": 9.7903341343786e-05, "loss": 0.7198, "step": 3568 }, { "epoch": 0.23958927552766685, "grad_norm": 1.469152808189392, "learning_rate": 9.790022595482924e-05, "loss": 0.7685, "step": 3570 }, { "epoch": 0.23972349921143585, "grad_norm": 1.1694190502166748, "learning_rate": 9.789710830268099e-05, "loss": 0.7018, "step": 3572 }, { "epoch": 0.23985772289520485, "grad_norm": 1.2232519388198853, "learning_rate": 9.789398838748856e-05, "loss": 0.7373, "step": 3574 }, { "epoch": 0.23999194657897385, "grad_norm": 1.0650718212127686, "learning_rate": 9.789086620939936e-05, "loss": 0.7908, "step": 3576 }, { "epoch": 0.24012617026274286, "grad_norm": 1.0550053119659424, "learning_rate": 9.78877417685609e-05, "loss": 0.7505, "step": 3578 }, { "epoch": 0.24026039394651186, "grad_norm": 1.0913403034210205, "learning_rate": 9.788461506512081e-05, "loss": 0.6891, "step": 3580 }, { "epoch": 0.24039461763028086, "grad_norm": 1.256256103515625, "learning_rate": 9.788148609922682e-05, "loss": 0.7299, "step": 3582 }, { "epoch": 0.24052884131404986, "grad_norm": 1.370922327041626, "learning_rate": 9.787835487102677e-05, "loss": 0.78, "step": 3584 }, { "epoch": 0.24066306499781887, "grad_norm": 1.1956902742385864, "learning_rate": 9.78752213806686e-05, "loss": 0.7117, "step": 3586 }, { "epoch": 0.24079728868158787, "grad_norm": 1.1800501346588135, "learning_rate": 9.787208562830036e-05, "loss": 0.6698, "step": 3588 }, { "epoch": 0.24093151236535687, "grad_norm": 1.1705611944198608, "learning_rate": 9.786894761407021e-05, "loss": 0.6684, "step": 3590 }, { "epoch": 0.24106573604912587, "grad_norm": 1.2129614353179932, "learning_rate": 9.786580733812643e-05, "loss": 0.7731, "step": 3592 }, { "epoch": 0.24119995973289488, "grad_norm": 1.265656590461731, "learning_rate": 9.786266480061737e-05, "loss": 0.7194, "step": 3594 }, { "epoch": 0.24133418341666388, "grad_norm": 1.2038565874099731, "learning_rate": 9.78595200016915e-05, "loss": 0.7071, "step": 3596 }, { "epoch": 0.24146840710043288, "grad_norm": 1.1142795085906982, "learning_rate": 9.785637294149743e-05, "loss": 0.7693, "step": 3598 }, { "epoch": 0.24160263078420188, "grad_norm": 1.061290979385376, "learning_rate": 9.785322362018385e-05, "loss": 0.7344, "step": 3600 }, { "epoch": 0.24173685446797089, "grad_norm": 1.1317564249038696, "learning_rate": 9.785007203789955e-05, "loss": 0.7455, "step": 3602 }, { "epoch": 0.2418710781517399, "grad_norm": 1.1659226417541504, "learning_rate": 9.784691819479343e-05, "loss": 0.7227, "step": 3604 }, { "epoch": 0.24200530183550886, "grad_norm": 1.1892750263214111, "learning_rate": 9.784376209101454e-05, "loss": 0.7811, "step": 3606 }, { "epoch": 0.24213952551927787, "grad_norm": 1.2177784442901611, "learning_rate": 9.784060372671195e-05, "loss": 0.7072, "step": 3608 }, { "epoch": 0.24227374920304687, "grad_norm": 1.257696270942688, "learning_rate": 9.783744310203491e-05, "loss": 0.7335, "step": 3610 }, { "epoch": 0.24240797288681587, "grad_norm": 1.164030909538269, "learning_rate": 9.783428021713274e-05, "loss": 0.7008, "step": 3612 }, { "epoch": 0.24254219657058487, "grad_norm": 1.138877034187317, "learning_rate": 9.783111507215491e-05, "loss": 0.7451, "step": 3614 }, { "epoch": 0.24267642025435388, "grad_norm": 1.120921015739441, "learning_rate": 9.782794766725094e-05, "loss": 0.8045, "step": 3616 }, { "epoch": 0.24281064393812288, "grad_norm": 1.1520529985427856, "learning_rate": 9.78247780025705e-05, "loss": 0.7086, "step": 3618 }, { "epoch": 0.24294486762189188, "grad_norm": 1.222719669342041, "learning_rate": 9.782160607826334e-05, "loss": 0.7644, "step": 3620 }, { "epoch": 0.24307909130566088, "grad_norm": 1.2890983819961548, "learning_rate": 9.781843189447933e-05, "loss": 0.7014, "step": 3622 }, { "epoch": 0.24321331498942989, "grad_norm": 1.2486456632614136, "learning_rate": 9.781525545136844e-05, "loss": 0.7205, "step": 3624 }, { "epoch": 0.2433475386731989, "grad_norm": 1.1125035285949707, "learning_rate": 9.781207674908076e-05, "loss": 0.7771, "step": 3626 }, { "epoch": 0.2434817623569679, "grad_norm": 1.0747028589248657, "learning_rate": 9.780889578776647e-05, "loss": 0.664, "step": 3628 }, { "epoch": 0.2436159860407369, "grad_norm": 1.0456690788269043, "learning_rate": 9.780571256757587e-05, "loss": 0.6862, "step": 3630 }, { "epoch": 0.2437502097245059, "grad_norm": 1.097830057144165, "learning_rate": 9.780252708865936e-05, "loss": 0.692, "step": 3632 }, { "epoch": 0.2438844334082749, "grad_norm": 1.0377219915390015, "learning_rate": 9.779933935116742e-05, "loss": 0.6649, "step": 3634 }, { "epoch": 0.2440186570920439, "grad_norm": 1.1227837800979614, "learning_rate": 9.779614935525073e-05, "loss": 0.7859, "step": 3636 }, { "epoch": 0.2441528807758129, "grad_norm": 1.0132102966308594, "learning_rate": 9.779295710105993e-05, "loss": 0.7911, "step": 3638 }, { "epoch": 0.2442871044595819, "grad_norm": 1.0197548866271973, "learning_rate": 9.77897625887459e-05, "loss": 0.6874, "step": 3640 }, { "epoch": 0.2444213281433509, "grad_norm": 1.0421596765518188, "learning_rate": 9.778656581845958e-05, "loss": 0.6859, "step": 3642 }, { "epoch": 0.24455555182711988, "grad_norm": 1.015177845954895, "learning_rate": 9.778336679035197e-05, "loss": 0.7016, "step": 3644 }, { "epoch": 0.24468977551088889, "grad_norm": 1.1596684455871582, "learning_rate": 9.778016550457425e-05, "loss": 0.6877, "step": 3646 }, { "epoch": 0.2448239991946579, "grad_norm": 1.020958662033081, "learning_rate": 9.777696196127766e-05, "loss": 0.697, "step": 3648 }, { "epoch": 0.2449582228784269, "grad_norm": 1.4052140712738037, "learning_rate": 9.777375616061359e-05, "loss": 0.7571, "step": 3650 }, { "epoch": 0.2450924465621959, "grad_norm": 1.1902647018432617, "learning_rate": 9.777054810273345e-05, "loss": 0.7328, "step": 3652 }, { "epoch": 0.2452266702459649, "grad_norm": 1.2707990407943726, "learning_rate": 9.776733778778888e-05, "loss": 0.7506, "step": 3654 }, { "epoch": 0.2453608939297339, "grad_norm": 1.1604020595550537, "learning_rate": 9.776412521593152e-05, "loss": 0.6732, "step": 3656 }, { "epoch": 0.2454951176135029, "grad_norm": 1.068184733390808, "learning_rate": 9.776091038731317e-05, "loss": 0.7242, "step": 3658 }, { "epoch": 0.2456293412972719, "grad_norm": 1.2311961650848389, "learning_rate": 9.775769330208571e-05, "loss": 0.6708, "step": 3660 }, { "epoch": 0.2457635649810409, "grad_norm": 1.0739384889602661, "learning_rate": 9.775447396040116e-05, "loss": 0.735, "step": 3662 }, { "epoch": 0.2458977886648099, "grad_norm": 1.1574194431304932, "learning_rate": 9.775125236241161e-05, "loss": 0.7217, "step": 3664 }, { "epoch": 0.2460320123485789, "grad_norm": 1.1732994318008423, "learning_rate": 9.77480285082693e-05, "loss": 0.7277, "step": 3666 }, { "epoch": 0.2461662360323479, "grad_norm": 1.008643627166748, "learning_rate": 9.774480239812653e-05, "loss": 0.6997, "step": 3668 }, { "epoch": 0.24630045971611692, "grad_norm": 1.0346351861953735, "learning_rate": 9.774157403213573e-05, "loss": 0.7228, "step": 3670 }, { "epoch": 0.24643468339988592, "grad_norm": 1.1072207689285278, "learning_rate": 9.773834341044944e-05, "loss": 0.7401, "step": 3672 }, { "epoch": 0.24656890708365492, "grad_norm": 1.3879812955856323, "learning_rate": 9.77351105332203e-05, "loss": 0.729, "step": 3674 }, { "epoch": 0.24670313076742392, "grad_norm": 1.1985526084899902, "learning_rate": 9.773187540060105e-05, "loss": 0.8118, "step": 3676 }, { "epoch": 0.24683735445119293, "grad_norm": 1.0382481813430786, "learning_rate": 9.772863801274455e-05, "loss": 0.7627, "step": 3678 }, { "epoch": 0.24697157813496193, "grad_norm": 1.1838608980178833, "learning_rate": 9.772539836980376e-05, "loss": 0.7657, "step": 3680 }, { "epoch": 0.2471058018187309, "grad_norm": 0.9234886169433594, "learning_rate": 9.772215647193174e-05, "loss": 0.6451, "step": 3682 }, { "epoch": 0.2472400255024999, "grad_norm": 1.0064630508422852, "learning_rate": 9.771891231928167e-05, "loss": 0.6762, "step": 3684 }, { "epoch": 0.2473742491862689, "grad_norm": 1.1998895406723022, "learning_rate": 9.771566591200682e-05, "loss": 0.659, "step": 3686 }, { "epoch": 0.2475084728700379, "grad_norm": 1.4650298357009888, "learning_rate": 9.77124172502606e-05, "loss": 0.8039, "step": 3688 }, { "epoch": 0.2476426965538069, "grad_norm": 1.1969459056854248, "learning_rate": 9.77091663341965e-05, "loss": 0.7041, "step": 3690 }, { "epoch": 0.24777692023757591, "grad_norm": 1.0319349765777588, "learning_rate": 9.770591316396807e-05, "loss": 0.7602, "step": 3692 }, { "epoch": 0.24791114392134492, "grad_norm": 1.1042534112930298, "learning_rate": 9.770265773972906e-05, "loss": 0.7217, "step": 3694 }, { "epoch": 0.24804536760511392, "grad_norm": 0.9951987266540527, "learning_rate": 9.769940006163329e-05, "loss": 0.8003, "step": 3696 }, { "epoch": 0.24817959128888292, "grad_norm": 1.2236394882202148, "learning_rate": 9.769614012983465e-05, "loss": 0.8061, "step": 3698 }, { "epoch": 0.24831381497265193, "grad_norm": 1.1980949640274048, "learning_rate": 9.769287794448721e-05, "loss": 0.6583, "step": 3700 }, { "epoch": 0.24844803865642093, "grad_norm": 1.1455384492874146, "learning_rate": 9.768961350574503e-05, "loss": 0.6776, "step": 3702 }, { "epoch": 0.24858226234018993, "grad_norm": 0.9626195430755615, "learning_rate": 9.768634681376243e-05, "loss": 0.6675, "step": 3704 }, { "epoch": 0.24871648602395893, "grad_norm": 1.0661108493804932, "learning_rate": 9.768307786869369e-05, "loss": 0.7096, "step": 3706 }, { "epoch": 0.24885070970772794, "grad_norm": 1.1644963026046753, "learning_rate": 9.767980667069328e-05, "loss": 0.7599, "step": 3708 }, { "epoch": 0.24898493339149694, "grad_norm": 1.3132227659225464, "learning_rate": 9.767653321991578e-05, "loss": 0.7179, "step": 3710 }, { "epoch": 0.24911915707526594, "grad_norm": 1.076185703277588, "learning_rate": 9.767325751651583e-05, "loss": 0.7005, "step": 3712 }, { "epoch": 0.24925338075903494, "grad_norm": 1.1602646112442017, "learning_rate": 9.76699795606482e-05, "loss": 0.681, "step": 3714 }, { "epoch": 0.24938760444280395, "grad_norm": 1.0248345136642456, "learning_rate": 9.766669935246778e-05, "loss": 0.6841, "step": 3716 }, { "epoch": 0.24952182812657295, "grad_norm": 1.0485111474990845, "learning_rate": 9.766341689212956e-05, "loss": 0.7047, "step": 3718 }, { "epoch": 0.24965605181034192, "grad_norm": 1.3126776218414307, "learning_rate": 9.76601321797886e-05, "loss": 0.697, "step": 3720 }, { "epoch": 0.24979027549411092, "grad_norm": 1.0859121084213257, "learning_rate": 9.765684521560012e-05, "loss": 0.7444, "step": 3722 }, { "epoch": 0.24992449917787993, "grad_norm": 1.1804007291793823, "learning_rate": 9.765355599971942e-05, "loss": 0.6579, "step": 3724 }, { "epoch": 0.25005872286164893, "grad_norm": 1.4063502550125122, "learning_rate": 9.765026453230191e-05, "loss": 0.6788, "step": 3726 }, { "epoch": 0.25019294654541796, "grad_norm": 1.0992275476455688, "learning_rate": 9.76469708135031e-05, "loss": 0.6705, "step": 3728 }, { "epoch": 0.25032717022918693, "grad_norm": 1.1454664468765259, "learning_rate": 9.764367484347861e-05, "loss": 0.6742, "step": 3730 }, { "epoch": 0.25046139391295597, "grad_norm": 1.3076444864273071, "learning_rate": 9.764037662238417e-05, "loss": 0.7585, "step": 3732 }, { "epoch": 0.25059561759672494, "grad_norm": 1.3327702283859253, "learning_rate": 9.763707615037561e-05, "loss": 0.7608, "step": 3734 }, { "epoch": 0.25072984128049397, "grad_norm": 1.3393803834915161, "learning_rate": 9.763377342760888e-05, "loss": 0.7539, "step": 3736 }, { "epoch": 0.25086406496426294, "grad_norm": 1.4489805698394775, "learning_rate": 9.763046845424002e-05, "loss": 0.8044, "step": 3738 }, { "epoch": 0.2509982886480319, "grad_norm": 1.1607704162597656, "learning_rate": 9.762716123042519e-05, "loss": 0.6865, "step": 3740 }, { "epoch": 0.25113251233180095, "grad_norm": 1.2858433723449707, "learning_rate": 9.762385175632065e-05, "loss": 0.7094, "step": 3742 }, { "epoch": 0.2512667360155699, "grad_norm": 1.336488127708435, "learning_rate": 9.762054003208276e-05, "loss": 0.8006, "step": 3744 }, { "epoch": 0.25140095969933896, "grad_norm": 1.2179168462753296, "learning_rate": 9.761722605786799e-05, "loss": 0.7414, "step": 3746 }, { "epoch": 0.25153518338310793, "grad_norm": 1.1243705749511719, "learning_rate": 9.761390983383294e-05, "loss": 0.6437, "step": 3748 }, { "epoch": 0.25166940706687696, "grad_norm": 1.0953590869903564, "learning_rate": 9.761059136013426e-05, "loss": 0.7216, "step": 3750 }, { "epoch": 0.25180363075064593, "grad_norm": 1.3589489459991455, "learning_rate": 9.760727063692878e-05, "loss": 0.8207, "step": 3752 }, { "epoch": 0.25193785443441497, "grad_norm": 1.2893167734146118, "learning_rate": 9.760394766437335e-05, "loss": 0.7309, "step": 3754 }, { "epoch": 0.25207207811818394, "grad_norm": 1.1073765754699707, "learning_rate": 9.760062244262502e-05, "loss": 0.7182, "step": 3756 }, { "epoch": 0.25220630180195297, "grad_norm": 0.994770884513855, "learning_rate": 9.759729497184089e-05, "loss": 0.7595, "step": 3758 }, { "epoch": 0.25234052548572194, "grad_norm": 1.1640686988830566, "learning_rate": 9.759396525217817e-05, "loss": 0.7536, "step": 3760 }, { "epoch": 0.252474749169491, "grad_norm": 1.36345636844635, "learning_rate": 9.759063328379416e-05, "loss": 0.7381, "step": 3762 }, { "epoch": 0.25260897285325995, "grad_norm": 1.3841369152069092, "learning_rate": 9.758729906684632e-05, "loss": 0.614, "step": 3764 }, { "epoch": 0.252743196537029, "grad_norm": 1.0968348979949951, "learning_rate": 9.758396260149219e-05, "loss": 0.6944, "step": 3766 }, { "epoch": 0.25287742022079795, "grad_norm": 1.3005449771881104, "learning_rate": 9.758062388788937e-05, "loss": 0.7201, "step": 3768 }, { "epoch": 0.253011643904567, "grad_norm": 1.2965426445007324, "learning_rate": 9.757728292619566e-05, "loss": 0.7133, "step": 3770 }, { "epoch": 0.25314586758833596, "grad_norm": 1.0011999607086182, "learning_rate": 9.757393971656888e-05, "loss": 0.7186, "step": 3772 }, { "epoch": 0.253280091272105, "grad_norm": 1.1350724697113037, "learning_rate": 9.7570594259167e-05, "loss": 0.7019, "step": 3774 }, { "epoch": 0.25341431495587396, "grad_norm": 1.1533780097961426, "learning_rate": 9.756724655414807e-05, "loss": 0.706, "step": 3776 }, { "epoch": 0.25354853863964294, "grad_norm": 1.064712405204773, "learning_rate": 9.75638966016703e-05, "loss": 0.7427, "step": 3778 }, { "epoch": 0.25368276232341197, "grad_norm": 1.1156611442565918, "learning_rate": 9.756054440189191e-05, "loss": 0.6738, "step": 3780 }, { "epoch": 0.25381698600718094, "grad_norm": 0.9972400665283203, "learning_rate": 9.755718995497136e-05, "loss": 0.6887, "step": 3782 }, { "epoch": 0.25395120969095, "grad_norm": 1.187756896018982, "learning_rate": 9.755383326106709e-05, "loss": 0.7279, "step": 3784 }, { "epoch": 0.25408543337471895, "grad_norm": 1.1233904361724854, "learning_rate": 9.75504743203377e-05, "loss": 0.7038, "step": 3786 }, { "epoch": 0.254219657058488, "grad_norm": 1.1424480676651, "learning_rate": 9.75471131329419e-05, "loss": 0.6967, "step": 3788 }, { "epoch": 0.25435388074225695, "grad_norm": 1.0352827310562134, "learning_rate": 9.754374969903852e-05, "loss": 0.6602, "step": 3790 }, { "epoch": 0.254488104426026, "grad_norm": 1.3024020195007324, "learning_rate": 9.754038401878645e-05, "loss": 0.7572, "step": 3792 }, { "epoch": 0.25462232810979496, "grad_norm": 1.093445062637329, "learning_rate": 9.753701609234471e-05, "loss": 0.751, "step": 3794 }, { "epoch": 0.254756551793564, "grad_norm": 1.159218192100525, "learning_rate": 9.753364591987244e-05, "loss": 0.7295, "step": 3796 }, { "epoch": 0.25489077547733296, "grad_norm": 1.127556562423706, "learning_rate": 9.753027350152888e-05, "loss": 0.7962, "step": 3798 }, { "epoch": 0.255024999161102, "grad_norm": 1.3137211799621582, "learning_rate": 9.752689883747335e-05, "loss": 0.738, "step": 3800 }, { "epoch": 0.25515922284487097, "grad_norm": 1.182447910308838, "learning_rate": 9.752352192786531e-05, "loss": 0.7343, "step": 3802 }, { "epoch": 0.25529344652864, "grad_norm": 1.095441222190857, "learning_rate": 9.752014277286432e-05, "loss": 0.6628, "step": 3804 }, { "epoch": 0.255427670212409, "grad_norm": 1.229159951210022, "learning_rate": 9.751676137263002e-05, "loss": 0.6822, "step": 3806 }, { "epoch": 0.255561893896178, "grad_norm": 1.1966331005096436, "learning_rate": 9.751337772732218e-05, "loss": 0.7565, "step": 3808 }, { "epoch": 0.255696117579947, "grad_norm": 1.3662197589874268, "learning_rate": 9.750999183710068e-05, "loss": 0.7229, "step": 3810 }, { "epoch": 0.255830341263716, "grad_norm": 1.0455504655838013, "learning_rate": 9.750660370212549e-05, "loss": 0.6916, "step": 3812 }, { "epoch": 0.255964564947485, "grad_norm": 1.1308659315109253, "learning_rate": 9.75032133225567e-05, "loss": 0.7304, "step": 3814 }, { "epoch": 0.25609878863125396, "grad_norm": 1.267609715461731, "learning_rate": 9.749982069855448e-05, "loss": 0.6578, "step": 3816 }, { "epoch": 0.256233012315023, "grad_norm": 1.3140232563018799, "learning_rate": 9.749642583027914e-05, "loss": 0.7273, "step": 3818 }, { "epoch": 0.25636723599879196, "grad_norm": 1.0035659074783325, "learning_rate": 9.749302871789107e-05, "loss": 0.6648, "step": 3820 }, { "epoch": 0.256501459682561, "grad_norm": 1.5064566135406494, "learning_rate": 9.748962936155079e-05, "loss": 0.8053, "step": 3822 }, { "epoch": 0.25663568336632997, "grad_norm": 1.0891283750534058, "learning_rate": 9.748622776141892e-05, "loss": 0.7495, "step": 3824 }, { "epoch": 0.256769907050099, "grad_norm": 1.0988272428512573, "learning_rate": 9.748282391765615e-05, "loss": 0.6811, "step": 3826 }, { "epoch": 0.256904130733868, "grad_norm": 1.0769731998443604, "learning_rate": 9.747941783042332e-05, "loss": 0.7194, "step": 3828 }, { "epoch": 0.257038354417637, "grad_norm": 1.1300938129425049, "learning_rate": 9.747600949988136e-05, "loss": 0.7761, "step": 3830 }, { "epoch": 0.257172578101406, "grad_norm": 1.0362656116485596, "learning_rate": 9.747259892619132e-05, "loss": 0.7048, "step": 3832 }, { "epoch": 0.257306801785175, "grad_norm": 1.1735504865646362, "learning_rate": 9.746918610951433e-05, "loss": 0.7157, "step": 3834 }, { "epoch": 0.257441025468944, "grad_norm": 1.0931183099746704, "learning_rate": 9.746577105001163e-05, "loss": 0.6933, "step": 3836 }, { "epoch": 0.257575249152713, "grad_norm": 1.193933367729187, "learning_rate": 9.74623537478446e-05, "loss": 0.7582, "step": 3838 }, { "epoch": 0.257709472836482, "grad_norm": 1.1191264390945435, "learning_rate": 9.745893420317469e-05, "loss": 0.6414, "step": 3840 }, { "epoch": 0.257843696520251, "grad_norm": 1.2166557312011719, "learning_rate": 9.745551241616344e-05, "loss": 0.7545, "step": 3842 }, { "epoch": 0.25797792020402, "grad_norm": 1.238582730293274, "learning_rate": 9.745208838697255e-05, "loss": 0.7233, "step": 3844 }, { "epoch": 0.258112143887789, "grad_norm": 1.1257667541503906, "learning_rate": 9.744866211576381e-05, "loss": 0.7227, "step": 3846 }, { "epoch": 0.258246367571558, "grad_norm": 1.0283488035202026, "learning_rate": 9.744523360269909e-05, "loss": 0.6334, "step": 3848 }, { "epoch": 0.25838059125532703, "grad_norm": 1.229082703590393, "learning_rate": 9.744180284794035e-05, "loss": 0.7463, "step": 3850 }, { "epoch": 0.258514814939096, "grad_norm": 1.1621450185775757, "learning_rate": 9.743836985164974e-05, "loss": 0.7193, "step": 3852 }, { "epoch": 0.258649038622865, "grad_norm": 1.0491615533828735, "learning_rate": 9.743493461398942e-05, "loss": 0.7096, "step": 3854 }, { "epoch": 0.258783262306634, "grad_norm": 1.0686851739883423, "learning_rate": 9.743149713512175e-05, "loss": 0.7385, "step": 3856 }, { "epoch": 0.258917485990403, "grad_norm": 1.1188812255859375, "learning_rate": 9.742805741520908e-05, "loss": 0.7903, "step": 3858 }, { "epoch": 0.259051709674172, "grad_norm": 1.0035182237625122, "learning_rate": 9.742461545441398e-05, "loss": 0.7312, "step": 3860 }, { "epoch": 0.259185933357941, "grad_norm": 1.1080131530761719, "learning_rate": 9.742117125289904e-05, "loss": 0.7621, "step": 3862 }, { "epoch": 0.25932015704171, "grad_norm": 1.1767661571502686, "learning_rate": 9.741772481082702e-05, "loss": 0.6926, "step": 3864 }, { "epoch": 0.259454380725479, "grad_norm": 1.1401045322418213, "learning_rate": 9.741427612836074e-05, "loss": 0.7, "step": 3866 }, { "epoch": 0.259588604409248, "grad_norm": 1.1692506074905396, "learning_rate": 9.741082520566314e-05, "loss": 0.7517, "step": 3868 }, { "epoch": 0.259722828093017, "grad_norm": 1.2039555311203003, "learning_rate": 9.740737204289729e-05, "loss": 0.7437, "step": 3870 }, { "epoch": 0.25985705177678603, "grad_norm": 1.1465946435928345, "learning_rate": 9.740391664022633e-05, "loss": 0.7207, "step": 3872 }, { "epoch": 0.259991275460555, "grad_norm": 1.2403197288513184, "learning_rate": 9.740045899781352e-05, "loss": 0.7601, "step": 3874 }, { "epoch": 0.26012549914432403, "grad_norm": 1.3701558113098145, "learning_rate": 9.739699911582225e-05, "loss": 0.8242, "step": 3876 }, { "epoch": 0.260259722828093, "grad_norm": 1.1910128593444824, "learning_rate": 9.739353699441596e-05, "loss": 0.6879, "step": 3878 }, { "epoch": 0.26039394651186204, "grad_norm": 1.1024234294891357, "learning_rate": 9.739007263375823e-05, "loss": 0.7299, "step": 3880 }, { "epoch": 0.260528170195631, "grad_norm": 1.1535654067993164, "learning_rate": 9.738660603401277e-05, "loss": 0.7267, "step": 3882 }, { "epoch": 0.26066239387940005, "grad_norm": 1.0920493602752686, "learning_rate": 9.738313719534337e-05, "loss": 0.7854, "step": 3884 }, { "epoch": 0.260796617563169, "grad_norm": 1.7549275159835815, "learning_rate": 9.73796661179139e-05, "loss": 0.682, "step": 3886 }, { "epoch": 0.260930841246938, "grad_norm": 1.212325930595398, "learning_rate": 9.737619280188837e-05, "loss": 0.7691, "step": 3888 }, { "epoch": 0.261065064930707, "grad_norm": 1.0737411975860596, "learning_rate": 9.737271724743088e-05, "loss": 0.652, "step": 3890 }, { "epoch": 0.261199288614476, "grad_norm": 1.1670502424240112, "learning_rate": 9.736923945470568e-05, "loss": 0.7321, "step": 3892 }, { "epoch": 0.26133351229824503, "grad_norm": 1.2263405323028564, "learning_rate": 9.736575942387706e-05, "loss": 0.7116, "step": 3894 }, { "epoch": 0.261467735982014, "grad_norm": 1.1557703018188477, "learning_rate": 9.736227715510944e-05, "loss": 0.7032, "step": 3896 }, { "epoch": 0.26160195966578303, "grad_norm": 1.3054964542388916, "learning_rate": 9.735879264856736e-05, "loss": 0.7081, "step": 3898 }, { "epoch": 0.261736183349552, "grad_norm": 1.119599461555481, "learning_rate": 9.735530590441545e-05, "loss": 0.6495, "step": 3900 }, { "epoch": 0.26187040703332104, "grad_norm": 1.9122085571289062, "learning_rate": 9.735181692281846e-05, "loss": 0.7446, "step": 3902 }, { "epoch": 0.26200463071709, "grad_norm": 1.135762095451355, "learning_rate": 9.734832570394124e-05, "loss": 0.7072, "step": 3904 }, { "epoch": 0.26213885440085904, "grad_norm": 1.079358696937561, "learning_rate": 9.734483224794872e-05, "loss": 0.764, "step": 3906 }, { "epoch": 0.262273078084628, "grad_norm": 1.1305729150772095, "learning_rate": 9.7341336555006e-05, "loss": 0.7986, "step": 3908 }, { "epoch": 0.26240730176839705, "grad_norm": 1.1853477954864502, "learning_rate": 9.73378386252782e-05, "loss": 0.6972, "step": 3910 }, { "epoch": 0.262541525452166, "grad_norm": 1.19010329246521, "learning_rate": 9.73343384589306e-05, "loss": 0.6651, "step": 3912 }, { "epoch": 0.26267574913593505, "grad_norm": 1.41134774684906, "learning_rate": 9.733083605612863e-05, "loss": 0.7109, "step": 3914 }, { "epoch": 0.26280997281970403, "grad_norm": 1.4223922491073608, "learning_rate": 9.732733141703769e-05, "loss": 0.6819, "step": 3916 }, { "epoch": 0.26294419650347306, "grad_norm": 1.5214942693710327, "learning_rate": 9.732382454182343e-05, "loss": 0.8066, "step": 3918 }, { "epoch": 0.26307842018724203, "grad_norm": 1.2094883918762207, "learning_rate": 9.73203154306515e-05, "loss": 0.6308, "step": 3920 }, { "epoch": 0.26321264387101106, "grad_norm": 1.1626900434494019, "learning_rate": 9.731680408368772e-05, "loss": 0.6826, "step": 3922 }, { "epoch": 0.26334686755478004, "grad_norm": 1.4367871284484863, "learning_rate": 9.7313290501098e-05, "loss": 0.7448, "step": 3924 }, { "epoch": 0.263481091238549, "grad_norm": 1.5802116394042969, "learning_rate": 9.730977468304834e-05, "loss": 0.7858, "step": 3926 }, { "epoch": 0.26361531492231804, "grad_norm": 1.0798490047454834, "learning_rate": 9.730625662970485e-05, "loss": 0.7485, "step": 3928 }, { "epoch": 0.263749538606087, "grad_norm": 1.8378727436065674, "learning_rate": 9.730273634123377e-05, "loss": 0.6474, "step": 3930 }, { "epoch": 0.26388376228985605, "grad_norm": 1.8537728786468506, "learning_rate": 9.72992138178014e-05, "loss": 0.7069, "step": 3932 }, { "epoch": 0.264017985973625, "grad_norm": 1.439657211303711, "learning_rate": 9.72956890595742e-05, "loss": 0.6436, "step": 3934 }, { "epoch": 0.26415220965739405, "grad_norm": 1.0422866344451904, "learning_rate": 9.729216206671868e-05, "loss": 0.6997, "step": 3936 }, { "epoch": 0.26428643334116303, "grad_norm": 1.206998586654663, "learning_rate": 9.728863283940151e-05, "loss": 0.7195, "step": 3938 }, { "epoch": 0.26442065702493206, "grad_norm": 1.0577679872512817, "learning_rate": 9.728510137778944e-05, "loss": 0.669, "step": 3940 }, { "epoch": 0.26455488070870103, "grad_norm": 1.1963422298431396, "learning_rate": 9.728156768204928e-05, "loss": 0.7314, "step": 3942 }, { "epoch": 0.26468910439247006, "grad_norm": 1.545266032218933, "learning_rate": 9.727803175234804e-05, "loss": 0.7109, "step": 3944 }, { "epoch": 0.26482332807623904, "grad_norm": 1.0092322826385498, "learning_rate": 9.727449358885276e-05, "loss": 0.753, "step": 3946 }, { "epoch": 0.26495755176000807, "grad_norm": 1.293685793876648, "learning_rate": 9.727095319173065e-05, "loss": 0.7639, "step": 3948 }, { "epoch": 0.26509177544377704, "grad_norm": 1.1440637111663818, "learning_rate": 9.726741056114892e-05, "loss": 0.8083, "step": 3950 }, { "epoch": 0.2652259991275461, "grad_norm": 1.0889347791671753, "learning_rate": 9.726386569727501e-05, "loss": 0.6597, "step": 3952 }, { "epoch": 0.26536022281131505, "grad_norm": 1.1356900930404663, "learning_rate": 9.726031860027637e-05, "loss": 0.7465, "step": 3954 }, { "epoch": 0.2654944464950841, "grad_norm": 1.1200767755508423, "learning_rate": 9.725676927032061e-05, "loss": 0.686, "step": 3956 }, { "epoch": 0.26562867017885305, "grad_norm": 1.100643515586853, "learning_rate": 9.725321770757545e-05, "loss": 0.7333, "step": 3958 }, { "epoch": 0.2657628938626221, "grad_norm": 1.022718906402588, "learning_rate": 9.724966391220865e-05, "loss": 0.792, "step": 3960 }, { "epoch": 0.26589711754639106, "grad_norm": 1.3356891870498657, "learning_rate": 9.724610788438815e-05, "loss": 0.709, "step": 3962 }, { "epoch": 0.26603134123016003, "grad_norm": 1.2211649417877197, "learning_rate": 9.724254962428196e-05, "loss": 0.6901, "step": 3964 }, { "epoch": 0.26616556491392906, "grad_norm": 1.1071388721466064, "learning_rate": 9.72389891320582e-05, "loss": 0.6893, "step": 3966 }, { "epoch": 0.26629978859769804, "grad_norm": 1.1717225313186646, "learning_rate": 9.723542640788509e-05, "loss": 0.7193, "step": 3968 }, { "epoch": 0.26643401228146707, "grad_norm": 1.0993868112564087, "learning_rate": 9.723186145193097e-05, "loss": 0.7268, "step": 3970 }, { "epoch": 0.26656823596523604, "grad_norm": 1.0340200662612915, "learning_rate": 9.722829426436427e-05, "loss": 0.7291, "step": 3972 }, { "epoch": 0.2667024596490051, "grad_norm": 0.938714325428009, "learning_rate": 9.722472484535354e-05, "loss": 0.716, "step": 3974 }, { "epoch": 0.26683668333277405, "grad_norm": 1.1499898433685303, "learning_rate": 9.722115319506743e-05, "loss": 0.7458, "step": 3976 }, { "epoch": 0.2669709070165431, "grad_norm": 1.1436411142349243, "learning_rate": 9.721757931367468e-05, "loss": 0.7299, "step": 3978 }, { "epoch": 0.26710513070031205, "grad_norm": 1.2237372398376465, "learning_rate": 9.721400320134415e-05, "loss": 0.7948, "step": 3980 }, { "epoch": 0.2672393543840811, "grad_norm": 1.1530967950820923, "learning_rate": 9.721042485824483e-05, "loss": 0.7577, "step": 3982 }, { "epoch": 0.26737357806785006, "grad_norm": 1.2369847297668457, "learning_rate": 9.720684428454576e-05, "loss": 0.6774, "step": 3984 }, { "epoch": 0.2675078017516191, "grad_norm": 1.1894441843032837, "learning_rate": 9.720326148041612e-05, "loss": 0.7308, "step": 3986 }, { "epoch": 0.26764202543538806, "grad_norm": 1.0823352336883545, "learning_rate": 9.719967644602521e-05, "loss": 0.7019, "step": 3988 }, { "epoch": 0.2677762491191571, "grad_norm": 0.9706602096557617, "learning_rate": 9.71960891815424e-05, "loss": 0.6352, "step": 3990 }, { "epoch": 0.26791047280292607, "grad_norm": 1.227729082107544, "learning_rate": 9.719249968713717e-05, "loss": 0.7034, "step": 3992 }, { "epoch": 0.2680446964866951, "grad_norm": 1.4342589378356934, "learning_rate": 9.718890796297914e-05, "loss": 0.7531, "step": 3994 }, { "epoch": 0.2681789201704641, "grad_norm": 1.027266025543213, "learning_rate": 9.7185314009238e-05, "loss": 0.6469, "step": 3996 }, { "epoch": 0.2683131438542331, "grad_norm": 1.0618791580200195, "learning_rate": 9.718171782608356e-05, "loss": 0.7386, "step": 3998 }, { "epoch": 0.2684473675380021, "grad_norm": 1.1263371706008911, "learning_rate": 9.717811941368574e-05, "loss": 0.696, "step": 4000 }, { "epoch": 0.26858159122177105, "grad_norm": 1.2227270603179932, "learning_rate": 9.717451877221453e-05, "loss": 0.7411, "step": 4002 }, { "epoch": 0.2687158149055401, "grad_norm": 1.1450586318969727, "learning_rate": 9.717091590184008e-05, "loss": 0.7522, "step": 4004 }, { "epoch": 0.26885003858930906, "grad_norm": 1.0368503332138062, "learning_rate": 9.71673108027326e-05, "loss": 0.718, "step": 4006 }, { "epoch": 0.2689842622730781, "grad_norm": 1.1287481784820557, "learning_rate": 9.716370347506247e-05, "loss": 0.7373, "step": 4008 }, { "epoch": 0.26911848595684706, "grad_norm": 1.2839189767837524, "learning_rate": 9.716009391900006e-05, "loss": 0.6749, "step": 4010 }, { "epoch": 0.2692527096406161, "grad_norm": 1.0585741996765137, "learning_rate": 9.715648213471597e-05, "loss": 0.7737, "step": 4012 }, { "epoch": 0.26938693332438507, "grad_norm": 0.9859800338745117, "learning_rate": 9.715286812238082e-05, "loss": 0.6506, "step": 4014 }, { "epoch": 0.2695211570081541, "grad_norm": 1.1902844905853271, "learning_rate": 9.714925188216537e-05, "loss": 0.8032, "step": 4016 }, { "epoch": 0.2696553806919231, "grad_norm": 2.5191569328308105, "learning_rate": 9.714563341424048e-05, "loss": 0.7063, "step": 4018 }, { "epoch": 0.2697896043756921, "grad_norm": 1.1799323558807373, "learning_rate": 9.714201271877713e-05, "loss": 0.7384, "step": 4020 }, { "epoch": 0.2699238280594611, "grad_norm": 1.308539867401123, "learning_rate": 9.713838979594638e-05, "loss": 0.6858, "step": 4022 }, { "epoch": 0.2700580517432301, "grad_norm": 1.099339485168457, "learning_rate": 9.71347646459194e-05, "loss": 0.679, "step": 4024 }, { "epoch": 0.2701922754269991, "grad_norm": 1.1468719244003296, "learning_rate": 9.713113726886747e-05, "loss": 0.6644, "step": 4026 }, { "epoch": 0.2703264991107681, "grad_norm": 1.1587730646133423, "learning_rate": 9.712750766496201e-05, "loss": 0.6816, "step": 4028 }, { "epoch": 0.2704607227945371, "grad_norm": 1.0334495306015015, "learning_rate": 9.712387583437445e-05, "loss": 0.7228, "step": 4030 }, { "epoch": 0.2705949464783061, "grad_norm": 1.2018488645553589, "learning_rate": 9.712024177727645e-05, "loss": 0.7465, "step": 4032 }, { "epoch": 0.2707291701620751, "grad_norm": 1.5213717222213745, "learning_rate": 9.711660549383967e-05, "loss": 0.7471, "step": 4034 }, { "epoch": 0.2708633938458441, "grad_norm": 1.8163784742355347, "learning_rate": 9.711296698423593e-05, "loss": 0.7182, "step": 4036 }, { "epoch": 0.2709976175296131, "grad_norm": 1.145598292350769, "learning_rate": 9.710932624863715e-05, "loss": 0.7038, "step": 4038 }, { "epoch": 0.2711318412133821, "grad_norm": 1.237273097038269, "learning_rate": 9.710568328721534e-05, "loss": 0.7144, "step": 4040 }, { "epoch": 0.2712660648971511, "grad_norm": 1.2341762781143188, "learning_rate": 9.710203810014262e-05, "loss": 0.778, "step": 4042 }, { "epoch": 0.2714002885809201, "grad_norm": 1.0736732482910156, "learning_rate": 9.709839068759123e-05, "loss": 0.6782, "step": 4044 }, { "epoch": 0.2715345122646891, "grad_norm": 1.2260222434997559, "learning_rate": 9.70947410497335e-05, "loss": 0.6977, "step": 4046 }, { "epoch": 0.2716687359484581, "grad_norm": 1.1734833717346191, "learning_rate": 9.709108918674185e-05, "loss": 0.7427, "step": 4048 }, { "epoch": 0.2718029596322271, "grad_norm": 1.1103187799453735, "learning_rate": 9.708743509878884e-05, "loss": 0.7538, "step": 4050 }, { "epoch": 0.2719371833159961, "grad_norm": 1.0368175506591797, "learning_rate": 9.70837787860471e-05, "loss": 0.7464, "step": 4052 }, { "epoch": 0.2720714069997651, "grad_norm": 1.0606738328933716, "learning_rate": 9.708012024868942e-05, "loss": 0.6899, "step": 4054 }, { "epoch": 0.2722056306835341, "grad_norm": 1.254096269607544, "learning_rate": 9.707645948688863e-05, "loss": 0.7723, "step": 4056 }, { "epoch": 0.2723398543673031, "grad_norm": 1.1519017219543457, "learning_rate": 9.70727965008177e-05, "loss": 0.6875, "step": 4058 }, { "epoch": 0.2724740780510721, "grad_norm": 2.8999814987182617, "learning_rate": 9.706913129064971e-05, "loss": 0.7196, "step": 4060 }, { "epoch": 0.27260830173484113, "grad_norm": 1.135886549949646, "learning_rate": 9.706546385655781e-05, "loss": 0.7366, "step": 4062 }, { "epoch": 0.2727425254186101, "grad_norm": 1.0506887435913086, "learning_rate": 9.706179419871531e-05, "loss": 0.6983, "step": 4064 }, { "epoch": 0.27287674910237913, "grad_norm": 1.1773637533187866, "learning_rate": 9.705812231729557e-05, "loss": 0.7328, "step": 4066 }, { "epoch": 0.2730109727861481, "grad_norm": 2.1838226318359375, "learning_rate": 9.705444821247208e-05, "loss": 0.7691, "step": 4068 }, { "epoch": 0.27314519646991714, "grad_norm": 1.373435139656067, "learning_rate": 9.705077188441844e-05, "loss": 0.6969, "step": 4070 }, { "epoch": 0.2732794201536861, "grad_norm": 1.2124875783920288, "learning_rate": 9.704709333330836e-05, "loss": 0.711, "step": 4072 }, { "epoch": 0.27341364383745514, "grad_norm": 1.3007575273513794, "learning_rate": 9.704341255931562e-05, "loss": 0.6932, "step": 4074 }, { "epoch": 0.2735478675212241, "grad_norm": 1.0921796560287476, "learning_rate": 9.703972956261416e-05, "loss": 0.7336, "step": 4076 }, { "epoch": 0.2736820912049931, "grad_norm": 1.2120072841644287, "learning_rate": 9.703604434337797e-05, "loss": 0.7159, "step": 4078 }, { "epoch": 0.2738163148887621, "grad_norm": 1.47358238697052, "learning_rate": 9.703235690178118e-05, "loss": 0.7175, "step": 4080 }, { "epoch": 0.2739505385725311, "grad_norm": 1.1698148250579834, "learning_rate": 9.7028667237998e-05, "loss": 0.6971, "step": 4082 }, { "epoch": 0.27408476225630013, "grad_norm": 1.2034008502960205, "learning_rate": 9.702497535220278e-05, "loss": 0.769, "step": 4084 }, { "epoch": 0.2742189859400691, "grad_norm": 1.183003544807434, "learning_rate": 9.702128124456996e-05, "loss": 0.7371, "step": 4086 }, { "epoch": 0.27435320962383813, "grad_norm": 1.3367705345153809, "learning_rate": 9.701758491527404e-05, "loss": 0.8031, "step": 4088 }, { "epoch": 0.2744874333076071, "grad_norm": 1.158776044845581, "learning_rate": 9.701388636448969e-05, "loss": 0.7452, "step": 4090 }, { "epoch": 0.27462165699137614, "grad_norm": 1.1235275268554688, "learning_rate": 9.701018559239167e-05, "loss": 0.7204, "step": 4092 }, { "epoch": 0.2747558806751451, "grad_norm": 1.0379655361175537, "learning_rate": 9.700648259915481e-05, "loss": 0.6754, "step": 4094 }, { "epoch": 0.27489010435891414, "grad_norm": 1.151607632637024, "learning_rate": 9.700277738495409e-05, "loss": 0.7668, "step": 4096 }, { "epoch": 0.2750243280426831, "grad_norm": 1.1385493278503418, "learning_rate": 9.699906994996457e-05, "loss": 0.7487, "step": 4098 }, { "epoch": 0.27515855172645215, "grad_norm": 1.1888922452926636, "learning_rate": 9.69953602943614e-05, "loss": 0.7085, "step": 4100 }, { "epoch": 0.2752927754102211, "grad_norm": 1.007498860359192, "learning_rate": 9.699164841831989e-05, "loss": 0.7052, "step": 4102 }, { "epoch": 0.27542699909399015, "grad_norm": 1.2998079061508179, "learning_rate": 9.698793432201538e-05, "loss": 0.7781, "step": 4104 }, { "epoch": 0.27556122277775913, "grad_norm": 1.1245256662368774, "learning_rate": 9.698421800562338e-05, "loss": 0.7311, "step": 4106 }, { "epoch": 0.27569544646152816, "grad_norm": 1.0242562294006348, "learning_rate": 9.698049946931947e-05, "loss": 0.7246, "step": 4108 }, { "epoch": 0.27582967014529713, "grad_norm": 1.1672991514205933, "learning_rate": 9.697677871327933e-05, "loss": 0.6993, "step": 4110 }, { "epoch": 0.27596389382906616, "grad_norm": 1.0333305597305298, "learning_rate": 9.697305573767879e-05, "loss": 0.7259, "step": 4112 }, { "epoch": 0.27609811751283514, "grad_norm": 1.0109241008758545, "learning_rate": 9.696933054269372e-05, "loss": 0.6522, "step": 4114 }, { "epoch": 0.2762323411966041, "grad_norm": 1.0484569072723389, "learning_rate": 9.696560312850015e-05, "loss": 0.7308, "step": 4116 }, { "epoch": 0.27636656488037314, "grad_norm": 1.2440683841705322, "learning_rate": 9.69618734952742e-05, "loss": 0.7118, "step": 4118 }, { "epoch": 0.2765007885641421, "grad_norm": 1.4809832572937012, "learning_rate": 9.695814164319204e-05, "loss": 0.7568, "step": 4120 }, { "epoch": 0.27663501224791115, "grad_norm": 1.076040267944336, "learning_rate": 9.695440757243005e-05, "loss": 0.6586, "step": 4122 }, { "epoch": 0.2767692359316801, "grad_norm": 1.0714213848114014, "learning_rate": 9.695067128316463e-05, "loss": 0.6833, "step": 4124 }, { "epoch": 0.27690345961544915, "grad_norm": 1.0056148767471313, "learning_rate": 9.694693277557232e-05, "loss": 0.681, "step": 4126 }, { "epoch": 0.27703768329921813, "grad_norm": 1.216422438621521, "learning_rate": 9.694319204982974e-05, "loss": 0.7266, "step": 4128 }, { "epoch": 0.27717190698298716, "grad_norm": 1.3905694484710693, "learning_rate": 9.693944910611365e-05, "loss": 0.6919, "step": 4130 }, { "epoch": 0.27730613066675613, "grad_norm": 1.0062493085861206, "learning_rate": 9.693570394460091e-05, "loss": 0.6494, "step": 4132 }, { "epoch": 0.27744035435052516, "grad_norm": 1.5734044313430786, "learning_rate": 9.693195656546843e-05, "loss": 0.685, "step": 4134 }, { "epoch": 0.27757457803429414, "grad_norm": 1.1906888484954834, "learning_rate": 9.692820696889331e-05, "loss": 0.7182, "step": 4136 }, { "epoch": 0.27770880171806317, "grad_norm": 1.1630241870880127, "learning_rate": 9.692445515505268e-05, "loss": 0.6852, "step": 4138 }, { "epoch": 0.27784302540183214, "grad_norm": 1.123489499092102, "learning_rate": 9.692070112412382e-05, "loss": 0.6891, "step": 4140 }, { "epoch": 0.2779772490856012, "grad_norm": 1.1171189546585083, "learning_rate": 9.69169448762841e-05, "loss": 0.7437, "step": 4142 }, { "epoch": 0.27811147276937015, "grad_norm": 1.5453418493270874, "learning_rate": 9.691318641171099e-05, "loss": 0.7297, "step": 4144 }, { "epoch": 0.2782456964531392, "grad_norm": 1.3269058465957642, "learning_rate": 9.690942573058207e-05, "loss": 0.6759, "step": 4146 }, { "epoch": 0.27837992013690815, "grad_norm": 1.1371320486068726, "learning_rate": 9.690566283307503e-05, "loss": 0.707, "step": 4148 }, { "epoch": 0.2785141438206772, "grad_norm": 1.9872580766677856, "learning_rate": 9.690189771936766e-05, "loss": 0.6828, "step": 4150 }, { "epoch": 0.27864836750444616, "grad_norm": 1.037161111831665, "learning_rate": 9.689813038963784e-05, "loss": 0.6583, "step": 4152 }, { "epoch": 0.27878259118821513, "grad_norm": 1.2190905809402466, "learning_rate": 9.68943608440636e-05, "loss": 0.7327, "step": 4154 }, { "epoch": 0.27891681487198416, "grad_norm": 1.0231003761291504, "learning_rate": 9.6890589082823e-05, "loss": 0.714, "step": 4156 }, { "epoch": 0.27905103855575314, "grad_norm": 1.1530766487121582, "learning_rate": 9.68868151060943e-05, "loss": 0.7366, "step": 4158 }, { "epoch": 0.27918526223952217, "grad_norm": 1.200112223625183, "learning_rate": 9.688303891405576e-05, "loss": 0.7856, "step": 4160 }, { "epoch": 0.27931948592329114, "grad_norm": 1.2123029232025146, "learning_rate": 9.687926050688583e-05, "loss": 0.748, "step": 4162 }, { "epoch": 0.2794537096070602, "grad_norm": 1.2193679809570312, "learning_rate": 9.687547988476303e-05, "loss": 0.7167, "step": 4164 }, { "epoch": 0.27958793329082915, "grad_norm": 1.103472113609314, "learning_rate": 9.687169704786599e-05, "loss": 0.683, "step": 4166 }, { "epoch": 0.2797221569745982, "grad_norm": 1.1211434602737427, "learning_rate": 9.686791199637342e-05, "loss": 0.6931, "step": 4168 }, { "epoch": 0.27985638065836715, "grad_norm": 1.15460205078125, "learning_rate": 9.686412473046418e-05, "loss": 0.7261, "step": 4170 }, { "epoch": 0.2799906043421362, "grad_norm": 0.9894868731498718, "learning_rate": 9.686033525031719e-05, "loss": 0.7467, "step": 4172 }, { "epoch": 0.28012482802590516, "grad_norm": 1.2631585597991943, "learning_rate": 9.685654355611151e-05, "loss": 0.7588, "step": 4174 }, { "epoch": 0.2802590517096742, "grad_norm": 1.0129464864730835, "learning_rate": 9.685274964802629e-05, "loss": 0.75, "step": 4176 }, { "epoch": 0.28039327539344316, "grad_norm": 1.4937212467193604, "learning_rate": 9.684895352624077e-05, "loss": 0.667, "step": 4178 }, { "epoch": 0.2805274990772122, "grad_norm": 1.0413357019424438, "learning_rate": 9.684515519093433e-05, "loss": 0.6708, "step": 4180 }, { "epoch": 0.28066172276098117, "grad_norm": 1.2190228700637817, "learning_rate": 9.684135464228643e-05, "loss": 0.7621, "step": 4182 }, { "epoch": 0.2807959464447502, "grad_norm": 1.024556040763855, "learning_rate": 9.683755188047663e-05, "loss": 0.701, "step": 4184 }, { "epoch": 0.2809301701285192, "grad_norm": 1.073548436164856, "learning_rate": 9.68337469056846e-05, "loss": 0.7019, "step": 4186 }, { "epoch": 0.2810643938122882, "grad_norm": 1.0884594917297363, "learning_rate": 9.682993971809012e-05, "loss": 0.7868, "step": 4188 }, { "epoch": 0.2811986174960572, "grad_norm": 1.0661922693252563, "learning_rate": 9.682613031787308e-05, "loss": 0.7261, "step": 4190 }, { "epoch": 0.28133284117982615, "grad_norm": 1.2550932168960571, "learning_rate": 9.682231870521347e-05, "loss": 0.726, "step": 4192 }, { "epoch": 0.2814670648635952, "grad_norm": 1.5789484977722168, "learning_rate": 9.681850488029136e-05, "loss": 0.6917, "step": 4194 }, { "epoch": 0.28160128854736416, "grad_norm": 1.1192291975021362, "learning_rate": 9.681468884328697e-05, "loss": 0.6583, "step": 4196 }, { "epoch": 0.2817355122311332, "grad_norm": 1.3032783269882202, "learning_rate": 9.681087059438059e-05, "loss": 0.6452, "step": 4198 }, { "epoch": 0.28186973591490216, "grad_norm": 1.3085541725158691, "learning_rate": 9.680705013375261e-05, "loss": 0.7515, "step": 4200 }, { "epoch": 0.2820039595986712, "grad_norm": 1.0685060024261475, "learning_rate": 9.680322746158357e-05, "loss": 0.7283, "step": 4202 }, { "epoch": 0.28213818328244017, "grad_norm": 0.9785438179969788, "learning_rate": 9.679940257805404e-05, "loss": 0.7339, "step": 4204 }, { "epoch": 0.2822724069662092, "grad_norm": 1.0365419387817383, "learning_rate": 9.679557548334479e-05, "loss": 0.6785, "step": 4206 }, { "epoch": 0.2824066306499782, "grad_norm": 1.2138546705245972, "learning_rate": 9.679174617763662e-05, "loss": 0.7403, "step": 4208 }, { "epoch": 0.2825408543337472, "grad_norm": 1.4012144804000854, "learning_rate": 9.678791466111044e-05, "loss": 0.8049, "step": 4210 }, { "epoch": 0.2826750780175162, "grad_norm": 1.1018271446228027, "learning_rate": 9.67840809339473e-05, "loss": 0.7087, "step": 4212 }, { "epoch": 0.2828093017012852, "grad_norm": 1.0963165760040283, "learning_rate": 9.678024499632834e-05, "loss": 0.7312, "step": 4214 }, { "epoch": 0.2829435253850542, "grad_norm": 1.218801498413086, "learning_rate": 9.677640684843478e-05, "loss": 0.7484, "step": 4216 }, { "epoch": 0.2830777490688232, "grad_norm": 1.1199160814285278, "learning_rate": 9.6772566490448e-05, "loss": 0.6566, "step": 4218 }, { "epoch": 0.2832119727525922, "grad_norm": 1.5163363218307495, "learning_rate": 9.676872392254941e-05, "loss": 0.6935, "step": 4220 }, { "epoch": 0.2833461964363612, "grad_norm": 1.098585844039917, "learning_rate": 9.67648791449206e-05, "loss": 0.7089, "step": 4222 }, { "epoch": 0.2834804201201302, "grad_norm": 1.1444215774536133, "learning_rate": 9.676103215774319e-05, "loss": 0.6433, "step": 4224 }, { "epoch": 0.2836146438038992, "grad_norm": 1.0866196155548096, "learning_rate": 9.675718296119898e-05, "loss": 0.7273, "step": 4226 }, { "epoch": 0.2837488674876682, "grad_norm": 1.1252814531326294, "learning_rate": 9.675333155546979e-05, "loss": 0.6798, "step": 4228 }, { "epoch": 0.2838830911714372, "grad_norm": 1.1624395847320557, "learning_rate": 9.674947794073765e-05, "loss": 0.7036, "step": 4230 }, { "epoch": 0.2840173148552062, "grad_norm": 1.0577150583267212, "learning_rate": 9.674562211718459e-05, "loss": 0.62, "step": 4232 }, { "epoch": 0.2841515385389752, "grad_norm": 1.045554280281067, "learning_rate": 9.674176408499281e-05, "loss": 0.6564, "step": 4234 }, { "epoch": 0.2842857622227442, "grad_norm": 1.1526782512664795, "learning_rate": 9.67379038443446e-05, "loss": 0.7117, "step": 4236 }, { "epoch": 0.2844199859065132, "grad_norm": 1.0404531955718994, "learning_rate": 9.673404139542233e-05, "loss": 0.7032, "step": 4238 }, { "epoch": 0.2845542095902822, "grad_norm": 1.041201114654541, "learning_rate": 9.67301767384085e-05, "loss": 0.7019, "step": 4240 }, { "epoch": 0.2846884332740512, "grad_norm": 1.1981208324432373, "learning_rate": 9.672630987348573e-05, "loss": 0.7029, "step": 4242 }, { "epoch": 0.2848226569578202, "grad_norm": 1.1003867387771606, "learning_rate": 9.672244080083668e-05, "loss": 0.7255, "step": 4244 }, { "epoch": 0.2849568806415892, "grad_norm": 2.205414056777954, "learning_rate": 9.671856952064418e-05, "loss": 0.7091, "step": 4246 }, { "epoch": 0.2850911043253582, "grad_norm": 1.0728976726531982, "learning_rate": 9.671469603309113e-05, "loss": 0.617, "step": 4248 }, { "epoch": 0.2852253280091272, "grad_norm": 1.2097114324569702, "learning_rate": 9.671082033836057e-05, "loss": 0.7732, "step": 4250 }, { "epoch": 0.28535955169289623, "grad_norm": 1.173802375793457, "learning_rate": 9.67069424366356e-05, "loss": 0.7405, "step": 4252 }, { "epoch": 0.2854937753766652, "grad_norm": 1.1673674583435059, "learning_rate": 9.670306232809945e-05, "loss": 0.7246, "step": 4254 }, { "epoch": 0.28562799906043423, "grad_norm": 1.9087454080581665, "learning_rate": 9.669918001293543e-05, "loss": 0.7625, "step": 4256 }, { "epoch": 0.2857622227442032, "grad_norm": 1.0080323219299316, "learning_rate": 9.669529549132699e-05, "loss": 0.6792, "step": 4258 }, { "epoch": 0.28589644642797224, "grad_norm": 1.1245241165161133, "learning_rate": 9.669140876345765e-05, "loss": 0.7519, "step": 4260 }, { "epoch": 0.2860306701117412, "grad_norm": 1.1891417503356934, "learning_rate": 9.668751982951108e-05, "loss": 0.7294, "step": 4262 }, { "epoch": 0.28616489379551024, "grad_norm": 0.9736157655715942, "learning_rate": 9.6683628689671e-05, "loss": 0.7211, "step": 4264 }, { "epoch": 0.2862991174792792, "grad_norm": 1.0116606950759888, "learning_rate": 9.667973534412125e-05, "loss": 0.6763, "step": 4266 }, { "epoch": 0.2864333411630482, "grad_norm": 1.6707595586776733, "learning_rate": 9.66758397930458e-05, "loss": 0.7135, "step": 4268 }, { "epoch": 0.2865675648468172, "grad_norm": 1.379185676574707, "learning_rate": 9.667194203662873e-05, "loss": 0.6941, "step": 4270 }, { "epoch": 0.2867017885305862, "grad_norm": 2.073378086090088, "learning_rate": 9.666804207505414e-05, "loss": 0.7383, "step": 4272 }, { "epoch": 0.28683601221435523, "grad_norm": 1.1643736362457275, "learning_rate": 9.666413990850635e-05, "loss": 0.6586, "step": 4274 }, { "epoch": 0.2869702358981242, "grad_norm": 0.9920051693916321, "learning_rate": 9.666023553716971e-05, "loss": 0.6988, "step": 4276 }, { "epoch": 0.28710445958189323, "grad_norm": 1.1142268180847168, "learning_rate": 9.665632896122869e-05, "loss": 0.6993, "step": 4278 }, { "epoch": 0.2872386832656622, "grad_norm": 1.9538670778274536, "learning_rate": 9.665242018086786e-05, "loss": 0.7178, "step": 4280 }, { "epoch": 0.28737290694943124, "grad_norm": 2.505720376968384, "learning_rate": 9.664850919627193e-05, "loss": 0.7012, "step": 4282 }, { "epoch": 0.2875071306332002, "grad_norm": 0.9926694631576538, "learning_rate": 9.664459600762568e-05, "loss": 0.7295, "step": 4284 }, { "epoch": 0.28764135431696924, "grad_norm": 0.9303231239318848, "learning_rate": 9.664068061511397e-05, "loss": 0.6918, "step": 4286 }, { "epoch": 0.2877755780007382, "grad_norm": 1.1642636060714722, "learning_rate": 9.663676301892182e-05, "loss": 0.7, "step": 4288 }, { "epoch": 0.28790980168450725, "grad_norm": 1.8136674165725708, "learning_rate": 9.663284321923434e-05, "loss": 0.71, "step": 4290 }, { "epoch": 0.2880440253682762, "grad_norm": 2.031097888946533, "learning_rate": 9.66289212162367e-05, "loss": 0.785, "step": 4292 }, { "epoch": 0.28817824905204525, "grad_norm": 1.2717703580856323, "learning_rate": 9.662499701011424e-05, "loss": 0.7158, "step": 4294 }, { "epoch": 0.28831247273581423, "grad_norm": 1.3626692295074463, "learning_rate": 9.662107060105234e-05, "loss": 0.663, "step": 4296 }, { "epoch": 0.28844669641958326, "grad_norm": 1.123256802558899, "learning_rate": 9.661714198923654e-05, "loss": 0.6952, "step": 4298 }, { "epoch": 0.28858092010335223, "grad_norm": 1.127862811088562, "learning_rate": 9.661321117485244e-05, "loss": 0.7416, "step": 4300 }, { "epoch": 0.28871514378712126, "grad_norm": 1.430205225944519, "learning_rate": 9.66092781580858e-05, "loss": 0.6815, "step": 4302 }, { "epoch": 0.28884936747089024, "grad_norm": 1.0847445726394653, "learning_rate": 9.66053429391224e-05, "loss": 0.7995, "step": 4304 }, { "epoch": 0.2889835911546592, "grad_norm": 1.3557614088058472, "learning_rate": 9.660140551814817e-05, "loss": 0.7073, "step": 4306 }, { "epoch": 0.28911781483842824, "grad_norm": 1.0713691711425781, "learning_rate": 9.65974658953492e-05, "loss": 0.6449, "step": 4308 }, { "epoch": 0.2892520385221972, "grad_norm": 1.2485902309417725, "learning_rate": 9.659352407091159e-05, "loss": 0.6684, "step": 4310 }, { "epoch": 0.28938626220596625, "grad_norm": 1.1673352718353271, "learning_rate": 9.658958004502158e-05, "loss": 0.7202, "step": 4312 }, { "epoch": 0.2895204858897352, "grad_norm": 0.8654835224151611, "learning_rate": 9.658563381786554e-05, "loss": 0.61, "step": 4314 }, { "epoch": 0.28965470957350425, "grad_norm": 1.395545244216919, "learning_rate": 9.65816853896299e-05, "loss": 0.6299, "step": 4316 }, { "epoch": 0.28978893325727323, "grad_norm": 1.2566602230072021, "learning_rate": 9.657773476050123e-05, "loss": 0.7165, "step": 4318 }, { "epoch": 0.28992315694104226, "grad_norm": 1.102645754814148, "learning_rate": 9.657378193066617e-05, "loss": 0.7139, "step": 4320 }, { "epoch": 0.29005738062481123, "grad_norm": 1.0785845518112183, "learning_rate": 9.65698269003115e-05, "loss": 0.6792, "step": 4322 }, { "epoch": 0.29019160430858026, "grad_norm": 1.2975529432296753, "learning_rate": 9.656586966962411e-05, "loss": 0.6619, "step": 4324 }, { "epoch": 0.29032582799234924, "grad_norm": 1.0212528705596924, "learning_rate": 9.656191023879092e-05, "loss": 0.7072, "step": 4326 }, { "epoch": 0.29046005167611827, "grad_norm": 0.9956158995628357, "learning_rate": 9.655794860799904e-05, "loss": 0.7268, "step": 4328 }, { "epoch": 0.29059427535988724, "grad_norm": 1.0832117795944214, "learning_rate": 9.655398477743565e-05, "loss": 0.7002, "step": 4330 }, { "epoch": 0.2907284990436563, "grad_norm": 1.1135586500167847, "learning_rate": 9.6550018747288e-05, "loss": 0.714, "step": 4332 }, { "epoch": 0.29086272272742525, "grad_norm": 1.170379400253296, "learning_rate": 9.654605051774352e-05, "loss": 0.6517, "step": 4334 }, { "epoch": 0.2909969464111943, "grad_norm": 1.1077511310577393, "learning_rate": 9.654208008898968e-05, "loss": 0.7301, "step": 4336 }, { "epoch": 0.29113117009496325, "grad_norm": 1.5347226858139038, "learning_rate": 9.653810746121407e-05, "loss": 0.7131, "step": 4338 }, { "epoch": 0.2912653937787323, "grad_norm": 1.0661876201629639, "learning_rate": 9.65341326346044e-05, "loss": 0.7085, "step": 4340 }, { "epoch": 0.29139961746250126, "grad_norm": 1.0869132280349731, "learning_rate": 9.653015560934846e-05, "loss": 0.7179, "step": 4342 }, { "epoch": 0.29153384114627023, "grad_norm": 2.238389015197754, "learning_rate": 9.652617638563417e-05, "loss": 0.8169, "step": 4344 }, { "epoch": 0.29166806483003926, "grad_norm": 1.0856190919876099, "learning_rate": 9.652219496364954e-05, "loss": 0.721, "step": 4346 }, { "epoch": 0.29180228851380824, "grad_norm": 1.2137451171875, "learning_rate": 9.651821134358268e-05, "loss": 0.7266, "step": 4348 }, { "epoch": 0.29193651219757727, "grad_norm": 1.2103468179702759, "learning_rate": 9.651422552562181e-05, "loss": 0.6986, "step": 4350 }, { "epoch": 0.29207073588134624, "grad_norm": 1.219045877456665, "learning_rate": 9.651023750995525e-05, "loss": 0.7148, "step": 4352 }, { "epoch": 0.2922049595651153, "grad_norm": 1.0039916038513184, "learning_rate": 9.650624729677141e-05, "loss": 0.7019, "step": 4354 }, { "epoch": 0.29233918324888425, "grad_norm": 1.551803469657898, "learning_rate": 9.650225488625886e-05, "loss": 0.7499, "step": 4356 }, { "epoch": 0.2924734069326533, "grad_norm": 1.1392501592636108, "learning_rate": 9.649826027860619e-05, "loss": 0.6889, "step": 4358 }, { "epoch": 0.29260763061642225, "grad_norm": 1.0280663967132568, "learning_rate": 9.649426347400217e-05, "loss": 0.679, "step": 4360 }, { "epoch": 0.2927418543001913, "grad_norm": 1.082579493522644, "learning_rate": 9.649026447263561e-05, "loss": 0.7187, "step": 4362 }, { "epoch": 0.29287607798396026, "grad_norm": 0.9832870960235596, "learning_rate": 9.648626327469549e-05, "loss": 0.7013, "step": 4364 }, { "epoch": 0.2930103016677293, "grad_norm": 1.9217787981033325, "learning_rate": 9.648225988037083e-05, "loss": 0.6945, "step": 4366 }, { "epoch": 0.29314452535149826, "grad_norm": 1.3493105173110962, "learning_rate": 9.64782542898508e-05, "loss": 0.7189, "step": 4368 }, { "epoch": 0.2932787490352673, "grad_norm": 0.9947900772094727, "learning_rate": 9.647424650332467e-05, "loss": 0.668, "step": 4370 }, { "epoch": 0.29341297271903627, "grad_norm": 1.1923651695251465, "learning_rate": 9.647023652098174e-05, "loss": 0.7023, "step": 4372 }, { "epoch": 0.2935471964028053, "grad_norm": 1.172889232635498, "learning_rate": 9.646622434301154e-05, "loss": 0.7749, "step": 4374 }, { "epoch": 0.2936814200865743, "grad_norm": 1.2299495935440063, "learning_rate": 9.64622099696036e-05, "loss": 0.7138, "step": 4376 }, { "epoch": 0.2938156437703433, "grad_norm": 1.1374621391296387, "learning_rate": 9.645819340094762e-05, "loss": 0.684, "step": 4378 }, { "epoch": 0.2939498674541123, "grad_norm": 1.3243783712387085, "learning_rate": 9.645417463723335e-05, "loss": 0.7502, "step": 4380 }, { "epoch": 0.29408409113788125, "grad_norm": 1.0015956163406372, "learning_rate": 9.645015367865067e-05, "loss": 0.7016, "step": 4382 }, { "epoch": 0.2942183148216503, "grad_norm": 1.0454777479171753, "learning_rate": 9.644613052538957e-05, "loss": 0.7469, "step": 4384 }, { "epoch": 0.29435253850541926, "grad_norm": 1.108790397644043, "learning_rate": 9.644210517764014e-05, "loss": 0.6956, "step": 4386 }, { "epoch": 0.2944867621891883, "grad_norm": 1.1916991472244263, "learning_rate": 9.643807763559258e-05, "loss": 0.6804, "step": 4388 }, { "epoch": 0.29462098587295726, "grad_norm": 1.013068437576294, "learning_rate": 9.643404789943713e-05, "loss": 0.699, "step": 4390 }, { "epoch": 0.2947552095567263, "grad_norm": 0.9876919984817505, "learning_rate": 9.643001596936427e-05, "loss": 0.671, "step": 4392 }, { "epoch": 0.29488943324049527, "grad_norm": 1.1245033740997314, "learning_rate": 9.642598184556442e-05, "loss": 0.7324, "step": 4394 }, { "epoch": 0.2950236569242643, "grad_norm": 1.155880331993103, "learning_rate": 9.642194552822823e-05, "loss": 0.7868, "step": 4396 }, { "epoch": 0.2951578806080333, "grad_norm": 1.117091417312622, "learning_rate": 9.64179070175464e-05, "loss": 0.7233, "step": 4398 }, { "epoch": 0.2952921042918023, "grad_norm": 1.099109411239624, "learning_rate": 9.641386631370976e-05, "loss": 0.7429, "step": 4400 }, { "epoch": 0.2954263279755713, "grad_norm": 1.0578813552856445, "learning_rate": 9.640982341690918e-05, "loss": 0.6906, "step": 4402 }, { "epoch": 0.2955605516593403, "grad_norm": 1.064645528793335, "learning_rate": 9.640577832733571e-05, "loss": 0.7581, "step": 4404 }, { "epoch": 0.2956947753431093, "grad_norm": 0.9522479772567749, "learning_rate": 9.640173104518047e-05, "loss": 0.692, "step": 4406 }, { "epoch": 0.2958289990268783, "grad_norm": 1.1719841957092285, "learning_rate": 9.63976815706347e-05, "loss": 0.7418, "step": 4408 }, { "epoch": 0.2959632227106473, "grad_norm": 1.7045280933380127, "learning_rate": 9.639362990388969e-05, "loss": 0.7086, "step": 4410 }, { "epoch": 0.2960974463944163, "grad_norm": 1.1607673168182373, "learning_rate": 9.63895760451369e-05, "loss": 0.7272, "step": 4412 }, { "epoch": 0.2962316700781853, "grad_norm": 1.157461404800415, "learning_rate": 9.638551999456786e-05, "loss": 0.729, "step": 4414 }, { "epoch": 0.2963658937619543, "grad_norm": 1.1352097988128662, "learning_rate": 9.638146175237421e-05, "loss": 0.718, "step": 4416 }, { "epoch": 0.2965001174457233, "grad_norm": 1.4608018398284912, "learning_rate": 9.637740131874771e-05, "loss": 0.7127, "step": 4418 }, { "epoch": 0.2966343411294923, "grad_norm": 1.2369742393493652, "learning_rate": 9.63733386938802e-05, "loss": 0.7211, "step": 4420 }, { "epoch": 0.2967685648132613, "grad_norm": 1.0688071250915527, "learning_rate": 9.636927387796361e-05, "loss": 0.6845, "step": 4422 }, { "epoch": 0.2969027884970303, "grad_norm": 1.10688316822052, "learning_rate": 9.636520687119002e-05, "loss": 0.6995, "step": 4424 }, { "epoch": 0.2970370121807993, "grad_norm": 0.9537732005119324, "learning_rate": 9.636113767375158e-05, "loss": 0.6519, "step": 4426 }, { "epoch": 0.2971712358645683, "grad_norm": 1.0720096826553345, "learning_rate": 9.635706628584054e-05, "loss": 0.6166, "step": 4428 }, { "epoch": 0.2973054595483373, "grad_norm": 1.4682979583740234, "learning_rate": 9.63529927076493e-05, "loss": 0.6336, "step": 4430 }, { "epoch": 0.2974396832321063, "grad_norm": 1.0823240280151367, "learning_rate": 9.634891693937026e-05, "loss": 0.6681, "step": 4432 }, { "epoch": 0.2975739069158753, "grad_norm": 1.2542070150375366, "learning_rate": 9.634483898119608e-05, "loss": 0.6513, "step": 4434 }, { "epoch": 0.2977081305996443, "grad_norm": 2.113034248352051, "learning_rate": 9.634075883331937e-05, "loss": 0.7436, "step": 4436 }, { "epoch": 0.2978423542834133, "grad_norm": 1.283736228942871, "learning_rate": 9.633667649593294e-05, "loss": 0.6824, "step": 4438 }, { "epoch": 0.2979765779671823, "grad_norm": 1.3601338863372803, "learning_rate": 9.633259196922966e-05, "loss": 0.6553, "step": 4440 }, { "epoch": 0.29811080165095133, "grad_norm": 1.5369936227798462, "learning_rate": 9.632850525340251e-05, "loss": 0.7418, "step": 4442 }, { "epoch": 0.2982450253347203, "grad_norm": 1.0400851964950562, "learning_rate": 9.63244163486446e-05, "loss": 0.6815, "step": 4444 }, { "epoch": 0.29837924901848933, "grad_norm": 1.124594807624817, "learning_rate": 9.63203252551491e-05, "loss": 0.6522, "step": 4446 }, { "epoch": 0.2985134727022583, "grad_norm": 1.2869653701782227, "learning_rate": 9.63162319731093e-05, "loss": 0.798, "step": 4448 }, { "epoch": 0.29864769638602734, "grad_norm": 1.1588083505630493, "learning_rate": 9.631213650271864e-05, "loss": 0.7211, "step": 4450 }, { "epoch": 0.2987819200697963, "grad_norm": 0.9154360890388489, "learning_rate": 9.630803884417061e-05, "loss": 0.6761, "step": 4452 }, { "epoch": 0.29891614375356534, "grad_norm": 0.9952070713043213, "learning_rate": 9.630393899765878e-05, "loss": 0.7719, "step": 4454 }, { "epoch": 0.2990503674373343, "grad_norm": 1.147867202758789, "learning_rate": 9.62998369633769e-05, "loss": 0.7057, "step": 4456 }, { "epoch": 0.2991845911211033, "grad_norm": 1.7422637939453125, "learning_rate": 9.629573274151876e-05, "loss": 0.6712, "step": 4458 }, { "epoch": 0.2993188148048723, "grad_norm": 1.0809777975082397, "learning_rate": 9.62916263322783e-05, "loss": 0.7134, "step": 4460 }, { "epoch": 0.2994530384886413, "grad_norm": 1.0853334665298462, "learning_rate": 9.628751773584951e-05, "loss": 0.6778, "step": 4462 }, { "epoch": 0.29958726217241033, "grad_norm": 1.066697120666504, "learning_rate": 9.628340695242652e-05, "loss": 0.7024, "step": 4464 }, { "epoch": 0.2997214858561793, "grad_norm": 1.0515681505203247, "learning_rate": 9.627929398220358e-05, "loss": 0.6421, "step": 4466 }, { "epoch": 0.29985570953994833, "grad_norm": 1.065488338470459, "learning_rate": 9.6275178825375e-05, "loss": 0.6713, "step": 4468 }, { "epoch": 0.2999899332237173, "grad_norm": 1.2193208932876587, "learning_rate": 9.627106148213522e-05, "loss": 0.6997, "step": 4470 }, { "epoch": 0.30012415690748634, "grad_norm": 1.091118335723877, "learning_rate": 9.626694195267876e-05, "loss": 0.6888, "step": 4472 }, { "epoch": 0.3002583805912553, "grad_norm": 1.1276484727859497, "learning_rate": 9.626282023720028e-05, "loss": 0.7017, "step": 4474 }, { "epoch": 0.30039260427502434, "grad_norm": 1.1557211875915527, "learning_rate": 9.625869633589453e-05, "loss": 0.6885, "step": 4476 }, { "epoch": 0.3005268279587933, "grad_norm": 1.1216962337493896, "learning_rate": 9.625457024895632e-05, "loss": 0.7308, "step": 4478 }, { "epoch": 0.30066105164256235, "grad_norm": 1.0295031070709229, "learning_rate": 9.625044197658063e-05, "loss": 0.6237, "step": 4480 }, { "epoch": 0.3007952753263313, "grad_norm": 1.1262531280517578, "learning_rate": 9.624631151896251e-05, "loss": 0.6779, "step": 4482 }, { "epoch": 0.30092949901010035, "grad_norm": 1.1350386142730713, "learning_rate": 9.62421788762971e-05, "loss": 0.6781, "step": 4484 }, { "epoch": 0.30106372269386933, "grad_norm": 1.081994652748108, "learning_rate": 9.623804404877967e-05, "loss": 0.7855, "step": 4486 }, { "epoch": 0.30119794637763836, "grad_norm": 1.0471010208129883, "learning_rate": 9.623390703660559e-05, "loss": 0.758, "step": 4488 }, { "epoch": 0.30133217006140733, "grad_norm": 1.1623107194900513, "learning_rate": 9.62297678399703e-05, "loss": 0.8105, "step": 4490 }, { "epoch": 0.3014663937451763, "grad_norm": 0.9763181209564209, "learning_rate": 9.62256264590694e-05, "loss": 0.6587, "step": 4492 }, { "epoch": 0.30160061742894534, "grad_norm": 1.1519794464111328, "learning_rate": 9.622148289409855e-05, "loss": 0.6951, "step": 4494 }, { "epoch": 0.3017348411127143, "grad_norm": 1.0183149576187134, "learning_rate": 9.621733714525353e-05, "loss": 0.6649, "step": 4496 }, { "epoch": 0.30186906479648334, "grad_norm": 1.2081207036972046, "learning_rate": 9.621318921273021e-05, "loss": 0.7239, "step": 4498 }, { "epoch": 0.3020032884802523, "grad_norm": 1.100765585899353, "learning_rate": 9.620903909672457e-05, "loss": 0.6723, "step": 4500 }, { "epoch": 0.30213751216402135, "grad_norm": 1.1360970735549927, "learning_rate": 9.620488679743269e-05, "loss": 0.6997, "step": 4502 }, { "epoch": 0.3022717358477903, "grad_norm": 1.055786371231079, "learning_rate": 9.620073231505078e-05, "loss": 0.6099, "step": 4504 }, { "epoch": 0.30240595953155935, "grad_norm": 1.0756354331970215, "learning_rate": 9.61965756497751e-05, "loss": 0.6755, "step": 4506 }, { "epoch": 0.30254018321532833, "grad_norm": 1.2146291732788086, "learning_rate": 9.619241680180209e-05, "loss": 0.6447, "step": 4508 }, { "epoch": 0.30267440689909736, "grad_norm": 1.0501223802566528, "learning_rate": 9.61882557713282e-05, "loss": 0.7336, "step": 4510 }, { "epoch": 0.30280863058286633, "grad_norm": 1.0172021389007568, "learning_rate": 9.618409255855006e-05, "loss": 0.6122, "step": 4512 }, { "epoch": 0.30294285426663536, "grad_norm": 1.4607539176940918, "learning_rate": 9.617992716366435e-05, "loss": 0.7468, "step": 4514 }, { "epoch": 0.30307707795040434, "grad_norm": 1.0794657468795776, "learning_rate": 9.61757595868679e-05, "loss": 0.7288, "step": 4516 }, { "epoch": 0.30321130163417337, "grad_norm": 1.0251612663269043, "learning_rate": 9.617158982835761e-05, "loss": 0.7091, "step": 4518 }, { "epoch": 0.30334552531794234, "grad_norm": 0.9694096446037292, "learning_rate": 9.61674178883305e-05, "loss": 0.7073, "step": 4520 }, { "epoch": 0.3034797490017114, "grad_norm": 1.239879846572876, "learning_rate": 9.616324376698366e-05, "loss": 0.6995, "step": 4522 }, { "epoch": 0.30361397268548035, "grad_norm": 1.0317822694778442, "learning_rate": 9.615906746451435e-05, "loss": 0.6413, "step": 4524 }, { "epoch": 0.3037481963692494, "grad_norm": 1.1398820877075195, "learning_rate": 9.615488898111985e-05, "loss": 0.6995, "step": 4526 }, { "epoch": 0.30388242005301835, "grad_norm": 1.0622879266738892, "learning_rate": 9.615070831699762e-05, "loss": 0.6748, "step": 4528 }, { "epoch": 0.30401664373678733, "grad_norm": 1.4225927591323853, "learning_rate": 9.614652547234516e-05, "loss": 0.7386, "step": 4530 }, { "epoch": 0.30415086742055636, "grad_norm": 1.211207389831543, "learning_rate": 9.614234044736012e-05, "loss": 0.7205, "step": 4532 }, { "epoch": 0.30428509110432533, "grad_norm": 1.1871228218078613, "learning_rate": 9.613815324224023e-05, "loss": 0.743, "step": 4534 }, { "epoch": 0.30441931478809436, "grad_norm": 1.1464089155197144, "learning_rate": 9.613396385718334e-05, "loss": 0.7394, "step": 4536 }, { "epoch": 0.30455353847186334, "grad_norm": 1.1008272171020508, "learning_rate": 9.612977229238735e-05, "loss": 0.7076, "step": 4538 }, { "epoch": 0.30468776215563237, "grad_norm": 1.11635422706604, "learning_rate": 9.612557854805036e-05, "loss": 0.769, "step": 4540 }, { "epoch": 0.30482198583940134, "grad_norm": 1.2722200155258179, "learning_rate": 9.612138262437046e-05, "loss": 0.7641, "step": 4542 }, { "epoch": 0.3049562095231704, "grad_norm": 2.5558648109436035, "learning_rate": 9.611718452154594e-05, "loss": 0.7306, "step": 4544 }, { "epoch": 0.30509043320693935, "grad_norm": 1.1128522157669067, "learning_rate": 9.611298423977512e-05, "loss": 0.7284, "step": 4546 }, { "epoch": 0.3052246568907084, "grad_norm": 0.9909085035324097, "learning_rate": 9.610878177925648e-05, "loss": 0.7257, "step": 4548 }, { "epoch": 0.30535888057447735, "grad_norm": 1.0871301889419556, "learning_rate": 9.610457714018857e-05, "loss": 0.7338, "step": 4550 }, { "epoch": 0.3054931042582464, "grad_norm": 1.0989935398101807, "learning_rate": 9.610037032277007e-05, "loss": 0.6772, "step": 4552 }, { "epoch": 0.30562732794201536, "grad_norm": 1.3562226295471191, "learning_rate": 9.609616132719971e-05, "loss": 0.7434, "step": 4554 }, { "epoch": 0.3057615516257844, "grad_norm": 0.993280827999115, "learning_rate": 9.609195015367636e-05, "loss": 0.6481, "step": 4556 }, { "epoch": 0.30589577530955336, "grad_norm": 1.5082858800888062, "learning_rate": 9.608773680239902e-05, "loss": 0.7755, "step": 4558 }, { "epoch": 0.3060299989933224, "grad_norm": 1.336323857307434, "learning_rate": 9.608352127356672e-05, "loss": 0.6694, "step": 4560 }, { "epoch": 0.30616422267709137, "grad_norm": 0.9614841938018799, "learning_rate": 9.607930356737869e-05, "loss": 0.6919, "step": 4562 }, { "epoch": 0.3062984463608604, "grad_norm": 1.0784282684326172, "learning_rate": 9.607508368403415e-05, "loss": 0.6726, "step": 4564 }, { "epoch": 0.3064326700446294, "grad_norm": 2.376894474029541, "learning_rate": 9.607086162373253e-05, "loss": 0.7332, "step": 4566 }, { "epoch": 0.30656689372839835, "grad_norm": 1.072051763534546, "learning_rate": 9.606663738667328e-05, "loss": 0.7185, "step": 4568 }, { "epoch": 0.3067011174121674, "grad_norm": 1.0158129930496216, "learning_rate": 9.6062410973056e-05, "loss": 0.6662, "step": 4570 }, { "epoch": 0.30683534109593635, "grad_norm": 1.059302806854248, "learning_rate": 9.605818238308038e-05, "loss": 0.6646, "step": 4572 }, { "epoch": 0.3069695647797054, "grad_norm": 1.0993070602416992, "learning_rate": 9.605395161694621e-05, "loss": 0.7113, "step": 4574 }, { "epoch": 0.30710378846347436, "grad_norm": 1.1712669134140015, "learning_rate": 9.60497186748534e-05, "loss": 0.8254, "step": 4576 }, { "epoch": 0.3072380121472434, "grad_norm": 1.1879925727844238, "learning_rate": 9.604548355700194e-05, "loss": 0.734, "step": 4578 }, { "epoch": 0.30737223583101236, "grad_norm": 1.0698761940002441, "learning_rate": 9.60412462635919e-05, "loss": 0.6339, "step": 4580 }, { "epoch": 0.3075064595147814, "grad_norm": 0.9980600476264954, "learning_rate": 9.603700679482352e-05, "loss": 0.6412, "step": 4582 }, { "epoch": 0.30764068319855037, "grad_norm": 0.9460391402244568, "learning_rate": 9.603276515089711e-05, "loss": 0.6941, "step": 4584 }, { "epoch": 0.3077749068823194, "grad_norm": 1.068634033203125, "learning_rate": 9.602852133201305e-05, "loss": 0.6952, "step": 4586 }, { "epoch": 0.3079091305660884, "grad_norm": 1.2626107931137085, "learning_rate": 9.602427533837188e-05, "loss": 0.7765, "step": 4588 }, { "epoch": 0.3080433542498574, "grad_norm": 1.057005763053894, "learning_rate": 9.60200271701742e-05, "loss": 0.7368, "step": 4590 }, { "epoch": 0.3081775779336264, "grad_norm": 0.9701996445655823, "learning_rate": 9.601577682762072e-05, "loss": 0.6842, "step": 4592 }, { "epoch": 0.3083118016173954, "grad_norm": 1.1067408323287964, "learning_rate": 9.60115243109123e-05, "loss": 0.7384, "step": 4594 }, { "epoch": 0.3084460253011644, "grad_norm": 1.0144931077957153, "learning_rate": 9.60072696202498e-05, "loss": 0.727, "step": 4596 }, { "epoch": 0.3085802489849334, "grad_norm": 1.1220800876617432, "learning_rate": 9.60030127558343e-05, "loss": 0.673, "step": 4598 }, { "epoch": 0.3087144726687024, "grad_norm": 1.1260027885437012, "learning_rate": 9.59987537178669e-05, "loss": 0.6752, "step": 4600 }, { "epoch": 0.3088486963524714, "grad_norm": 1.149194359779358, "learning_rate": 9.599449250654884e-05, "loss": 0.7066, "step": 4602 }, { "epoch": 0.3089829200362404, "grad_norm": 0.9491012096405029, "learning_rate": 9.599022912208145e-05, "loss": 0.6873, "step": 4604 }, { "epoch": 0.30911714372000937, "grad_norm": 1.353082537651062, "learning_rate": 9.598596356466618e-05, "loss": 0.6817, "step": 4606 }, { "epoch": 0.3092513674037784, "grad_norm": 1.115274429321289, "learning_rate": 9.598169583450455e-05, "loss": 0.6553, "step": 4608 }, { "epoch": 0.3093855910875474, "grad_norm": 1.2805801630020142, "learning_rate": 9.597742593179822e-05, "loss": 0.6605, "step": 4610 }, { "epoch": 0.3095198147713164, "grad_norm": 1.1053426265716553, "learning_rate": 9.597315385674893e-05, "loss": 0.6783, "step": 4612 }, { "epoch": 0.3096540384550854, "grad_norm": 0.9258078932762146, "learning_rate": 9.596887960955849e-05, "loss": 0.6579, "step": 4614 }, { "epoch": 0.3097882621388544, "grad_norm": 1.1637978553771973, "learning_rate": 9.596460319042891e-05, "loss": 0.6987, "step": 4616 }, { "epoch": 0.3099224858226234, "grad_norm": 1.219173789024353, "learning_rate": 9.596032459956222e-05, "loss": 0.7033, "step": 4618 }, { "epoch": 0.3100567095063924, "grad_norm": 1.0405182838439941, "learning_rate": 9.595604383716055e-05, "loss": 0.7149, "step": 4620 }, { "epoch": 0.3101909331901614, "grad_norm": 1.7751362323760986, "learning_rate": 9.59517609034262e-05, "loss": 0.7148, "step": 4622 }, { "epoch": 0.3103251568739304, "grad_norm": 1.0881564617156982, "learning_rate": 9.594747579856149e-05, "loss": 0.7243, "step": 4624 }, { "epoch": 0.3104593805576994, "grad_norm": 1.2595254182815552, "learning_rate": 9.59431885227689e-05, "loss": 0.685, "step": 4626 }, { "epoch": 0.3105936042414684, "grad_norm": 1.3012375831604004, "learning_rate": 9.5938899076251e-05, "loss": 0.68, "step": 4628 }, { "epoch": 0.3107278279252374, "grad_norm": 1.1929621696472168, "learning_rate": 9.593460745921046e-05, "loss": 0.7239, "step": 4630 }, { "epoch": 0.31086205160900643, "grad_norm": 1.0787216424942017, "learning_rate": 9.593031367185003e-05, "loss": 0.6862, "step": 4632 }, { "epoch": 0.3109962752927754, "grad_norm": 1.2381178140640259, "learning_rate": 9.592601771437261e-05, "loss": 0.6973, "step": 4634 }, { "epoch": 0.31113049897654443, "grad_norm": 1.4702860116958618, "learning_rate": 9.592171958698115e-05, "loss": 0.752, "step": 4636 }, { "epoch": 0.3112647226603134, "grad_norm": 1.1221622228622437, "learning_rate": 9.591741928987876e-05, "loss": 0.7334, "step": 4638 }, { "epoch": 0.31139894634408244, "grad_norm": 1.3333992958068848, "learning_rate": 9.591311682326859e-05, "loss": 0.7177, "step": 4640 }, { "epoch": 0.3115331700278514, "grad_norm": 0.9714601039886475, "learning_rate": 9.590881218735394e-05, "loss": 0.6296, "step": 4642 }, { "epoch": 0.3116673937116204, "grad_norm": 1.1566107273101807, "learning_rate": 9.590450538233817e-05, "loss": 0.6886, "step": 4644 }, { "epoch": 0.3118016173953894, "grad_norm": 1.1195974349975586, "learning_rate": 9.590019640842482e-05, "loss": 0.6605, "step": 4646 }, { "epoch": 0.3119358410791584, "grad_norm": 1.0662175416946411, "learning_rate": 9.589588526581741e-05, "loss": 0.7264, "step": 4648 }, { "epoch": 0.3120700647629274, "grad_norm": 1.082502007484436, "learning_rate": 9.58915719547197e-05, "loss": 0.7297, "step": 4650 }, { "epoch": 0.3122042884466964, "grad_norm": 1.0617055892944336, "learning_rate": 9.588725647533545e-05, "loss": 0.7096, "step": 4652 }, { "epoch": 0.31233851213046543, "grad_norm": 0.9861075282096863, "learning_rate": 9.588293882786857e-05, "loss": 0.6781, "step": 4654 }, { "epoch": 0.3124727358142344, "grad_norm": 1.15412175655365, "learning_rate": 9.587861901252305e-05, "loss": 0.648, "step": 4656 }, { "epoch": 0.31260695949800343, "grad_norm": 1.5281078815460205, "learning_rate": 9.5874297029503e-05, "loss": 0.7575, "step": 4658 }, { "epoch": 0.3127411831817724, "grad_norm": 1.079062819480896, "learning_rate": 9.586997287901262e-05, "loss": 0.7238, "step": 4660 }, { "epoch": 0.31287540686554144, "grad_norm": 1.133266806602478, "learning_rate": 9.586564656125623e-05, "loss": 0.7354, "step": 4662 }, { "epoch": 0.3130096305493104, "grad_norm": 1.4366016387939453, "learning_rate": 9.586131807643822e-05, "loss": 0.6718, "step": 4664 }, { "epoch": 0.31314385423307944, "grad_norm": 0.9520801305770874, "learning_rate": 9.585698742476311e-05, "loss": 0.6562, "step": 4666 }, { "epoch": 0.3132780779168484, "grad_norm": 1.0416120290756226, "learning_rate": 9.585265460643553e-05, "loss": 0.7258, "step": 4668 }, { "epoch": 0.31341230160061745, "grad_norm": 1.092458724975586, "learning_rate": 9.584831962166017e-05, "loss": 0.6946, "step": 4670 }, { "epoch": 0.3135465252843864, "grad_norm": 1.0674761533737183, "learning_rate": 9.584398247064188e-05, "loss": 0.6967, "step": 4672 }, { "epoch": 0.31368074896815545, "grad_norm": 1.082927942276001, "learning_rate": 9.583964315358555e-05, "loss": 0.6804, "step": 4674 }, { "epoch": 0.31381497265192443, "grad_norm": 1.2659660577774048, "learning_rate": 9.583530167069626e-05, "loss": 0.7403, "step": 4676 }, { "epoch": 0.31394919633569346, "grad_norm": 0.9640771150588989, "learning_rate": 9.583095802217905e-05, "loss": 0.6655, "step": 4678 }, { "epoch": 0.31408342001946243, "grad_norm": 1.516160011291504, "learning_rate": 9.582661220823922e-05, "loss": 0.6648, "step": 4680 }, { "epoch": 0.3142176437032314, "grad_norm": 1.02597177028656, "learning_rate": 9.582226422908207e-05, "loss": 0.7132, "step": 4682 }, { "epoch": 0.31435186738700044, "grad_norm": 1.0739665031433105, "learning_rate": 9.581791408491305e-05, "loss": 0.6212, "step": 4684 }, { "epoch": 0.3144860910707694, "grad_norm": 1.6776080131530762, "learning_rate": 9.581356177593767e-05, "loss": 0.7355, "step": 4686 }, { "epoch": 0.31462031475453844, "grad_norm": 1.3557602167129517, "learning_rate": 9.58092073023616e-05, "loss": 0.7082, "step": 4688 }, { "epoch": 0.3147545384383074, "grad_norm": 1.0693676471710205, "learning_rate": 9.580485066439056e-05, "loss": 0.6973, "step": 4690 }, { "epoch": 0.31488876212207645, "grad_norm": 1.0692919492721558, "learning_rate": 9.58004918622304e-05, "loss": 0.6769, "step": 4692 }, { "epoch": 0.3150229858058454, "grad_norm": 1.411782145500183, "learning_rate": 9.579613089608705e-05, "loss": 0.6878, "step": 4694 }, { "epoch": 0.31515720948961445, "grad_norm": 1.0561375617980957, "learning_rate": 9.579176776616658e-05, "loss": 0.6816, "step": 4696 }, { "epoch": 0.31529143317338343, "grad_norm": 1.1592587232589722, "learning_rate": 9.578740247267514e-05, "loss": 0.7211, "step": 4698 }, { "epoch": 0.31542565685715246, "grad_norm": 0.9945803880691528, "learning_rate": 9.578303501581895e-05, "loss": 0.6684, "step": 4700 }, { "epoch": 0.31555988054092143, "grad_norm": 1.0777732133865356, "learning_rate": 9.57786653958044e-05, "loss": 0.6935, "step": 4702 }, { "epoch": 0.31569410422469046, "grad_norm": 0.9888481497764587, "learning_rate": 9.577429361283792e-05, "loss": 0.6799, "step": 4704 }, { "epoch": 0.31582832790845944, "grad_norm": 1.0442769527435303, "learning_rate": 9.576991966712607e-05, "loss": 0.6754, "step": 4706 }, { "epoch": 0.31596255159222847, "grad_norm": 1.0910941362380981, "learning_rate": 9.576554355887554e-05, "loss": 0.6806, "step": 4708 }, { "epoch": 0.31609677527599744, "grad_norm": 1.2258208990097046, "learning_rate": 9.576116528829306e-05, "loss": 0.726, "step": 4710 }, { "epoch": 0.3162309989597665, "grad_norm": 1.060171127319336, "learning_rate": 9.575678485558551e-05, "loss": 0.7486, "step": 4712 }, { "epoch": 0.31636522264353545, "grad_norm": 1.1440684795379639, "learning_rate": 9.575240226095984e-05, "loss": 0.6969, "step": 4714 }, { "epoch": 0.3164994463273045, "grad_norm": 1.1293154954910278, "learning_rate": 9.574801750462315e-05, "loss": 0.7284, "step": 4716 }, { "epoch": 0.31663367001107345, "grad_norm": 1.2209240198135376, "learning_rate": 9.574363058678257e-05, "loss": 0.74, "step": 4718 }, { "epoch": 0.31676789369484243, "grad_norm": 1.7266669273376465, "learning_rate": 9.573924150764541e-05, "loss": 0.7009, "step": 4720 }, { "epoch": 0.31690211737861146, "grad_norm": 1.1648743152618408, "learning_rate": 9.573485026741902e-05, "loss": 0.6951, "step": 4722 }, { "epoch": 0.31703634106238043, "grad_norm": 1.3447242975234985, "learning_rate": 9.57304568663109e-05, "loss": 0.6777, "step": 4724 }, { "epoch": 0.31717056474614946, "grad_norm": 1.0942497253417969, "learning_rate": 9.572606130452862e-05, "loss": 0.6868, "step": 4726 }, { "epoch": 0.31730478842991844, "grad_norm": 1.0130993127822876, "learning_rate": 9.572166358227985e-05, "loss": 0.6962, "step": 4728 }, { "epoch": 0.31743901211368747, "grad_norm": 1.2938907146453857, "learning_rate": 9.571726369977239e-05, "loss": 0.7458, "step": 4730 }, { "epoch": 0.31757323579745644, "grad_norm": 1.0879647731781006, "learning_rate": 9.571286165721412e-05, "loss": 0.7154, "step": 4732 }, { "epoch": 0.3177074594812255, "grad_norm": 1.0983290672302246, "learning_rate": 9.570845745481303e-05, "loss": 0.717, "step": 4734 }, { "epoch": 0.31784168316499445, "grad_norm": 0.8874186277389526, "learning_rate": 9.570405109277719e-05, "loss": 0.6164, "step": 4736 }, { "epoch": 0.3179759068487635, "grad_norm": 1.0488560199737549, "learning_rate": 9.569964257131484e-05, "loss": 0.7797, "step": 4738 }, { "epoch": 0.31811013053253245, "grad_norm": 1.1975808143615723, "learning_rate": 9.56952318906342e-05, "loss": 0.7539, "step": 4740 }, { "epoch": 0.3182443542163015, "grad_norm": 1.1216825246810913, "learning_rate": 9.569081905094375e-05, "loss": 0.6643, "step": 4742 }, { "epoch": 0.31837857790007046, "grad_norm": 1.1932557821273804, "learning_rate": 9.568640405245192e-05, "loss": 0.6596, "step": 4744 }, { "epoch": 0.3185128015838395, "grad_norm": 1.1482906341552734, "learning_rate": 9.568198689536734e-05, "loss": 0.7082, "step": 4746 }, { "epoch": 0.31864702526760846, "grad_norm": 1.2021764516830444, "learning_rate": 9.567756757989872e-05, "loss": 0.6848, "step": 4748 }, { "epoch": 0.3187812489513775, "grad_norm": 2.35219144821167, "learning_rate": 9.567314610625485e-05, "loss": 0.6858, "step": 4750 }, { "epoch": 0.31891547263514647, "grad_norm": 1.1158368587493896, "learning_rate": 9.566872247464464e-05, "loss": 0.6739, "step": 4752 }, { "epoch": 0.3190496963189155, "grad_norm": 1.0408796072006226, "learning_rate": 9.56642966852771e-05, "loss": 0.7017, "step": 4754 }, { "epoch": 0.3191839200026845, "grad_norm": 1.1878182888031006, "learning_rate": 9.565986873836132e-05, "loss": 0.7339, "step": 4756 }, { "epoch": 0.31931814368645345, "grad_norm": 1.1423181295394897, "learning_rate": 9.565543863410654e-05, "loss": 0.6595, "step": 4758 }, { "epoch": 0.3194523673702225, "grad_norm": 1.1221919059753418, "learning_rate": 9.565100637272206e-05, "loss": 0.6839, "step": 4760 }, { "epoch": 0.31958659105399145, "grad_norm": 1.0802282094955444, "learning_rate": 9.564657195441731e-05, "loss": 0.6891, "step": 4762 }, { "epoch": 0.3197208147377605, "grad_norm": 1.8912112712860107, "learning_rate": 9.564213537940177e-05, "loss": 0.7711, "step": 4764 }, { "epoch": 0.31985503842152946, "grad_norm": 1.4157387018203735, "learning_rate": 9.563769664788511e-05, "loss": 0.6955, "step": 4766 }, { "epoch": 0.3199892621052985, "grad_norm": 1.271424412727356, "learning_rate": 9.563325576007701e-05, "loss": 0.7484, "step": 4768 }, { "epoch": 0.32012348578906746, "grad_norm": 1.1220924854278564, "learning_rate": 9.562881271618732e-05, "loss": 0.6604, "step": 4770 }, { "epoch": 0.3202577094728365, "grad_norm": 1.0575189590454102, "learning_rate": 9.562436751642593e-05, "loss": 0.6773, "step": 4772 }, { "epoch": 0.32039193315660547, "grad_norm": 1.2651679515838623, "learning_rate": 9.561992016100293e-05, "loss": 0.6567, "step": 4774 }, { "epoch": 0.3205261568403745, "grad_norm": 1.2005423307418823, "learning_rate": 9.561547065012839e-05, "loss": 0.7591, "step": 4776 }, { "epoch": 0.3206603805241435, "grad_norm": 1.3069040775299072, "learning_rate": 9.561101898401255e-05, "loss": 0.7619, "step": 4778 }, { "epoch": 0.3207946042079125, "grad_norm": 1.2713435888290405, "learning_rate": 9.560656516286577e-05, "loss": 0.7043, "step": 4780 }, { "epoch": 0.3209288278916815, "grad_norm": 1.2976057529449463, "learning_rate": 9.560210918689847e-05, "loss": 0.7565, "step": 4782 }, { "epoch": 0.3210630515754505, "grad_norm": 1.0996086597442627, "learning_rate": 9.559765105632117e-05, "loss": 0.7015, "step": 4784 }, { "epoch": 0.3211972752592195, "grad_norm": 1.0666567087173462, "learning_rate": 9.559319077134453e-05, "loss": 0.6563, "step": 4786 }, { "epoch": 0.3213314989429885, "grad_norm": 1.154266357421875, "learning_rate": 9.558872833217927e-05, "loss": 0.6679, "step": 4788 }, { "epoch": 0.3214657226267575, "grad_norm": 1.0351626873016357, "learning_rate": 9.558426373903626e-05, "loss": 0.6951, "step": 4790 }, { "epoch": 0.3215999463105265, "grad_norm": 1.0043928623199463, "learning_rate": 9.557979699212642e-05, "loss": 0.6525, "step": 4792 }, { "epoch": 0.3217341699942955, "grad_norm": 1.2971715927124023, "learning_rate": 9.557532809166079e-05, "loss": 0.7109, "step": 4794 }, { "epoch": 0.32186839367806447, "grad_norm": 1.2481286525726318, "learning_rate": 9.557085703785054e-05, "loss": 0.7872, "step": 4796 }, { "epoch": 0.3220026173618335, "grad_norm": 1.089882493019104, "learning_rate": 9.55663838309069e-05, "loss": 0.7012, "step": 4798 }, { "epoch": 0.3221368410456025, "grad_norm": 1.3476847410202026, "learning_rate": 9.556190847104123e-05, "loss": 0.7281, "step": 4800 }, { "epoch": 0.3222710647293715, "grad_norm": 1.4910799264907837, "learning_rate": 9.555743095846497e-05, "loss": 0.6774, "step": 4802 }, { "epoch": 0.3224052884131405, "grad_norm": 1.1183050870895386, "learning_rate": 9.555295129338969e-05, "loss": 0.635, "step": 4804 }, { "epoch": 0.3225395120969095, "grad_norm": 1.0580859184265137, "learning_rate": 9.554846947602704e-05, "loss": 0.7013, "step": 4806 }, { "epoch": 0.3226737357806785, "grad_norm": 1.1748672723770142, "learning_rate": 9.554398550658876e-05, "loss": 0.6827, "step": 4808 }, { "epoch": 0.3228079594644475, "grad_norm": 1.1253255605697632, "learning_rate": 9.553949938528675e-05, "loss": 0.6665, "step": 4810 }, { "epoch": 0.3229421831482165, "grad_norm": 1.010771632194519, "learning_rate": 9.553501111233292e-05, "loss": 0.7021, "step": 4812 }, { "epoch": 0.3230764068319855, "grad_norm": 1.1532608270645142, "learning_rate": 9.553052068793937e-05, "loss": 0.7278, "step": 4814 }, { "epoch": 0.3232106305157545, "grad_norm": 1.0641151666641235, "learning_rate": 9.552602811231824e-05, "loss": 0.7161, "step": 4816 }, { "epoch": 0.3233448541995235, "grad_norm": 1.0781819820404053, "learning_rate": 9.552153338568181e-05, "loss": 0.7942, "step": 4818 }, { "epoch": 0.3234790778832925, "grad_norm": 1.099227786064148, "learning_rate": 9.551703650824243e-05, "loss": 0.7583, "step": 4820 }, { "epoch": 0.32361330156706153, "grad_norm": 1.1302250623703003, "learning_rate": 9.551253748021259e-05, "loss": 0.7083, "step": 4822 }, { "epoch": 0.3237475252508305, "grad_norm": 1.2145006656646729, "learning_rate": 9.550803630180485e-05, "loss": 0.7189, "step": 4824 }, { "epoch": 0.32388174893459953, "grad_norm": 1.0137630701065063, "learning_rate": 9.55035329732319e-05, "loss": 0.6715, "step": 4826 }, { "epoch": 0.3240159726183685, "grad_norm": 0.9891653060913086, "learning_rate": 9.549902749470646e-05, "loss": 0.6566, "step": 4828 }, { "epoch": 0.32415019630213754, "grad_norm": 1.0337892770767212, "learning_rate": 9.549451986644147e-05, "loss": 0.6769, "step": 4830 }, { "epoch": 0.3242844199859065, "grad_norm": 1.083341360092163, "learning_rate": 9.549001008864987e-05, "loss": 0.7068, "step": 4832 }, { "epoch": 0.3244186436696755, "grad_norm": 1.1820749044418335, "learning_rate": 9.548549816154473e-05, "loss": 0.7483, "step": 4834 }, { "epoch": 0.3245528673534445, "grad_norm": 0.8702498078346252, "learning_rate": 9.548098408533926e-05, "loss": 0.6265, "step": 4836 }, { "epoch": 0.3246870910372135, "grad_norm": 1.1651930809020996, "learning_rate": 9.547646786024673e-05, "loss": 0.7112, "step": 4838 }, { "epoch": 0.3248213147209825, "grad_norm": 1.0169198513031006, "learning_rate": 9.547194948648051e-05, "loss": 0.6766, "step": 4840 }, { "epoch": 0.3249555384047515, "grad_norm": 1.4236356019973755, "learning_rate": 9.546742896425409e-05, "loss": 0.7752, "step": 4842 }, { "epoch": 0.32508976208852053, "grad_norm": 1.0329314470291138, "learning_rate": 9.546290629378107e-05, "loss": 0.7007, "step": 4844 }, { "epoch": 0.3252239857722895, "grad_norm": 0.963914155960083, "learning_rate": 9.545838147527512e-05, "loss": 0.6976, "step": 4846 }, { "epoch": 0.32535820945605853, "grad_norm": 1.0919042825698853, "learning_rate": 9.545385450895003e-05, "loss": 0.7028, "step": 4848 }, { "epoch": 0.3254924331398275, "grad_norm": 1.2006242275238037, "learning_rate": 9.544932539501971e-05, "loss": 0.6799, "step": 4850 }, { "epoch": 0.32562665682359654, "grad_norm": 1.0634151697158813, "learning_rate": 9.544479413369814e-05, "loss": 0.7073, "step": 4852 }, { "epoch": 0.3257608805073655, "grad_norm": 1.1870208978652954, "learning_rate": 9.54402607251994e-05, "loss": 0.7336, "step": 4854 }, { "epoch": 0.32589510419113454, "grad_norm": 1.11781907081604, "learning_rate": 9.543572516973769e-05, "loss": 0.6152, "step": 4856 }, { "epoch": 0.3260293278749035, "grad_norm": 1.5316169261932373, "learning_rate": 9.543118746752733e-05, "loss": 0.6583, "step": 4858 }, { "epoch": 0.32616355155867255, "grad_norm": 1.018977165222168, "learning_rate": 9.542664761878269e-05, "loss": 0.691, "step": 4860 }, { "epoch": 0.3262977752424415, "grad_norm": 1.1612788438796997, "learning_rate": 9.542210562371828e-05, "loss": 0.685, "step": 4862 }, { "epoch": 0.32643199892621055, "grad_norm": 1.6843246221542358, "learning_rate": 9.541756148254867e-05, "loss": 0.6789, "step": 4864 }, { "epoch": 0.32656622260997953, "grad_norm": 1.2154302597045898, "learning_rate": 9.541301519548861e-05, "loss": 0.7006, "step": 4866 }, { "epoch": 0.32670044629374856, "grad_norm": 0.9375450015068054, "learning_rate": 9.54084667627529e-05, "loss": 0.6265, "step": 4868 }, { "epoch": 0.32683466997751753, "grad_norm": 1.1680549383163452, "learning_rate": 9.540391618455641e-05, "loss": 0.7199, "step": 4870 }, { "epoch": 0.3269688936612865, "grad_norm": 1.1197963953018188, "learning_rate": 9.539936346111416e-05, "loss": 0.7492, "step": 4872 }, { "epoch": 0.32710311734505554, "grad_norm": 1.323502540588379, "learning_rate": 9.539480859264128e-05, "loss": 0.7417, "step": 4874 }, { "epoch": 0.3272373410288245, "grad_norm": 1.6119531393051147, "learning_rate": 9.539025157935292e-05, "loss": 0.7183, "step": 4876 }, { "epoch": 0.32737156471259354, "grad_norm": 1.082277774810791, "learning_rate": 9.538569242146447e-05, "loss": 0.7381, "step": 4878 }, { "epoch": 0.3275057883963625, "grad_norm": 1.2153085470199585, "learning_rate": 9.53811311191913e-05, "loss": 0.6995, "step": 4880 }, { "epoch": 0.32764001208013155, "grad_norm": 1.1135528087615967, "learning_rate": 9.53765676727489e-05, "loss": 0.6961, "step": 4882 }, { "epoch": 0.3277742357639005, "grad_norm": 1.0991860628128052, "learning_rate": 9.537200208235291e-05, "loss": 0.723, "step": 4884 }, { "epoch": 0.32790845944766955, "grad_norm": 1.1873701810836792, "learning_rate": 9.536743434821904e-05, "loss": 0.667, "step": 4886 }, { "epoch": 0.32804268313143853, "grad_norm": 1.292967677116394, "learning_rate": 9.536286447056311e-05, "loss": 0.7801, "step": 4888 }, { "epoch": 0.32817690681520756, "grad_norm": 1.1023390293121338, "learning_rate": 9.535829244960104e-05, "loss": 0.7174, "step": 4890 }, { "epoch": 0.32831113049897653, "grad_norm": 1.0848222970962524, "learning_rate": 9.535371828554884e-05, "loss": 0.6969, "step": 4892 }, { "epoch": 0.32844535418274556, "grad_norm": 1.024688482284546, "learning_rate": 9.534914197862266e-05, "loss": 0.6672, "step": 4894 }, { "epoch": 0.32857957786651454, "grad_norm": 1.1134281158447266, "learning_rate": 9.534456352903866e-05, "loss": 0.6991, "step": 4896 }, { "epoch": 0.32871380155028357, "grad_norm": 1.0506900548934937, "learning_rate": 9.533998293701323e-05, "loss": 0.6822, "step": 4898 }, { "epoch": 0.32884802523405254, "grad_norm": 1.1279277801513672, "learning_rate": 9.533540020276274e-05, "loss": 0.719, "step": 4900 }, { "epoch": 0.3289822489178216, "grad_norm": 1.0671149492263794, "learning_rate": 9.533081532650375e-05, "loss": 0.6608, "step": 4902 }, { "epoch": 0.32911647260159055, "grad_norm": 1.2312167882919312, "learning_rate": 9.532622830845287e-05, "loss": 0.6822, "step": 4904 }, { "epoch": 0.3292506962853596, "grad_norm": 1.6732779741287231, "learning_rate": 9.532163914882685e-05, "loss": 0.6851, "step": 4906 }, { "epoch": 0.32938491996912855, "grad_norm": 1.053810477256775, "learning_rate": 9.531704784784248e-05, "loss": 0.644, "step": 4908 }, { "epoch": 0.32951914365289753, "grad_norm": 0.9314039945602417, "learning_rate": 9.531245440571672e-05, "loss": 0.6335, "step": 4910 }, { "epoch": 0.32965336733666656, "grad_norm": 1.0545871257781982, "learning_rate": 9.53078588226666e-05, "loss": 0.6814, "step": 4912 }, { "epoch": 0.32978759102043553, "grad_norm": 1.5250245332717896, "learning_rate": 9.530326109890924e-05, "loss": 0.6458, "step": 4914 }, { "epoch": 0.32992181470420456, "grad_norm": 1.0881139039993286, "learning_rate": 9.529866123466187e-05, "loss": 0.6858, "step": 4916 }, { "epoch": 0.33005603838797354, "grad_norm": 1.2018465995788574, "learning_rate": 9.529405923014183e-05, "loss": 0.6176, "step": 4918 }, { "epoch": 0.33019026207174257, "grad_norm": 1.7062848806381226, "learning_rate": 9.528945508556656e-05, "loss": 0.7139, "step": 4920 }, { "epoch": 0.33032448575551154, "grad_norm": 0.940497100353241, "learning_rate": 9.528484880115361e-05, "loss": 0.6638, "step": 4922 }, { "epoch": 0.3304587094392806, "grad_norm": 1.4594491720199585, "learning_rate": 9.52802403771206e-05, "loss": 0.8803, "step": 4924 }, { "epoch": 0.33059293312304955, "grad_norm": 1.187603235244751, "learning_rate": 9.527562981368525e-05, "loss": 0.6924, "step": 4926 }, { "epoch": 0.3307271568068186, "grad_norm": 1.187653660774231, "learning_rate": 9.527101711106546e-05, "loss": 0.6564, "step": 4928 }, { "epoch": 0.33086138049058755, "grad_norm": 1.141374945640564, "learning_rate": 9.52664022694791e-05, "loss": 0.7499, "step": 4930 }, { "epoch": 0.3309956041743566, "grad_norm": 1.122341275215149, "learning_rate": 9.526178528914425e-05, "loss": 0.6705, "step": 4932 }, { "epoch": 0.33112982785812556, "grad_norm": 1.0938045978546143, "learning_rate": 9.525716617027906e-05, "loss": 0.6893, "step": 4934 }, { "epoch": 0.3312640515418946, "grad_norm": 1.0191075801849365, "learning_rate": 9.525254491310176e-05, "loss": 0.6375, "step": 4936 }, { "epoch": 0.33139827522566356, "grad_norm": 1.026029109954834, "learning_rate": 9.524792151783069e-05, "loss": 0.6443, "step": 4938 }, { "epoch": 0.3315324989094326, "grad_norm": 1.1697235107421875, "learning_rate": 9.524329598468431e-05, "loss": 0.6307, "step": 4940 }, { "epoch": 0.33166672259320157, "grad_norm": 1.4177303314208984, "learning_rate": 9.523866831388116e-05, "loss": 0.727, "step": 4942 }, { "epoch": 0.3318009462769706, "grad_norm": 1.0223388671875, "learning_rate": 9.52340385056399e-05, "loss": 0.6492, "step": 4944 }, { "epoch": 0.3319351699607396, "grad_norm": 1.0136746168136597, "learning_rate": 9.522940656017926e-05, "loss": 0.7031, "step": 4946 }, { "epoch": 0.33206939364450855, "grad_norm": 1.5003118515014648, "learning_rate": 9.52247724777181e-05, "loss": 0.6505, "step": 4948 }, { "epoch": 0.3322036173282776, "grad_norm": 1.07683265209198, "learning_rate": 9.522013625847537e-05, "loss": 0.7223, "step": 4950 }, { "epoch": 0.33233784101204655, "grad_norm": 1.5895766019821167, "learning_rate": 9.521549790267013e-05, "loss": 0.6667, "step": 4952 }, { "epoch": 0.3324720646958156, "grad_norm": 1.0356003046035767, "learning_rate": 9.521085741052152e-05, "loss": 0.6584, "step": 4954 }, { "epoch": 0.33260628837958456, "grad_norm": 1.148694634437561, "learning_rate": 9.52062147822488e-05, "loss": 0.6871, "step": 4956 }, { "epoch": 0.3327405120633536, "grad_norm": 1.2603338956832886, "learning_rate": 9.520157001807133e-05, "loss": 0.7459, "step": 4958 }, { "epoch": 0.33287473574712256, "grad_norm": 1.1903748512268066, "learning_rate": 9.519692311820856e-05, "loss": 0.6372, "step": 4960 }, { "epoch": 0.3330089594308916, "grad_norm": 0.9854759573936462, "learning_rate": 9.519227408288006e-05, "loss": 0.6916, "step": 4962 }, { "epoch": 0.33314318311466057, "grad_norm": 1.0852856636047363, "learning_rate": 9.518762291230546e-05, "loss": 0.7023, "step": 4964 }, { "epoch": 0.3332774067984296, "grad_norm": 0.9276344776153564, "learning_rate": 9.518296960670455e-05, "loss": 0.6828, "step": 4966 }, { "epoch": 0.3334116304821986, "grad_norm": 1.0925185680389404, "learning_rate": 9.517831416629716e-05, "loss": 0.6997, "step": 4968 }, { "epoch": 0.3335458541659676, "grad_norm": 1.4964567422866821, "learning_rate": 9.517365659130326e-05, "loss": 0.671, "step": 4970 }, { "epoch": 0.3336800778497366, "grad_norm": 1.0770866870880127, "learning_rate": 9.516899688194294e-05, "loss": 0.6594, "step": 4972 }, { "epoch": 0.3338143015335056, "grad_norm": 1.1555149555206299, "learning_rate": 9.516433503843631e-05, "loss": 0.7312, "step": 4974 }, { "epoch": 0.3339485252172746, "grad_norm": 1.0918995141983032, "learning_rate": 9.515967106100368e-05, "loss": 0.7366, "step": 4976 }, { "epoch": 0.3340827489010436, "grad_norm": 1.065820574760437, "learning_rate": 9.515500494986541e-05, "loss": 0.7003, "step": 4978 }, { "epoch": 0.3342169725848126, "grad_norm": 1.0946167707443237, "learning_rate": 9.515033670524192e-05, "loss": 0.7005, "step": 4980 }, { "epoch": 0.3343511962685816, "grad_norm": 1.5920342206954956, "learning_rate": 9.514566632735382e-05, "loss": 0.6515, "step": 4982 }, { "epoch": 0.3344854199523506, "grad_norm": 1.3612632751464844, "learning_rate": 9.514099381642175e-05, "loss": 0.7155, "step": 4984 }, { "epoch": 0.33461964363611957, "grad_norm": 1.3884814977645874, "learning_rate": 9.51363191726665e-05, "loss": 0.7438, "step": 4986 }, { "epoch": 0.3347538673198886, "grad_norm": 0.9352781772613525, "learning_rate": 9.513164239630891e-05, "loss": 0.6303, "step": 4988 }, { "epoch": 0.3348880910036576, "grad_norm": 1.0428996086120605, "learning_rate": 9.512696348756997e-05, "loss": 0.6896, "step": 4990 }, { "epoch": 0.3350223146874266, "grad_norm": 1.1056615114212036, "learning_rate": 9.512228244667076e-05, "loss": 0.668, "step": 4992 }, { "epoch": 0.3351565383711956, "grad_norm": 1.1554951667785645, "learning_rate": 9.511759927383243e-05, "loss": 0.6924, "step": 4994 }, { "epoch": 0.3352907620549646, "grad_norm": 1.0461509227752686, "learning_rate": 9.511291396927625e-05, "loss": 0.6861, "step": 4996 }, { "epoch": 0.3354249857387336, "grad_norm": 1.0818904638290405, "learning_rate": 9.510822653322359e-05, "loss": 0.6708, "step": 4998 }, { "epoch": 0.3355592094225026, "grad_norm": 1.0191214084625244, "learning_rate": 9.510353696589593e-05, "loss": 0.6678, "step": 5000 }, { "epoch": 0.3356934331062716, "grad_norm": 1.0327562093734741, "learning_rate": 9.509884526751485e-05, "loss": 0.6998, "step": 5002 }, { "epoch": 0.3358276567900406, "grad_norm": 1.0621801614761353, "learning_rate": 9.5094151438302e-05, "loss": 0.7418, "step": 5004 }, { "epoch": 0.3359618804738096, "grad_norm": 1.079667329788208, "learning_rate": 9.508945547847916e-05, "loss": 0.7226, "step": 5006 }, { "epoch": 0.3360961041575786, "grad_norm": 1.0097723007202148, "learning_rate": 9.508475738826823e-05, "loss": 0.748, "step": 5008 }, { "epoch": 0.3362303278413476, "grad_norm": 1.095018744468689, "learning_rate": 9.508005716789117e-05, "loss": 0.7191, "step": 5010 }, { "epoch": 0.33636455152511663, "grad_norm": 1.254999041557312, "learning_rate": 9.507535481757005e-05, "loss": 0.7584, "step": 5012 }, { "epoch": 0.3364987752088856, "grad_norm": 1.6734964847564697, "learning_rate": 9.507065033752704e-05, "loss": 0.7268, "step": 5014 }, { "epoch": 0.33663299889265463, "grad_norm": 1.0575109720230103, "learning_rate": 9.506594372798446e-05, "loss": 0.7239, "step": 5016 }, { "epoch": 0.3367672225764236, "grad_norm": 1.4216424226760864, "learning_rate": 9.506123498916463e-05, "loss": 0.6463, "step": 5018 }, { "epoch": 0.33690144626019264, "grad_norm": 0.9909161329269409, "learning_rate": 9.505652412129008e-05, "loss": 0.6506, "step": 5020 }, { "epoch": 0.3370356699439616, "grad_norm": 1.262615442276001, "learning_rate": 9.505181112458335e-05, "loss": 0.7405, "step": 5022 }, { "epoch": 0.3371698936277306, "grad_norm": 1.5525568723678589, "learning_rate": 9.504709599926712e-05, "loss": 0.6885, "step": 5024 }, { "epoch": 0.3373041173114996, "grad_norm": 1.0351464748382568, "learning_rate": 9.504237874556421e-05, "loss": 0.6916, "step": 5026 }, { "epoch": 0.3374383409952686, "grad_norm": 1.0131417512893677, "learning_rate": 9.503765936369746e-05, "loss": 0.6773, "step": 5028 }, { "epoch": 0.3375725646790376, "grad_norm": 1.277593731880188, "learning_rate": 9.503293785388987e-05, "loss": 0.7126, "step": 5030 }, { "epoch": 0.3377067883628066, "grad_norm": 1.1509110927581787, "learning_rate": 9.502821421636454e-05, "loss": 0.7054, "step": 5032 }, { "epoch": 0.33784101204657563, "grad_norm": 1.7441537380218506, "learning_rate": 9.502348845134461e-05, "loss": 0.68, "step": 5034 }, { "epoch": 0.3379752357303446, "grad_norm": 1.1158937215805054, "learning_rate": 9.501876055905339e-05, "loss": 0.7326, "step": 5036 }, { "epoch": 0.33810945941411363, "grad_norm": 1.407645583152771, "learning_rate": 9.501403053971427e-05, "loss": 0.6986, "step": 5038 }, { "epoch": 0.3382436830978826, "grad_norm": 1.029934048652649, "learning_rate": 9.500929839355071e-05, "loss": 0.6293, "step": 5040 }, { "epoch": 0.33837790678165164, "grad_norm": 1.1010067462921143, "learning_rate": 9.500456412078631e-05, "loss": 0.7056, "step": 5042 }, { "epoch": 0.3385121304654206, "grad_norm": 1.2496752738952637, "learning_rate": 9.499982772164475e-05, "loss": 0.7108, "step": 5044 }, { "epoch": 0.33864635414918964, "grad_norm": 1.099348783493042, "learning_rate": 9.499508919634983e-05, "loss": 0.6907, "step": 5046 }, { "epoch": 0.3387805778329586, "grad_norm": 1.123132348060608, "learning_rate": 9.499034854512542e-05, "loss": 0.713, "step": 5048 }, { "epoch": 0.33891480151672765, "grad_norm": 1.031817078590393, "learning_rate": 9.49856057681955e-05, "loss": 0.7342, "step": 5050 }, { "epoch": 0.3390490252004966, "grad_norm": 1.2734098434448242, "learning_rate": 9.498086086578418e-05, "loss": 0.7487, "step": 5052 }, { "epoch": 0.33918324888426565, "grad_norm": 1.0864171981811523, "learning_rate": 9.497611383811564e-05, "loss": 0.6994, "step": 5054 }, { "epoch": 0.33931747256803463, "grad_norm": 1.0995585918426514, "learning_rate": 9.497136468541415e-05, "loss": 0.6424, "step": 5056 }, { "epoch": 0.33945169625180366, "grad_norm": 1.089625358581543, "learning_rate": 9.496661340790411e-05, "loss": 0.7, "step": 5058 }, { "epoch": 0.33958591993557263, "grad_norm": 1.0574417114257812, "learning_rate": 9.496186000581002e-05, "loss": 0.7276, "step": 5060 }, { "epoch": 0.3397201436193416, "grad_norm": 1.0373708009719849, "learning_rate": 9.495710447935646e-05, "loss": 0.681, "step": 5062 }, { "epoch": 0.33985436730311064, "grad_norm": 1.1043150424957275, "learning_rate": 9.49523468287681e-05, "loss": 0.7475, "step": 5064 }, { "epoch": 0.3399885909868796, "grad_norm": 1.123861312866211, "learning_rate": 9.494758705426978e-05, "loss": 0.7078, "step": 5066 }, { "epoch": 0.34012281467064864, "grad_norm": 1.1560646295547485, "learning_rate": 9.494282515608632e-05, "loss": 0.6853, "step": 5068 }, { "epoch": 0.3402570383544176, "grad_norm": 1.0215603113174438, "learning_rate": 9.493806113444277e-05, "loss": 0.732, "step": 5070 }, { "epoch": 0.34039126203818665, "grad_norm": 0.9919569492340088, "learning_rate": 9.493329498956421e-05, "loss": 0.6517, "step": 5072 }, { "epoch": 0.3405254857219556, "grad_norm": 1.1932328939437866, "learning_rate": 9.492852672167578e-05, "loss": 0.6514, "step": 5074 }, { "epoch": 0.34065970940572465, "grad_norm": 1.3314809799194336, "learning_rate": 9.492375633100283e-05, "loss": 0.6408, "step": 5076 }, { "epoch": 0.34079393308949363, "grad_norm": 1.1061758995056152, "learning_rate": 9.491898381777074e-05, "loss": 0.6561, "step": 5078 }, { "epoch": 0.34092815677326266, "grad_norm": 1.0593011379241943, "learning_rate": 9.4914209182205e-05, "loss": 0.7318, "step": 5080 }, { "epoch": 0.34106238045703163, "grad_norm": 1.0847229957580566, "learning_rate": 9.490943242453118e-05, "loss": 0.693, "step": 5082 }, { "epoch": 0.34119660414080066, "grad_norm": 1.0208863019943237, "learning_rate": 9.4904653544975e-05, "loss": 0.6298, "step": 5084 }, { "epoch": 0.34133082782456964, "grad_norm": 1.1964168548583984, "learning_rate": 9.489987254376222e-05, "loss": 0.7066, "step": 5086 }, { "epoch": 0.34146505150833867, "grad_norm": 1.3700635433197021, "learning_rate": 9.489508942111878e-05, "loss": 0.7003, "step": 5088 }, { "epoch": 0.34159927519210764, "grad_norm": 1.1736687421798706, "learning_rate": 9.489030417727063e-05, "loss": 0.7113, "step": 5090 }, { "epoch": 0.3417334988758767, "grad_norm": 1.2480403184890747, "learning_rate": 9.488551681244388e-05, "loss": 0.7044, "step": 5092 }, { "epoch": 0.34186772255964565, "grad_norm": 0.9867148995399475, "learning_rate": 9.488072732686474e-05, "loss": 0.6922, "step": 5094 }, { "epoch": 0.3420019462434146, "grad_norm": 0.9716460704803467, "learning_rate": 9.487593572075948e-05, "loss": 0.6429, "step": 5096 }, { "epoch": 0.34213616992718365, "grad_norm": 1.0517593622207642, "learning_rate": 9.48711419943545e-05, "loss": 0.7262, "step": 5098 }, { "epoch": 0.34227039361095263, "grad_norm": 1.0895615816116333, "learning_rate": 9.486634614787631e-05, "loss": 0.6671, "step": 5100 }, { "epoch": 0.34240461729472166, "grad_norm": 1.1830127239227295, "learning_rate": 9.486154818155146e-05, "loss": 0.7024, "step": 5102 }, { "epoch": 0.34253884097849063, "grad_norm": 0.9760610461235046, "learning_rate": 9.485674809560669e-05, "loss": 0.6577, "step": 5104 }, { "epoch": 0.34267306466225966, "grad_norm": 1.0928126573562622, "learning_rate": 9.485194589026878e-05, "loss": 0.6874, "step": 5106 }, { "epoch": 0.34280728834602864, "grad_norm": 1.0870157480239868, "learning_rate": 9.484714156576464e-05, "loss": 0.7024, "step": 5108 }, { "epoch": 0.34294151202979767, "grad_norm": 1.1574848890304565, "learning_rate": 9.484233512232123e-05, "loss": 0.727, "step": 5110 }, { "epoch": 0.34307573571356664, "grad_norm": 1.5628575086593628, "learning_rate": 9.483752656016567e-05, "loss": 0.7628, "step": 5112 }, { "epoch": 0.3432099593973357, "grad_norm": 1.4135750532150269, "learning_rate": 9.483271587952515e-05, "loss": 0.6959, "step": 5114 }, { "epoch": 0.34334418308110465, "grad_norm": 1.5119282007217407, "learning_rate": 9.482790308062697e-05, "loss": 0.7533, "step": 5116 }, { "epoch": 0.3434784067648737, "grad_norm": 1.1340827941894531, "learning_rate": 9.48230881636985e-05, "loss": 0.7502, "step": 5118 }, { "epoch": 0.34361263044864265, "grad_norm": 1.1876959800720215, "learning_rate": 9.481827112896727e-05, "loss": 0.7444, "step": 5120 }, { "epoch": 0.3437468541324117, "grad_norm": 1.0453839302062988, "learning_rate": 9.481345197666087e-05, "loss": 0.6689, "step": 5122 }, { "epoch": 0.34388107781618066, "grad_norm": 1.0681465864181519, "learning_rate": 9.480863070700696e-05, "loss": 0.6624, "step": 5124 }, { "epoch": 0.3440153014999497, "grad_norm": 1.1562570333480835, "learning_rate": 9.480380732023338e-05, "loss": 0.7213, "step": 5126 }, { "epoch": 0.34414952518371866, "grad_norm": 1.0983346700668335, "learning_rate": 9.479898181656801e-05, "loss": 0.6863, "step": 5128 }, { "epoch": 0.3442837488674877, "grad_norm": 1.2765334844589233, "learning_rate": 9.479415419623883e-05, "loss": 0.7889, "step": 5130 }, { "epoch": 0.34441797255125667, "grad_norm": 1.126746654510498, "learning_rate": 9.478932445947395e-05, "loss": 0.6746, "step": 5132 }, { "epoch": 0.34455219623502564, "grad_norm": 1.1382261514663696, "learning_rate": 9.478449260650158e-05, "loss": 0.7237, "step": 5134 }, { "epoch": 0.3446864199187947, "grad_norm": 1.141695261001587, "learning_rate": 9.477965863754998e-05, "loss": 0.6959, "step": 5136 }, { "epoch": 0.34482064360256365, "grad_norm": 1.0475116968154907, "learning_rate": 9.477482255284757e-05, "loss": 0.7596, "step": 5138 }, { "epoch": 0.3449548672863327, "grad_norm": 1.4221386909484863, "learning_rate": 9.476998435262284e-05, "loss": 0.6576, "step": 5140 }, { "epoch": 0.34508909097010165, "grad_norm": 0.9887280464172363, "learning_rate": 9.476514403710439e-05, "loss": 0.7574, "step": 5142 }, { "epoch": 0.3452233146538707, "grad_norm": 0.9547103047370911, "learning_rate": 9.476030160652091e-05, "loss": 0.62, "step": 5144 }, { "epoch": 0.34535753833763966, "grad_norm": 1.191341757774353, "learning_rate": 9.475545706110119e-05, "loss": 0.7415, "step": 5146 }, { "epoch": 0.3454917620214087, "grad_norm": 1.0935919284820557, "learning_rate": 9.475061040107414e-05, "loss": 0.7202, "step": 5148 }, { "epoch": 0.34562598570517766, "grad_norm": 1.2453701496124268, "learning_rate": 9.474576162666874e-05, "loss": 0.7048, "step": 5150 }, { "epoch": 0.3457602093889467, "grad_norm": 1.120084524154663, "learning_rate": 9.474091073811409e-05, "loss": 0.6792, "step": 5152 }, { "epoch": 0.34589443307271567, "grad_norm": 1.1592801809310913, "learning_rate": 9.47360577356394e-05, "loss": 0.7244, "step": 5154 }, { "epoch": 0.3460286567564847, "grad_norm": 1.1230835914611816, "learning_rate": 9.473120261947395e-05, "loss": 0.7103, "step": 5156 }, { "epoch": 0.3461628804402537, "grad_norm": 0.9806675910949707, "learning_rate": 9.472634538984712e-05, "loss": 0.6911, "step": 5158 }, { "epoch": 0.3462971041240227, "grad_norm": 1.457768201828003, "learning_rate": 9.472148604698843e-05, "loss": 0.7743, "step": 5160 }, { "epoch": 0.3464313278077917, "grad_norm": 1.0284276008605957, "learning_rate": 9.471662459112747e-05, "loss": 0.7379, "step": 5162 }, { "epoch": 0.3465655514915607, "grad_norm": 1.2664660215377808, "learning_rate": 9.471176102249393e-05, "loss": 0.587, "step": 5164 }, { "epoch": 0.3466997751753297, "grad_norm": 1.3976407051086426, "learning_rate": 9.47068953413176e-05, "loss": 0.7207, "step": 5166 }, { "epoch": 0.3468339988590987, "grad_norm": 1.140342354774475, "learning_rate": 9.470202754782837e-05, "loss": 0.6351, "step": 5168 }, { "epoch": 0.3469682225428677, "grad_norm": 1.0120048522949219, "learning_rate": 9.469715764225626e-05, "loss": 0.6729, "step": 5170 }, { "epoch": 0.34710244622663666, "grad_norm": 1.1098517179489136, "learning_rate": 9.469228562483132e-05, "loss": 0.6339, "step": 5172 }, { "epoch": 0.3472366699104057, "grad_norm": 0.980413556098938, "learning_rate": 9.468741149578379e-05, "loss": 0.6917, "step": 5174 }, { "epoch": 0.34737089359417467, "grad_norm": 1.0559087991714478, "learning_rate": 9.468253525534393e-05, "loss": 0.7193, "step": 5176 }, { "epoch": 0.3475051172779437, "grad_norm": 1.0527195930480957, "learning_rate": 9.467765690374214e-05, "loss": 0.6606, "step": 5178 }, { "epoch": 0.3476393409617127, "grad_norm": 1.15939462184906, "learning_rate": 9.467277644120893e-05, "loss": 0.7285, "step": 5180 }, { "epoch": 0.3477735646454817, "grad_norm": 1.1253271102905273, "learning_rate": 9.466789386797486e-05, "loss": 0.7025, "step": 5182 }, { "epoch": 0.3479077883292507, "grad_norm": 1.0979676246643066, "learning_rate": 9.466300918427065e-05, "loss": 0.7205, "step": 5184 }, { "epoch": 0.3480420120130197, "grad_norm": 0.9581397771835327, "learning_rate": 9.465812239032708e-05, "loss": 0.6445, "step": 5186 }, { "epoch": 0.3481762356967887, "grad_norm": 1.073517918586731, "learning_rate": 9.465323348637505e-05, "loss": 0.6466, "step": 5188 }, { "epoch": 0.3483104593805577, "grad_norm": 1.308125376701355, "learning_rate": 9.464834247264553e-05, "loss": 0.6924, "step": 5190 }, { "epoch": 0.3484446830643267, "grad_norm": 1.2113579511642456, "learning_rate": 9.464344934936964e-05, "loss": 0.6725, "step": 5192 }, { "epoch": 0.3485789067480957, "grad_norm": 1.140137791633606, "learning_rate": 9.463855411677856e-05, "loss": 0.6954, "step": 5194 }, { "epoch": 0.3487131304318647, "grad_norm": 1.337956190109253, "learning_rate": 9.463365677510357e-05, "loss": 0.7545, "step": 5196 }, { "epoch": 0.3488473541156337, "grad_norm": 1.0394006967544556, "learning_rate": 9.462875732457606e-05, "loss": 0.6959, "step": 5198 }, { "epoch": 0.3489815777994027, "grad_norm": 1.1676454544067383, "learning_rate": 9.462385576542752e-05, "loss": 0.6698, "step": 5200 }, { "epoch": 0.34911580148317173, "grad_norm": 1.0103447437286377, "learning_rate": 9.461895209788956e-05, "loss": 0.6978, "step": 5202 }, { "epoch": 0.3492500251669407, "grad_norm": 1.0410352945327759, "learning_rate": 9.461404632219384e-05, "loss": 0.6244, "step": 5204 }, { "epoch": 0.34938424885070973, "grad_norm": 1.0641223192214966, "learning_rate": 9.460913843857217e-05, "loss": 0.7342, "step": 5206 }, { "epoch": 0.3495184725344787, "grad_norm": 1.1432600021362305, "learning_rate": 9.460422844725642e-05, "loss": 0.6896, "step": 5208 }, { "epoch": 0.3496526962182477, "grad_norm": 1.2629886865615845, "learning_rate": 9.459931634847859e-05, "loss": 0.7591, "step": 5210 }, { "epoch": 0.3497869199020167, "grad_norm": 1.0171542167663574, "learning_rate": 9.459440214247077e-05, "loss": 0.6011, "step": 5212 }, { "epoch": 0.3499211435857857, "grad_norm": 1.1639806032180786, "learning_rate": 9.458948582946514e-05, "loss": 0.6729, "step": 5214 }, { "epoch": 0.3500553672695547, "grad_norm": 1.0054445266723633, "learning_rate": 9.458456740969397e-05, "loss": 0.644, "step": 5216 }, { "epoch": 0.3501895909533237, "grad_norm": 0.9366871118545532, "learning_rate": 9.457964688338967e-05, "loss": 0.6568, "step": 5218 }, { "epoch": 0.3503238146370927, "grad_norm": 1.5994693040847778, "learning_rate": 9.457472425078473e-05, "loss": 0.7491, "step": 5220 }, { "epoch": 0.3504580383208617, "grad_norm": 1.1239392757415771, "learning_rate": 9.456979951211172e-05, "loss": 0.5976, "step": 5222 }, { "epoch": 0.35059226200463073, "grad_norm": 1.031184196472168, "learning_rate": 9.45648726676033e-05, "loss": 0.6977, "step": 5224 }, { "epoch": 0.3507264856883997, "grad_norm": 1.017400860786438, "learning_rate": 9.455994371749231e-05, "loss": 0.6405, "step": 5226 }, { "epoch": 0.35086070937216873, "grad_norm": 1.146761417388916, "learning_rate": 9.45550126620116e-05, "loss": 0.7208, "step": 5228 }, { "epoch": 0.3509949330559377, "grad_norm": 1.1883848905563354, "learning_rate": 9.455007950139412e-05, "loss": 0.6999, "step": 5230 }, { "epoch": 0.35112915673970674, "grad_norm": 1.1775646209716797, "learning_rate": 9.454514423587301e-05, "loss": 0.6619, "step": 5232 }, { "epoch": 0.3512633804234757, "grad_norm": 1.1438260078430176, "learning_rate": 9.454020686568143e-05, "loss": 0.727, "step": 5234 }, { "epoch": 0.35139760410724474, "grad_norm": 1.1475664377212524, "learning_rate": 9.453526739105267e-05, "loss": 0.6989, "step": 5236 }, { "epoch": 0.3515318277910137, "grad_norm": 1.0647696256637573, "learning_rate": 9.45303258122201e-05, "loss": 0.6298, "step": 5238 }, { "epoch": 0.35166605147478275, "grad_norm": 2.985783338546753, "learning_rate": 9.452538212941719e-05, "loss": 0.6947, "step": 5240 }, { "epoch": 0.3518002751585517, "grad_norm": 0.9413696527481079, "learning_rate": 9.452043634287753e-05, "loss": 0.714, "step": 5242 }, { "epoch": 0.35193449884232075, "grad_norm": 1.4837936162948608, "learning_rate": 9.45154884528348e-05, "loss": 0.7204, "step": 5244 }, { "epoch": 0.35206872252608973, "grad_norm": 1.350868821144104, "learning_rate": 9.451053845952278e-05, "loss": 0.6835, "step": 5246 }, { "epoch": 0.3522029462098587, "grad_norm": 1.0339003801345825, "learning_rate": 9.450558636317533e-05, "loss": 0.6535, "step": 5248 }, { "epoch": 0.35233716989362773, "grad_norm": 1.4247851371765137, "learning_rate": 9.450063216402644e-05, "loss": 0.7949, "step": 5250 }, { "epoch": 0.3524713935773967, "grad_norm": 1.0391966104507446, "learning_rate": 9.44956758623102e-05, "loss": 0.6573, "step": 5252 }, { "epoch": 0.35260561726116574, "grad_norm": 1.1045339107513428, "learning_rate": 9.44907174582608e-05, "loss": 0.7038, "step": 5254 }, { "epoch": 0.3527398409449347, "grad_norm": 1.0392239093780518, "learning_rate": 9.448575695211244e-05, "loss": 0.7089, "step": 5256 }, { "epoch": 0.35287406462870374, "grad_norm": 1.0227470397949219, "learning_rate": 9.448079434409956e-05, "loss": 0.6423, "step": 5258 }, { "epoch": 0.3530082883124727, "grad_norm": 1.0412817001342773, "learning_rate": 9.447582963445663e-05, "loss": 0.6564, "step": 5260 }, { "epoch": 0.35314251199624175, "grad_norm": 1.1213363409042358, "learning_rate": 9.447086282341818e-05, "loss": 0.6377, "step": 5262 }, { "epoch": 0.3532767356800107, "grad_norm": 1.0754303932189941, "learning_rate": 9.446589391121893e-05, "loss": 0.7076, "step": 5264 }, { "epoch": 0.35341095936377975, "grad_norm": 1.133129596710205, "learning_rate": 9.446092289809361e-05, "loss": 0.6513, "step": 5266 }, { "epoch": 0.35354518304754873, "grad_norm": 1.0103533267974854, "learning_rate": 9.445594978427714e-05, "loss": 0.652, "step": 5268 }, { "epoch": 0.35367940673131776, "grad_norm": 1.0910634994506836, "learning_rate": 9.445097457000444e-05, "loss": 0.7455, "step": 5270 }, { "epoch": 0.35381363041508673, "grad_norm": 1.0608628988265991, "learning_rate": 9.444599725551061e-05, "loss": 0.7294, "step": 5272 }, { "epoch": 0.35394785409885576, "grad_norm": 1.2099720239639282, "learning_rate": 9.444101784103082e-05, "loss": 0.6946, "step": 5274 }, { "epoch": 0.35408207778262474, "grad_norm": 1.0302938222885132, "learning_rate": 9.443603632680031e-05, "loss": 0.6494, "step": 5276 }, { "epoch": 0.35421630146639377, "grad_norm": 1.0468542575836182, "learning_rate": 9.443105271305445e-05, "loss": 0.6824, "step": 5278 }, { "epoch": 0.35435052515016274, "grad_norm": 1.090032696723938, "learning_rate": 9.442606700002874e-05, "loss": 0.6674, "step": 5280 }, { "epoch": 0.3544847488339318, "grad_norm": 1.011366844177246, "learning_rate": 9.442107918795873e-05, "loss": 0.7143, "step": 5282 }, { "epoch": 0.35461897251770075, "grad_norm": 1.0054188966751099, "learning_rate": 9.441608927708006e-05, "loss": 0.746, "step": 5284 }, { "epoch": 0.3547531962014697, "grad_norm": 1.2857420444488525, "learning_rate": 9.441109726762852e-05, "loss": 0.7379, "step": 5286 }, { "epoch": 0.35488741988523875, "grad_norm": 1.0563364028930664, "learning_rate": 9.440610315983998e-05, "loss": 0.6357, "step": 5288 }, { "epoch": 0.3550216435690077, "grad_norm": 1.1448321342468262, "learning_rate": 9.440110695395037e-05, "loss": 0.6954, "step": 5290 }, { "epoch": 0.35515586725277676, "grad_norm": 1.1138883829116821, "learning_rate": 9.439610865019577e-05, "loss": 0.6362, "step": 5292 }, { "epoch": 0.35529009093654573, "grad_norm": 1.0992587804794312, "learning_rate": 9.439110824881232e-05, "loss": 0.7334, "step": 5294 }, { "epoch": 0.35542431462031476, "grad_norm": 1.2044397592544556, "learning_rate": 9.438610575003632e-05, "loss": 0.708, "step": 5296 }, { "epoch": 0.35555853830408374, "grad_norm": 1.0396875143051147, "learning_rate": 9.43811011541041e-05, "loss": 0.6138, "step": 5298 }, { "epoch": 0.35569276198785277, "grad_norm": 1.1773802042007446, "learning_rate": 9.437609446125211e-05, "loss": 0.7468, "step": 5300 }, { "epoch": 0.35582698567162174, "grad_norm": 1.0798813104629517, "learning_rate": 9.437108567171693e-05, "loss": 0.7499, "step": 5302 }, { "epoch": 0.3559612093553908, "grad_norm": 1.066114902496338, "learning_rate": 9.436607478573522e-05, "loss": 0.6672, "step": 5304 }, { "epoch": 0.35609543303915975, "grad_norm": 1.2825912237167358, "learning_rate": 9.436106180354369e-05, "loss": 0.7209, "step": 5306 }, { "epoch": 0.3562296567229288, "grad_norm": 1.0812764167785645, "learning_rate": 9.435604672537924e-05, "loss": 0.6479, "step": 5308 }, { "epoch": 0.35636388040669775, "grad_norm": 1.2417733669281006, "learning_rate": 9.43510295514788e-05, "loss": 0.631, "step": 5310 }, { "epoch": 0.3564981040904668, "grad_norm": 0.9054554104804993, "learning_rate": 9.434601028207942e-05, "loss": 0.5859, "step": 5312 }, { "epoch": 0.35663232777423576, "grad_norm": 1.10965895652771, "learning_rate": 9.434098891741827e-05, "loss": 0.6654, "step": 5314 }, { "epoch": 0.3567665514580048, "grad_norm": 1.175485372543335, "learning_rate": 9.433596545773258e-05, "loss": 0.7067, "step": 5316 }, { "epoch": 0.35690077514177376, "grad_norm": 1.0619306564331055, "learning_rate": 9.43309399032597e-05, "loss": 0.7069, "step": 5318 }, { "epoch": 0.3570349988255428, "grad_norm": 1.0990415811538696, "learning_rate": 9.432591225423708e-05, "loss": 0.7396, "step": 5320 }, { "epoch": 0.35716922250931177, "grad_norm": 1.2514533996582031, "learning_rate": 9.432088251090228e-05, "loss": 0.6848, "step": 5322 }, { "epoch": 0.35730344619308074, "grad_norm": 1.0155177116394043, "learning_rate": 9.431585067349293e-05, "loss": 0.6888, "step": 5324 }, { "epoch": 0.3574376698768498, "grad_norm": 1.1702946424484253, "learning_rate": 9.431081674224677e-05, "loss": 0.6642, "step": 5326 }, { "epoch": 0.35757189356061875, "grad_norm": 1.1770367622375488, "learning_rate": 9.430578071740167e-05, "loss": 0.6797, "step": 5328 }, { "epoch": 0.3577061172443878, "grad_norm": 1.0894922018051147, "learning_rate": 9.430074259919554e-05, "loss": 0.6688, "step": 5330 }, { "epoch": 0.35784034092815675, "grad_norm": 1.1621495485305786, "learning_rate": 9.429570238786645e-05, "loss": 0.6791, "step": 5332 }, { "epoch": 0.3579745646119258, "grad_norm": 1.0528497695922852, "learning_rate": 9.429066008365251e-05, "loss": 0.7114, "step": 5334 }, { "epoch": 0.35810878829569476, "grad_norm": 1.1805702447891235, "learning_rate": 9.428561568679199e-05, "loss": 0.7519, "step": 5336 }, { "epoch": 0.3582430119794638, "grad_norm": 1.0912284851074219, "learning_rate": 9.428056919752319e-05, "loss": 0.7414, "step": 5338 }, { "epoch": 0.35837723566323276, "grad_norm": 1.135192632675171, "learning_rate": 9.42755206160846e-05, "loss": 0.6375, "step": 5340 }, { "epoch": 0.3585114593470018, "grad_norm": 1.0422782897949219, "learning_rate": 9.427046994271471e-05, "loss": 0.7249, "step": 5342 }, { "epoch": 0.35864568303077077, "grad_norm": 1.1697744131088257, "learning_rate": 9.426541717765216e-05, "loss": 0.6435, "step": 5344 }, { "epoch": 0.3587799067145398, "grad_norm": 1.6935707330703735, "learning_rate": 9.426036232113571e-05, "loss": 0.7125, "step": 5346 }, { "epoch": 0.3589141303983088, "grad_norm": 1.2502233982086182, "learning_rate": 9.425530537340417e-05, "loss": 0.7024, "step": 5348 }, { "epoch": 0.3590483540820778, "grad_norm": 0.9787608981132507, "learning_rate": 9.425024633469647e-05, "loss": 0.6551, "step": 5350 }, { "epoch": 0.3591825777658468, "grad_norm": 1.5133332014083862, "learning_rate": 9.424518520525165e-05, "loss": 0.7225, "step": 5352 }, { "epoch": 0.3593168014496158, "grad_norm": 1.0226023197174072, "learning_rate": 9.424012198530882e-05, "loss": 0.6953, "step": 5354 }, { "epoch": 0.3594510251333848, "grad_norm": 1.0732944011688232, "learning_rate": 9.423505667510724e-05, "loss": 0.6278, "step": 5356 }, { "epoch": 0.3595852488171538, "grad_norm": 1.168982744216919, "learning_rate": 9.42299892748862e-05, "loss": 0.6749, "step": 5358 }, { "epoch": 0.3597194725009228, "grad_norm": 1.084213376045227, "learning_rate": 9.422491978488515e-05, "loss": 0.7054, "step": 5360 }, { "epoch": 0.35985369618469176, "grad_norm": 0.9596426486968994, "learning_rate": 9.42198482053436e-05, "loss": 0.6319, "step": 5362 }, { "epoch": 0.3599879198684608, "grad_norm": 1.0419789552688599, "learning_rate": 9.421477453650118e-05, "loss": 0.5885, "step": 5364 }, { "epoch": 0.36012214355222977, "grad_norm": 1.0027025938034058, "learning_rate": 9.420969877859761e-05, "loss": 0.661, "step": 5366 }, { "epoch": 0.3602563672359988, "grad_norm": 0.8679866194725037, "learning_rate": 9.420462093187271e-05, "loss": 0.6068, "step": 5368 }, { "epoch": 0.3603905909197678, "grad_norm": 2.679222583770752, "learning_rate": 9.419954099656638e-05, "loss": 0.7255, "step": 5370 }, { "epoch": 0.3605248146035368, "grad_norm": 1.037379264831543, "learning_rate": 9.419445897291867e-05, "loss": 0.7083, "step": 5372 }, { "epoch": 0.3606590382873058, "grad_norm": 1.331444501876831, "learning_rate": 9.418937486116968e-05, "loss": 0.673, "step": 5374 }, { "epoch": 0.3607932619710748, "grad_norm": 1.0762038230895996, "learning_rate": 9.418428866155961e-05, "loss": 0.7332, "step": 5376 }, { "epoch": 0.3609274856548438, "grad_norm": 1.0162066221237183, "learning_rate": 9.417920037432879e-05, "loss": 0.5862, "step": 5378 }, { "epoch": 0.3610617093386128, "grad_norm": 1.0344343185424805, "learning_rate": 9.417410999971762e-05, "loss": 0.6889, "step": 5380 }, { "epoch": 0.3611959330223818, "grad_norm": 1.0645183324813843, "learning_rate": 9.416901753796663e-05, "loss": 0.6554, "step": 5382 }, { "epoch": 0.3613301567061508, "grad_norm": 1.1452938318252563, "learning_rate": 9.41639229893164e-05, "loss": 0.7387, "step": 5384 }, { "epoch": 0.3614643803899198, "grad_norm": 1.0661026239395142, "learning_rate": 9.415882635400768e-05, "loss": 0.7022, "step": 5386 }, { "epoch": 0.3615986040736888, "grad_norm": 1.194131851196289, "learning_rate": 9.415372763228123e-05, "loss": 0.7464, "step": 5388 }, { "epoch": 0.3617328277574578, "grad_norm": 1.0507525205612183, "learning_rate": 9.414862682437797e-05, "loss": 0.6617, "step": 5390 }, { "epoch": 0.36186705144122683, "grad_norm": 1.1302318572998047, "learning_rate": 9.414352393053891e-05, "loss": 0.6155, "step": 5392 }, { "epoch": 0.3620012751249958, "grad_norm": 1.166967511177063, "learning_rate": 9.413841895100515e-05, "loss": 0.6621, "step": 5394 }, { "epoch": 0.36213549880876483, "grad_norm": 1.0846643447875977, "learning_rate": 9.413331188601791e-05, "loss": 0.7087, "step": 5396 }, { "epoch": 0.3622697224925338, "grad_norm": 1.0447102785110474, "learning_rate": 9.412820273581844e-05, "loss": 0.6652, "step": 5398 }, { "epoch": 0.3624039461763028, "grad_norm": 1.3294551372528076, "learning_rate": 9.412309150064817e-05, "loss": 0.651, "step": 5400 }, { "epoch": 0.3625381698600718, "grad_norm": 1.0539995431900024, "learning_rate": 9.411797818074861e-05, "loss": 0.6983, "step": 5402 }, { "epoch": 0.3626723935438408, "grad_norm": 1.0341007709503174, "learning_rate": 9.411286277636131e-05, "loss": 0.6941, "step": 5404 }, { "epoch": 0.3628066172276098, "grad_norm": 1.1077064275741577, "learning_rate": 9.410774528772802e-05, "loss": 0.6889, "step": 5406 }, { "epoch": 0.3629408409113788, "grad_norm": 1.3066242933273315, "learning_rate": 9.410262571509046e-05, "loss": 0.7075, "step": 5408 }, { "epoch": 0.3630750645951478, "grad_norm": 1.058754563331604, "learning_rate": 9.409750405869058e-05, "loss": 0.6989, "step": 5410 }, { "epoch": 0.3632092882789168, "grad_norm": 1.1940826177597046, "learning_rate": 9.409238031877034e-05, "loss": 0.7646, "step": 5412 }, { "epoch": 0.36334351196268583, "grad_norm": 1.1361353397369385, "learning_rate": 9.408725449557184e-05, "loss": 0.7342, "step": 5414 }, { "epoch": 0.3634777356464548, "grad_norm": 1.0520503520965576, "learning_rate": 9.408212658933726e-05, "loss": 0.6822, "step": 5416 }, { "epoch": 0.36361195933022383, "grad_norm": 1.1402872800827026, "learning_rate": 9.407699660030888e-05, "loss": 0.7148, "step": 5418 }, { "epoch": 0.3637461830139928, "grad_norm": 1.1878187656402588, "learning_rate": 9.407186452872908e-05, "loss": 0.7087, "step": 5420 }, { "epoch": 0.36388040669776184, "grad_norm": 1.087986707687378, "learning_rate": 9.406673037484035e-05, "loss": 0.721, "step": 5422 }, { "epoch": 0.3640146303815308, "grad_norm": 0.9601892232894897, "learning_rate": 9.406159413888527e-05, "loss": 0.6224, "step": 5424 }, { "epoch": 0.36414885406529984, "grad_norm": 1.0169936418533325, "learning_rate": 9.40564558211065e-05, "loss": 0.6989, "step": 5426 }, { "epoch": 0.3642830777490688, "grad_norm": 1.0120033025741577, "learning_rate": 9.405131542174684e-05, "loss": 0.73, "step": 5428 }, { "epoch": 0.36441730143283785, "grad_norm": 1.2148478031158447, "learning_rate": 9.404617294104911e-05, "loss": 0.8287, "step": 5430 }, { "epoch": 0.3645515251166068, "grad_norm": 1.4467118978500366, "learning_rate": 9.404102837925637e-05, "loss": 0.7313, "step": 5432 }, { "epoch": 0.36468574880037585, "grad_norm": 1.2116284370422363, "learning_rate": 9.403588173661162e-05, "loss": 0.6943, "step": 5434 }, { "epoch": 0.3648199724841448, "grad_norm": 1.098501443862915, "learning_rate": 9.403073301335805e-05, "loss": 0.7122, "step": 5436 }, { "epoch": 0.3649541961679138, "grad_norm": 1.039089322090149, "learning_rate": 9.402558220973892e-05, "loss": 0.6847, "step": 5438 }, { "epoch": 0.36508841985168283, "grad_norm": 1.4157084226608276, "learning_rate": 9.402042932599762e-05, "loss": 0.693, "step": 5440 }, { "epoch": 0.3652226435354518, "grad_norm": 1.2092018127441406, "learning_rate": 9.401527436237758e-05, "loss": 0.7003, "step": 5442 }, { "epoch": 0.36535686721922084, "grad_norm": 1.0189353227615356, "learning_rate": 9.40101173191224e-05, "loss": 0.609, "step": 5444 }, { "epoch": 0.3654910909029898, "grad_norm": 1.053718090057373, "learning_rate": 9.40049581964757e-05, "loss": 0.6704, "step": 5446 }, { "epoch": 0.36562531458675884, "grad_norm": 1.7575159072875977, "learning_rate": 9.399979699468126e-05, "loss": 0.6315, "step": 5448 }, { "epoch": 0.3657595382705278, "grad_norm": 1.2005894184112549, "learning_rate": 9.399463371398295e-05, "loss": 0.7318, "step": 5450 }, { "epoch": 0.36589376195429685, "grad_norm": 0.9507780075073242, "learning_rate": 9.398946835462469e-05, "loss": 0.6205, "step": 5452 }, { "epoch": 0.3660279856380658, "grad_norm": 0.9422311186790466, "learning_rate": 9.398430091685056e-05, "loss": 0.6896, "step": 5454 }, { "epoch": 0.36616220932183485, "grad_norm": 1.461273193359375, "learning_rate": 9.397913140090471e-05, "loss": 0.6858, "step": 5456 }, { "epoch": 0.3662964330056038, "grad_norm": 1.0395759344100952, "learning_rate": 9.397395980703137e-05, "loss": 0.7179, "step": 5458 }, { "epoch": 0.36643065668937286, "grad_norm": 0.9786043763160706, "learning_rate": 9.39687861354749e-05, "loss": 0.6653, "step": 5460 }, { "epoch": 0.36656488037314183, "grad_norm": 1.1348570585250854, "learning_rate": 9.396361038647976e-05, "loss": 0.7807, "step": 5462 }, { "epoch": 0.36669910405691086, "grad_norm": 1.1227689981460571, "learning_rate": 9.395843256029047e-05, "loss": 0.715, "step": 5464 }, { "epoch": 0.36683332774067984, "grad_norm": 1.1113135814666748, "learning_rate": 9.39532526571517e-05, "loss": 0.7552, "step": 5466 }, { "epoch": 0.36696755142444887, "grad_norm": 1.094340205192566, "learning_rate": 9.394807067730814e-05, "loss": 0.6482, "step": 5468 }, { "epoch": 0.36710177510821784, "grad_norm": 1.2362637519836426, "learning_rate": 9.394288662100467e-05, "loss": 0.7001, "step": 5470 }, { "epoch": 0.3672359987919869, "grad_norm": 0.9833071827888489, "learning_rate": 9.393770048848622e-05, "loss": 0.6472, "step": 5472 }, { "epoch": 0.36737022247575585, "grad_norm": 1.1718744039535522, "learning_rate": 9.393251227999784e-05, "loss": 0.6698, "step": 5474 }, { "epoch": 0.3675044461595248, "grad_norm": 1.0212267637252808, "learning_rate": 9.392732199578462e-05, "loss": 0.6415, "step": 5476 }, { "epoch": 0.36763866984329385, "grad_norm": 1.1257481575012207, "learning_rate": 9.392212963609183e-05, "loss": 0.6981, "step": 5478 }, { "epoch": 0.3677728935270628, "grad_norm": 1.2375941276550293, "learning_rate": 9.391693520116477e-05, "loss": 0.7004, "step": 5480 }, { "epoch": 0.36790711721083186, "grad_norm": 1.1024177074432373, "learning_rate": 9.391173869124889e-05, "loss": 0.6756, "step": 5482 }, { "epoch": 0.36804134089460083, "grad_norm": 1.1458148956298828, "learning_rate": 9.390654010658971e-05, "loss": 0.6584, "step": 5484 }, { "epoch": 0.36817556457836986, "grad_norm": 1.274086594581604, "learning_rate": 9.390133944743284e-05, "loss": 0.6454, "step": 5486 }, { "epoch": 0.36830978826213884, "grad_norm": 1.0998162031173706, "learning_rate": 9.389613671402402e-05, "loss": 0.6255, "step": 5488 }, { "epoch": 0.36844401194590787, "grad_norm": 1.1039185523986816, "learning_rate": 9.389093190660905e-05, "loss": 0.7152, "step": 5490 }, { "epoch": 0.36857823562967684, "grad_norm": 0.9752369523048401, "learning_rate": 9.388572502543384e-05, "loss": 0.6572, "step": 5492 }, { "epoch": 0.3687124593134459, "grad_norm": 1.011757254600525, "learning_rate": 9.388051607074445e-05, "loss": 0.7229, "step": 5494 }, { "epoch": 0.36884668299721485, "grad_norm": 1.072043776512146, "learning_rate": 9.387530504278695e-05, "loss": 0.6519, "step": 5496 }, { "epoch": 0.3689809066809839, "grad_norm": 1.2205461263656616, "learning_rate": 9.387009194180755e-05, "loss": 0.6945, "step": 5498 }, { "epoch": 0.36911513036475285, "grad_norm": 0.9959816336631775, "learning_rate": 9.38648767680526e-05, "loss": 0.6862, "step": 5500 }, { "epoch": 0.3692493540485219, "grad_norm": 1.0489308834075928, "learning_rate": 9.385965952176847e-05, "loss": 0.6726, "step": 5502 }, { "epoch": 0.36938357773229086, "grad_norm": 1.0031033754348755, "learning_rate": 9.385444020320166e-05, "loss": 0.6779, "step": 5504 }, { "epoch": 0.3695178014160599, "grad_norm": 1.0508573055267334, "learning_rate": 9.38492188125988e-05, "loss": 0.6764, "step": 5506 }, { "epoch": 0.36965202509982886, "grad_norm": 1.0428763628005981, "learning_rate": 9.384399535020657e-05, "loss": 0.6462, "step": 5508 }, { "epoch": 0.3697862487835979, "grad_norm": 1.3005393743515015, "learning_rate": 9.383876981627178e-05, "loss": 0.7182, "step": 5510 }, { "epoch": 0.36992047246736687, "grad_norm": 1.1237488985061646, "learning_rate": 9.383354221104132e-05, "loss": 0.7415, "step": 5512 }, { "epoch": 0.37005469615113584, "grad_norm": 1.087746024131775, "learning_rate": 9.382831253476219e-05, "loss": 0.6635, "step": 5514 }, { "epoch": 0.3701889198349049, "grad_norm": 1.189430832862854, "learning_rate": 9.382308078768146e-05, "loss": 0.6528, "step": 5516 }, { "epoch": 0.37032314351867385, "grad_norm": 1.0363761186599731, "learning_rate": 9.381784697004636e-05, "loss": 0.6923, "step": 5518 }, { "epoch": 0.3704573672024429, "grad_norm": 1.0994951725006104, "learning_rate": 9.381261108210412e-05, "loss": 0.6299, "step": 5520 }, { "epoch": 0.37059159088621185, "grad_norm": 1.5011996030807495, "learning_rate": 9.380737312410219e-05, "loss": 0.591, "step": 5522 }, { "epoch": 0.3707258145699809, "grad_norm": 1.8518561124801636, "learning_rate": 9.380213309628803e-05, "loss": 0.7117, "step": 5524 }, { "epoch": 0.37086003825374986, "grad_norm": 1.2186483144760132, "learning_rate": 9.379689099890921e-05, "loss": 0.6424, "step": 5526 }, { "epoch": 0.3709942619375189, "grad_norm": 1.0989634990692139, "learning_rate": 9.37916468322134e-05, "loss": 0.5955, "step": 5528 }, { "epoch": 0.37112848562128786, "grad_norm": 1.1953972578048706, "learning_rate": 9.378640059644839e-05, "loss": 0.6873, "step": 5530 }, { "epoch": 0.3712627093050569, "grad_norm": 1.0643166303634644, "learning_rate": 9.378115229186207e-05, "loss": 0.6677, "step": 5532 }, { "epoch": 0.37139693298882587, "grad_norm": 1.1248139142990112, "learning_rate": 9.377590191870236e-05, "loss": 0.6331, "step": 5534 }, { "epoch": 0.3715311566725949, "grad_norm": 1.00499427318573, "learning_rate": 9.37706494772174e-05, "loss": 0.6585, "step": 5536 }, { "epoch": 0.3716653803563639, "grad_norm": 1.1205849647521973, "learning_rate": 9.376539496765531e-05, "loss": 0.6161, "step": 5538 }, { "epoch": 0.3717996040401329, "grad_norm": 0.9915878772735596, "learning_rate": 9.376013839026437e-05, "loss": 0.6394, "step": 5540 }, { "epoch": 0.3719338277239019, "grad_norm": 1.1141587495803833, "learning_rate": 9.375487974529296e-05, "loss": 0.6941, "step": 5542 }, { "epoch": 0.3720680514076709, "grad_norm": 1.0914654731750488, "learning_rate": 9.37496190329895e-05, "loss": 0.6707, "step": 5544 }, { "epoch": 0.3722022750914399, "grad_norm": 1.206699252128601, "learning_rate": 9.374435625360259e-05, "loss": 0.6699, "step": 5546 }, { "epoch": 0.3723364987752089, "grad_norm": 0.985154390335083, "learning_rate": 9.373909140738084e-05, "loss": 0.6467, "step": 5548 }, { "epoch": 0.3724707224589779, "grad_norm": 1.2127586603164673, "learning_rate": 9.373382449457304e-05, "loss": 0.7294, "step": 5550 }, { "epoch": 0.37260494614274686, "grad_norm": 1.0873113870620728, "learning_rate": 9.372855551542805e-05, "loss": 0.7035, "step": 5552 }, { "epoch": 0.3727391698265159, "grad_norm": 1.1156877279281616, "learning_rate": 9.372328447019478e-05, "loss": 0.7254, "step": 5554 }, { "epoch": 0.37287339351028487, "grad_norm": 1.3185495138168335, "learning_rate": 9.37180113591223e-05, "loss": 0.7596, "step": 5556 }, { "epoch": 0.3730076171940539, "grad_norm": 2.6542375087738037, "learning_rate": 9.371273618245976e-05, "loss": 0.6865, "step": 5558 }, { "epoch": 0.3731418408778229, "grad_norm": 1.883700966835022, "learning_rate": 9.370745894045639e-05, "loss": 0.6406, "step": 5560 }, { "epoch": 0.3732760645615919, "grad_norm": 1.2540524005889893, "learning_rate": 9.370217963336152e-05, "loss": 0.6706, "step": 5562 }, { "epoch": 0.3734102882453609, "grad_norm": 1.111669898033142, "learning_rate": 9.36968982614246e-05, "loss": 0.7454, "step": 5564 }, { "epoch": 0.3735445119291299, "grad_norm": 1.7078750133514404, "learning_rate": 9.369161482489519e-05, "loss": 0.6821, "step": 5566 }, { "epoch": 0.3736787356128989, "grad_norm": 1.2800893783569336, "learning_rate": 9.368632932402287e-05, "loss": 0.7193, "step": 5568 }, { "epoch": 0.3738129592966679, "grad_norm": 1.2646973133087158, "learning_rate": 9.368104175905741e-05, "loss": 0.7339, "step": 5570 }, { "epoch": 0.3739471829804369, "grad_norm": 1.9984244108200073, "learning_rate": 9.367575213024861e-05, "loss": 0.7086, "step": 5572 }, { "epoch": 0.3740814066642059, "grad_norm": 1.0493110418319702, "learning_rate": 9.36704604378464e-05, "loss": 0.7568, "step": 5574 }, { "epoch": 0.3742156303479749, "grad_norm": 1.1971096992492676, "learning_rate": 9.366516668210083e-05, "loss": 0.7471, "step": 5576 }, { "epoch": 0.3743498540317439, "grad_norm": 1.066906213760376, "learning_rate": 9.365987086326198e-05, "loss": 0.6573, "step": 5578 }, { "epoch": 0.3744840777155129, "grad_norm": 1.0150924921035767, "learning_rate": 9.365457298158009e-05, "loss": 0.6744, "step": 5580 }, { "epoch": 0.3746183013992819, "grad_norm": 1.3731441497802734, "learning_rate": 9.364927303730549e-05, "loss": 0.7055, "step": 5582 }, { "epoch": 0.3747525250830509, "grad_norm": 2.7943010330200195, "learning_rate": 9.364397103068854e-05, "loss": 0.5967, "step": 5584 }, { "epoch": 0.37488674876681993, "grad_norm": 1.1104239225387573, "learning_rate": 9.36386669619798e-05, "loss": 0.6855, "step": 5586 }, { "epoch": 0.3750209724505889, "grad_norm": 1.2128210067749023, "learning_rate": 9.363336083142986e-05, "loss": 0.6556, "step": 5588 }, { "epoch": 0.3751551961343579, "grad_norm": 1.0144624710083008, "learning_rate": 9.36280526392894e-05, "loss": 0.6796, "step": 5590 }, { "epoch": 0.3752894198181269, "grad_norm": 1.0004570484161377, "learning_rate": 9.362274238580926e-05, "loss": 0.6382, "step": 5592 }, { "epoch": 0.3754236435018959, "grad_norm": 1.1422725915908813, "learning_rate": 9.361743007124032e-05, "loss": 0.7098, "step": 5594 }, { "epoch": 0.3755578671856649, "grad_norm": 1.0474408864974976, "learning_rate": 9.36121156958336e-05, "loss": 0.7866, "step": 5596 }, { "epoch": 0.3756920908694339, "grad_norm": 1.7418471574783325, "learning_rate": 9.360679925984013e-05, "loss": 0.6917, "step": 5598 }, { "epoch": 0.3758263145532029, "grad_norm": 1.1903367042541504, "learning_rate": 9.360148076351117e-05, "loss": 0.5844, "step": 5600 }, { "epoch": 0.3759605382369719, "grad_norm": 1.1024620532989502, "learning_rate": 9.359616020709798e-05, "loss": 0.6292, "step": 5602 }, { "epoch": 0.3760947619207409, "grad_norm": 1.4756131172180176, "learning_rate": 9.359083759085195e-05, "loss": 0.6778, "step": 5604 }, { "epoch": 0.3762289856045099, "grad_norm": 1.1311850547790527, "learning_rate": 9.358551291502456e-05, "loss": 0.6545, "step": 5606 }, { "epoch": 0.37636320928827893, "grad_norm": 1.1814398765563965, "learning_rate": 9.358018617986739e-05, "loss": 0.6401, "step": 5608 }, { "epoch": 0.3764974329720479, "grad_norm": 1.0626928806304932, "learning_rate": 9.357485738563212e-05, "loss": 0.6659, "step": 5610 }, { "epoch": 0.37663165665581694, "grad_norm": 1.2566537857055664, "learning_rate": 9.356952653257051e-05, "loss": 0.6408, "step": 5612 }, { "epoch": 0.3767658803395859, "grad_norm": 1.0836503505706787, "learning_rate": 9.356419362093449e-05, "loss": 0.7051, "step": 5614 }, { "epoch": 0.37690010402335494, "grad_norm": 1.1769944429397583, "learning_rate": 9.355885865097595e-05, "loss": 0.632, "step": 5616 }, { "epoch": 0.3770343277071239, "grad_norm": 1.2091494798660278, "learning_rate": 9.3553521622947e-05, "loss": 0.741, "step": 5618 }, { "epoch": 0.37716855139089295, "grad_norm": 1.059908390045166, "learning_rate": 9.354818253709981e-05, "loss": 0.7112, "step": 5620 }, { "epoch": 0.3773027750746619, "grad_norm": 1.0945123434066772, "learning_rate": 9.354284139368662e-05, "loss": 0.7043, "step": 5622 }, { "epoch": 0.37743699875843095, "grad_norm": 1.1156584024429321, "learning_rate": 9.35374981929598e-05, "loss": 0.7091, "step": 5624 }, { "epoch": 0.3775712224421999, "grad_norm": 0.9157266616821289, "learning_rate": 9.35321529351718e-05, "loss": 0.6328, "step": 5626 }, { "epoch": 0.3777054461259689, "grad_norm": 1.1768842935562134, "learning_rate": 9.352680562057516e-05, "loss": 0.6926, "step": 5628 }, { "epoch": 0.37783966980973793, "grad_norm": 1.0927475690841675, "learning_rate": 9.352145624942256e-05, "loss": 0.6667, "step": 5630 }, { "epoch": 0.3779738934935069, "grad_norm": 1.2773330211639404, "learning_rate": 9.351610482196676e-05, "loss": 0.656, "step": 5632 }, { "epoch": 0.37810811717727594, "grad_norm": 1.2487109899520874, "learning_rate": 9.351075133846054e-05, "loss": 0.7294, "step": 5634 }, { "epoch": 0.3782423408610449, "grad_norm": 0.9611471891403198, "learning_rate": 9.35053957991569e-05, "loss": 0.5967, "step": 5636 }, { "epoch": 0.37837656454481394, "grad_norm": 0.9779294729232788, "learning_rate": 9.350003820430885e-05, "loss": 0.6705, "step": 5638 }, { "epoch": 0.3785107882285829, "grad_norm": 1.0110894441604614, "learning_rate": 9.349467855416953e-05, "loss": 0.6288, "step": 5640 }, { "epoch": 0.37864501191235195, "grad_norm": 1.2623742818832397, "learning_rate": 9.348931684899219e-05, "loss": 0.6304, "step": 5642 }, { "epoch": 0.3787792355961209, "grad_norm": 1.0891246795654297, "learning_rate": 9.348395308903014e-05, "loss": 0.7221, "step": 5644 }, { "epoch": 0.37891345927988995, "grad_norm": 2.084601402282715, "learning_rate": 9.347858727453682e-05, "loss": 0.7258, "step": 5646 }, { "epoch": 0.3790476829636589, "grad_norm": 1.037649154663086, "learning_rate": 9.347321940576575e-05, "loss": 0.6064, "step": 5648 }, { "epoch": 0.37918190664742796, "grad_norm": 1.1743371486663818, "learning_rate": 9.346784948297054e-05, "loss": 0.6693, "step": 5650 }, { "epoch": 0.37931613033119693, "grad_norm": 1.0375677347183228, "learning_rate": 9.346247750640491e-05, "loss": 0.6843, "step": 5652 }, { "epoch": 0.37945035401496596, "grad_norm": 1.1350818872451782, "learning_rate": 9.34571034763227e-05, "loss": 0.6466, "step": 5654 }, { "epoch": 0.37958457769873494, "grad_norm": 1.1146163940429688, "learning_rate": 9.34517273929778e-05, "loss": 0.7157, "step": 5656 }, { "epoch": 0.37971880138250397, "grad_norm": 1.2034099102020264, "learning_rate": 9.34463492566242e-05, "loss": 0.725, "step": 5658 }, { "epoch": 0.37985302506627294, "grad_norm": 1.1982332468032837, "learning_rate": 9.344096906751607e-05, "loss": 0.659, "step": 5660 }, { "epoch": 0.379987248750042, "grad_norm": 1.1427327394485474, "learning_rate": 9.343558682590756e-05, "loss": 0.7168, "step": 5662 }, { "epoch": 0.38012147243381095, "grad_norm": 1.4069221019744873, "learning_rate": 9.343020253205298e-05, "loss": 0.6388, "step": 5664 }, { "epoch": 0.3802556961175799, "grad_norm": 1.046435832977295, "learning_rate": 9.342481618620673e-05, "loss": 0.6666, "step": 5666 }, { "epoch": 0.38038991980134895, "grad_norm": 1.6005109548568726, "learning_rate": 9.341942778862331e-05, "loss": 0.6731, "step": 5668 }, { "epoch": 0.3805241434851179, "grad_norm": 1.3098866939544678, "learning_rate": 9.341403733955732e-05, "loss": 0.7199, "step": 5670 }, { "epoch": 0.38065836716888696, "grad_norm": 1.012797474861145, "learning_rate": 9.340864483926343e-05, "loss": 0.7605, "step": 5672 }, { "epoch": 0.38079259085265593, "grad_norm": 1.3212889432907104, "learning_rate": 9.340325028799642e-05, "loss": 0.6812, "step": 5674 }, { "epoch": 0.38092681453642496, "grad_norm": 1.115822672843933, "learning_rate": 9.339785368601119e-05, "loss": 0.6059, "step": 5676 }, { "epoch": 0.38106103822019394, "grad_norm": 0.9674661159515381, "learning_rate": 9.339245503356271e-05, "loss": 0.6786, "step": 5678 }, { "epoch": 0.38119526190396297, "grad_norm": 1.1214262247085571, "learning_rate": 9.338705433090607e-05, "loss": 0.6805, "step": 5680 }, { "epoch": 0.38132948558773194, "grad_norm": 1.024985909461975, "learning_rate": 9.338165157829641e-05, "loss": 0.653, "step": 5682 }, { "epoch": 0.381463709271501, "grad_norm": 0.968446671962738, "learning_rate": 9.337624677598903e-05, "loss": 0.6053, "step": 5684 }, { "epoch": 0.38159793295526995, "grad_norm": 1.6247999668121338, "learning_rate": 9.337083992423927e-05, "loss": 0.6536, "step": 5686 }, { "epoch": 0.381732156639039, "grad_norm": 1.0137710571289062, "learning_rate": 9.336543102330263e-05, "loss": 0.6477, "step": 5688 }, { "epoch": 0.38186638032280795, "grad_norm": 1.031615138053894, "learning_rate": 9.336002007343464e-05, "loss": 0.6506, "step": 5690 }, { "epoch": 0.382000604006577, "grad_norm": 1.087575912475586, "learning_rate": 9.335460707489097e-05, "loss": 0.7036, "step": 5692 }, { "epoch": 0.38213482769034596, "grad_norm": 0.9327123761177063, "learning_rate": 9.334919202792736e-05, "loss": 0.6221, "step": 5694 }, { "epoch": 0.382269051374115, "grad_norm": 1.122715950012207, "learning_rate": 9.334377493279968e-05, "loss": 0.6935, "step": 5696 }, { "epoch": 0.38240327505788396, "grad_norm": 1.0195330381393433, "learning_rate": 9.333835578976385e-05, "loss": 0.6045, "step": 5698 }, { "epoch": 0.38253749874165294, "grad_norm": 1.1630983352661133, "learning_rate": 9.333293459907595e-05, "loss": 0.6921, "step": 5700 }, { "epoch": 0.38267172242542197, "grad_norm": 1.014896273612976, "learning_rate": 9.33275113609921e-05, "loss": 0.6407, "step": 5702 }, { "epoch": 0.38280594610919094, "grad_norm": 1.0407695770263672, "learning_rate": 9.332208607576851e-05, "loss": 0.6618, "step": 5704 }, { "epoch": 0.38294016979296, "grad_norm": 1.3244556188583374, "learning_rate": 9.331665874366156e-05, "loss": 0.7152, "step": 5706 }, { "epoch": 0.38307439347672895, "grad_norm": 1.077439785003662, "learning_rate": 9.331122936492766e-05, "loss": 0.6275, "step": 5708 }, { "epoch": 0.383208617160498, "grad_norm": 1.2128117084503174, "learning_rate": 9.330579793982335e-05, "loss": 0.7595, "step": 5710 }, { "epoch": 0.38334284084426695, "grad_norm": 1.2303240299224854, "learning_rate": 9.330036446860524e-05, "loss": 0.7631, "step": 5712 }, { "epoch": 0.383477064528036, "grad_norm": 1.2520685195922852, "learning_rate": 9.329492895153006e-05, "loss": 0.6768, "step": 5714 }, { "epoch": 0.38361128821180496, "grad_norm": 1.1363435983657837, "learning_rate": 9.328949138885461e-05, "loss": 0.6893, "step": 5716 }, { "epoch": 0.383745511895574, "grad_norm": 1.4869840145111084, "learning_rate": 9.328405178083584e-05, "loss": 0.6519, "step": 5718 }, { "epoch": 0.38387973557934296, "grad_norm": 1.1456031799316406, "learning_rate": 9.327861012773071e-05, "loss": 0.7197, "step": 5720 }, { "epoch": 0.384013959263112, "grad_norm": 1.1507480144500732, "learning_rate": 9.327316642979638e-05, "loss": 0.7817, "step": 5722 }, { "epoch": 0.38414818294688097, "grad_norm": 0.9522068500518799, "learning_rate": 9.326772068729001e-05, "loss": 0.6497, "step": 5724 }, { "epoch": 0.38428240663065, "grad_norm": 1.0786789655685425, "learning_rate": 9.326227290046892e-05, "loss": 0.6879, "step": 5726 }, { "epoch": 0.384416630314419, "grad_norm": 0.8930708765983582, "learning_rate": 9.325682306959051e-05, "loss": 0.6191, "step": 5728 }, { "epoch": 0.384550853998188, "grad_norm": 1.2715744972229004, "learning_rate": 9.325137119491227e-05, "loss": 0.7194, "step": 5730 }, { "epoch": 0.384685077681957, "grad_norm": 1.288090467453003, "learning_rate": 9.324591727669181e-05, "loss": 0.6743, "step": 5732 }, { "epoch": 0.384819301365726, "grad_norm": 1.147087812423706, "learning_rate": 9.324046131518678e-05, "loss": 0.6997, "step": 5734 }, { "epoch": 0.384953525049495, "grad_norm": 1.0619357824325562, "learning_rate": 9.323500331065498e-05, "loss": 0.6485, "step": 5736 }, { "epoch": 0.38508774873326396, "grad_norm": 1.1486061811447144, "learning_rate": 9.322954326335429e-05, "loss": 0.7399, "step": 5738 }, { "epoch": 0.385221972417033, "grad_norm": 0.9960038661956787, "learning_rate": 9.322408117354271e-05, "loss": 0.6371, "step": 5740 }, { "epoch": 0.38535619610080196, "grad_norm": 1.1259183883666992, "learning_rate": 9.321861704147827e-05, "loss": 0.6672, "step": 5742 }, { "epoch": 0.385490419784571, "grad_norm": 1.2630796432495117, "learning_rate": 9.321315086741916e-05, "loss": 0.6638, "step": 5744 }, { "epoch": 0.38562464346833997, "grad_norm": 1.0901516675949097, "learning_rate": 9.320768265162366e-05, "loss": 0.6766, "step": 5746 }, { "epoch": 0.385758867152109, "grad_norm": 1.1709264516830444, "learning_rate": 9.320221239435012e-05, "loss": 0.616, "step": 5748 }, { "epoch": 0.38589309083587797, "grad_norm": 1.2259521484375, "learning_rate": 9.319674009585699e-05, "loss": 0.5973, "step": 5750 }, { "epoch": 0.386027314519647, "grad_norm": 1.7706619501113892, "learning_rate": 9.319126575640283e-05, "loss": 0.6522, "step": 5752 }, { "epoch": 0.386161538203416, "grad_norm": 1.1718233823776245, "learning_rate": 9.318578937624629e-05, "loss": 0.6479, "step": 5754 }, { "epoch": 0.386295761887185, "grad_norm": 1.0942471027374268, "learning_rate": 9.318031095564613e-05, "loss": 0.6747, "step": 5756 }, { "epoch": 0.386429985570954, "grad_norm": 1.0590591430664062, "learning_rate": 9.31748304948612e-05, "loss": 0.6361, "step": 5758 }, { "epoch": 0.386564209254723, "grad_norm": 0.9840893745422363, "learning_rate": 9.316934799415041e-05, "loss": 0.67, "step": 5760 }, { "epoch": 0.386698432938492, "grad_norm": 1.2189501523971558, "learning_rate": 9.316386345377281e-05, "loss": 0.6674, "step": 5762 }, { "epoch": 0.386832656622261, "grad_norm": 1.0591330528259277, "learning_rate": 9.315837687398756e-05, "loss": 0.6833, "step": 5764 }, { "epoch": 0.38696688030603, "grad_norm": 1.0182664394378662, "learning_rate": 9.315288825505387e-05, "loss": 0.6918, "step": 5766 }, { "epoch": 0.387101103989799, "grad_norm": 1.3563718795776367, "learning_rate": 9.314739759723105e-05, "loss": 0.7408, "step": 5768 }, { "epoch": 0.387235327673568, "grad_norm": 1.2516956329345703, "learning_rate": 9.314190490077857e-05, "loss": 0.6319, "step": 5770 }, { "epoch": 0.387369551357337, "grad_norm": 1.058447241783142, "learning_rate": 9.313641016595588e-05, "loss": 0.6691, "step": 5772 }, { "epoch": 0.387503775041106, "grad_norm": 1.2977604866027832, "learning_rate": 9.313091339302267e-05, "loss": 0.7333, "step": 5774 }, { "epoch": 0.387637998724875, "grad_norm": 1.0655789375305176, "learning_rate": 9.312541458223858e-05, "loss": 0.7506, "step": 5776 }, { "epoch": 0.387772222408644, "grad_norm": 1.0868114233016968, "learning_rate": 9.311991373386349e-05, "loss": 0.6913, "step": 5778 }, { "epoch": 0.387906446092413, "grad_norm": 1.147434949874878, "learning_rate": 9.311441084815724e-05, "loss": 0.6807, "step": 5780 }, { "epoch": 0.388040669776182, "grad_norm": 1.7061388492584229, "learning_rate": 9.310890592537987e-05, "loss": 0.6832, "step": 5782 }, { "epoch": 0.388174893459951, "grad_norm": 1.2586777210235596, "learning_rate": 9.310339896579145e-05, "loss": 0.669, "step": 5784 }, { "epoch": 0.38830911714372, "grad_norm": 0.9657377600669861, "learning_rate": 9.30978899696522e-05, "loss": 0.6274, "step": 5786 }, { "epoch": 0.388443340827489, "grad_norm": 1.2519251108169556, "learning_rate": 9.30923789372224e-05, "loss": 0.6726, "step": 5788 }, { "epoch": 0.388577564511258, "grad_norm": 1.0089620351791382, "learning_rate": 9.308686586876243e-05, "loss": 0.6184, "step": 5790 }, { "epoch": 0.388711788195027, "grad_norm": 1.0273845195770264, "learning_rate": 9.308135076453277e-05, "loss": 0.6022, "step": 5792 }, { "epoch": 0.388846011878796, "grad_norm": 1.6973581314086914, "learning_rate": 9.307583362479402e-05, "loss": 0.6645, "step": 5794 }, { "epoch": 0.388980235562565, "grad_norm": 1.1413888931274414, "learning_rate": 9.307031444980681e-05, "loss": 0.6716, "step": 5796 }, { "epoch": 0.38911445924633403, "grad_norm": 1.123063087463379, "learning_rate": 9.306479323983195e-05, "loss": 0.6733, "step": 5798 }, { "epoch": 0.389248682930103, "grad_norm": 1.134331226348877, "learning_rate": 9.305926999513029e-05, "loss": 0.7427, "step": 5800 }, { "epoch": 0.38938290661387204, "grad_norm": 1.0449963808059692, "learning_rate": 9.30537447159628e-05, "loss": 0.6834, "step": 5802 }, { "epoch": 0.389517130297641, "grad_norm": 0.9834620952606201, "learning_rate": 9.304821740259053e-05, "loss": 0.6828, "step": 5804 }, { "epoch": 0.38965135398141004, "grad_norm": 1.0525524616241455, "learning_rate": 9.304268805527464e-05, "loss": 0.7005, "step": 5806 }, { "epoch": 0.389785577665179, "grad_norm": 1.0604172945022583, "learning_rate": 9.303715667427639e-05, "loss": 0.6736, "step": 5808 }, { "epoch": 0.38991980134894805, "grad_norm": 1.0922068357467651, "learning_rate": 9.30316232598571e-05, "loss": 0.7082, "step": 5810 }, { "epoch": 0.390054025032717, "grad_norm": 1.2460064888000488, "learning_rate": 9.302608781227823e-05, "loss": 0.7062, "step": 5812 }, { "epoch": 0.390188248716486, "grad_norm": 1.2911078929901123, "learning_rate": 9.302055033180133e-05, "loss": 0.6592, "step": 5814 }, { "epoch": 0.390322472400255, "grad_norm": 1.1606218814849854, "learning_rate": 9.3015010818688e-05, "loss": 0.6507, "step": 5816 }, { "epoch": 0.390456696084024, "grad_norm": 1.1166902780532837, "learning_rate": 9.300946927320001e-05, "loss": 0.5937, "step": 5818 }, { "epoch": 0.39059091976779303, "grad_norm": 1.0168946981430054, "learning_rate": 9.300392569559917e-05, "loss": 0.7135, "step": 5820 }, { "epoch": 0.390725143451562, "grad_norm": 1.0619525909423828, "learning_rate": 9.29983800861474e-05, "loss": 0.7102, "step": 5822 }, { "epoch": 0.39085936713533104, "grad_norm": 1.3438724279403687, "learning_rate": 9.299283244510674e-05, "loss": 0.7085, "step": 5824 }, { "epoch": 0.3909935908191, "grad_norm": 1.0972028970718384, "learning_rate": 9.298728277273927e-05, "loss": 0.6942, "step": 5826 }, { "epoch": 0.39112781450286904, "grad_norm": 1.483447551727295, "learning_rate": 9.298173106930723e-05, "loss": 0.5973, "step": 5828 }, { "epoch": 0.391262038186638, "grad_norm": 1.2423871755599976, "learning_rate": 9.297617733507291e-05, "loss": 0.7694, "step": 5830 }, { "epoch": 0.39139626187040705, "grad_norm": 0.929648756980896, "learning_rate": 9.297062157029872e-05, "loss": 0.6573, "step": 5832 }, { "epoch": 0.391530485554176, "grad_norm": 1.2835553884506226, "learning_rate": 9.296506377524716e-05, "loss": 0.7118, "step": 5834 }, { "epoch": 0.39166470923794505, "grad_norm": 1.2994301319122314, "learning_rate": 9.295950395018084e-05, "loss": 0.7424, "step": 5836 }, { "epoch": 0.391798932921714, "grad_norm": 1.0983941555023193, "learning_rate": 9.295394209536242e-05, "loss": 0.758, "step": 5838 }, { "epoch": 0.39193315660548306, "grad_norm": 1.120267391204834, "learning_rate": 9.29483782110547e-05, "loss": 0.7137, "step": 5840 }, { "epoch": 0.39206738028925203, "grad_norm": 1.349510908126831, "learning_rate": 9.294281229752057e-05, "loss": 0.6632, "step": 5842 }, { "epoch": 0.39220160397302106, "grad_norm": 1.0613240003585815, "learning_rate": 9.2937244355023e-05, "loss": 0.6693, "step": 5844 }, { "epoch": 0.39233582765679004, "grad_norm": 1.2015703916549683, "learning_rate": 9.293167438382506e-05, "loss": 0.7193, "step": 5846 }, { "epoch": 0.39247005134055907, "grad_norm": 1.14816415309906, "learning_rate": 9.292610238418992e-05, "loss": 0.6654, "step": 5848 }, { "epoch": 0.39260427502432804, "grad_norm": 1.147283911705017, "learning_rate": 9.292052835638088e-05, "loss": 0.762, "step": 5850 }, { "epoch": 0.392738498708097, "grad_norm": 1.0627939701080322, "learning_rate": 9.291495230066125e-05, "loss": 0.6812, "step": 5852 }, { "epoch": 0.39287272239186605, "grad_norm": 1.2463438510894775, "learning_rate": 9.290937421729454e-05, "loss": 0.6162, "step": 5854 }, { "epoch": 0.393006946075635, "grad_norm": 1.1472399234771729, "learning_rate": 9.290379410654425e-05, "loss": 0.6816, "step": 5856 }, { "epoch": 0.39314116975940405, "grad_norm": 1.1549139022827148, "learning_rate": 9.289821196867405e-05, "loss": 0.7213, "step": 5858 }, { "epoch": 0.393275393443173, "grad_norm": 1.0665532350540161, "learning_rate": 9.289262780394772e-05, "loss": 0.672, "step": 5860 }, { "epoch": 0.39340961712694206, "grad_norm": 1.235317587852478, "learning_rate": 9.288704161262904e-05, "loss": 0.6619, "step": 5862 }, { "epoch": 0.39354384081071103, "grad_norm": 1.0959912538528442, "learning_rate": 9.2881453394982e-05, "loss": 0.6623, "step": 5864 }, { "epoch": 0.39367806449448006, "grad_norm": 1.1718889474868774, "learning_rate": 9.28758631512706e-05, "loss": 0.6712, "step": 5866 }, { "epoch": 0.39381228817824904, "grad_norm": 1.1043980121612549, "learning_rate": 9.287027088175898e-05, "loss": 0.7072, "step": 5868 }, { "epoch": 0.39394651186201807, "grad_norm": 1.0995796918869019, "learning_rate": 9.286467658671136e-05, "loss": 0.6571, "step": 5870 }, { "epoch": 0.39408073554578704, "grad_norm": 1.475175380706787, "learning_rate": 9.285908026639207e-05, "loss": 0.7097, "step": 5872 }, { "epoch": 0.3942149592295561, "grad_norm": 1.5103731155395508, "learning_rate": 9.285348192106551e-05, "loss": 0.7555, "step": 5874 }, { "epoch": 0.39434918291332505, "grad_norm": 1.0760539770126343, "learning_rate": 9.28478815509962e-05, "loss": 0.6706, "step": 5876 }, { "epoch": 0.3944834065970941, "grad_norm": 0.9949193596839905, "learning_rate": 9.284227915644872e-05, "loss": 0.7143, "step": 5878 }, { "epoch": 0.39461763028086305, "grad_norm": 1.2008304595947266, "learning_rate": 9.283667473768782e-05, "loss": 0.6664, "step": 5880 }, { "epoch": 0.3947518539646321, "grad_norm": 1.1227686405181885, "learning_rate": 9.283106829497828e-05, "loss": 0.7323, "step": 5882 }, { "epoch": 0.39488607764840106, "grad_norm": 1.1749778985977173, "learning_rate": 9.282545982858496e-05, "loss": 0.6511, "step": 5884 }, { "epoch": 0.3950203013321701, "grad_norm": 0.969157874584198, "learning_rate": 9.28198493387729e-05, "loss": 0.687, "step": 5886 }, { "epoch": 0.39515452501593906, "grad_norm": 1.1056897640228271, "learning_rate": 9.281423682580714e-05, "loss": 0.7228, "step": 5888 }, { "epoch": 0.39528874869970804, "grad_norm": 1.1171948909759521, "learning_rate": 9.280862228995291e-05, "loss": 0.7403, "step": 5890 }, { "epoch": 0.39542297238347707, "grad_norm": 1.099923014640808, "learning_rate": 9.280300573147542e-05, "loss": 0.6672, "step": 5892 }, { "epoch": 0.39555719606724604, "grad_norm": 1.0958002805709839, "learning_rate": 9.27973871506401e-05, "loss": 0.7194, "step": 5894 }, { "epoch": 0.39569141975101507, "grad_norm": 1.0290558338165283, "learning_rate": 9.27917665477124e-05, "loss": 0.6969, "step": 5896 }, { "epoch": 0.39582564343478405, "grad_norm": 1.0672677755355835, "learning_rate": 9.278614392295786e-05, "loss": 0.6818, "step": 5898 }, { "epoch": 0.3959598671185531, "grad_norm": 1.061200737953186, "learning_rate": 9.278051927664217e-05, "loss": 0.6052, "step": 5900 }, { "epoch": 0.39609409080232205, "grad_norm": 1.2648062705993652, "learning_rate": 9.277489260903104e-05, "loss": 0.6926, "step": 5902 }, { "epoch": 0.3962283144860911, "grad_norm": 1.0376838445663452, "learning_rate": 9.276926392039038e-05, "loss": 0.6976, "step": 5904 }, { "epoch": 0.39636253816986006, "grad_norm": 1.3055962324142456, "learning_rate": 9.276363321098609e-05, "loss": 0.6908, "step": 5906 }, { "epoch": 0.3964967618536291, "grad_norm": 1.1608704328536987, "learning_rate": 9.275800048108423e-05, "loss": 0.6734, "step": 5908 }, { "epoch": 0.39663098553739806, "grad_norm": 1.1547906398773193, "learning_rate": 9.27523657309509e-05, "loss": 0.7042, "step": 5910 }, { "epoch": 0.3967652092211671, "grad_norm": 1.0295121669769287, "learning_rate": 9.27467289608524e-05, "loss": 0.6748, "step": 5912 }, { "epoch": 0.39689943290493607, "grad_norm": 1.1790900230407715, "learning_rate": 9.274109017105497e-05, "loss": 0.6604, "step": 5914 }, { "epoch": 0.3970336565887051, "grad_norm": 1.0177485942840576, "learning_rate": 9.27354493618251e-05, "loss": 0.6853, "step": 5916 }, { "epoch": 0.39716788027247407, "grad_norm": 1.054178237915039, "learning_rate": 9.272980653342929e-05, "loss": 0.6572, "step": 5918 }, { "epoch": 0.3973021039562431, "grad_norm": 0.9216815233230591, "learning_rate": 9.272416168613414e-05, "loss": 0.6137, "step": 5920 }, { "epoch": 0.3974363276400121, "grad_norm": 1.038726806640625, "learning_rate": 9.271851482020635e-05, "loss": 0.6604, "step": 5922 }, { "epoch": 0.3975705513237811, "grad_norm": 1.3206719160079956, "learning_rate": 9.271286593591275e-05, "loss": 0.7159, "step": 5924 }, { "epoch": 0.3977047750075501, "grad_norm": 1.3097436428070068, "learning_rate": 9.270721503352022e-05, "loss": 0.7292, "step": 5926 }, { "epoch": 0.39783899869131906, "grad_norm": 1.3265849351882935, "learning_rate": 9.270156211329578e-05, "loss": 0.6842, "step": 5928 }, { "epoch": 0.3979732223750881, "grad_norm": 0.9673488736152649, "learning_rate": 9.269590717550647e-05, "loss": 0.6519, "step": 5930 }, { "epoch": 0.39810744605885706, "grad_norm": 1.0370547771453857, "learning_rate": 9.269025022041953e-05, "loss": 0.6785, "step": 5932 }, { "epoch": 0.3982416697426261, "grad_norm": 1.2383445501327515, "learning_rate": 9.268459124830218e-05, "loss": 0.6802, "step": 5934 }, { "epoch": 0.39837589342639507, "grad_norm": 1.4812686443328857, "learning_rate": 9.267893025942186e-05, "loss": 0.6809, "step": 5936 }, { "epoch": 0.3985101171101641, "grad_norm": 1.1149955987930298, "learning_rate": 9.267326725404599e-05, "loss": 0.712, "step": 5938 }, { "epoch": 0.39864434079393307, "grad_norm": 1.1058670282363892, "learning_rate": 9.266760223244218e-05, "loss": 0.6745, "step": 5940 }, { "epoch": 0.3987785644777021, "grad_norm": 1.172837257385254, "learning_rate": 9.266193519487805e-05, "loss": 0.698, "step": 5942 }, { "epoch": 0.3989127881614711, "grad_norm": 1.115440845489502, "learning_rate": 9.265626614162137e-05, "loss": 0.693, "step": 5944 }, { "epoch": 0.3990470118452401, "grad_norm": 1.037589430809021, "learning_rate": 9.265059507294001e-05, "loss": 0.7305, "step": 5946 }, { "epoch": 0.3991812355290091, "grad_norm": 1.1750209331512451, "learning_rate": 9.264492198910189e-05, "loss": 0.7095, "step": 5948 }, { "epoch": 0.3993154592127781, "grad_norm": 1.1001358032226562, "learning_rate": 9.263924689037505e-05, "loss": 0.6478, "step": 5950 }, { "epoch": 0.3994496828965471, "grad_norm": 1.6089013814926147, "learning_rate": 9.263356977702766e-05, "loss": 0.7067, "step": 5952 }, { "epoch": 0.3995839065803161, "grad_norm": 1.092583179473877, "learning_rate": 9.262789064932794e-05, "loss": 0.6607, "step": 5954 }, { "epoch": 0.3997181302640851, "grad_norm": 1.0944645404815674, "learning_rate": 9.262220950754419e-05, "loss": 0.6722, "step": 5956 }, { "epoch": 0.3998523539478541, "grad_norm": 1.061228632926941, "learning_rate": 9.261652635194487e-05, "loss": 0.6982, "step": 5958 }, { "epoch": 0.3999865776316231, "grad_norm": 0.9771525263786316, "learning_rate": 9.261084118279847e-05, "loss": 0.6908, "step": 5960 }, { "epoch": 0.4001208013153921, "grad_norm": 1.602399468421936, "learning_rate": 9.260515400037362e-05, "loss": 0.6764, "step": 5962 }, { "epoch": 0.4002550249991611, "grad_norm": 1.0964266061782837, "learning_rate": 9.259946480493902e-05, "loss": 0.685, "step": 5964 }, { "epoch": 0.4003892486829301, "grad_norm": 1.1376968622207642, "learning_rate": 9.259377359676348e-05, "loss": 0.7389, "step": 5966 }, { "epoch": 0.4005234723666991, "grad_norm": 1.1643990278244019, "learning_rate": 9.25880803761159e-05, "loss": 0.5875, "step": 5968 }, { "epoch": 0.4006576960504681, "grad_norm": 1.0562384128570557, "learning_rate": 9.258238514326525e-05, "loss": 0.605, "step": 5970 }, { "epoch": 0.4007919197342371, "grad_norm": 1.1646043062210083, "learning_rate": 9.257668789848067e-05, "loss": 0.6567, "step": 5972 }, { "epoch": 0.4009261434180061, "grad_norm": 1.1420818567276, "learning_rate": 9.257098864203128e-05, "loss": 0.6783, "step": 5974 }, { "epoch": 0.4010603671017751, "grad_norm": 1.2688438892364502, "learning_rate": 9.25652873741864e-05, "loss": 0.6516, "step": 5976 }, { "epoch": 0.4011945907855441, "grad_norm": 1.054469108581543, "learning_rate": 9.255958409521538e-05, "loss": 0.6125, "step": 5978 }, { "epoch": 0.4013288144693131, "grad_norm": 1.1423333883285522, "learning_rate": 9.25538788053877e-05, "loss": 0.5963, "step": 5980 }, { "epoch": 0.4014630381530821, "grad_norm": 1.0816683769226074, "learning_rate": 9.254817150497295e-05, "loss": 0.6697, "step": 5982 }, { "epoch": 0.4015972618368511, "grad_norm": 0.975953221321106, "learning_rate": 9.254246219424075e-05, "loss": 0.6628, "step": 5984 }, { "epoch": 0.4017314855206201, "grad_norm": 1.167656421661377, "learning_rate": 9.253675087346087e-05, "loss": 0.6356, "step": 5986 }, { "epoch": 0.40186570920438913, "grad_norm": 1.33633553981781, "learning_rate": 9.253103754290315e-05, "loss": 0.6431, "step": 5988 }, { "epoch": 0.4019999328881581, "grad_norm": 1.1282232999801636, "learning_rate": 9.252532220283754e-05, "loss": 0.6035, "step": 5990 }, { "epoch": 0.40213415657192714, "grad_norm": 0.9276664853096008, "learning_rate": 9.251960485353408e-05, "loss": 0.6548, "step": 5992 }, { "epoch": 0.4022683802556961, "grad_norm": 1.0622806549072266, "learning_rate": 9.251388549526292e-05, "loss": 0.6835, "step": 5994 }, { "epoch": 0.40240260393946514, "grad_norm": 1.0541036128997803, "learning_rate": 9.250816412829425e-05, "loss": 0.7287, "step": 5996 }, { "epoch": 0.4025368276232341, "grad_norm": 1.134677767753601, "learning_rate": 9.250244075289843e-05, "loss": 0.6533, "step": 5998 }, { "epoch": 0.40267105130700315, "grad_norm": 1.189555048942566, "learning_rate": 9.249671536934585e-05, "loss": 0.6717, "step": 6000 }, { "epoch": 0.4028052749907721, "grad_norm": 1.0003716945648193, "learning_rate": 9.249098797790702e-05, "loss": 0.6487, "step": 6002 }, { "epoch": 0.4029394986745411, "grad_norm": 1.3909540176391602, "learning_rate": 9.248525857885259e-05, "loss": 0.6151, "step": 6004 }, { "epoch": 0.4030737223583101, "grad_norm": 0.9759494066238403, "learning_rate": 9.247952717245321e-05, "loss": 0.7081, "step": 6006 }, { "epoch": 0.4032079460420791, "grad_norm": 1.2941428422927856, "learning_rate": 9.247379375897974e-05, "loss": 0.7397, "step": 6008 }, { "epoch": 0.40334216972584813, "grad_norm": 1.346400260925293, "learning_rate": 9.2468058338703e-05, "loss": 0.7726, "step": 6010 }, { "epoch": 0.4034763934096171, "grad_norm": 1.0712957382202148, "learning_rate": 9.246232091189402e-05, "loss": 0.6759, "step": 6012 }, { "epoch": 0.40361061709338614, "grad_norm": 1.012364387512207, "learning_rate": 9.245658147882388e-05, "loss": 0.6428, "step": 6014 }, { "epoch": 0.4037448407771551, "grad_norm": 1.3964303731918335, "learning_rate": 9.245084003976377e-05, "loss": 0.7316, "step": 6016 }, { "epoch": 0.40387906446092414, "grad_norm": 1.0563886165618896, "learning_rate": 9.244509659498493e-05, "loss": 0.7132, "step": 6018 }, { "epoch": 0.4040132881446931, "grad_norm": 0.9809919595718384, "learning_rate": 9.243935114475872e-05, "loss": 0.656, "step": 6020 }, { "epoch": 0.40414751182846215, "grad_norm": 1.0551362037658691, "learning_rate": 9.243360368935666e-05, "loss": 0.6481, "step": 6022 }, { "epoch": 0.4042817355122311, "grad_norm": 1.0678483247756958, "learning_rate": 9.242785422905025e-05, "loss": 0.6739, "step": 6024 }, { "epoch": 0.40441595919600015, "grad_norm": 0.9836398959159851, "learning_rate": 9.242210276411115e-05, "loss": 0.6616, "step": 6026 }, { "epoch": 0.4045501828797691, "grad_norm": 1.125788927078247, "learning_rate": 9.241634929481112e-05, "loss": 0.6742, "step": 6028 }, { "epoch": 0.40468440656353816, "grad_norm": 1.3970268964767456, "learning_rate": 9.2410593821422e-05, "loss": 0.638, "step": 6030 }, { "epoch": 0.40481863024730713, "grad_norm": 1.178939938545227, "learning_rate": 9.24048363442157e-05, "loss": 0.6704, "step": 6032 }, { "epoch": 0.40495285393107616, "grad_norm": 1.0103930234909058, "learning_rate": 9.239907686346429e-05, "loss": 0.643, "step": 6034 }, { "epoch": 0.40508707761484514, "grad_norm": 1.0354652404785156, "learning_rate": 9.239331537943987e-05, "loss": 0.6443, "step": 6036 }, { "epoch": 0.40522130129861417, "grad_norm": 1.1033591032028198, "learning_rate": 9.238755189241466e-05, "loss": 0.713, "step": 6038 }, { "epoch": 0.40535552498238314, "grad_norm": 0.9593886733055115, "learning_rate": 9.238178640266096e-05, "loss": 0.638, "step": 6040 }, { "epoch": 0.4054897486661521, "grad_norm": 0.9968031048774719, "learning_rate": 9.23760189104512e-05, "loss": 0.6368, "step": 6042 }, { "epoch": 0.40562397234992115, "grad_norm": 1.142154574394226, "learning_rate": 9.237024941605789e-05, "loss": 0.6719, "step": 6044 }, { "epoch": 0.4057581960336901, "grad_norm": 1.0583053827285767, "learning_rate": 9.23644779197536e-05, "loss": 0.6468, "step": 6046 }, { "epoch": 0.40589241971745915, "grad_norm": 1.1052918434143066, "learning_rate": 9.235870442181104e-05, "loss": 0.6907, "step": 6048 }, { "epoch": 0.4060266434012281, "grad_norm": 1.1614830493927002, "learning_rate": 9.235292892250298e-05, "loss": 0.7555, "step": 6050 }, { "epoch": 0.40616086708499716, "grad_norm": 1.1473082304000854, "learning_rate": 9.234715142210233e-05, "loss": 0.6158, "step": 6052 }, { "epoch": 0.40629509076876613, "grad_norm": 1.0020947456359863, "learning_rate": 9.234137192088202e-05, "loss": 0.6192, "step": 6054 }, { "epoch": 0.40642931445253516, "grad_norm": 1.0485420227050781, "learning_rate": 9.233559041911517e-05, "loss": 0.6748, "step": 6056 }, { "epoch": 0.40656353813630414, "grad_norm": 0.935448944568634, "learning_rate": 9.232980691707491e-05, "loss": 0.5965, "step": 6058 }, { "epoch": 0.40669776182007317, "grad_norm": 1.0142656564712524, "learning_rate": 9.232402141503452e-05, "loss": 0.7221, "step": 6060 }, { "epoch": 0.40683198550384214, "grad_norm": 1.1718939542770386, "learning_rate": 9.231823391326734e-05, "loss": 0.627, "step": 6062 }, { "epoch": 0.40696620918761117, "grad_norm": 1.0399384498596191, "learning_rate": 9.231244441204683e-05, "loss": 0.6466, "step": 6064 }, { "epoch": 0.40710043287138015, "grad_norm": 1.2102640867233276, "learning_rate": 9.230665291164652e-05, "loss": 0.6899, "step": 6066 }, { "epoch": 0.4072346565551492, "grad_norm": 1.152836561203003, "learning_rate": 9.230085941234006e-05, "loss": 0.7211, "step": 6068 }, { "epoch": 0.40736888023891815, "grad_norm": 1.1507648229599, "learning_rate": 9.229506391440115e-05, "loss": 0.7227, "step": 6070 }, { "epoch": 0.4075031039226872, "grad_norm": 1.045613408088684, "learning_rate": 9.228926641810367e-05, "loss": 0.6501, "step": 6072 }, { "epoch": 0.40763732760645616, "grad_norm": 1.1735528707504272, "learning_rate": 9.22834669237215e-05, "loss": 0.6691, "step": 6074 }, { "epoch": 0.4077715512902252, "grad_norm": 1.136020541191101, "learning_rate": 9.227766543152864e-05, "loss": 0.6165, "step": 6076 }, { "epoch": 0.40790577497399416, "grad_norm": 1.159848928451538, "learning_rate": 9.227186194179925e-05, "loss": 0.6709, "step": 6078 }, { "epoch": 0.40803999865776314, "grad_norm": 1.3651721477508545, "learning_rate": 9.22660564548075e-05, "loss": 0.7813, "step": 6080 }, { "epoch": 0.40817422234153217, "grad_norm": 1.0709141492843628, "learning_rate": 9.22602489708277e-05, "loss": 0.6063, "step": 6082 }, { "epoch": 0.40830844602530114, "grad_norm": 1.0377559661865234, "learning_rate": 9.225443949013424e-05, "loss": 0.6476, "step": 6084 }, { "epoch": 0.40844266970907017, "grad_norm": 1.231859803199768, "learning_rate": 9.224862801300159e-05, "loss": 0.7013, "step": 6086 }, { "epoch": 0.40857689339283915, "grad_norm": 1.0269285440444946, "learning_rate": 9.224281453970436e-05, "loss": 0.6628, "step": 6088 }, { "epoch": 0.4087111170766082, "grad_norm": 1.080043077468872, "learning_rate": 9.22369990705172e-05, "loss": 0.6303, "step": 6090 }, { "epoch": 0.40884534076037715, "grad_norm": 1.0308558940887451, "learning_rate": 9.223118160571489e-05, "loss": 0.6574, "step": 6092 }, { "epoch": 0.4089795644441462, "grad_norm": 1.0971088409423828, "learning_rate": 9.22253621455723e-05, "loss": 0.6433, "step": 6094 }, { "epoch": 0.40911378812791516, "grad_norm": 0.9573982954025269, "learning_rate": 9.221954069036438e-05, "loss": 0.6165, "step": 6096 }, { "epoch": 0.4092480118116842, "grad_norm": 1.0471969842910767, "learning_rate": 9.221371724036619e-05, "loss": 0.7094, "step": 6098 }, { "epoch": 0.40938223549545316, "grad_norm": 1.1849780082702637, "learning_rate": 9.220789179585286e-05, "loss": 0.6669, "step": 6100 }, { "epoch": 0.4095164591792222, "grad_norm": 1.0062451362609863, "learning_rate": 9.220206435709963e-05, "loss": 0.6624, "step": 6102 }, { "epoch": 0.40965068286299117, "grad_norm": 1.4470155239105225, "learning_rate": 9.219623492438185e-05, "loss": 0.6079, "step": 6104 }, { "epoch": 0.4097849065467602, "grad_norm": 1.346421241760254, "learning_rate": 9.219040349797495e-05, "loss": 0.7008, "step": 6106 }, { "epoch": 0.40991913023052917, "grad_norm": 1.0121887922286987, "learning_rate": 9.218457007815447e-05, "loss": 0.6691, "step": 6108 }, { "epoch": 0.4100533539142982, "grad_norm": 2.2194879055023193, "learning_rate": 9.217873466519597e-05, "loss": 0.7418, "step": 6110 }, { "epoch": 0.4101875775980672, "grad_norm": 0.9041343331336975, "learning_rate": 9.217289725937521e-05, "loss": 0.615, "step": 6112 }, { "epoch": 0.4103218012818362, "grad_norm": 1.1116787195205688, "learning_rate": 9.2167057860968e-05, "loss": 0.681, "step": 6114 }, { "epoch": 0.4104560249656052, "grad_norm": 1.2395095825195312, "learning_rate": 9.216121647025021e-05, "loss": 0.7903, "step": 6116 }, { "epoch": 0.41059024864937416, "grad_norm": 1.2876553535461426, "learning_rate": 9.215537308749784e-05, "loss": 0.6527, "step": 6118 }, { "epoch": 0.4107244723331432, "grad_norm": 1.1910831928253174, "learning_rate": 9.214952771298701e-05, "loss": 0.6842, "step": 6120 }, { "epoch": 0.41085869601691216, "grad_norm": 0.9551767706871033, "learning_rate": 9.214368034699387e-05, "loss": 0.6568, "step": 6122 }, { "epoch": 0.4109929197006812, "grad_norm": 1.0212019681930542, "learning_rate": 9.213783098979469e-05, "loss": 0.7006, "step": 6124 }, { "epoch": 0.41112714338445017, "grad_norm": 1.1502602100372314, "learning_rate": 9.213197964166587e-05, "loss": 0.6756, "step": 6126 }, { "epoch": 0.4112613670682192, "grad_norm": 1.0294241905212402, "learning_rate": 9.212612630288386e-05, "loss": 0.6463, "step": 6128 }, { "epoch": 0.41139559075198817, "grad_norm": 1.4980016946792603, "learning_rate": 9.212027097372522e-05, "loss": 0.7198, "step": 6130 }, { "epoch": 0.4115298144357572, "grad_norm": 1.0276916027069092, "learning_rate": 9.21144136544666e-05, "loss": 0.6685, "step": 6132 }, { "epoch": 0.4116640381195262, "grad_norm": 0.9319644570350647, "learning_rate": 9.210855434538477e-05, "loss": 0.6197, "step": 6134 }, { "epoch": 0.4117982618032952, "grad_norm": 1.0256725549697876, "learning_rate": 9.210269304675652e-05, "loss": 0.6934, "step": 6136 }, { "epoch": 0.4119324854870642, "grad_norm": 2.123508930206299, "learning_rate": 9.209682975885882e-05, "loss": 0.6992, "step": 6138 }, { "epoch": 0.4120667091708332, "grad_norm": 1.2907676696777344, "learning_rate": 9.209096448196872e-05, "loss": 0.692, "step": 6140 }, { "epoch": 0.4122009328546022, "grad_norm": 1.0666604042053223, "learning_rate": 9.208509721636328e-05, "loss": 0.6574, "step": 6142 }, { "epoch": 0.4123351565383712, "grad_norm": 1.0116019248962402, "learning_rate": 9.207922796231977e-05, "loss": 0.6908, "step": 6144 }, { "epoch": 0.4124693802221402, "grad_norm": 1.1583993434906006, "learning_rate": 9.20733567201155e-05, "loss": 0.6884, "step": 6146 }, { "epoch": 0.4126036039059092, "grad_norm": 1.0761034488677979, "learning_rate": 9.206748349002782e-05, "loss": 0.663, "step": 6148 }, { "epoch": 0.4127378275896782, "grad_norm": 1.0891854763031006, "learning_rate": 9.20616082723343e-05, "loss": 0.6682, "step": 6150 }, { "epoch": 0.4128720512734472, "grad_norm": 1.028387427330017, "learning_rate": 9.20557310673125e-05, "loss": 0.7085, "step": 6152 }, { "epoch": 0.4130062749572162, "grad_norm": 1.0554955005645752, "learning_rate": 9.20498518752401e-05, "loss": 0.586, "step": 6154 }, { "epoch": 0.4131404986409852, "grad_norm": 0.9535325169563293, "learning_rate": 9.204397069639486e-05, "loss": 0.6625, "step": 6156 }, { "epoch": 0.4132747223247542, "grad_norm": 1.0058443546295166, "learning_rate": 9.203808753105471e-05, "loss": 0.6522, "step": 6158 }, { "epoch": 0.4134089460085232, "grad_norm": 1.4530055522918701, "learning_rate": 9.203220237949758e-05, "loss": 0.6723, "step": 6160 }, { "epoch": 0.4135431696922922, "grad_norm": 0.9904844760894775, "learning_rate": 9.202631524200153e-05, "loss": 0.5549, "step": 6162 }, { "epoch": 0.4136773933760612, "grad_norm": 1.3306118249893188, "learning_rate": 9.202042611884475e-05, "loss": 0.6845, "step": 6164 }, { "epoch": 0.4138116170598302, "grad_norm": 0.8913776278495789, "learning_rate": 9.201453501030546e-05, "loss": 0.6191, "step": 6166 }, { "epoch": 0.4139458407435992, "grad_norm": 1.0931206941604614, "learning_rate": 9.200864191666199e-05, "loss": 0.623, "step": 6168 }, { "epoch": 0.4140800644273682, "grad_norm": 1.4609962701797485, "learning_rate": 9.200274683819282e-05, "loss": 0.653, "step": 6170 }, { "epoch": 0.4142142881111372, "grad_norm": 1.0243514776229858, "learning_rate": 9.199684977517645e-05, "loss": 0.6721, "step": 6172 }, { "epoch": 0.4143485117949062, "grad_norm": 1.3936306238174438, "learning_rate": 9.199095072789149e-05, "loss": 0.7181, "step": 6174 }, { "epoch": 0.4144827354786752, "grad_norm": 1.211826205253601, "learning_rate": 9.19850496966167e-05, "loss": 0.6401, "step": 6176 }, { "epoch": 0.41461695916244423, "grad_norm": 1.2502893209457397, "learning_rate": 9.197914668163085e-05, "loss": 0.6879, "step": 6178 }, { "epoch": 0.4147511828462132, "grad_norm": 1.0616451501846313, "learning_rate": 9.19732416832129e-05, "loss": 0.6886, "step": 6180 }, { "epoch": 0.41488540652998224, "grad_norm": 1.069578766822815, "learning_rate": 9.19673347016418e-05, "loss": 0.5796, "step": 6182 }, { "epoch": 0.4150196302137512, "grad_norm": 1.216532588005066, "learning_rate": 9.196142573719666e-05, "loss": 0.734, "step": 6184 }, { "epoch": 0.41515385389752024, "grad_norm": 1.1617928743362427, "learning_rate": 9.195551479015667e-05, "loss": 0.6501, "step": 6186 }, { "epoch": 0.4152880775812892, "grad_norm": 1.0069537162780762, "learning_rate": 9.19496018608011e-05, "loss": 0.6785, "step": 6188 }, { "epoch": 0.41542230126505825, "grad_norm": 1.260364294052124, "learning_rate": 9.194368694940935e-05, "loss": 0.6827, "step": 6190 }, { "epoch": 0.4155565249488272, "grad_norm": 1.0657905340194702, "learning_rate": 9.193777005626086e-05, "loss": 0.6688, "step": 6192 }, { "epoch": 0.4156907486325962, "grad_norm": 1.3277440071105957, "learning_rate": 9.193185118163521e-05, "loss": 0.6361, "step": 6194 }, { "epoch": 0.4158249723163652, "grad_norm": 1.1172504425048828, "learning_rate": 9.192593032581203e-05, "loss": 0.6907, "step": 6196 }, { "epoch": 0.4159591960001342, "grad_norm": 1.1119155883789062, "learning_rate": 9.19200074890711e-05, "loss": 0.6752, "step": 6198 }, { "epoch": 0.41609341968390323, "grad_norm": 1.1108487844467163, "learning_rate": 9.191408267169226e-05, "loss": 0.7023, "step": 6200 }, { "epoch": 0.4162276433676722, "grad_norm": 0.9239070415496826, "learning_rate": 9.19081558739554e-05, "loss": 0.6321, "step": 6202 }, { "epoch": 0.41636186705144124, "grad_norm": 0.999114453792572, "learning_rate": 9.190222709614061e-05, "loss": 0.6482, "step": 6204 }, { "epoch": 0.4164960907352102, "grad_norm": 1.2064570188522339, "learning_rate": 9.189629633852799e-05, "loss": 0.6746, "step": 6206 }, { "epoch": 0.41663031441897924, "grad_norm": 1.2322992086410522, "learning_rate": 9.189036360139773e-05, "loss": 0.6422, "step": 6208 }, { "epoch": 0.4167645381027482, "grad_norm": 1.1882107257843018, "learning_rate": 9.188442888503018e-05, "loss": 0.6634, "step": 6210 }, { "epoch": 0.41689876178651725, "grad_norm": 1.0807157754898071, "learning_rate": 9.187849218970572e-05, "loss": 0.7188, "step": 6212 }, { "epoch": 0.4170329854702862, "grad_norm": 1.0926183462142944, "learning_rate": 9.187255351570487e-05, "loss": 0.6573, "step": 6214 }, { "epoch": 0.41716720915405525, "grad_norm": 1.1089380979537964, "learning_rate": 9.18666128633082e-05, "loss": 0.6824, "step": 6216 }, { "epoch": 0.4173014328378242, "grad_norm": 1.1253862380981445, "learning_rate": 9.186067023279639e-05, "loss": 0.6919, "step": 6218 }, { "epoch": 0.41743565652159326, "grad_norm": 0.9917927384376526, "learning_rate": 9.185472562445022e-05, "loss": 0.6455, "step": 6220 }, { "epoch": 0.41756988020536223, "grad_norm": 1.724709153175354, "learning_rate": 9.184877903855058e-05, "loss": 0.7087, "step": 6222 }, { "epoch": 0.41770410388913126, "grad_norm": 1.1316773891448975, "learning_rate": 9.184283047537843e-05, "loss": 0.6598, "step": 6224 }, { "epoch": 0.41783832757290024, "grad_norm": 0.9227393865585327, "learning_rate": 9.18368799352148e-05, "loss": 0.6334, "step": 6226 }, { "epoch": 0.41797255125666927, "grad_norm": 1.0776231288909912, "learning_rate": 9.183092741834087e-05, "loss": 0.693, "step": 6228 }, { "epoch": 0.41810677494043824, "grad_norm": 1.0032908916473389, "learning_rate": 9.182497292503789e-05, "loss": 0.6563, "step": 6230 }, { "epoch": 0.4182409986242072, "grad_norm": 1.2029250860214233, "learning_rate": 9.181901645558717e-05, "loss": 0.6592, "step": 6232 }, { "epoch": 0.41837522230797625, "grad_norm": 1.0548990964889526, "learning_rate": 9.181305801027015e-05, "loss": 0.6138, "step": 6234 }, { "epoch": 0.4185094459917452, "grad_norm": 0.9585376977920532, "learning_rate": 9.180709758936839e-05, "loss": 0.6225, "step": 6236 }, { "epoch": 0.41864366967551425, "grad_norm": 1.2465183734893799, "learning_rate": 9.180113519316345e-05, "loss": 0.7468, "step": 6238 }, { "epoch": 0.4187778933592832, "grad_norm": 1.2814899682998657, "learning_rate": 9.179517082193709e-05, "loss": 0.6708, "step": 6240 }, { "epoch": 0.41891211704305226, "grad_norm": 1.0063660144805908, "learning_rate": 9.178920447597108e-05, "loss": 0.6204, "step": 6242 }, { "epoch": 0.41904634072682123, "grad_norm": 1.1164515018463135, "learning_rate": 9.178323615554733e-05, "loss": 0.6057, "step": 6244 }, { "epoch": 0.41918056441059026, "grad_norm": 0.9850273132324219, "learning_rate": 9.177726586094785e-05, "loss": 0.6054, "step": 6246 }, { "epoch": 0.41931478809435924, "grad_norm": 1.115423321723938, "learning_rate": 9.177129359245471e-05, "loss": 0.6518, "step": 6248 }, { "epoch": 0.41944901177812827, "grad_norm": 1.0860360860824585, "learning_rate": 9.176531935035009e-05, "loss": 0.6096, "step": 6250 }, { "epoch": 0.41958323546189724, "grad_norm": 1.2603168487548828, "learning_rate": 9.175934313491625e-05, "loss": 0.7101, "step": 6252 }, { "epoch": 0.41971745914566627, "grad_norm": 1.0043116807937622, "learning_rate": 9.175336494643557e-05, "loss": 0.7133, "step": 6254 }, { "epoch": 0.41985168282943525, "grad_norm": 1.0733429193496704, "learning_rate": 9.174738478519047e-05, "loss": 0.6137, "step": 6256 }, { "epoch": 0.4199859065132043, "grad_norm": 1.2858965396881104, "learning_rate": 9.174140265146356e-05, "loss": 0.6648, "step": 6258 }, { "epoch": 0.42012013019697325, "grad_norm": 1.0367754697799683, "learning_rate": 9.173541854553745e-05, "loss": 0.6619, "step": 6260 }, { "epoch": 0.4202543538807423, "grad_norm": 1.035481333732605, "learning_rate": 9.172943246769489e-05, "loss": 0.7092, "step": 6262 }, { "epoch": 0.42038857756451126, "grad_norm": 0.9812489151954651, "learning_rate": 9.17234444182187e-05, "loss": 0.6362, "step": 6264 }, { "epoch": 0.42052280124828023, "grad_norm": 1.0386441946029663, "learning_rate": 9.17174543973918e-05, "loss": 0.6318, "step": 6266 }, { "epoch": 0.42065702493204926, "grad_norm": 1.134960651397705, "learning_rate": 9.171146240549722e-05, "loss": 0.7133, "step": 6268 }, { "epoch": 0.42079124861581824, "grad_norm": 1.1966896057128906, "learning_rate": 9.170546844281807e-05, "loss": 0.7314, "step": 6270 }, { "epoch": 0.42092547229958727, "grad_norm": 0.960020124912262, "learning_rate": 9.169947250963753e-05, "loss": 0.6676, "step": 6272 }, { "epoch": 0.42105969598335624, "grad_norm": 1.9024641513824463, "learning_rate": 9.169347460623892e-05, "loss": 0.598, "step": 6274 }, { "epoch": 0.42119391966712527, "grad_norm": 1.1077617406845093, "learning_rate": 9.168747473290562e-05, "loss": 0.7323, "step": 6276 }, { "epoch": 0.42132814335089425, "grad_norm": 1.0826691389083862, "learning_rate": 9.168147288992112e-05, "loss": 0.7214, "step": 6278 }, { "epoch": 0.4214623670346633, "grad_norm": 1.00456702709198, "learning_rate": 9.167546907756898e-05, "loss": 0.6772, "step": 6280 }, { "epoch": 0.42159659071843225, "grad_norm": 1.0859038829803467, "learning_rate": 9.166946329613288e-05, "loss": 0.6402, "step": 6282 }, { "epoch": 0.4217308144022013, "grad_norm": 1.00326669216156, "learning_rate": 9.166345554589658e-05, "loss": 0.5797, "step": 6284 }, { "epoch": 0.42186503808597026, "grad_norm": 1.0553573369979858, "learning_rate": 9.165744582714393e-05, "loss": 0.6244, "step": 6286 }, { "epoch": 0.4219992617697393, "grad_norm": 1.1167014837265015, "learning_rate": 9.165143414015889e-05, "loss": 0.7148, "step": 6288 }, { "epoch": 0.42213348545350826, "grad_norm": 1.0160471200942993, "learning_rate": 9.164542048522549e-05, "loss": 0.6077, "step": 6290 }, { "epoch": 0.4222677091372773, "grad_norm": 0.9857290983200073, "learning_rate": 9.163940486262785e-05, "loss": 0.6954, "step": 6292 }, { "epoch": 0.42240193282104627, "grad_norm": 1.2125552892684937, "learning_rate": 9.163338727265022e-05, "loss": 0.6373, "step": 6294 }, { "epoch": 0.4225361565048153, "grad_norm": 1.0877693891525269, "learning_rate": 9.162736771557692e-05, "loss": 0.7392, "step": 6296 }, { "epoch": 0.42267038018858427, "grad_norm": 1.0825779438018799, "learning_rate": 9.162134619169233e-05, "loss": 0.6596, "step": 6298 }, { "epoch": 0.4228046038723533, "grad_norm": 1.0345951318740845, "learning_rate": 9.161532270128099e-05, "loss": 0.6746, "step": 6300 }, { "epoch": 0.4229388275561223, "grad_norm": 1.1396832466125488, "learning_rate": 9.16092972446275e-05, "loss": 0.7129, "step": 6302 }, { "epoch": 0.42307305123989125, "grad_norm": 0.9923988580703735, "learning_rate": 9.160326982201652e-05, "loss": 0.6696, "step": 6304 }, { "epoch": 0.4232072749236603, "grad_norm": 1.2008332014083862, "learning_rate": 9.159724043373284e-05, "loss": 0.6764, "step": 6306 }, { "epoch": 0.42334149860742926, "grad_norm": 1.0264410972595215, "learning_rate": 9.159120908006135e-05, "loss": 0.6352, "step": 6308 }, { "epoch": 0.4234757222911983, "grad_norm": 1.1287544965744019, "learning_rate": 9.158517576128705e-05, "loss": 0.642, "step": 6310 }, { "epoch": 0.42360994597496726, "grad_norm": 1.1886465549468994, "learning_rate": 9.157914047769493e-05, "loss": 0.6993, "step": 6312 }, { "epoch": 0.4237441696587363, "grad_norm": 1.4661725759506226, "learning_rate": 9.15731032295702e-05, "loss": 0.6129, "step": 6314 }, { "epoch": 0.42387839334250527, "grad_norm": 1.1720086336135864, "learning_rate": 9.15670640171981e-05, "loss": 0.6844, "step": 6316 }, { "epoch": 0.4240126170262743, "grad_norm": 2.242379903793335, "learning_rate": 9.156102284086394e-05, "loss": 0.7241, "step": 6318 }, { "epoch": 0.42414684071004327, "grad_norm": 1.0734869241714478, "learning_rate": 9.15549797008532e-05, "loss": 0.6562, "step": 6320 }, { "epoch": 0.4242810643938123, "grad_norm": 1.0534396171569824, "learning_rate": 9.154893459745138e-05, "loss": 0.6223, "step": 6322 }, { "epoch": 0.4244152880775813, "grad_norm": 0.8479713201522827, "learning_rate": 9.154288753094408e-05, "loss": 0.5982, "step": 6324 }, { "epoch": 0.4245495117613503, "grad_norm": 1.071876049041748, "learning_rate": 9.153683850161706e-05, "loss": 0.6977, "step": 6326 }, { "epoch": 0.4246837354451193, "grad_norm": 1.4902291297912598, "learning_rate": 9.15307875097561e-05, "loss": 0.6598, "step": 6328 }, { "epoch": 0.4248179591288883, "grad_norm": 1.0573935508728027, "learning_rate": 9.152473455564708e-05, "loss": 0.7148, "step": 6330 }, { "epoch": 0.4249521828126573, "grad_norm": 1.0465713739395142, "learning_rate": 9.151867963957601e-05, "loss": 0.6745, "step": 6332 }, { "epoch": 0.4250864064964263, "grad_norm": 1.1158103942871094, "learning_rate": 9.151262276182898e-05, "loss": 0.6817, "step": 6334 }, { "epoch": 0.4252206301801953, "grad_norm": 1.0501621961593628, "learning_rate": 9.150656392269215e-05, "loss": 0.7074, "step": 6336 }, { "epoch": 0.4253548538639643, "grad_norm": 1.1173028945922852, "learning_rate": 9.15005031224518e-05, "loss": 0.7018, "step": 6338 }, { "epoch": 0.4254890775477333, "grad_norm": 1.1124805212020874, "learning_rate": 9.149444036139427e-05, "loss": 0.6565, "step": 6340 }, { "epoch": 0.42562330123150227, "grad_norm": 1.1531590223312378, "learning_rate": 9.148837563980606e-05, "loss": 0.6702, "step": 6342 }, { "epoch": 0.4257575249152713, "grad_norm": 1.3122427463531494, "learning_rate": 9.148230895797366e-05, "loss": 0.6935, "step": 6344 }, { "epoch": 0.4258917485990403, "grad_norm": 1.274470329284668, "learning_rate": 9.147624031618373e-05, "loss": 0.6396, "step": 6346 }, { "epoch": 0.4260259722828093, "grad_norm": 1.0647002458572388, "learning_rate": 9.147016971472299e-05, "loss": 0.6394, "step": 6348 }, { "epoch": 0.4261601959665783, "grad_norm": 2.0520029067993164, "learning_rate": 9.146409715387832e-05, "loss": 0.6615, "step": 6350 }, { "epoch": 0.4262944196503473, "grad_norm": 1.0095345973968506, "learning_rate": 9.145802263393657e-05, "loss": 0.5958, "step": 6352 }, { "epoch": 0.4264286433341163, "grad_norm": 1.1976794004440308, "learning_rate": 9.145194615518477e-05, "loss": 0.6894, "step": 6354 }, { "epoch": 0.4265628670178853, "grad_norm": 1.1440362930297852, "learning_rate": 9.144586771791003e-05, "loss": 0.6281, "step": 6356 }, { "epoch": 0.4266970907016543, "grad_norm": 1.0716235637664795, "learning_rate": 9.143978732239955e-05, "loss": 0.7333, "step": 6358 }, { "epoch": 0.4268313143854233, "grad_norm": 1.2851024866104126, "learning_rate": 9.143370496894061e-05, "loss": 0.7129, "step": 6360 }, { "epoch": 0.4269655380691923, "grad_norm": 1.1375380754470825, "learning_rate": 9.142762065782058e-05, "loss": 0.658, "step": 6362 }, { "epoch": 0.4270997617529613, "grad_norm": 0.9954057931900024, "learning_rate": 9.142153438932693e-05, "loss": 0.657, "step": 6364 }, { "epoch": 0.4272339854367303, "grad_norm": 0.9236775636672974, "learning_rate": 9.141544616374724e-05, "loss": 0.6579, "step": 6366 }, { "epoch": 0.42736820912049933, "grad_norm": 1.0957763195037842, "learning_rate": 9.140935598136914e-05, "loss": 0.7027, "step": 6368 }, { "epoch": 0.4275024328042683, "grad_norm": 1.032928705215454, "learning_rate": 9.140326384248042e-05, "loss": 0.7045, "step": 6370 }, { "epoch": 0.42763665648803734, "grad_norm": 1.0739259719848633, "learning_rate": 9.139716974736889e-05, "loss": 0.6771, "step": 6372 }, { "epoch": 0.4277708801718063, "grad_norm": 1.215862512588501, "learning_rate": 9.13910736963225e-05, "loss": 0.6626, "step": 6374 }, { "epoch": 0.42790510385557534, "grad_norm": 1.0815331935882568, "learning_rate": 9.138497568962927e-05, "loss": 0.6489, "step": 6376 }, { "epoch": 0.4280393275393443, "grad_norm": 0.9747056365013123, "learning_rate": 9.137887572757732e-05, "loss": 0.6308, "step": 6378 }, { "epoch": 0.4281735512231133, "grad_norm": 1.240158200263977, "learning_rate": 9.137277381045486e-05, "loss": 0.5727, "step": 6380 }, { "epoch": 0.4283077749068823, "grad_norm": 1.1604822874069214, "learning_rate": 9.136666993855018e-05, "loss": 0.6793, "step": 6382 }, { "epoch": 0.4284419985906513, "grad_norm": 1.0272921323776245, "learning_rate": 9.13605641121517e-05, "loss": 0.6959, "step": 6384 }, { "epoch": 0.4285762222744203, "grad_norm": 1.115839958190918, "learning_rate": 9.135445633154789e-05, "loss": 0.684, "step": 6386 }, { "epoch": 0.4287104459581893, "grad_norm": 1.3776438236236572, "learning_rate": 9.134834659702736e-05, "loss": 0.7394, "step": 6388 }, { "epoch": 0.42884466964195833, "grad_norm": 1.1234407424926758, "learning_rate": 9.134223490887875e-05, "loss": 0.6542, "step": 6390 }, { "epoch": 0.4289788933257273, "grad_norm": 1.107689619064331, "learning_rate": 9.133612126739082e-05, "loss": 0.6801, "step": 6392 }, { "epoch": 0.42911311700949634, "grad_norm": 1.1047486066818237, "learning_rate": 9.133000567285245e-05, "loss": 0.6355, "step": 6394 }, { "epoch": 0.4292473406932653, "grad_norm": 1.3371986150741577, "learning_rate": 9.13238881255526e-05, "loss": 0.7247, "step": 6396 }, { "epoch": 0.42938156437703434, "grad_norm": 0.9681903719902039, "learning_rate": 9.131776862578027e-05, "loss": 0.6958, "step": 6398 }, { "epoch": 0.4295157880608033, "grad_norm": 1.271085262298584, "learning_rate": 9.131164717382466e-05, "loss": 0.7858, "step": 6400 }, { "epoch": 0.42965001174457235, "grad_norm": 1.1457453966140747, "learning_rate": 9.130552376997492e-05, "loss": 0.6198, "step": 6402 }, { "epoch": 0.4297842354283413, "grad_norm": 1.0122690200805664, "learning_rate": 9.129939841452042e-05, "loss": 0.652, "step": 6404 }, { "epoch": 0.42991845911211035, "grad_norm": 0.8656542301177979, "learning_rate": 9.129327110775056e-05, "loss": 0.5501, "step": 6406 }, { "epoch": 0.4300526827958793, "grad_norm": 1.0972250699996948, "learning_rate": 9.128714184995483e-05, "loss": 0.62, "step": 6408 }, { "epoch": 0.43018690647964836, "grad_norm": 1.1215505599975586, "learning_rate": 9.128101064142285e-05, "loss": 0.6519, "step": 6410 }, { "epoch": 0.43032113016341733, "grad_norm": 1.1611055135726929, "learning_rate": 9.127487748244427e-05, "loss": 0.7007, "step": 6412 }, { "epoch": 0.43045535384718636, "grad_norm": 1.1495709419250488, "learning_rate": 9.126874237330891e-05, "loss": 0.7069, "step": 6414 }, { "epoch": 0.43058957753095534, "grad_norm": 1.1690107583999634, "learning_rate": 9.126260531430662e-05, "loss": 0.6409, "step": 6416 }, { "epoch": 0.4307238012147243, "grad_norm": 1.081002116203308, "learning_rate": 9.125646630572737e-05, "loss": 0.6799, "step": 6418 }, { "epoch": 0.43085802489849334, "grad_norm": 1.0002063512802124, "learning_rate": 9.125032534786122e-05, "loss": 0.6398, "step": 6420 }, { "epoch": 0.4309922485822623, "grad_norm": 1.1421399116516113, "learning_rate": 9.12441824409983e-05, "loss": 0.6651, "step": 6422 }, { "epoch": 0.43112647226603135, "grad_norm": 0.9874308109283447, "learning_rate": 9.123803758542888e-05, "loss": 0.5975, "step": 6424 }, { "epoch": 0.4312606959498003, "grad_norm": 1.8552358150482178, "learning_rate": 9.123189078144326e-05, "loss": 0.6139, "step": 6426 }, { "epoch": 0.43139491963356935, "grad_norm": 1.4092941284179688, "learning_rate": 9.122574202933188e-05, "loss": 0.7107, "step": 6428 }, { "epoch": 0.4315291433173383, "grad_norm": 1.1505235433578491, "learning_rate": 9.121959132938527e-05, "loss": 0.6559, "step": 6430 }, { "epoch": 0.43166336700110736, "grad_norm": 0.9634313583374023, "learning_rate": 9.1213438681894e-05, "loss": 0.6344, "step": 6432 }, { "epoch": 0.43179759068487633, "grad_norm": 0.8595042824745178, "learning_rate": 9.120728408714882e-05, "loss": 0.6284, "step": 6434 }, { "epoch": 0.43193181436864536, "grad_norm": 0.9590168595314026, "learning_rate": 9.120112754544047e-05, "loss": 0.654, "step": 6436 }, { "epoch": 0.43206603805241434, "grad_norm": 0.9372844099998474, "learning_rate": 9.119496905705989e-05, "loss": 0.5882, "step": 6438 }, { "epoch": 0.43220026173618337, "grad_norm": 1.245493769645691, "learning_rate": 9.118880862229802e-05, "loss": 0.7153, "step": 6440 }, { "epoch": 0.43233448541995234, "grad_norm": 1.0889474153518677, "learning_rate": 9.118264624144594e-05, "loss": 0.6479, "step": 6442 }, { "epoch": 0.43246870910372137, "grad_norm": 1.0484486818313599, "learning_rate": 9.11764819147948e-05, "loss": 0.6561, "step": 6444 }, { "epoch": 0.43260293278749035, "grad_norm": 1.3504307270050049, "learning_rate": 9.117031564263584e-05, "loss": 0.7183, "step": 6446 }, { "epoch": 0.4327371564712594, "grad_norm": 1.2466015815734863, "learning_rate": 9.116414742526047e-05, "loss": 0.6879, "step": 6448 }, { "epoch": 0.43287138015502835, "grad_norm": 1.1774563789367676, "learning_rate": 9.115797726296004e-05, "loss": 0.6828, "step": 6450 }, { "epoch": 0.4330056038387974, "grad_norm": 0.9712786674499512, "learning_rate": 9.115180515602614e-05, "loss": 0.6084, "step": 6452 }, { "epoch": 0.43313982752256636, "grad_norm": 1.1016490459442139, "learning_rate": 9.114563110475036e-05, "loss": 0.6523, "step": 6454 }, { "epoch": 0.43327405120633533, "grad_norm": 1.0148249864578247, "learning_rate": 9.113945510942443e-05, "loss": 0.6723, "step": 6456 }, { "epoch": 0.43340827489010436, "grad_norm": 0.9622500538825989, "learning_rate": 9.113327717034013e-05, "loss": 0.6633, "step": 6458 }, { "epoch": 0.43354249857387334, "grad_norm": 1.0464283227920532, "learning_rate": 9.112709728778937e-05, "loss": 0.6835, "step": 6460 }, { "epoch": 0.43367672225764237, "grad_norm": 1.1165943145751953, "learning_rate": 9.112091546206414e-05, "loss": 0.5918, "step": 6462 }, { "epoch": 0.43381094594141134, "grad_norm": 1.076734185218811, "learning_rate": 9.111473169345652e-05, "loss": 0.6961, "step": 6464 }, { "epoch": 0.43394516962518037, "grad_norm": 1.0434982776641846, "learning_rate": 9.110854598225867e-05, "loss": 0.638, "step": 6466 }, { "epoch": 0.43407939330894935, "grad_norm": 1.0204663276672363, "learning_rate": 9.110235832876286e-05, "loss": 0.6514, "step": 6468 }, { "epoch": 0.4342136169927184, "grad_norm": 1.104552984237671, "learning_rate": 9.109616873326144e-05, "loss": 0.6957, "step": 6470 }, { "epoch": 0.43434784067648735, "grad_norm": 1.1974577903747559, "learning_rate": 9.108997719604687e-05, "loss": 0.6785, "step": 6472 }, { "epoch": 0.4344820643602564, "grad_norm": 0.9422672986984253, "learning_rate": 9.108378371741167e-05, "loss": 0.6053, "step": 6474 }, { "epoch": 0.43461628804402536, "grad_norm": 1.0764330625534058, "learning_rate": 9.107758829764848e-05, "loss": 0.6145, "step": 6476 }, { "epoch": 0.4347505117277944, "grad_norm": 0.8954029679298401, "learning_rate": 9.107139093705001e-05, "loss": 0.6287, "step": 6478 }, { "epoch": 0.43488473541156336, "grad_norm": 1.232624888420105, "learning_rate": 9.106519163590907e-05, "loss": 0.7467, "step": 6480 }, { "epoch": 0.4350189590953324, "grad_norm": 1.0775651931762695, "learning_rate": 9.10589903945186e-05, "loss": 0.6934, "step": 6482 }, { "epoch": 0.43515318277910137, "grad_norm": 0.9535127282142639, "learning_rate": 9.105278721317157e-05, "loss": 0.6415, "step": 6484 }, { "epoch": 0.4352874064628704, "grad_norm": 1.1277869939804077, "learning_rate": 9.104658209216108e-05, "loss": 0.7023, "step": 6486 }, { "epoch": 0.43542163014663937, "grad_norm": 1.0773403644561768, "learning_rate": 9.104037503178028e-05, "loss": 0.6835, "step": 6488 }, { "epoch": 0.4355558538304084, "grad_norm": 1.0601844787597656, "learning_rate": 9.103416603232246e-05, "loss": 0.6951, "step": 6490 }, { "epoch": 0.4356900775141774, "grad_norm": 0.9843427538871765, "learning_rate": 9.102795509408099e-05, "loss": 0.6212, "step": 6492 }, { "epoch": 0.43582430119794635, "grad_norm": 1.3423608541488647, "learning_rate": 9.102174221734934e-05, "loss": 0.7103, "step": 6494 }, { "epoch": 0.4359585248817154, "grad_norm": 0.9906916618347168, "learning_rate": 9.101552740242102e-05, "loss": 0.6194, "step": 6496 }, { "epoch": 0.43609274856548436, "grad_norm": 1.5656930208206177, "learning_rate": 9.100931064958968e-05, "loss": 0.6622, "step": 6498 }, { "epoch": 0.4362269722492534, "grad_norm": 1.0899085998535156, "learning_rate": 9.100309195914907e-05, "loss": 0.6361, "step": 6500 }, { "epoch": 0.43636119593302236, "grad_norm": 0.9742255806922913, "learning_rate": 9.099687133139298e-05, "loss": 0.5813, "step": 6502 }, { "epoch": 0.4364954196167914, "grad_norm": 1.050657033920288, "learning_rate": 9.099064876661533e-05, "loss": 0.6577, "step": 6504 }, { "epoch": 0.43662964330056037, "grad_norm": 1.137790560722351, "learning_rate": 9.098442426511014e-05, "loss": 0.6901, "step": 6506 }, { "epoch": 0.4367638669843294, "grad_norm": 1.0315080881118774, "learning_rate": 9.09781978271715e-05, "loss": 0.6881, "step": 6508 }, { "epoch": 0.43689809066809837, "grad_norm": 0.9549081325531006, "learning_rate": 9.097196945309359e-05, "loss": 0.6294, "step": 6510 }, { "epoch": 0.4370323143518674, "grad_norm": 1.1911718845367432, "learning_rate": 9.096573914317068e-05, "loss": 0.6654, "step": 6512 }, { "epoch": 0.4371665380356364, "grad_norm": 1.0279988050460815, "learning_rate": 9.095950689769716e-05, "loss": 0.6934, "step": 6514 }, { "epoch": 0.4373007617194054, "grad_norm": 1.69241201877594, "learning_rate": 9.095327271696749e-05, "loss": 0.6478, "step": 6516 }, { "epoch": 0.4374349854031744, "grad_norm": 1.1771697998046875, "learning_rate": 9.094703660127622e-05, "loss": 0.6668, "step": 6518 }, { "epoch": 0.4375692090869434, "grad_norm": 1.045859932899475, "learning_rate": 9.094079855091797e-05, "loss": 0.7428, "step": 6520 }, { "epoch": 0.4377034327707124, "grad_norm": 1.130373239517212, "learning_rate": 9.09345585661875e-05, "loss": 0.6426, "step": 6522 }, { "epoch": 0.4378376564544814, "grad_norm": 1.051623821258545, "learning_rate": 9.092831664737964e-05, "loss": 0.7218, "step": 6524 }, { "epoch": 0.4379718801382504, "grad_norm": 1.0282305479049683, "learning_rate": 9.092207279478929e-05, "loss": 0.6457, "step": 6526 }, { "epoch": 0.4381061038220194, "grad_norm": 1.3251056671142578, "learning_rate": 9.091582700871148e-05, "loss": 0.6205, "step": 6528 }, { "epoch": 0.4382403275057884, "grad_norm": 1.111969232559204, "learning_rate": 9.090957928944129e-05, "loss": 0.6603, "step": 6530 }, { "epoch": 0.43837455118955737, "grad_norm": 1.0762337446212769, "learning_rate": 9.090332963727393e-05, "loss": 0.6861, "step": 6532 }, { "epoch": 0.4385087748733264, "grad_norm": 1.1892808675765991, "learning_rate": 9.089707805250468e-05, "loss": 0.6627, "step": 6534 }, { "epoch": 0.4386429985570954, "grad_norm": 0.946892499923706, "learning_rate": 9.089082453542891e-05, "loss": 0.6566, "step": 6536 }, { "epoch": 0.4387772222408644, "grad_norm": 1.0272513628005981, "learning_rate": 9.088456908634209e-05, "loss": 0.6747, "step": 6538 }, { "epoch": 0.4389114459246334, "grad_norm": 1.255175232887268, "learning_rate": 9.087831170553978e-05, "loss": 0.6437, "step": 6540 }, { "epoch": 0.4390456696084024, "grad_norm": 0.9205337166786194, "learning_rate": 9.087205239331762e-05, "loss": 0.6389, "step": 6542 }, { "epoch": 0.4391798932921714, "grad_norm": 1.233696699142456, "learning_rate": 9.086579114997136e-05, "loss": 0.6947, "step": 6544 }, { "epoch": 0.4393141169759404, "grad_norm": 1.099592924118042, "learning_rate": 9.085952797579682e-05, "loss": 0.6951, "step": 6546 }, { "epoch": 0.4394483406597094, "grad_norm": 1.2203367948532104, "learning_rate": 9.085326287108995e-05, "loss": 0.7214, "step": 6548 }, { "epoch": 0.4395825643434784, "grad_norm": 1.1897811889648438, "learning_rate": 9.084699583614673e-05, "loss": 0.6923, "step": 6550 }, { "epoch": 0.4397167880272474, "grad_norm": 1.0716934204101562, "learning_rate": 9.084072687126327e-05, "loss": 0.7315, "step": 6552 }, { "epoch": 0.4398510117110164, "grad_norm": 1.0627871751785278, "learning_rate": 9.083445597673578e-05, "loss": 0.602, "step": 6554 }, { "epoch": 0.4399852353947854, "grad_norm": 0.977236270904541, "learning_rate": 9.082818315286055e-05, "loss": 0.6165, "step": 6556 }, { "epoch": 0.44011945907855443, "grad_norm": 0.9908711910247803, "learning_rate": 9.082190839993395e-05, "loss": 0.6062, "step": 6558 }, { "epoch": 0.4402536827623234, "grad_norm": 1.441779613494873, "learning_rate": 9.081563171825245e-05, "loss": 0.6575, "step": 6560 }, { "epoch": 0.44038790644609244, "grad_norm": 0.9981227517127991, "learning_rate": 9.08093531081126e-05, "loss": 0.686, "step": 6562 }, { "epoch": 0.4405221301298614, "grad_norm": 1.0307862758636475, "learning_rate": 9.080307256981109e-05, "loss": 0.6502, "step": 6564 }, { "epoch": 0.44065635381363044, "grad_norm": 1.0696779489517212, "learning_rate": 9.079679010364461e-05, "loss": 0.6779, "step": 6566 }, { "epoch": 0.4407905774973994, "grad_norm": 0.9596073627471924, "learning_rate": 9.079050570991004e-05, "loss": 0.6073, "step": 6568 }, { "epoch": 0.4409248011811684, "grad_norm": 1.1085256338119507, "learning_rate": 9.078421938890426e-05, "loss": 0.623, "step": 6570 }, { "epoch": 0.4410590248649374, "grad_norm": 0.9325559139251709, "learning_rate": 9.077793114092435e-05, "loss": 0.6415, "step": 6572 }, { "epoch": 0.4411932485487064, "grad_norm": 1.551746129989624, "learning_rate": 9.077164096626736e-05, "loss": 0.6947, "step": 6574 }, { "epoch": 0.4413274722324754, "grad_norm": 1.1572011709213257, "learning_rate": 9.07653488652305e-05, "loss": 0.6643, "step": 6576 }, { "epoch": 0.4414616959162444, "grad_norm": 1.0599066019058228, "learning_rate": 9.075905483811109e-05, "loss": 0.6904, "step": 6578 }, { "epoch": 0.44159591960001343, "grad_norm": 1.0249269008636475, "learning_rate": 9.075275888520647e-05, "loss": 0.617, "step": 6580 }, { "epoch": 0.4417301432837824, "grad_norm": 1.1034280061721802, "learning_rate": 9.074646100681413e-05, "loss": 0.6757, "step": 6582 }, { "epoch": 0.44186436696755144, "grad_norm": 0.9876016974449158, "learning_rate": 9.074016120323163e-05, "loss": 0.632, "step": 6584 }, { "epoch": 0.4419985906513204, "grad_norm": 1.3852652311325073, "learning_rate": 9.073385947475664e-05, "loss": 0.6871, "step": 6586 }, { "epoch": 0.44213281433508944, "grad_norm": 1.137299656867981, "learning_rate": 9.072755582168688e-05, "loss": 0.6558, "step": 6588 }, { "epoch": 0.4422670380188584, "grad_norm": 0.9819977283477783, "learning_rate": 9.07212502443202e-05, "loss": 0.6194, "step": 6590 }, { "epoch": 0.44240126170262745, "grad_norm": 1.3733234405517578, "learning_rate": 9.071494274295452e-05, "loss": 0.6471, "step": 6592 }, { "epoch": 0.4425354853863964, "grad_norm": 1.1099233627319336, "learning_rate": 9.070863331788785e-05, "loss": 0.679, "step": 6594 }, { "epoch": 0.44266970907016545, "grad_norm": 1.0077110528945923, "learning_rate": 9.07023219694183e-05, "loss": 0.6723, "step": 6596 }, { "epoch": 0.4428039327539344, "grad_norm": 0.9958982467651367, "learning_rate": 9.06960086978441e-05, "loss": 0.6577, "step": 6598 }, { "epoch": 0.44293815643770346, "grad_norm": 1.4362531900405884, "learning_rate": 9.068969350346349e-05, "loss": 0.6941, "step": 6600 }, { "epoch": 0.44307238012147243, "grad_norm": 2.014660358428955, "learning_rate": 9.068337638657489e-05, "loss": 0.6327, "step": 6602 }, { "epoch": 0.44320660380524146, "grad_norm": 0.9608994126319885, "learning_rate": 9.067705734747674e-05, "loss": 0.6199, "step": 6604 }, { "epoch": 0.44334082748901044, "grad_norm": 1.7143373489379883, "learning_rate": 9.067073638646763e-05, "loss": 0.7387, "step": 6606 }, { "epoch": 0.4434750511727794, "grad_norm": 0.9786049723625183, "learning_rate": 9.06644135038462e-05, "loss": 0.6506, "step": 6608 }, { "epoch": 0.44360927485654844, "grad_norm": 3.5571486949920654, "learning_rate": 9.06580886999112e-05, "loss": 0.6904, "step": 6610 }, { "epoch": 0.4437434985403174, "grad_norm": 1.0379436016082764, "learning_rate": 9.065176197496146e-05, "loss": 0.6766, "step": 6612 }, { "epoch": 0.44387772222408645, "grad_norm": 1.0329163074493408, "learning_rate": 9.064543332929589e-05, "loss": 0.6724, "step": 6614 }, { "epoch": 0.4440119459078554, "grad_norm": 1.0833524465560913, "learning_rate": 9.063910276321354e-05, "loss": 0.6512, "step": 6616 }, { "epoch": 0.44414616959162445, "grad_norm": 1.4965224266052246, "learning_rate": 9.063277027701349e-05, "loss": 0.6345, "step": 6618 }, { "epoch": 0.4442803932753934, "grad_norm": 0.9639096856117249, "learning_rate": 9.062643587099495e-05, "loss": 0.6275, "step": 6620 }, { "epoch": 0.44441461695916246, "grad_norm": 1.1125866174697876, "learning_rate": 9.062009954545719e-05, "loss": 0.5863, "step": 6622 }, { "epoch": 0.44454884064293143, "grad_norm": 1.0979225635528564, "learning_rate": 9.061376130069961e-05, "loss": 0.696, "step": 6624 }, { "epoch": 0.44468306432670046, "grad_norm": 0.9352512359619141, "learning_rate": 9.060742113702168e-05, "loss": 0.6487, "step": 6626 }, { "epoch": 0.44481728801046944, "grad_norm": 1.1315749883651733, "learning_rate": 9.060107905472294e-05, "loss": 0.7242, "step": 6628 }, { "epoch": 0.44495151169423847, "grad_norm": 1.142823338508606, "learning_rate": 9.059473505410305e-05, "loss": 0.6645, "step": 6630 }, { "epoch": 0.44508573537800744, "grad_norm": 1.253380298614502, "learning_rate": 9.058838913546178e-05, "loss": 0.6615, "step": 6632 }, { "epoch": 0.44521995906177647, "grad_norm": 1.1333796977996826, "learning_rate": 9.058204129909891e-05, "loss": 0.6809, "step": 6634 }, { "epoch": 0.44535418274554545, "grad_norm": 1.2247251272201538, "learning_rate": 9.05756915453144e-05, "loss": 0.6851, "step": 6636 }, { "epoch": 0.4454884064293145, "grad_norm": 1.5173280239105225, "learning_rate": 9.056933987440825e-05, "loss": 0.7467, "step": 6638 }, { "epoch": 0.44562263011308345, "grad_norm": 1.110931158065796, "learning_rate": 9.056298628668056e-05, "loss": 0.722, "step": 6640 }, { "epoch": 0.4457568537968525, "grad_norm": 1.1217375993728638, "learning_rate": 9.055663078243156e-05, "loss": 0.6467, "step": 6642 }, { "epoch": 0.44589107748062146, "grad_norm": 1.1028882265090942, "learning_rate": 9.055027336196146e-05, "loss": 0.6725, "step": 6644 }, { "epoch": 0.44602530116439043, "grad_norm": 1.1858761310577393, "learning_rate": 9.054391402557072e-05, "loss": 0.6381, "step": 6646 }, { "epoch": 0.44615952484815946, "grad_norm": 1.0766549110412598, "learning_rate": 9.053755277355976e-05, "loss": 0.6382, "step": 6648 }, { "epoch": 0.44629374853192844, "grad_norm": 1.1438759565353394, "learning_rate": 9.053118960622915e-05, "loss": 0.7142, "step": 6650 }, { "epoch": 0.44642797221569747, "grad_norm": 1.0107367038726807, "learning_rate": 9.052482452387953e-05, "loss": 0.5961, "step": 6652 }, { "epoch": 0.44656219589946644, "grad_norm": 0.9405455589294434, "learning_rate": 9.051845752681163e-05, "loss": 0.6392, "step": 6654 }, { "epoch": 0.44669641958323547, "grad_norm": 1.139909029006958, "learning_rate": 9.051208861532629e-05, "loss": 0.6692, "step": 6656 }, { "epoch": 0.44683064326700445, "grad_norm": 1.0859870910644531, "learning_rate": 9.050571778972443e-05, "loss": 0.7422, "step": 6658 }, { "epoch": 0.4469648669507735, "grad_norm": 1.1717115640640259, "learning_rate": 9.049934505030705e-05, "loss": 0.599, "step": 6660 }, { "epoch": 0.44709909063454245, "grad_norm": 1.094762921333313, "learning_rate": 9.049297039737528e-05, "loss": 0.6316, "step": 6662 }, { "epoch": 0.4472333143183115, "grad_norm": 1.4241257905960083, "learning_rate": 9.048659383123026e-05, "loss": 0.6412, "step": 6664 }, { "epoch": 0.44736753800208046, "grad_norm": 1.7000561952590942, "learning_rate": 9.04802153521733e-05, "loss": 0.7028, "step": 6666 }, { "epoch": 0.4475017616858495, "grad_norm": 1.068803310394287, "learning_rate": 9.047383496050576e-05, "loss": 0.6925, "step": 6668 }, { "epoch": 0.44763598536961846, "grad_norm": 0.993427574634552, "learning_rate": 9.046745265652912e-05, "loss": 0.69, "step": 6670 }, { "epoch": 0.4477702090533875, "grad_norm": 1.00606369972229, "learning_rate": 9.046106844054491e-05, "loss": 0.6426, "step": 6672 }, { "epoch": 0.44790443273715647, "grad_norm": 1.0299243927001953, "learning_rate": 9.045468231285477e-05, "loss": 0.6346, "step": 6674 }, { "epoch": 0.4480386564209255, "grad_norm": 1.0653523206710815, "learning_rate": 9.044829427376046e-05, "loss": 0.7017, "step": 6676 }, { "epoch": 0.44817288010469447, "grad_norm": 1.4495142698287964, "learning_rate": 9.044190432356377e-05, "loss": 0.6282, "step": 6678 }, { "epoch": 0.4483071037884635, "grad_norm": 1.0920124053955078, "learning_rate": 9.043551246256664e-05, "loss": 0.6573, "step": 6680 }, { "epoch": 0.4484413274722325, "grad_norm": 0.9434648752212524, "learning_rate": 9.042911869107105e-05, "loss": 0.6084, "step": 6682 }, { "epoch": 0.44857555115600145, "grad_norm": 1.13992440700531, "learning_rate": 9.04227230093791e-05, "loss": 0.671, "step": 6684 }, { "epoch": 0.4487097748397705, "grad_norm": 0.9534379243850708, "learning_rate": 9.041632541779298e-05, "loss": 0.6019, "step": 6686 }, { "epoch": 0.44884399852353946, "grad_norm": 1.0360051393508911, "learning_rate": 9.040992591661495e-05, "loss": 0.6658, "step": 6688 }, { "epoch": 0.4489782222073085, "grad_norm": 1.1044501066207886, "learning_rate": 9.04035245061474e-05, "loss": 0.6831, "step": 6690 }, { "epoch": 0.44911244589107746, "grad_norm": 1.1218234300613403, "learning_rate": 9.039712118669276e-05, "loss": 0.7078, "step": 6692 }, { "epoch": 0.4492466695748465, "grad_norm": 1.112831711769104, "learning_rate": 9.039071595855357e-05, "loss": 0.7103, "step": 6694 }, { "epoch": 0.44938089325861547, "grad_norm": 1.0213844776153564, "learning_rate": 9.038430882203249e-05, "loss": 0.6505, "step": 6696 }, { "epoch": 0.4495151169423845, "grad_norm": 1.2835336923599243, "learning_rate": 9.037789977743223e-05, "loss": 0.6708, "step": 6698 }, { "epoch": 0.44964934062615347, "grad_norm": 1.0042214393615723, "learning_rate": 9.03714888250556e-05, "loss": 0.7034, "step": 6700 }, { "epoch": 0.4497835643099225, "grad_norm": 1.216511845588684, "learning_rate": 9.036507596520551e-05, "loss": 0.6803, "step": 6702 }, { "epoch": 0.4499177879936915, "grad_norm": 0.9730432033538818, "learning_rate": 9.035866119818495e-05, "loss": 0.625, "step": 6704 }, { "epoch": 0.4500520116774605, "grad_norm": 1.2708172798156738, "learning_rate": 9.035224452429703e-05, "loss": 0.6832, "step": 6706 }, { "epoch": 0.4501862353612295, "grad_norm": 1.033936858177185, "learning_rate": 9.034582594384488e-05, "loss": 0.6413, "step": 6708 }, { "epoch": 0.4503204590449985, "grad_norm": 1.223681092262268, "learning_rate": 9.033940545713182e-05, "loss": 0.6432, "step": 6710 }, { "epoch": 0.4504546827287675, "grad_norm": 1.0408893823623657, "learning_rate": 9.033298306446115e-05, "loss": 0.615, "step": 6712 }, { "epoch": 0.4505889064125365, "grad_norm": 1.0709335803985596, "learning_rate": 9.032655876613636e-05, "loss": 0.6871, "step": 6714 }, { "epoch": 0.4507231300963055, "grad_norm": 1.0904982089996338, "learning_rate": 9.032013256246094e-05, "loss": 0.6562, "step": 6716 }, { "epoch": 0.4508573537800745, "grad_norm": 1.2430102825164795, "learning_rate": 9.031370445373856e-05, "loss": 0.6927, "step": 6718 }, { "epoch": 0.4509915774638435, "grad_norm": 1.2163711786270142, "learning_rate": 9.030727444027294e-05, "loss": 0.6461, "step": 6720 }, { "epoch": 0.45112580114761247, "grad_norm": 0.965525209903717, "learning_rate": 9.030084252236783e-05, "loss": 0.6482, "step": 6722 }, { "epoch": 0.4512600248313815, "grad_norm": 1.178641438484192, "learning_rate": 9.029440870032718e-05, "loss": 0.6777, "step": 6724 }, { "epoch": 0.4513942485151505, "grad_norm": 1.0191656351089478, "learning_rate": 9.028797297445495e-05, "loss": 0.5993, "step": 6726 }, { "epoch": 0.4515284721989195, "grad_norm": 1.122101902961731, "learning_rate": 9.028153534505522e-05, "loss": 0.6529, "step": 6728 }, { "epoch": 0.4516626958826885, "grad_norm": 1.0204761028289795, "learning_rate": 9.027509581243214e-05, "loss": 0.6585, "step": 6730 }, { "epoch": 0.4517969195664575, "grad_norm": 1.0416470766067505, "learning_rate": 9.026865437688998e-05, "loss": 0.669, "step": 6732 }, { "epoch": 0.4519311432502265, "grad_norm": 1.1773619651794434, "learning_rate": 9.026221103873312e-05, "loss": 0.6587, "step": 6734 }, { "epoch": 0.4520653669339955, "grad_norm": 1.0783255100250244, "learning_rate": 9.025576579826593e-05, "loss": 0.647, "step": 6736 }, { "epoch": 0.4521995906177645, "grad_norm": 1.2282097339630127, "learning_rate": 9.024931865579296e-05, "loss": 0.6124, "step": 6738 }, { "epoch": 0.4523338143015335, "grad_norm": 0.9424877762794495, "learning_rate": 9.024286961161885e-05, "loss": 0.6388, "step": 6740 }, { "epoch": 0.4524680379853025, "grad_norm": 1.1589195728302002, "learning_rate": 9.023641866604829e-05, "loss": 0.6858, "step": 6742 }, { "epoch": 0.4526022616690715, "grad_norm": 1.2010003328323364, "learning_rate": 9.022996581938605e-05, "loss": 0.6983, "step": 6744 }, { "epoch": 0.4527364853528405, "grad_norm": 0.9364185333251953, "learning_rate": 9.022351107193704e-05, "loss": 0.6288, "step": 6746 }, { "epoch": 0.45287070903660953, "grad_norm": 0.9412966370582581, "learning_rate": 9.021705442400623e-05, "loss": 0.6178, "step": 6748 }, { "epoch": 0.4530049327203785, "grad_norm": 1.013659119606018, "learning_rate": 9.021059587589869e-05, "loss": 0.6267, "step": 6750 }, { "epoch": 0.45313915640414754, "grad_norm": 1.0910239219665527, "learning_rate": 9.020413542791955e-05, "loss": 0.6773, "step": 6752 }, { "epoch": 0.4532733800879165, "grad_norm": 1.3612470626831055, "learning_rate": 9.019767308037407e-05, "loss": 0.5651, "step": 6754 }, { "epoch": 0.45340760377168554, "grad_norm": 1.2523924112319946, "learning_rate": 9.01912088335676e-05, "loss": 0.673, "step": 6756 }, { "epoch": 0.4535418274554545, "grad_norm": 1.0892006158828735, "learning_rate": 9.018474268780553e-05, "loss": 0.6062, "step": 6758 }, { "epoch": 0.4536760511392235, "grad_norm": 1.0809118747711182, "learning_rate": 9.017827464339338e-05, "loss": 0.6266, "step": 6760 }, { "epoch": 0.4538102748229925, "grad_norm": 1.1393630504608154, "learning_rate": 9.017180470063679e-05, "loss": 0.6653, "step": 6762 }, { "epoch": 0.4539444985067615, "grad_norm": 1.0781728029251099, "learning_rate": 9.01653328598414e-05, "loss": 0.6775, "step": 6764 }, { "epoch": 0.4540787221905305, "grad_norm": 1.149721622467041, "learning_rate": 9.015885912131302e-05, "loss": 0.7004, "step": 6766 }, { "epoch": 0.4542129458742995, "grad_norm": 1.207176923751831, "learning_rate": 9.015238348535751e-05, "loss": 0.714, "step": 6768 }, { "epoch": 0.45434716955806853, "grad_norm": 1.212417483329773, "learning_rate": 9.014590595228086e-05, "loss": 0.6742, "step": 6770 }, { "epoch": 0.4544813932418375, "grad_norm": 1.082265019416809, "learning_rate": 9.013942652238908e-05, "loss": 0.6742, "step": 6772 }, { "epoch": 0.45461561692560654, "grad_norm": 0.9447269439697266, "learning_rate": 9.01329451959883e-05, "loss": 0.6373, "step": 6774 }, { "epoch": 0.4547498406093755, "grad_norm": 0.9641191959381104, "learning_rate": 9.012646197338481e-05, "loss": 0.654, "step": 6776 }, { "epoch": 0.45488406429314454, "grad_norm": 0.9599052667617798, "learning_rate": 9.011997685488489e-05, "loss": 0.6235, "step": 6778 }, { "epoch": 0.4550182879769135, "grad_norm": 1.0892101526260376, "learning_rate": 9.011348984079496e-05, "loss": 0.6639, "step": 6780 }, { "epoch": 0.45515251166068255, "grad_norm": 1.0836642980575562, "learning_rate": 9.010700093142151e-05, "loss": 0.6323, "step": 6782 }, { "epoch": 0.4552867353444515, "grad_norm": 1.3505182266235352, "learning_rate": 9.010051012707114e-05, "loss": 0.6974, "step": 6784 }, { "epoch": 0.45542095902822055, "grad_norm": 0.9923969507217407, "learning_rate": 9.009401742805052e-05, "loss": 0.6549, "step": 6786 }, { "epoch": 0.4555551827119895, "grad_norm": 0.9913681149482727, "learning_rate": 9.008752283466641e-05, "loss": 0.635, "step": 6788 }, { "epoch": 0.45568940639575856, "grad_norm": 1.0151026248931885, "learning_rate": 9.008102634722568e-05, "loss": 0.615, "step": 6790 }, { "epoch": 0.45582363007952753, "grad_norm": 1.1850115060806274, "learning_rate": 9.007452796603526e-05, "loss": 0.5593, "step": 6792 }, { "epoch": 0.45595785376329656, "grad_norm": 1.1839693784713745, "learning_rate": 9.006802769140221e-05, "loss": 0.6833, "step": 6794 }, { "epoch": 0.45609207744706554, "grad_norm": 1.0648049116134644, "learning_rate": 9.006152552363363e-05, "loss": 0.7479, "step": 6796 }, { "epoch": 0.4562263011308345, "grad_norm": 1.0856976509094238, "learning_rate": 9.005502146303676e-05, "loss": 0.5972, "step": 6798 }, { "epoch": 0.45636052481460354, "grad_norm": 1.185408353805542, "learning_rate": 9.004851550991888e-05, "loss": 0.6914, "step": 6800 }, { "epoch": 0.4564947484983725, "grad_norm": 1.1428455114364624, "learning_rate": 9.004200766458742e-05, "loss": 0.6366, "step": 6802 }, { "epoch": 0.45662897218214155, "grad_norm": 1.0512735843658447, "learning_rate": 9.00354979273498e-05, "loss": 0.6416, "step": 6804 }, { "epoch": 0.4567631958659105, "grad_norm": 1.0457926988601685, "learning_rate": 9.002898629851364e-05, "loss": 0.6299, "step": 6806 }, { "epoch": 0.45689741954967955, "grad_norm": 0.984696090221405, "learning_rate": 9.00224727783866e-05, "loss": 0.6399, "step": 6808 }, { "epoch": 0.4570316432334485, "grad_norm": 1.4410641193389893, "learning_rate": 9.001595736727642e-05, "loss": 0.6988, "step": 6810 }, { "epoch": 0.45716586691721756, "grad_norm": 1.2012965679168701, "learning_rate": 9.000944006549095e-05, "loss": 0.7485, "step": 6812 }, { "epoch": 0.45730009060098653, "grad_norm": 1.048329472541809, "learning_rate": 9.00029208733381e-05, "loss": 0.6663, "step": 6814 }, { "epoch": 0.45743431428475556, "grad_norm": 1.1386462450027466, "learning_rate": 8.99963997911259e-05, "loss": 0.7471, "step": 6816 }, { "epoch": 0.45756853796852454, "grad_norm": 1.0403379201889038, "learning_rate": 8.998987681916246e-05, "loss": 0.628, "step": 6818 }, { "epoch": 0.45770276165229357, "grad_norm": 1.0118883848190308, "learning_rate": 8.998335195775599e-05, "loss": 0.6844, "step": 6820 }, { "epoch": 0.45783698533606254, "grad_norm": 1.1694881916046143, "learning_rate": 8.997682520721476e-05, "loss": 0.6405, "step": 6822 }, { "epoch": 0.45797120901983157, "grad_norm": 1.1148027181625366, "learning_rate": 8.997029656784715e-05, "loss": 0.653, "step": 6824 }, { "epoch": 0.45810543270360055, "grad_norm": 1.0107674598693848, "learning_rate": 8.996376603996161e-05, "loss": 0.6555, "step": 6826 }, { "epoch": 0.4582396563873696, "grad_norm": 0.9955770969390869, "learning_rate": 8.995723362386672e-05, "loss": 0.6387, "step": 6828 }, { "epoch": 0.45837388007113855, "grad_norm": 1.1577054262161255, "learning_rate": 8.995069931987113e-05, "loss": 0.6875, "step": 6830 }, { "epoch": 0.4585081037549076, "grad_norm": 1.1350963115692139, "learning_rate": 8.994416312828354e-05, "loss": 0.6236, "step": 6832 }, { "epoch": 0.45864232743867656, "grad_norm": 1.0730098485946655, "learning_rate": 8.993762504941277e-05, "loss": 0.7111, "step": 6834 }, { "epoch": 0.45877655112244553, "grad_norm": 1.1010090112686157, "learning_rate": 8.993108508356779e-05, "loss": 0.6476, "step": 6836 }, { "epoch": 0.45891077480621456, "grad_norm": 1.0718328952789307, "learning_rate": 8.992454323105752e-05, "loss": 0.7306, "step": 6838 }, { "epoch": 0.45904499848998354, "grad_norm": 1.111706018447876, "learning_rate": 8.991799949219112e-05, "loss": 0.624, "step": 6840 }, { "epoch": 0.45917922217375257, "grad_norm": 1.0027763843536377, "learning_rate": 8.991145386727773e-05, "loss": 0.6267, "step": 6842 }, { "epoch": 0.45931344585752154, "grad_norm": 1.1393334865570068, "learning_rate": 8.990490635662663e-05, "loss": 0.6908, "step": 6844 }, { "epoch": 0.45944766954129057, "grad_norm": 1.5427532196044922, "learning_rate": 8.989835696054718e-05, "loss": 0.7297, "step": 6846 }, { "epoch": 0.45958189322505955, "grad_norm": 1.078510046005249, "learning_rate": 8.989180567934881e-05, "loss": 0.7104, "step": 6848 }, { "epoch": 0.4597161169088286, "grad_norm": 1.0581449270248413, "learning_rate": 8.988525251334106e-05, "loss": 0.6171, "step": 6850 }, { "epoch": 0.45985034059259755, "grad_norm": 1.0170172452926636, "learning_rate": 8.987869746283358e-05, "loss": 0.643, "step": 6852 }, { "epoch": 0.4599845642763666, "grad_norm": 0.9322235584259033, "learning_rate": 8.987214052813604e-05, "loss": 0.6311, "step": 6854 }, { "epoch": 0.46011878796013556, "grad_norm": 1.0942978858947754, "learning_rate": 8.986558170955828e-05, "loss": 0.6775, "step": 6856 }, { "epoch": 0.4602530116439046, "grad_norm": 1.0409022569656372, "learning_rate": 8.985902100741018e-05, "loss": 0.5689, "step": 6858 }, { "epoch": 0.46038723532767356, "grad_norm": 1.0800971984863281, "learning_rate": 8.98524584220017e-05, "loss": 0.69, "step": 6860 }, { "epoch": 0.4605214590114426, "grad_norm": 1.049082636833191, "learning_rate": 8.984589395364294e-05, "loss": 0.649, "step": 6862 }, { "epoch": 0.46065568269521157, "grad_norm": 1.1251814365386963, "learning_rate": 8.983932760264405e-05, "loss": 0.6719, "step": 6864 }, { "epoch": 0.4607899063789806, "grad_norm": 1.1507809162139893, "learning_rate": 8.983275936931526e-05, "loss": 0.727, "step": 6866 }, { "epoch": 0.46092413006274957, "grad_norm": 0.9597031474113464, "learning_rate": 8.982618925396691e-05, "loss": 0.63, "step": 6868 }, { "epoch": 0.46105835374651855, "grad_norm": 1.1732813119888306, "learning_rate": 8.981961725690943e-05, "loss": 0.6308, "step": 6870 }, { "epoch": 0.4611925774302876, "grad_norm": 1.0450263023376465, "learning_rate": 8.981304337845337e-05, "loss": 0.5938, "step": 6872 }, { "epoch": 0.46132680111405655, "grad_norm": 1.109165906906128, "learning_rate": 8.980646761890928e-05, "loss": 0.6669, "step": 6874 }, { "epoch": 0.4614610247978256, "grad_norm": 0.9738684296607971, "learning_rate": 8.979988997858785e-05, "loss": 0.6748, "step": 6876 }, { "epoch": 0.46159524848159456, "grad_norm": 1.1553380489349365, "learning_rate": 8.97933104577999e-05, "loss": 0.6687, "step": 6878 }, { "epoch": 0.4617294721653636, "grad_norm": 1.0648709535598755, "learning_rate": 8.978672905685629e-05, "loss": 0.6825, "step": 6880 }, { "epoch": 0.46186369584913256, "grad_norm": 1.3166885375976562, "learning_rate": 8.978014577606797e-05, "loss": 0.6791, "step": 6882 }, { "epoch": 0.4619979195329016, "grad_norm": 0.9989602565765381, "learning_rate": 8.977356061574597e-05, "loss": 0.6538, "step": 6884 }, { "epoch": 0.46213214321667057, "grad_norm": 1.5037658214569092, "learning_rate": 8.976697357620145e-05, "loss": 0.6675, "step": 6886 }, { "epoch": 0.4622663669004396, "grad_norm": 1.0083638429641724, "learning_rate": 8.976038465774563e-05, "loss": 0.5991, "step": 6888 }, { "epoch": 0.46240059058420857, "grad_norm": 1.194140911102295, "learning_rate": 8.975379386068981e-05, "loss": 0.7116, "step": 6890 }, { "epoch": 0.4625348142679776, "grad_norm": 1.4568372964859009, "learning_rate": 8.974720118534541e-05, "loss": 0.6345, "step": 6892 }, { "epoch": 0.4626690379517466, "grad_norm": 1.069860577583313, "learning_rate": 8.974060663202392e-05, "loss": 0.7056, "step": 6894 }, { "epoch": 0.4628032616355156, "grad_norm": 1.410301923751831, "learning_rate": 8.97340102010369e-05, "loss": 0.6908, "step": 6896 }, { "epoch": 0.4629374853192846, "grad_norm": 1.1220239400863647, "learning_rate": 8.972741189269605e-05, "loss": 0.6735, "step": 6898 }, { "epoch": 0.4630717090030536, "grad_norm": 2.3627586364746094, "learning_rate": 8.972081170731307e-05, "loss": 0.6223, "step": 6900 }, { "epoch": 0.4632059326868226, "grad_norm": 1.1087870597839355, "learning_rate": 8.971420964519988e-05, "loss": 0.6791, "step": 6902 }, { "epoch": 0.4633401563705916, "grad_norm": 1.1223286390304565, "learning_rate": 8.970760570666839e-05, "loss": 0.7227, "step": 6904 }, { "epoch": 0.4634743800543606, "grad_norm": 0.9709183573722839, "learning_rate": 8.970099989203058e-05, "loss": 0.6467, "step": 6906 }, { "epoch": 0.46360860373812957, "grad_norm": 1.0744274854660034, "learning_rate": 8.96943922015986e-05, "loss": 0.6976, "step": 6908 }, { "epoch": 0.4637428274218986, "grad_norm": 1.0004558563232422, "learning_rate": 8.968778263568465e-05, "loss": 0.613, "step": 6910 }, { "epoch": 0.46387705110566757, "grad_norm": 1.0825237035751343, "learning_rate": 8.968117119460103e-05, "loss": 0.7359, "step": 6912 }, { "epoch": 0.4640112747894366, "grad_norm": 1.074562668800354, "learning_rate": 8.967455787866007e-05, "loss": 0.6747, "step": 6914 }, { "epoch": 0.4641454984732056, "grad_norm": 1.0532876253128052, "learning_rate": 8.966794268817427e-05, "loss": 0.6445, "step": 6916 }, { "epoch": 0.4642797221569746, "grad_norm": 1.5095680952072144, "learning_rate": 8.96613256234562e-05, "loss": 0.7003, "step": 6918 }, { "epoch": 0.4644139458407436, "grad_norm": 2.5291311740875244, "learning_rate": 8.965470668481848e-05, "loss": 0.6705, "step": 6920 }, { "epoch": 0.4645481695245126, "grad_norm": 1.0325822830200195, "learning_rate": 8.964808587257386e-05, "loss": 0.651, "step": 6922 }, { "epoch": 0.4646823932082816, "grad_norm": 1.061231255531311, "learning_rate": 8.964146318703512e-05, "loss": 0.5798, "step": 6924 }, { "epoch": 0.4648166168920506, "grad_norm": 1.285365104675293, "learning_rate": 8.96348386285152e-05, "loss": 0.7, "step": 6926 }, { "epoch": 0.4649508405758196, "grad_norm": 1.1364279985427856, "learning_rate": 8.962821219732711e-05, "loss": 0.656, "step": 6928 }, { "epoch": 0.4650850642595886, "grad_norm": 1.1411540508270264, "learning_rate": 8.96215838937839e-05, "loss": 0.6316, "step": 6930 }, { "epoch": 0.4652192879433576, "grad_norm": 1.1594171524047852, "learning_rate": 8.961495371819877e-05, "loss": 0.723, "step": 6932 }, { "epoch": 0.4653535116271266, "grad_norm": 1.0198676586151123, "learning_rate": 8.960832167088498e-05, "loss": 0.6404, "step": 6934 }, { "epoch": 0.4654877353108956, "grad_norm": 0.9912198781967163, "learning_rate": 8.960168775215588e-05, "loss": 0.6453, "step": 6936 }, { "epoch": 0.46562195899466463, "grad_norm": 1.1035021543502808, "learning_rate": 8.95950519623249e-05, "loss": 0.6325, "step": 6938 }, { "epoch": 0.4657561826784336, "grad_norm": 0.9891259074211121, "learning_rate": 8.95884143017056e-05, "loss": 0.6341, "step": 6940 }, { "epoch": 0.46589040636220264, "grad_norm": 0.9204497337341309, "learning_rate": 8.958177477061154e-05, "loss": 0.6293, "step": 6942 }, { "epoch": 0.4660246300459716, "grad_norm": 1.120475172996521, "learning_rate": 8.957513336935646e-05, "loss": 0.6751, "step": 6944 }, { "epoch": 0.4661588537297406, "grad_norm": 1.1103843450546265, "learning_rate": 8.956849009825417e-05, "loss": 0.6358, "step": 6946 }, { "epoch": 0.4662930774135096, "grad_norm": 1.2845054864883423, "learning_rate": 8.95618449576185e-05, "loss": 0.7357, "step": 6948 }, { "epoch": 0.4664273010972786, "grad_norm": 1.0931564569473267, "learning_rate": 8.955519794776348e-05, "loss": 0.6829, "step": 6950 }, { "epoch": 0.4665615247810476, "grad_norm": 1.0615421533584595, "learning_rate": 8.954854906900312e-05, "loss": 0.6505, "step": 6952 }, { "epoch": 0.4666957484648166, "grad_norm": 1.0076024532318115, "learning_rate": 8.954189832165159e-05, "loss": 0.6809, "step": 6954 }, { "epoch": 0.4668299721485856, "grad_norm": 1.1049742698669434, "learning_rate": 8.953524570602313e-05, "loss": 0.7196, "step": 6956 }, { "epoch": 0.4669641958323546, "grad_norm": 1.5184391736984253, "learning_rate": 8.952859122243204e-05, "loss": 0.7042, "step": 6958 }, { "epoch": 0.46709841951612363, "grad_norm": 1.5310966968536377, "learning_rate": 8.952193487119276e-05, "loss": 0.6192, "step": 6960 }, { "epoch": 0.4672326431998926, "grad_norm": 1.4648321866989136, "learning_rate": 8.951527665261976e-05, "loss": 0.6901, "step": 6962 }, { "epoch": 0.46736686688366164, "grad_norm": 1.162032961845398, "learning_rate": 8.950861656702764e-05, "loss": 0.7611, "step": 6964 }, { "epoch": 0.4675010905674306, "grad_norm": 1.1246246099472046, "learning_rate": 8.950195461473109e-05, "loss": 0.6745, "step": 6966 }, { "epoch": 0.46763531425119964, "grad_norm": 1.2168980836868286, "learning_rate": 8.949529079604485e-05, "loss": 0.6573, "step": 6968 }, { "epoch": 0.4677695379349686, "grad_norm": 1.2611074447631836, "learning_rate": 8.94886251112838e-05, "loss": 0.6753, "step": 6970 }, { "epoch": 0.46790376161873765, "grad_norm": 1.0567899942398071, "learning_rate": 8.948195756076285e-05, "loss": 0.7266, "step": 6972 }, { "epoch": 0.4680379853025066, "grad_norm": 1.4233819246292114, "learning_rate": 8.947528814479704e-05, "loss": 0.6134, "step": 6974 }, { "epoch": 0.46817220898627565, "grad_norm": 1.2391204833984375, "learning_rate": 8.94686168637015e-05, "loss": 0.663, "step": 6976 }, { "epoch": 0.4683064326700446, "grad_norm": 1.310097575187683, "learning_rate": 8.946194371779142e-05, "loss": 0.6546, "step": 6978 }, { "epoch": 0.46844065635381366, "grad_norm": 1.034376621246338, "learning_rate": 8.94552687073821e-05, "loss": 0.6731, "step": 6980 }, { "epoch": 0.46857488003758263, "grad_norm": 1.1510058641433716, "learning_rate": 8.944859183278891e-05, "loss": 0.6442, "step": 6982 }, { "epoch": 0.4687091037213516, "grad_norm": 1.1591176986694336, "learning_rate": 8.944191309432735e-05, "loss": 0.6933, "step": 6984 }, { "epoch": 0.46884332740512064, "grad_norm": 1.083055853843689, "learning_rate": 8.943523249231293e-05, "loss": 0.6641, "step": 6986 }, { "epoch": 0.4689775510888896, "grad_norm": 1.4717953205108643, "learning_rate": 8.942855002706134e-05, "loss": 0.6953, "step": 6988 }, { "epoch": 0.46911177477265864, "grad_norm": 1.2020087242126465, "learning_rate": 8.942186569888829e-05, "loss": 0.6102, "step": 6990 }, { "epoch": 0.4692459984564276, "grad_norm": 1.069995403289795, "learning_rate": 8.94151795081096e-05, "loss": 0.6819, "step": 6992 }, { "epoch": 0.46938022214019665, "grad_norm": 1.1206103563308716, "learning_rate": 8.940849145504118e-05, "loss": 0.683, "step": 6994 }, { "epoch": 0.4695144458239656, "grad_norm": 1.803853988647461, "learning_rate": 8.940180153999904e-05, "loss": 0.6418, "step": 6996 }, { "epoch": 0.46964866950773465, "grad_norm": 2.395535707473755, "learning_rate": 8.939510976329927e-05, "loss": 0.6804, "step": 6998 }, { "epoch": 0.4697828931915036, "grad_norm": 1.123642921447754, "learning_rate": 8.938841612525801e-05, "loss": 0.6524, "step": 7000 }, { "epoch": 0.46991711687527266, "grad_norm": 1.076267123222351, "learning_rate": 8.938172062619155e-05, "loss": 0.6429, "step": 7002 }, { "epoch": 0.47005134055904163, "grad_norm": 1.2277798652648926, "learning_rate": 8.937502326641622e-05, "loss": 0.7775, "step": 7004 }, { "epoch": 0.47018556424281066, "grad_norm": 0.8746130466461182, "learning_rate": 8.936832404624848e-05, "loss": 0.6474, "step": 7006 }, { "epoch": 0.47031978792657964, "grad_norm": 1.0472476482391357, "learning_rate": 8.936162296600486e-05, "loss": 0.6987, "step": 7008 }, { "epoch": 0.47045401161034867, "grad_norm": 1.4213472604751587, "learning_rate": 8.935492002600194e-05, "loss": 0.5801, "step": 7010 }, { "epoch": 0.47058823529411764, "grad_norm": 1.0555846691131592, "learning_rate": 8.934821522655642e-05, "loss": 0.7152, "step": 7012 }, { "epoch": 0.47072245897788667, "grad_norm": 1.1034077405929565, "learning_rate": 8.934150856798514e-05, "loss": 0.6107, "step": 7014 }, { "epoch": 0.47085668266165565, "grad_norm": 1.109114408493042, "learning_rate": 8.933480005060492e-05, "loss": 0.7219, "step": 7016 }, { "epoch": 0.4709909063454247, "grad_norm": 1.1220183372497559, "learning_rate": 8.932808967473274e-05, "loss": 0.641, "step": 7018 }, { "epoch": 0.47112513002919365, "grad_norm": 0.983274519443512, "learning_rate": 8.932137744068567e-05, "loss": 0.604, "step": 7020 }, { "epoch": 0.4712593537129626, "grad_norm": 1.0105772018432617, "learning_rate": 8.931466334878085e-05, "loss": 0.6437, "step": 7022 }, { "epoch": 0.47139357739673166, "grad_norm": 1.0627435445785522, "learning_rate": 8.930794739933547e-05, "loss": 0.633, "step": 7024 }, { "epoch": 0.47152780108050063, "grad_norm": 1.4478068351745605, "learning_rate": 8.930122959266689e-05, "loss": 0.6532, "step": 7026 }, { "epoch": 0.47166202476426966, "grad_norm": 0.9602782130241394, "learning_rate": 8.929450992909248e-05, "loss": 0.7098, "step": 7028 }, { "epoch": 0.47179624844803864, "grad_norm": 1.1455880403518677, "learning_rate": 8.928778840892975e-05, "loss": 0.6132, "step": 7030 }, { "epoch": 0.47193047213180767, "grad_norm": 1.0074069499969482, "learning_rate": 8.928106503249628e-05, "loss": 0.6814, "step": 7032 }, { "epoch": 0.47206469581557664, "grad_norm": 1.0082672834396362, "learning_rate": 8.927433980010973e-05, "loss": 0.5948, "step": 7034 }, { "epoch": 0.47219891949934567, "grad_norm": 1.070309042930603, "learning_rate": 8.926761271208785e-05, "loss": 0.7057, "step": 7036 }, { "epoch": 0.47233314318311465, "grad_norm": 1.1839375495910645, "learning_rate": 8.926088376874849e-05, "loss": 0.6744, "step": 7038 }, { "epoch": 0.4724673668668837, "grad_norm": 1.0738030672073364, "learning_rate": 8.925415297040957e-05, "loss": 0.6774, "step": 7040 }, { "epoch": 0.47260159055065265, "grad_norm": 1.1747937202453613, "learning_rate": 8.924742031738911e-05, "loss": 0.5994, "step": 7042 }, { "epoch": 0.4727358142344217, "grad_norm": 0.9965662360191345, "learning_rate": 8.924068581000521e-05, "loss": 0.6201, "step": 7044 }, { "epoch": 0.47287003791819066, "grad_norm": 1.0046024322509766, "learning_rate": 8.923394944857609e-05, "loss": 0.6126, "step": 7046 }, { "epoch": 0.4730042616019597, "grad_norm": 0.9661778807640076, "learning_rate": 8.922721123341999e-05, "loss": 0.6749, "step": 7048 }, { "epoch": 0.47313848528572866, "grad_norm": 1.0183511972427368, "learning_rate": 8.922047116485532e-05, "loss": 0.6765, "step": 7050 }, { "epoch": 0.4732727089694977, "grad_norm": 0.9786390662193298, "learning_rate": 8.921372924320048e-05, "loss": 0.6522, "step": 7052 }, { "epoch": 0.47340693265326667, "grad_norm": 1.198760747909546, "learning_rate": 8.920698546877406e-05, "loss": 0.6494, "step": 7054 }, { "epoch": 0.4735411563370357, "grad_norm": 0.9723764061927795, "learning_rate": 8.920023984189468e-05, "loss": 0.6539, "step": 7056 }, { "epoch": 0.47367538002080467, "grad_norm": 1.0999858379364014, "learning_rate": 8.919349236288105e-05, "loss": 0.6382, "step": 7058 }, { "epoch": 0.47380960370457365, "grad_norm": 1.075578212738037, "learning_rate": 8.918674303205197e-05, "loss": 0.7138, "step": 7060 }, { "epoch": 0.4739438273883427, "grad_norm": 0.970807671546936, "learning_rate": 8.917999184972634e-05, "loss": 0.654, "step": 7062 }, { "epoch": 0.47407805107211165, "grad_norm": 1.1560735702514648, "learning_rate": 8.917323881622314e-05, "loss": 0.6951, "step": 7064 }, { "epoch": 0.4742122747558807, "grad_norm": 1.1930887699127197, "learning_rate": 8.916648393186143e-05, "loss": 0.5947, "step": 7066 }, { "epoch": 0.47434649843964966, "grad_norm": 1.014479637145996, "learning_rate": 8.915972719696037e-05, "loss": 0.6198, "step": 7068 }, { "epoch": 0.4744807221234187, "grad_norm": 0.9746931791305542, "learning_rate": 8.915296861183923e-05, "loss": 0.7069, "step": 7070 }, { "epoch": 0.47461494580718766, "grad_norm": 1.0965828895568848, "learning_rate": 8.914620817681729e-05, "loss": 0.6967, "step": 7072 }, { "epoch": 0.4747491694909567, "grad_norm": 1.3043853044509888, "learning_rate": 8.9139445892214e-05, "loss": 0.6357, "step": 7074 }, { "epoch": 0.47488339317472567, "grad_norm": 1.6396334171295166, "learning_rate": 8.913268175834886e-05, "loss": 0.7468, "step": 7076 }, { "epoch": 0.4750176168584947, "grad_norm": 1.0765565633773804, "learning_rate": 8.912591577554143e-05, "loss": 0.6436, "step": 7078 }, { "epoch": 0.47515184054226367, "grad_norm": 1.0507439374923706, "learning_rate": 8.911914794411144e-05, "loss": 0.6137, "step": 7080 }, { "epoch": 0.4752860642260327, "grad_norm": 1.0619193315505981, "learning_rate": 8.911237826437865e-05, "loss": 0.6245, "step": 7082 }, { "epoch": 0.4754202879098017, "grad_norm": 1.102828025817871, "learning_rate": 8.910560673666289e-05, "loss": 0.7056, "step": 7084 }, { "epoch": 0.4755545115935707, "grad_norm": 1.1942352056503296, "learning_rate": 8.909883336128408e-05, "loss": 0.6584, "step": 7086 }, { "epoch": 0.4756887352773397, "grad_norm": 0.9986782073974609, "learning_rate": 8.909205813856232e-05, "loss": 0.6168, "step": 7088 }, { "epoch": 0.4758229589611087, "grad_norm": 0.9641457200050354, "learning_rate": 8.908528106881765e-05, "loss": 0.6498, "step": 7090 }, { "epoch": 0.4759571826448777, "grad_norm": 1.2766104936599731, "learning_rate": 8.907850215237032e-05, "loss": 0.7591, "step": 7092 }, { "epoch": 0.4760914063286467, "grad_norm": 0.9906812310218811, "learning_rate": 8.907172138954061e-05, "loss": 0.6064, "step": 7094 }, { "epoch": 0.4762256300124157, "grad_norm": 1.2280668020248413, "learning_rate": 8.90649387806489e-05, "loss": 0.6616, "step": 7096 }, { "epoch": 0.47635985369618467, "grad_norm": 0.9963435530662537, "learning_rate": 8.905815432601566e-05, "loss": 0.6256, "step": 7098 }, { "epoch": 0.4764940773799537, "grad_norm": 1.073029637336731, "learning_rate": 8.905136802596142e-05, "loss": 0.624, "step": 7100 }, { "epoch": 0.47662830106372267, "grad_norm": 0.966012716293335, "learning_rate": 8.904457988080681e-05, "loss": 0.651, "step": 7102 }, { "epoch": 0.4767625247474917, "grad_norm": 1.126132607460022, "learning_rate": 8.90377898908726e-05, "loss": 0.6854, "step": 7104 }, { "epoch": 0.4768967484312607, "grad_norm": 1.112329125404358, "learning_rate": 8.903099805647959e-05, "loss": 0.6875, "step": 7106 }, { "epoch": 0.4770309721150297, "grad_norm": 0.9800264239311218, "learning_rate": 8.902420437794865e-05, "loss": 0.6274, "step": 7108 }, { "epoch": 0.4771651957987987, "grad_norm": 1.270875096321106, "learning_rate": 8.901740885560082e-05, "loss": 0.6694, "step": 7110 }, { "epoch": 0.4772994194825677, "grad_norm": 1.0601409673690796, "learning_rate": 8.901061148975711e-05, "loss": 0.6756, "step": 7112 }, { "epoch": 0.4774336431663367, "grad_norm": 1.0308822393417358, "learning_rate": 8.900381228073875e-05, "loss": 0.6611, "step": 7114 }, { "epoch": 0.4775678668501057, "grad_norm": 1.0516387224197388, "learning_rate": 8.899701122886695e-05, "loss": 0.6754, "step": 7116 }, { "epoch": 0.4777020905338747, "grad_norm": 1.080004096031189, "learning_rate": 8.899020833446304e-05, "loss": 0.6428, "step": 7118 }, { "epoch": 0.4778363142176437, "grad_norm": 1.009861707687378, "learning_rate": 8.898340359784847e-05, "loss": 0.5901, "step": 7120 }, { "epoch": 0.4779705379014127, "grad_norm": 1.106646180152893, "learning_rate": 8.897659701934474e-05, "loss": 0.6663, "step": 7122 }, { "epoch": 0.4781047615851817, "grad_norm": 1.251336693763733, "learning_rate": 8.896978859927343e-05, "loss": 0.6562, "step": 7124 }, { "epoch": 0.4782389852689507, "grad_norm": 1.4340730905532837, "learning_rate": 8.896297833795625e-05, "loss": 0.5836, "step": 7126 }, { "epoch": 0.47837320895271973, "grad_norm": 1.22294282913208, "learning_rate": 8.895616623571497e-05, "loss": 0.6578, "step": 7128 }, { "epoch": 0.4785074326364887, "grad_norm": 1.2244223356246948, "learning_rate": 8.894935229287142e-05, "loss": 0.6836, "step": 7130 }, { "epoch": 0.47864165632025774, "grad_norm": 0.978756844997406, "learning_rate": 8.894253650974757e-05, "loss": 0.6477, "step": 7132 }, { "epoch": 0.4787758800040267, "grad_norm": 1.2327920198440552, "learning_rate": 8.893571888666545e-05, "loss": 0.6362, "step": 7134 }, { "epoch": 0.4789101036877957, "grad_norm": 0.9988105893135071, "learning_rate": 8.892889942394719e-05, "loss": 0.6637, "step": 7136 }, { "epoch": 0.4790443273715647, "grad_norm": 0.9424437880516052, "learning_rate": 8.892207812191497e-05, "loss": 0.5791, "step": 7138 }, { "epoch": 0.4791785510553337, "grad_norm": 1.1618750095367432, "learning_rate": 8.89152549808911e-05, "loss": 0.6781, "step": 7140 }, { "epoch": 0.4793127747391027, "grad_norm": 1.0431857109069824, "learning_rate": 8.890843000119795e-05, "loss": 0.6332, "step": 7142 }, { "epoch": 0.4794469984228717, "grad_norm": 1.1647588014602661, "learning_rate": 8.890160318315798e-05, "loss": 0.7326, "step": 7144 }, { "epoch": 0.4795812221066407, "grad_norm": 1.0543895959854126, "learning_rate": 8.889477452709378e-05, "loss": 0.6853, "step": 7146 }, { "epoch": 0.4797154457904097, "grad_norm": 0.9098528027534485, "learning_rate": 8.888794403332797e-05, "loss": 0.5688, "step": 7148 }, { "epoch": 0.47984966947417873, "grad_norm": 1.0702451467514038, "learning_rate": 8.888111170218325e-05, "loss": 0.6047, "step": 7150 }, { "epoch": 0.4799838931579477, "grad_norm": 1.1213874816894531, "learning_rate": 8.887427753398248e-05, "loss": 0.6557, "step": 7152 }, { "epoch": 0.48011811684171674, "grad_norm": 1.6079059839248657, "learning_rate": 8.886744152904851e-05, "loss": 0.6407, "step": 7154 }, { "epoch": 0.4802523405254857, "grad_norm": 1.0884034633636475, "learning_rate": 8.886060368770439e-05, "loss": 0.6665, "step": 7156 }, { "epoch": 0.48038656420925474, "grad_norm": 1.1436165571212769, "learning_rate": 8.885376401027315e-05, "loss": 0.7486, "step": 7158 }, { "epoch": 0.4805207878930237, "grad_norm": 0.9711717963218689, "learning_rate": 8.884692249707795e-05, "loss": 0.6311, "step": 7160 }, { "epoch": 0.48065501157679275, "grad_norm": 1.055991768836975, "learning_rate": 8.884007914844208e-05, "loss": 0.6775, "step": 7162 }, { "epoch": 0.4807892352605617, "grad_norm": 0.982055127620697, "learning_rate": 8.883323396468882e-05, "loss": 0.5728, "step": 7164 }, { "epoch": 0.48092345894433075, "grad_norm": 1.1967780590057373, "learning_rate": 8.882638694614163e-05, "loss": 0.6189, "step": 7166 }, { "epoch": 0.4810576826280997, "grad_norm": 2.022672176361084, "learning_rate": 8.8819538093124e-05, "loss": 0.6913, "step": 7168 }, { "epoch": 0.48119190631186876, "grad_norm": 1.0841823816299438, "learning_rate": 8.881268740595954e-05, "loss": 0.6773, "step": 7170 }, { "epoch": 0.48132612999563773, "grad_norm": 1.0845587253570557, "learning_rate": 8.880583488497192e-05, "loss": 0.658, "step": 7172 }, { "epoch": 0.4814603536794067, "grad_norm": 1.6183933019638062, "learning_rate": 8.87989805304849e-05, "loss": 0.5913, "step": 7174 }, { "epoch": 0.48159457736317574, "grad_norm": 1.0903329849243164, "learning_rate": 8.879212434282235e-05, "loss": 0.6679, "step": 7176 }, { "epoch": 0.4817288010469447, "grad_norm": 1.0819309949874878, "learning_rate": 8.878526632230819e-05, "loss": 0.6312, "step": 7178 }, { "epoch": 0.48186302473071374, "grad_norm": 1.1857296228408813, "learning_rate": 8.87784064692665e-05, "loss": 0.6358, "step": 7180 }, { "epoch": 0.4819972484144827, "grad_norm": 1.0356576442718506, "learning_rate": 8.877154478402131e-05, "loss": 0.6143, "step": 7182 }, { "epoch": 0.48213147209825175, "grad_norm": 1.0137851238250732, "learning_rate": 8.876468126689692e-05, "loss": 0.7352, "step": 7184 }, { "epoch": 0.4822656957820207, "grad_norm": 1.004191279411316, "learning_rate": 8.875781591821754e-05, "loss": 0.6567, "step": 7186 }, { "epoch": 0.48239991946578975, "grad_norm": 1.1644058227539062, "learning_rate": 8.875094873830758e-05, "loss": 0.6588, "step": 7188 }, { "epoch": 0.4825341431495587, "grad_norm": 1.0665473937988281, "learning_rate": 8.87440797274915e-05, "loss": 0.6818, "step": 7190 }, { "epoch": 0.48266836683332776, "grad_norm": 1.214895248413086, "learning_rate": 8.873720888609382e-05, "loss": 0.6378, "step": 7192 }, { "epoch": 0.48280259051709673, "grad_norm": 1.0585546493530273, "learning_rate": 8.873033621443921e-05, "loss": 0.6605, "step": 7194 }, { "epoch": 0.48293681420086576, "grad_norm": 1.1378669738769531, "learning_rate": 8.872346171285237e-05, "loss": 0.6071, "step": 7196 }, { "epoch": 0.48307103788463474, "grad_norm": 1.0823408365249634, "learning_rate": 8.871658538165811e-05, "loss": 0.661, "step": 7198 }, { "epoch": 0.48320526156840377, "grad_norm": 1.0443452596664429, "learning_rate": 8.870970722118132e-05, "loss": 0.6439, "step": 7200 }, { "epoch": 0.48333948525217274, "grad_norm": 1.0604746341705322, "learning_rate": 8.870282723174699e-05, "loss": 0.5891, "step": 7202 }, { "epoch": 0.48347370893594177, "grad_norm": 1.1016902923583984, "learning_rate": 8.869594541368017e-05, "loss": 0.6408, "step": 7204 }, { "epoch": 0.48360793261971075, "grad_norm": 1.0672671794891357, "learning_rate": 8.868906176730602e-05, "loss": 0.7246, "step": 7206 }, { "epoch": 0.4837421563034798, "grad_norm": 1.0549395084381104, "learning_rate": 8.868217629294979e-05, "loss": 0.6581, "step": 7208 }, { "epoch": 0.48387637998724875, "grad_norm": 1.0876435041427612, "learning_rate": 8.86752889909368e-05, "loss": 0.6263, "step": 7210 }, { "epoch": 0.4840106036710177, "grad_norm": 0.9596108794212341, "learning_rate": 8.866839986159244e-05, "loss": 0.5984, "step": 7212 }, { "epoch": 0.48414482735478676, "grad_norm": 2.628302574157715, "learning_rate": 8.866150890524224e-05, "loss": 0.6294, "step": 7214 }, { "epoch": 0.48427905103855573, "grad_norm": 1.0163933038711548, "learning_rate": 8.865461612221176e-05, "loss": 0.6496, "step": 7216 }, { "epoch": 0.48441327472232476, "grad_norm": 1.1279231309890747, "learning_rate": 8.864772151282668e-05, "loss": 0.6772, "step": 7218 }, { "epoch": 0.48454749840609374, "grad_norm": 1.2656511068344116, "learning_rate": 8.864082507741276e-05, "loss": 0.6073, "step": 7220 }, { "epoch": 0.48468172208986277, "grad_norm": 0.9335100650787354, "learning_rate": 8.863392681629583e-05, "loss": 0.5735, "step": 7222 }, { "epoch": 0.48481594577363174, "grad_norm": 1.1231383085250854, "learning_rate": 8.862702672980185e-05, "loss": 0.6264, "step": 7224 }, { "epoch": 0.48495016945740077, "grad_norm": 1.1034128665924072, "learning_rate": 8.862012481825679e-05, "loss": 0.5951, "step": 7226 }, { "epoch": 0.48508439314116975, "grad_norm": 1.06645929813385, "learning_rate": 8.861322108198678e-05, "loss": 0.6167, "step": 7228 }, { "epoch": 0.4852186168249388, "grad_norm": 1.1769659519195557, "learning_rate": 8.860631552131801e-05, "loss": 0.6778, "step": 7230 }, { "epoch": 0.48535284050870775, "grad_norm": 1.1352919340133667, "learning_rate": 8.859940813657675e-05, "loss": 0.638, "step": 7232 }, { "epoch": 0.4854870641924768, "grad_norm": 1.0001920461654663, "learning_rate": 8.859249892808935e-05, "loss": 0.6859, "step": 7234 }, { "epoch": 0.48562128787624576, "grad_norm": 1.0392810106277466, "learning_rate": 8.858558789618228e-05, "loss": 0.6792, "step": 7236 }, { "epoch": 0.4857555115600148, "grad_norm": 1.0945895910263062, "learning_rate": 8.857867504118204e-05, "loss": 0.6026, "step": 7238 }, { "epoch": 0.48588973524378376, "grad_norm": 0.982492208480835, "learning_rate": 8.857176036341526e-05, "loss": 0.619, "step": 7240 }, { "epoch": 0.4860239589275528, "grad_norm": 1.1728482246398926, "learning_rate": 8.856484386320867e-05, "loss": 0.6795, "step": 7242 }, { "epoch": 0.48615818261132177, "grad_norm": 1.0668323040008545, "learning_rate": 8.855792554088903e-05, "loss": 0.61, "step": 7244 }, { "epoch": 0.4862924062950908, "grad_norm": 1.027684211730957, "learning_rate": 8.855100539678324e-05, "loss": 0.6812, "step": 7246 }, { "epoch": 0.48642662997885977, "grad_norm": 1.0103657245635986, "learning_rate": 8.854408343121824e-05, "loss": 0.6584, "step": 7248 }, { "epoch": 0.48656085366262875, "grad_norm": 1.0728471279144287, "learning_rate": 8.85371596445211e-05, "loss": 0.725, "step": 7250 }, { "epoch": 0.4866950773463978, "grad_norm": 1.1303855180740356, "learning_rate": 8.853023403701894e-05, "loss": 0.7265, "step": 7252 }, { "epoch": 0.48682930103016675, "grad_norm": 0.9891423583030701, "learning_rate": 8.852330660903899e-05, "loss": 0.6167, "step": 7254 }, { "epoch": 0.4869635247139358, "grad_norm": 1.1069655418395996, "learning_rate": 8.851637736090857e-05, "loss": 0.6709, "step": 7256 }, { "epoch": 0.48709774839770476, "grad_norm": 1.0861634016036987, "learning_rate": 8.850944629295503e-05, "loss": 0.6759, "step": 7258 }, { "epoch": 0.4872319720814738, "grad_norm": 1.3577293157577515, "learning_rate": 8.850251340550591e-05, "loss": 0.6849, "step": 7260 }, { "epoch": 0.48736619576524276, "grad_norm": 1.1354912519454956, "learning_rate": 8.849557869888872e-05, "loss": 0.6488, "step": 7262 }, { "epoch": 0.4875004194490118, "grad_norm": 1.0820255279541016, "learning_rate": 8.848864217343114e-05, "loss": 0.5862, "step": 7264 }, { "epoch": 0.48763464313278077, "grad_norm": 0.9930956959724426, "learning_rate": 8.848170382946091e-05, "loss": 0.6174, "step": 7266 }, { "epoch": 0.4877688668165498, "grad_norm": 1.0091296434402466, "learning_rate": 8.847476366730584e-05, "loss": 0.6347, "step": 7268 }, { "epoch": 0.48790309050031877, "grad_norm": 0.9442898631095886, "learning_rate": 8.846782168729384e-05, "loss": 0.5684, "step": 7270 }, { "epoch": 0.4880373141840878, "grad_norm": 1.0996882915496826, "learning_rate": 8.846087788975292e-05, "loss": 0.7144, "step": 7272 }, { "epoch": 0.4881715378678568, "grad_norm": 1.119537591934204, "learning_rate": 8.845393227501114e-05, "loss": 0.6634, "step": 7274 }, { "epoch": 0.4883057615516258, "grad_norm": 1.0286153554916382, "learning_rate": 8.844698484339668e-05, "loss": 0.6733, "step": 7276 }, { "epoch": 0.4884399852353948, "grad_norm": 1.1285734176635742, "learning_rate": 8.844003559523779e-05, "loss": 0.6943, "step": 7278 }, { "epoch": 0.4885742089191638, "grad_norm": 1.0654995441436768, "learning_rate": 8.84330845308628e-05, "loss": 0.6218, "step": 7280 }, { "epoch": 0.4887084326029328, "grad_norm": 1.1435794830322266, "learning_rate": 8.842613165060015e-05, "loss": 0.6557, "step": 7282 }, { "epoch": 0.4888426562867018, "grad_norm": 1.215554118156433, "learning_rate": 8.841917695477835e-05, "loss": 0.6423, "step": 7284 }, { "epoch": 0.4889768799704708, "grad_norm": 1.0148378610610962, "learning_rate": 8.841222044372597e-05, "loss": 0.6413, "step": 7286 }, { "epoch": 0.48911110365423976, "grad_norm": 0.9803034067153931, "learning_rate": 8.840526211777172e-05, "loss": 0.674, "step": 7288 }, { "epoch": 0.4892453273380088, "grad_norm": 1.213284969329834, "learning_rate": 8.839830197724435e-05, "loss": 0.5822, "step": 7290 }, { "epoch": 0.48937955102177777, "grad_norm": 1.2273319959640503, "learning_rate": 8.839134002247272e-05, "loss": 0.6526, "step": 7292 }, { "epoch": 0.4895137747055468, "grad_norm": 0.9492382407188416, "learning_rate": 8.838437625378575e-05, "loss": 0.6438, "step": 7294 }, { "epoch": 0.4896479983893158, "grad_norm": 1.0803555250167847, "learning_rate": 8.83774106715125e-05, "loss": 0.6604, "step": 7296 }, { "epoch": 0.4897822220730848, "grad_norm": 1.1315208673477173, "learning_rate": 8.837044327598206e-05, "loss": 0.6355, "step": 7298 }, { "epoch": 0.4899164457568538, "grad_norm": 1.0323400497436523, "learning_rate": 8.836347406752363e-05, "loss": 0.6169, "step": 7300 }, { "epoch": 0.4900506694406228, "grad_norm": 1.054308295249939, "learning_rate": 8.835650304646648e-05, "loss": 0.6799, "step": 7302 }, { "epoch": 0.4901848931243918, "grad_norm": 1.666773796081543, "learning_rate": 8.834953021313999e-05, "loss": 0.6705, "step": 7304 }, { "epoch": 0.4903191168081608, "grad_norm": 1.113276720046997, "learning_rate": 8.834255556787361e-05, "loss": 0.7109, "step": 7306 }, { "epoch": 0.4904533404919298, "grad_norm": 1.0723297595977783, "learning_rate": 8.833557911099688e-05, "loss": 0.743, "step": 7308 }, { "epoch": 0.4905875641756988, "grad_norm": 0.982081413269043, "learning_rate": 8.832860084283942e-05, "loss": 0.6018, "step": 7310 }, { "epoch": 0.4907217878594678, "grad_norm": 1.1384668350219727, "learning_rate": 8.832162076373094e-05, "loss": 0.649, "step": 7312 }, { "epoch": 0.4908560115432368, "grad_norm": 1.1195060014724731, "learning_rate": 8.831463887400122e-05, "loss": 0.6211, "step": 7314 }, { "epoch": 0.4909902352270058, "grad_norm": 1.0770231485366821, "learning_rate": 8.830765517398017e-05, "loss": 0.6519, "step": 7316 }, { "epoch": 0.49112445891077483, "grad_norm": 1.090936541557312, "learning_rate": 8.830066966399775e-05, "loss": 0.7029, "step": 7318 }, { "epoch": 0.4912586825945438, "grad_norm": 1.2828911542892456, "learning_rate": 8.829368234438397e-05, "loss": 0.5789, "step": 7320 }, { "epoch": 0.49139290627831284, "grad_norm": 1.2686687707901, "learning_rate": 8.828669321546902e-05, "loss": 0.6519, "step": 7322 }, { "epoch": 0.4915271299620818, "grad_norm": 1.336713194847107, "learning_rate": 8.827970227758313e-05, "loss": 0.713, "step": 7324 }, { "epoch": 0.4916613536458508, "grad_norm": 1.0106137990951538, "learning_rate": 8.827270953105655e-05, "loss": 0.7171, "step": 7326 }, { "epoch": 0.4917955773296198, "grad_norm": 1.1118881702423096, "learning_rate": 8.826571497621972e-05, "loss": 0.6424, "step": 7328 }, { "epoch": 0.4919298010133888, "grad_norm": 0.9412743449211121, "learning_rate": 8.825871861340308e-05, "loss": 0.6053, "step": 7330 }, { "epoch": 0.4920640246971578, "grad_norm": 1.0979979038238525, "learning_rate": 8.825172044293725e-05, "loss": 0.6517, "step": 7332 }, { "epoch": 0.4921982483809268, "grad_norm": 1.1139260530471802, "learning_rate": 8.824472046515283e-05, "loss": 0.6765, "step": 7334 }, { "epoch": 0.4923324720646958, "grad_norm": 1.2965672016143799, "learning_rate": 8.823771868038058e-05, "loss": 0.6566, "step": 7336 }, { "epoch": 0.4924666957484648, "grad_norm": 1.1571414470672607, "learning_rate": 8.823071508895131e-05, "loss": 0.6023, "step": 7338 }, { "epoch": 0.49260091943223383, "grad_norm": 1.5252083539962769, "learning_rate": 8.822370969119592e-05, "loss": 0.7048, "step": 7340 }, { "epoch": 0.4927351431160028, "grad_norm": 1.0156404972076416, "learning_rate": 8.821670248744542e-05, "loss": 0.6371, "step": 7342 }, { "epoch": 0.49286936679977184, "grad_norm": 1.6277072429656982, "learning_rate": 8.820969347803088e-05, "loss": 0.7344, "step": 7344 }, { "epoch": 0.4930035904835408, "grad_norm": 1.3882527351379395, "learning_rate": 8.820268266328345e-05, "loss": 0.6687, "step": 7346 }, { "epoch": 0.49313781416730984, "grad_norm": 1.1527955532073975, "learning_rate": 8.81956700435344e-05, "loss": 0.6568, "step": 7348 }, { "epoch": 0.4932720378510788, "grad_norm": 1.0824893712997437, "learning_rate": 8.818865561911504e-05, "loss": 0.6134, "step": 7350 }, { "epoch": 0.49340626153484785, "grad_norm": 1.1220204830169678, "learning_rate": 8.81816393903568e-05, "loss": 0.7174, "step": 7352 }, { "epoch": 0.4935404852186168, "grad_norm": 1.1143301725387573, "learning_rate": 8.817462135759117e-05, "loss": 0.6575, "step": 7354 }, { "epoch": 0.49367470890238585, "grad_norm": 1.1951769590377808, "learning_rate": 8.816760152114976e-05, "loss": 0.679, "step": 7356 }, { "epoch": 0.4938089325861548, "grad_norm": 1.0683196783065796, "learning_rate": 8.816057988136422e-05, "loss": 0.6011, "step": 7358 }, { "epoch": 0.49394315626992386, "grad_norm": 1.1153167486190796, "learning_rate": 8.815355643856633e-05, "loss": 0.6792, "step": 7360 }, { "epoch": 0.49407737995369283, "grad_norm": 1.0507798194885254, "learning_rate": 8.814653119308794e-05, "loss": 0.6355, "step": 7362 }, { "epoch": 0.4942116036374618, "grad_norm": 1.233003854751587, "learning_rate": 8.813950414526093e-05, "loss": 0.6855, "step": 7364 }, { "epoch": 0.49434582732123084, "grad_norm": 0.9083017706871033, "learning_rate": 8.813247529541737e-05, "loss": 0.651, "step": 7366 }, { "epoch": 0.4944800510049998, "grad_norm": 1.1303188800811768, "learning_rate": 8.812544464388932e-05, "loss": 0.7534, "step": 7368 }, { "epoch": 0.49461427468876884, "grad_norm": 1.0329996347427368, "learning_rate": 8.8118412191009e-05, "loss": 0.5869, "step": 7370 }, { "epoch": 0.4947484983725378, "grad_norm": 1.0308709144592285, "learning_rate": 8.811137793710863e-05, "loss": 0.6149, "step": 7372 }, { "epoch": 0.49488272205630685, "grad_norm": 1.163661241531372, "learning_rate": 8.810434188252063e-05, "loss": 0.7506, "step": 7374 }, { "epoch": 0.4950169457400758, "grad_norm": 1.409184217453003, "learning_rate": 8.809730402757739e-05, "loss": 0.5969, "step": 7376 }, { "epoch": 0.49515116942384485, "grad_norm": 1.0647064447402954, "learning_rate": 8.809026437261145e-05, "loss": 0.6521, "step": 7378 }, { "epoch": 0.4952853931076138, "grad_norm": 1.0129282474517822, "learning_rate": 8.80832229179554e-05, "loss": 0.6498, "step": 7380 }, { "epoch": 0.49541961679138286, "grad_norm": 1.0250266790390015, "learning_rate": 8.8076179663942e-05, "loss": 0.6595, "step": 7382 }, { "epoch": 0.49555384047515183, "grad_norm": 1.0052516460418701, "learning_rate": 8.806913461090395e-05, "loss": 0.6798, "step": 7384 }, { "epoch": 0.49568806415892086, "grad_norm": 0.9906535744667053, "learning_rate": 8.806208775917417e-05, "loss": 0.6435, "step": 7386 }, { "epoch": 0.49582228784268984, "grad_norm": 1.077197551727295, "learning_rate": 8.805503910908557e-05, "loss": 0.6533, "step": 7388 }, { "epoch": 0.49595651152645887, "grad_norm": 1.6716821193695068, "learning_rate": 8.804798866097121e-05, "loss": 0.6529, "step": 7390 }, { "epoch": 0.49609073521022784, "grad_norm": 0.9911879301071167, "learning_rate": 8.80409364151642e-05, "loss": 0.6959, "step": 7392 }, { "epoch": 0.49622495889399687, "grad_norm": 1.0691368579864502, "learning_rate": 8.803388237199776e-05, "loss": 0.6926, "step": 7394 }, { "epoch": 0.49635918257776585, "grad_norm": 0.9496676921844482, "learning_rate": 8.802682653180516e-05, "loss": 0.5831, "step": 7396 }, { "epoch": 0.4964934062615349, "grad_norm": 1.2202308177947998, "learning_rate": 8.801976889491979e-05, "loss": 0.6378, "step": 7398 }, { "epoch": 0.49662762994530385, "grad_norm": 1.1562025547027588, "learning_rate": 8.80127094616751e-05, "loss": 0.6617, "step": 7400 }, { "epoch": 0.4967618536290728, "grad_norm": 0.94105064868927, "learning_rate": 8.800564823240464e-05, "loss": 0.6513, "step": 7402 }, { "epoch": 0.49689607731284186, "grad_norm": 1.05088472366333, "learning_rate": 8.799858520744201e-05, "loss": 0.6262, "step": 7404 }, { "epoch": 0.49703030099661083, "grad_norm": 1.0232298374176025, "learning_rate": 8.799152038712099e-05, "loss": 0.5954, "step": 7406 }, { "epoch": 0.49716452468037986, "grad_norm": 1.0250871181488037, "learning_rate": 8.798445377177531e-05, "loss": 0.6446, "step": 7408 }, { "epoch": 0.49729874836414883, "grad_norm": 0.9300850033760071, "learning_rate": 8.79773853617389e-05, "loss": 0.6087, "step": 7410 }, { "epoch": 0.49743297204791787, "grad_norm": 1.0377005338668823, "learning_rate": 8.797031515734571e-05, "loss": 0.6592, "step": 7412 }, { "epoch": 0.49756719573168684, "grad_norm": 1.0114991664886475, "learning_rate": 8.796324315892978e-05, "loss": 0.6064, "step": 7414 }, { "epoch": 0.49770141941545587, "grad_norm": 0.962981641292572, "learning_rate": 8.795616936682528e-05, "loss": 0.6048, "step": 7416 }, { "epoch": 0.49783564309922484, "grad_norm": 1.3811211585998535, "learning_rate": 8.794909378136639e-05, "loss": 0.6492, "step": 7418 }, { "epoch": 0.4979698667829939, "grad_norm": 0.9624200463294983, "learning_rate": 8.794201640288746e-05, "loss": 0.6221, "step": 7420 }, { "epoch": 0.49810409046676285, "grad_norm": 1.0491639375686646, "learning_rate": 8.793493723172286e-05, "loss": 0.6212, "step": 7422 }, { "epoch": 0.4982383141505319, "grad_norm": 1.1057180166244507, "learning_rate": 8.792785626820708e-05, "loss": 0.6459, "step": 7424 }, { "epoch": 0.49837253783430085, "grad_norm": 1.0114151239395142, "learning_rate": 8.792077351267466e-05, "loss": 0.6959, "step": 7426 }, { "epoch": 0.4985067615180699, "grad_norm": 1.1589282751083374, "learning_rate": 8.791368896546027e-05, "loss": 0.6812, "step": 7428 }, { "epoch": 0.49864098520183886, "grad_norm": 1.2169181108474731, "learning_rate": 8.790660262689863e-05, "loss": 0.7807, "step": 7430 }, { "epoch": 0.4987752088856079, "grad_norm": 1.1884033679962158, "learning_rate": 8.789951449732454e-05, "loss": 0.6644, "step": 7432 }, { "epoch": 0.49890943256937687, "grad_norm": 1.1186769008636475, "learning_rate": 8.789242457707294e-05, "loss": 0.6636, "step": 7434 }, { "epoch": 0.4990436562531459, "grad_norm": 1.0410146713256836, "learning_rate": 8.788533286647878e-05, "loss": 0.587, "step": 7436 }, { "epoch": 0.49917787993691487, "grad_norm": 1.07662832736969, "learning_rate": 8.787823936587714e-05, "loss": 0.6532, "step": 7438 }, { "epoch": 0.49931210362068384, "grad_norm": 0.859747588634491, "learning_rate": 8.787114407560317e-05, "loss": 0.6396, "step": 7440 }, { "epoch": 0.4994463273044529, "grad_norm": 1.036584496498108, "learning_rate": 8.786404699599211e-05, "loss": 0.7043, "step": 7442 }, { "epoch": 0.49958055098822185, "grad_norm": 1.0626755952835083, "learning_rate": 8.78569481273793e-05, "loss": 0.6924, "step": 7444 }, { "epoch": 0.4997147746719909, "grad_norm": 1.1623696088790894, "learning_rate": 8.784984747010012e-05, "loss": 0.6752, "step": 7446 }, { "epoch": 0.49984899835575985, "grad_norm": 1.1692180633544922, "learning_rate": 8.784274502449009e-05, "loss": 0.6518, "step": 7448 }, { "epoch": 0.4999832220395289, "grad_norm": 1.004929780960083, "learning_rate": 8.783564079088477e-05, "loss": 0.6102, "step": 7450 }, { "epoch": 0.5001174457232979, "grad_norm": 1.0000412464141846, "learning_rate": 8.782853476961981e-05, "loss": 0.626, "step": 7452 }, { "epoch": 0.5002516694070669, "grad_norm": 1.0467336177825928, "learning_rate": 8.782142696103098e-05, "loss": 0.675, "step": 7454 }, { "epoch": 0.5003858930908359, "grad_norm": 1.0724281072616577, "learning_rate": 8.781431736545409e-05, "loss": 0.6264, "step": 7456 }, { "epoch": 0.5005201167746048, "grad_norm": 0.9236459136009216, "learning_rate": 8.780720598322507e-05, "loss": 0.612, "step": 7458 }, { "epoch": 0.5006543404583739, "grad_norm": 1.2468255758285522, "learning_rate": 8.780009281467991e-05, "loss": 0.669, "step": 7460 }, { "epoch": 0.5007885641421429, "grad_norm": 1.2404422760009766, "learning_rate": 8.77929778601547e-05, "loss": 0.6444, "step": 7462 }, { "epoch": 0.5009227878259119, "grad_norm": 0.9682061076164246, "learning_rate": 8.778586111998561e-05, "loss": 0.5992, "step": 7464 }, { "epoch": 0.5010570115096808, "grad_norm": 1.0936394929885864, "learning_rate": 8.777874259450887e-05, "loss": 0.649, "step": 7466 }, { "epoch": 0.5011912351934499, "grad_norm": 1.0420470237731934, "learning_rate": 8.777162228406083e-05, "loss": 0.6183, "step": 7468 }, { "epoch": 0.5013254588772189, "grad_norm": 1.0366129875183105, "learning_rate": 8.776450018897792e-05, "loss": 0.6587, "step": 7470 }, { "epoch": 0.5014596825609879, "grad_norm": 1.1497166156768799, "learning_rate": 8.775737630959662e-05, "loss": 0.6408, "step": 7472 }, { "epoch": 0.5015939062447569, "grad_norm": 0.9807114601135254, "learning_rate": 8.775025064625355e-05, "loss": 0.6383, "step": 7474 }, { "epoch": 0.5017281299285259, "grad_norm": 1.0735164880752563, "learning_rate": 8.774312319928537e-05, "loss": 0.6411, "step": 7476 }, { "epoch": 0.5018623536122949, "grad_norm": 0.995513916015625, "learning_rate": 8.773599396902886e-05, "loss": 0.6344, "step": 7478 }, { "epoch": 0.5019965772960638, "grad_norm": 0.891380786895752, "learning_rate": 8.772886295582079e-05, "loss": 0.5634, "step": 7480 }, { "epoch": 0.5021308009798329, "grad_norm": 1.1450775861740112, "learning_rate": 8.772173015999816e-05, "loss": 0.5742, "step": 7482 }, { "epoch": 0.5022650246636019, "grad_norm": 0.9486705660820007, "learning_rate": 8.771459558189796e-05, "loss": 0.6125, "step": 7484 }, { "epoch": 0.5023992483473709, "grad_norm": 1.1187535524368286, "learning_rate": 8.770745922185728e-05, "loss": 0.64, "step": 7486 }, { "epoch": 0.5025334720311398, "grad_norm": 1.4808590412139893, "learning_rate": 8.770032108021331e-05, "loss": 0.671, "step": 7488 }, { "epoch": 0.5026676957149089, "grad_norm": 1.0867255926132202, "learning_rate": 8.76931811573033e-05, "loss": 0.6704, "step": 7490 }, { "epoch": 0.5028019193986779, "grad_norm": 0.9657959342002869, "learning_rate": 8.768603945346458e-05, "loss": 0.6457, "step": 7492 }, { "epoch": 0.5029361430824469, "grad_norm": 1.0877447128295898, "learning_rate": 8.767889596903462e-05, "loss": 0.6815, "step": 7494 }, { "epoch": 0.5030703667662159, "grad_norm": 1.057449221611023, "learning_rate": 8.767175070435092e-05, "loss": 0.6401, "step": 7496 }, { "epoch": 0.5032045904499849, "grad_norm": 5.432621002197266, "learning_rate": 8.766460365975107e-05, "loss": 0.5941, "step": 7498 }, { "epoch": 0.5033388141337539, "grad_norm": 1.0450407266616821, "learning_rate": 8.765745483557276e-05, "loss": 0.671, "step": 7500 }, { "epoch": 0.503473037817523, "grad_norm": 0.9949643015861511, "learning_rate": 8.765030423215377e-05, "loss": 0.6221, "step": 7502 }, { "epoch": 0.5036072615012919, "grad_norm": 1.0133390426635742, "learning_rate": 8.764315184983193e-05, "loss": 0.6609, "step": 7504 }, { "epoch": 0.5037414851850609, "grad_norm": 1.0307201147079468, "learning_rate": 8.76359976889452e-05, "loss": 0.6813, "step": 7506 }, { "epoch": 0.5038757088688299, "grad_norm": 1.0315346717834473, "learning_rate": 8.762884174983158e-05, "loss": 0.6635, "step": 7508 }, { "epoch": 0.504009932552599, "grad_norm": 1.23046875, "learning_rate": 8.762168403282917e-05, "loss": 0.7086, "step": 7510 }, { "epoch": 0.5041441562363679, "grad_norm": 1.2309722900390625, "learning_rate": 8.761452453827618e-05, "loss": 0.6886, "step": 7512 }, { "epoch": 0.5042783799201369, "grad_norm": 0.9779922962188721, "learning_rate": 8.760736326651087e-05, "loss": 0.6119, "step": 7514 }, { "epoch": 0.5044126036039059, "grad_norm": 0.9598320126533508, "learning_rate": 8.760020021787158e-05, "loss": 0.62, "step": 7516 }, { "epoch": 0.5045468272876749, "grad_norm": 2.2394909858703613, "learning_rate": 8.759303539269679e-05, "loss": 0.6471, "step": 7518 }, { "epoch": 0.5046810509714439, "grad_norm": 1.0633375644683838, "learning_rate": 8.7585868791325e-05, "loss": 0.6494, "step": 7520 }, { "epoch": 0.5048152746552129, "grad_norm": 1.1028856039047241, "learning_rate": 8.757870041409481e-05, "loss": 0.6536, "step": 7522 }, { "epoch": 0.504949498338982, "grad_norm": 1.4106289148330688, "learning_rate": 8.757153026134493e-05, "loss": 0.6635, "step": 7524 }, { "epoch": 0.5050837220227509, "grad_norm": 1.292608618736267, "learning_rate": 8.756435833341412e-05, "loss": 0.7536, "step": 7526 }, { "epoch": 0.5052179457065199, "grad_norm": 1.009282112121582, "learning_rate": 8.755718463064125e-05, "loss": 0.601, "step": 7528 }, { "epoch": 0.5053521693902889, "grad_norm": 0.946381688117981, "learning_rate": 8.755000915336527e-05, "loss": 0.6527, "step": 7530 }, { "epoch": 0.505486393074058, "grad_norm": 1.0703035593032837, "learning_rate": 8.754283190192516e-05, "loss": 0.6132, "step": 7532 }, { "epoch": 0.5056206167578269, "grad_norm": 1.0881116390228271, "learning_rate": 8.75356528766601e-05, "loss": 0.6117, "step": 7534 }, { "epoch": 0.5057548404415959, "grad_norm": 1.224095344543457, "learning_rate": 8.752847207790924e-05, "loss": 0.6924, "step": 7536 }, { "epoch": 0.5058890641253649, "grad_norm": 1.0547585487365723, "learning_rate": 8.752128950601189e-05, "loss": 0.6394, "step": 7538 }, { "epoch": 0.506023287809134, "grad_norm": 1.133722186088562, "learning_rate": 8.751410516130739e-05, "loss": 0.5899, "step": 7540 }, { "epoch": 0.5061575114929029, "grad_norm": 1.1165887117385864, "learning_rate": 8.750691904413515e-05, "loss": 0.7256, "step": 7542 }, { "epoch": 0.5062917351766719, "grad_norm": 1.232200026512146, "learning_rate": 8.749973115483478e-05, "loss": 0.6446, "step": 7544 }, { "epoch": 0.506425958860441, "grad_norm": 1.9479700326919556, "learning_rate": 8.749254149374583e-05, "loss": 0.6947, "step": 7546 }, { "epoch": 0.50656018254421, "grad_norm": 0.9247474074363708, "learning_rate": 8.748535006120804e-05, "loss": 0.6398, "step": 7548 }, { "epoch": 0.5066944062279789, "grad_norm": 1.1821279525756836, "learning_rate": 8.747815685756115e-05, "loss": 0.6105, "step": 7550 }, { "epoch": 0.5068286299117479, "grad_norm": 1.039718508720398, "learning_rate": 8.747096188314506e-05, "loss": 0.6645, "step": 7552 }, { "epoch": 0.506962853595517, "grad_norm": 1.0514421463012695, "learning_rate": 8.74637651382997e-05, "loss": 0.694, "step": 7554 }, { "epoch": 0.5070970772792859, "grad_norm": 1.110547423362732, "learning_rate": 8.745656662336511e-05, "loss": 0.6288, "step": 7556 }, { "epoch": 0.5072313009630549, "grad_norm": 1.2142102718353271, "learning_rate": 8.744936633868139e-05, "loss": 0.6727, "step": 7558 }, { "epoch": 0.5073655246468239, "grad_norm": 0.9618511199951172, "learning_rate": 8.744216428458878e-05, "loss": 0.6716, "step": 7560 }, { "epoch": 0.507499748330593, "grad_norm": 0.9959256052970886, "learning_rate": 8.743496046142752e-05, "loss": 0.6373, "step": 7562 }, { "epoch": 0.5076339720143619, "grad_norm": 1.234066128730774, "learning_rate": 8.742775486953799e-05, "loss": 0.5901, "step": 7564 }, { "epoch": 0.5077681956981309, "grad_norm": 1.105370044708252, "learning_rate": 8.742054750926063e-05, "loss": 0.6654, "step": 7566 }, { "epoch": 0.5079024193819, "grad_norm": 1.0729789733886719, "learning_rate": 8.7413338380936e-05, "loss": 0.7325, "step": 7568 }, { "epoch": 0.508036643065669, "grad_norm": 1.02970290184021, "learning_rate": 8.740612748490468e-05, "loss": 0.6863, "step": 7570 }, { "epoch": 0.5081708667494379, "grad_norm": 1.0629090070724487, "learning_rate": 8.739891482150741e-05, "loss": 0.6224, "step": 7572 }, { "epoch": 0.5083050904332069, "grad_norm": 1.0690653324127197, "learning_rate": 8.739170039108495e-05, "loss": 0.6694, "step": 7574 }, { "epoch": 0.508439314116976, "grad_norm": 1.087717056274414, "learning_rate": 8.738448419397818e-05, "loss": 0.6956, "step": 7576 }, { "epoch": 0.508573537800745, "grad_norm": 0.9817718863487244, "learning_rate": 8.737726623052805e-05, "loss": 0.6292, "step": 7578 }, { "epoch": 0.5087077614845139, "grad_norm": 2.2005228996276855, "learning_rate": 8.737004650107557e-05, "loss": 0.6386, "step": 7580 }, { "epoch": 0.5088419851682829, "grad_norm": 1.0479660034179688, "learning_rate": 8.736282500596188e-05, "loss": 0.6512, "step": 7582 }, { "epoch": 0.508976208852052, "grad_norm": 1.2043403387069702, "learning_rate": 8.735560174552821e-05, "loss": 0.7167, "step": 7584 }, { "epoch": 0.509110432535821, "grad_norm": 1.121976613998413, "learning_rate": 8.73483767201158e-05, "loss": 0.7294, "step": 7586 }, { "epoch": 0.5092446562195899, "grad_norm": 1.1195271015167236, "learning_rate": 8.734114993006602e-05, "loss": 0.6299, "step": 7588 }, { "epoch": 0.509378879903359, "grad_norm": 1.4202194213867188, "learning_rate": 8.733392137572033e-05, "loss": 0.6179, "step": 7590 }, { "epoch": 0.509513103587128, "grad_norm": 1.0699143409729004, "learning_rate": 8.732669105742028e-05, "loss": 0.6533, "step": 7592 }, { "epoch": 0.5096473272708969, "grad_norm": 1.2938518524169922, "learning_rate": 8.731945897550748e-05, "loss": 0.6644, "step": 7594 }, { "epoch": 0.5097815509546659, "grad_norm": 1.3335658311843872, "learning_rate": 8.731222513032362e-05, "loss": 0.6138, "step": 7596 }, { "epoch": 0.509915774638435, "grad_norm": 1.1592415571212769, "learning_rate": 8.730498952221051e-05, "loss": 0.6655, "step": 7598 }, { "epoch": 0.510049998322204, "grad_norm": 1.2052128314971924, "learning_rate": 8.729775215151e-05, "loss": 0.6133, "step": 7600 }, { "epoch": 0.5101842220059729, "grad_norm": 1.1826194524765015, "learning_rate": 8.729051301856402e-05, "loss": 0.5907, "step": 7602 }, { "epoch": 0.5103184456897419, "grad_norm": 1.1111299991607666, "learning_rate": 8.728327212371466e-05, "loss": 0.7097, "step": 7604 }, { "epoch": 0.510452669373511, "grad_norm": 1.5861774682998657, "learning_rate": 8.727602946730397e-05, "loss": 0.6082, "step": 7606 }, { "epoch": 0.51058689305728, "grad_norm": 1.4667607545852661, "learning_rate": 8.726878504967421e-05, "loss": 0.667, "step": 7608 }, { "epoch": 0.5107211167410489, "grad_norm": 0.9902428388595581, "learning_rate": 8.726153887116766e-05, "loss": 0.6366, "step": 7610 }, { "epoch": 0.510855340424818, "grad_norm": 0.9585524201393127, "learning_rate": 8.725429093212664e-05, "loss": 0.6093, "step": 7612 }, { "epoch": 0.510989564108587, "grad_norm": 1.0268088579177856, "learning_rate": 8.724704123289365e-05, "loss": 0.6577, "step": 7614 }, { "epoch": 0.511123787792356, "grad_norm": 0.9739425778388977, "learning_rate": 8.723978977381119e-05, "loss": 0.6113, "step": 7616 }, { "epoch": 0.5112580114761249, "grad_norm": 1.6853216886520386, "learning_rate": 8.72325365552219e-05, "loss": 0.6764, "step": 7618 }, { "epoch": 0.511392235159894, "grad_norm": 1.0694867372512817, "learning_rate": 8.722528157746847e-05, "loss": 0.6613, "step": 7620 }, { "epoch": 0.511526458843663, "grad_norm": 1.4094675779342651, "learning_rate": 8.721802484089368e-05, "loss": 0.6777, "step": 7622 }, { "epoch": 0.511660682527432, "grad_norm": 1.1949818134307861, "learning_rate": 8.721076634584042e-05, "loss": 0.6498, "step": 7624 }, { "epoch": 0.5117949062112009, "grad_norm": 1.0607755184173584, "learning_rate": 8.72035060926516e-05, "loss": 0.6183, "step": 7626 }, { "epoch": 0.51192912989497, "grad_norm": 1.1353689432144165, "learning_rate": 8.719624408167029e-05, "loss": 0.5894, "step": 7628 }, { "epoch": 0.512063353578739, "grad_norm": 0.9764440059661865, "learning_rate": 8.71889803132396e-05, "loss": 0.6442, "step": 7630 }, { "epoch": 0.5121975772625079, "grad_norm": 1.0443223714828491, "learning_rate": 8.718171478770269e-05, "loss": 0.6744, "step": 7632 }, { "epoch": 0.512331800946277, "grad_norm": 1.0360503196716309, "learning_rate": 8.71744475054029e-05, "loss": 0.6924, "step": 7634 }, { "epoch": 0.512466024630046, "grad_norm": 1.087961196899414, "learning_rate": 8.716717846668354e-05, "loss": 0.6285, "step": 7636 }, { "epoch": 0.512600248313815, "grad_norm": 1.0620614290237427, "learning_rate": 8.715990767188811e-05, "loss": 0.6147, "step": 7638 }, { "epoch": 0.5127344719975839, "grad_norm": 1.312974214553833, "learning_rate": 8.71526351213601e-05, "loss": 0.6432, "step": 7640 }, { "epoch": 0.512868695681353, "grad_norm": 1.0307996273040771, "learning_rate": 8.714536081544314e-05, "loss": 0.5833, "step": 7642 }, { "epoch": 0.513002919365122, "grad_norm": 1.0482127666473389, "learning_rate": 8.71380847544809e-05, "loss": 0.6352, "step": 7644 }, { "epoch": 0.513137143048891, "grad_norm": 1.0585205554962158, "learning_rate": 8.713080693881723e-05, "loss": 0.6624, "step": 7646 }, { "epoch": 0.5132713667326599, "grad_norm": 1.017253041267395, "learning_rate": 8.712352736879592e-05, "loss": 0.7177, "step": 7648 }, { "epoch": 0.513405590416429, "grad_norm": 0.938526451587677, "learning_rate": 8.711624604476094e-05, "loss": 0.6576, "step": 7650 }, { "epoch": 0.513539814100198, "grad_norm": 1.0195128917694092, "learning_rate": 8.710896296705634e-05, "loss": 0.6401, "step": 7652 }, { "epoch": 0.513674037783967, "grad_norm": 1.232199788093567, "learning_rate": 8.710167813602619e-05, "loss": 0.684, "step": 7654 }, { "epoch": 0.513808261467736, "grad_norm": 1.0564061403274536, "learning_rate": 8.70943915520147e-05, "loss": 0.6007, "step": 7656 }, { "epoch": 0.513942485151505, "grad_norm": 0.9898374676704407, "learning_rate": 8.708710321536617e-05, "loss": 0.5945, "step": 7658 }, { "epoch": 0.514076708835274, "grad_norm": 1.0006496906280518, "learning_rate": 8.707981312642494e-05, "loss": 0.6414, "step": 7660 }, { "epoch": 0.514210932519043, "grad_norm": 1.2971662282943726, "learning_rate": 8.707252128553544e-05, "loss": 0.6242, "step": 7662 }, { "epoch": 0.514345156202812, "grad_norm": 1.0045400857925415, "learning_rate": 8.706522769304224e-05, "loss": 0.6339, "step": 7664 }, { "epoch": 0.514479379886581, "grad_norm": 0.9474780559539795, "learning_rate": 8.70579323492899e-05, "loss": 0.6462, "step": 7666 }, { "epoch": 0.51461360357035, "grad_norm": 1.4041409492492676, "learning_rate": 8.705063525462312e-05, "loss": 0.6008, "step": 7668 }, { "epoch": 0.5147478272541189, "grad_norm": 0.9978337287902832, "learning_rate": 8.704333640938669e-05, "loss": 0.637, "step": 7670 }, { "epoch": 0.514882050937888, "grad_norm": 1.0927438735961914, "learning_rate": 8.703603581392546e-05, "loss": 0.646, "step": 7672 }, { "epoch": 0.515016274621657, "grad_norm": 1.1463407278060913, "learning_rate": 8.702873346858434e-05, "loss": 0.6386, "step": 7674 }, { "epoch": 0.515150498305426, "grad_norm": 0.9773491024971008, "learning_rate": 8.70214293737084e-05, "loss": 0.6463, "step": 7676 }, { "epoch": 0.515284721989195, "grad_norm": 1.1640729904174805, "learning_rate": 8.701412352964274e-05, "loss": 0.6623, "step": 7678 }, { "epoch": 0.515418945672964, "grad_norm": 1.3043346405029297, "learning_rate": 8.70068159367325e-05, "loss": 0.6581, "step": 7680 }, { "epoch": 0.515553169356733, "grad_norm": 1.5669822692871094, "learning_rate": 8.699950659532298e-05, "loss": 0.6526, "step": 7682 }, { "epoch": 0.515687393040502, "grad_norm": 1.0059089660644531, "learning_rate": 8.699219550575953e-05, "loss": 0.6044, "step": 7684 }, { "epoch": 0.515821616724271, "grad_norm": 1.0378648042678833, "learning_rate": 8.698488266838759e-05, "loss": 0.6286, "step": 7686 }, { "epoch": 0.51595584040804, "grad_norm": 1.1167110204696655, "learning_rate": 8.697756808355267e-05, "loss": 0.634, "step": 7688 }, { "epoch": 0.516090064091809, "grad_norm": 1.049477219581604, "learning_rate": 8.697025175160039e-05, "loss": 0.5723, "step": 7690 }, { "epoch": 0.516224287775578, "grad_norm": 1.1996572017669678, "learning_rate": 8.696293367287638e-05, "loss": 0.7119, "step": 7692 }, { "epoch": 0.516358511459347, "grad_norm": 1.0795083045959473, "learning_rate": 8.695561384772646e-05, "loss": 0.6275, "step": 7694 }, { "epoch": 0.516492735143116, "grad_norm": 1.0214886665344238, "learning_rate": 8.694829227649644e-05, "loss": 0.6757, "step": 7696 }, { "epoch": 0.516626958826885, "grad_norm": 1.1827250719070435, "learning_rate": 8.694096895953227e-05, "loss": 0.6685, "step": 7698 }, { "epoch": 0.5167611825106541, "grad_norm": 1.0351839065551758, "learning_rate": 8.693364389717998e-05, "loss": 0.6215, "step": 7700 }, { "epoch": 0.516895406194423, "grad_norm": 1.0110132694244385, "learning_rate": 8.692631708978562e-05, "loss": 0.6475, "step": 7702 }, { "epoch": 0.517029629878192, "grad_norm": 1.1591826677322388, "learning_rate": 8.69189885376954e-05, "loss": 0.6171, "step": 7704 }, { "epoch": 0.517163853561961, "grad_norm": 1.0220993757247925, "learning_rate": 8.691165824125556e-05, "loss": 0.6669, "step": 7706 }, { "epoch": 0.51729807724573, "grad_norm": 1.1127986907958984, "learning_rate": 8.690432620081246e-05, "loss": 0.6426, "step": 7708 }, { "epoch": 0.517432300929499, "grad_norm": 1.0023584365844727, "learning_rate": 8.689699241671252e-05, "loss": 0.5993, "step": 7710 }, { "epoch": 0.517566524613268, "grad_norm": 1.3513555526733398, "learning_rate": 8.688965688930224e-05, "loss": 0.5884, "step": 7712 }, { "epoch": 0.517700748297037, "grad_norm": 1.107893466949463, "learning_rate": 8.68823196189282e-05, "loss": 0.6851, "step": 7714 }, { "epoch": 0.517834971980806, "grad_norm": 1.0814416408538818, "learning_rate": 8.68749806059371e-05, "loss": 0.7228, "step": 7716 }, { "epoch": 0.517969195664575, "grad_norm": 1.3134574890136719, "learning_rate": 8.686763985067568e-05, "loss": 0.7092, "step": 7718 }, { "epoch": 0.518103419348344, "grad_norm": 1.3009024858474731, "learning_rate": 8.686029735349075e-05, "loss": 0.6243, "step": 7720 }, { "epoch": 0.5182376430321131, "grad_norm": 1.1668957471847534, "learning_rate": 8.685295311472927e-05, "loss": 0.6548, "step": 7722 }, { "epoch": 0.518371866715882, "grad_norm": 1.0200467109680176, "learning_rate": 8.684560713473822e-05, "loss": 0.6426, "step": 7724 }, { "epoch": 0.518506090399651, "grad_norm": 1.109863042831421, "learning_rate": 8.68382594138647e-05, "loss": 0.5925, "step": 7726 }, { "epoch": 0.51864031408342, "grad_norm": 1.043623447418213, "learning_rate": 8.683090995245584e-05, "loss": 0.6926, "step": 7728 }, { "epoch": 0.5187745377671891, "grad_norm": 1.005377173423767, "learning_rate": 8.682355875085893e-05, "loss": 0.5992, "step": 7730 }, { "epoch": 0.518908761450958, "grad_norm": 1.0924162864685059, "learning_rate": 8.681620580942127e-05, "loss": 0.6099, "step": 7732 }, { "epoch": 0.519042985134727, "grad_norm": 1.9004437923431396, "learning_rate": 8.680885112849028e-05, "loss": 0.6336, "step": 7734 }, { "epoch": 0.519177208818496, "grad_norm": 1.0223441123962402, "learning_rate": 8.680149470841346e-05, "loss": 0.6434, "step": 7736 }, { "epoch": 0.5193114325022651, "grad_norm": 1.0525568723678589, "learning_rate": 8.679413654953837e-05, "loss": 0.6352, "step": 7738 }, { "epoch": 0.519445656186034, "grad_norm": 0.9689558744430542, "learning_rate": 8.678677665221268e-05, "loss": 0.5739, "step": 7740 }, { "epoch": 0.519579879869803, "grad_norm": 0.9976610541343689, "learning_rate": 8.677941501678415e-05, "loss": 0.6204, "step": 7742 }, { "epoch": 0.5197141035535721, "grad_norm": 1.1912943124771118, "learning_rate": 8.677205164360059e-05, "loss": 0.7589, "step": 7744 }, { "epoch": 0.519848327237341, "grad_norm": 1.831667423248291, "learning_rate": 8.676468653300987e-05, "loss": 0.6202, "step": 7746 }, { "epoch": 0.51998255092111, "grad_norm": 1.085209608078003, "learning_rate": 8.675731968536002e-05, "loss": 0.6968, "step": 7748 }, { "epoch": 0.520116774604879, "grad_norm": 1.301666498184204, "learning_rate": 8.674995110099911e-05, "loss": 0.6577, "step": 7750 }, { "epoch": 0.5202509982886481, "grad_norm": 0.9996120929718018, "learning_rate": 8.674258078027525e-05, "loss": 0.6601, "step": 7752 }, { "epoch": 0.520385221972417, "grad_norm": 1.0176335573196411, "learning_rate": 8.673520872353671e-05, "loss": 0.6561, "step": 7754 }, { "epoch": 0.520519445656186, "grad_norm": 1.3511710166931152, "learning_rate": 8.672783493113179e-05, "loss": 0.7215, "step": 7756 }, { "epoch": 0.520653669339955, "grad_norm": 1.0890061855316162, "learning_rate": 8.672045940340889e-05, "loss": 0.6518, "step": 7758 }, { "epoch": 0.5207878930237241, "grad_norm": 1.0245070457458496, "learning_rate": 8.671308214071652e-05, "loss": 0.6538, "step": 7760 }, { "epoch": 0.520922116707493, "grad_norm": 1.0657215118408203, "learning_rate": 8.670570314340318e-05, "loss": 0.6401, "step": 7762 }, { "epoch": 0.521056340391262, "grad_norm": 1.4132921695709229, "learning_rate": 8.669832241181756e-05, "loss": 0.67, "step": 7764 }, { "epoch": 0.5211905640750311, "grad_norm": 1.603489875793457, "learning_rate": 8.669093994630835e-05, "loss": 0.7102, "step": 7766 }, { "epoch": 0.5213247877588001, "grad_norm": 1.1183873414993286, "learning_rate": 8.66835557472244e-05, "loss": 0.6822, "step": 7768 }, { "epoch": 0.521459011442569, "grad_norm": 1.2292097806930542, "learning_rate": 8.667616981491458e-05, "loss": 0.7148, "step": 7770 }, { "epoch": 0.521593235126338, "grad_norm": 1.140467643737793, "learning_rate": 8.666878214972783e-05, "loss": 0.6574, "step": 7772 }, { "epoch": 0.5217274588101071, "grad_norm": 1.070495367050171, "learning_rate": 8.666139275201325e-05, "loss": 0.6431, "step": 7774 }, { "epoch": 0.521861682493876, "grad_norm": 1.0301347970962524, "learning_rate": 8.665400162211995e-05, "loss": 0.6536, "step": 7776 }, { "epoch": 0.521995906177645, "grad_norm": 1.162405252456665, "learning_rate": 8.664660876039715e-05, "loss": 0.6408, "step": 7778 }, { "epoch": 0.522130129861414, "grad_norm": 1.1394559144973755, "learning_rate": 8.663921416719415e-05, "loss": 0.7394, "step": 7780 }, { "epoch": 0.5222643535451831, "grad_norm": 1.1406190395355225, "learning_rate": 8.663181784286032e-05, "loss": 0.6864, "step": 7782 }, { "epoch": 0.522398577228952, "grad_norm": 0.9947435259819031, "learning_rate": 8.662441978774514e-05, "loss": 0.6074, "step": 7784 }, { "epoch": 0.522532800912721, "grad_norm": 1.0343014001846313, "learning_rate": 8.661702000219814e-05, "loss": 0.6589, "step": 7786 }, { "epoch": 0.5226670245964901, "grad_norm": 0.998464822769165, "learning_rate": 8.660961848656895e-05, "loss": 0.6141, "step": 7788 }, { "epoch": 0.5228012482802591, "grad_norm": 1.8650562763214111, "learning_rate": 8.660221524120727e-05, "loss": 0.7038, "step": 7790 }, { "epoch": 0.522935471964028, "grad_norm": 1.15292227268219, "learning_rate": 8.65948102664629e-05, "loss": 0.5998, "step": 7792 }, { "epoch": 0.523069695647797, "grad_norm": 1.2766640186309814, "learning_rate": 8.658740356268571e-05, "loss": 0.6357, "step": 7794 }, { "epoch": 0.5232039193315661, "grad_norm": 1.0240525007247925, "learning_rate": 8.657999513022563e-05, "loss": 0.6269, "step": 7796 }, { "epoch": 0.5233381430153351, "grad_norm": 1.5196304321289062, "learning_rate": 8.657258496943274e-05, "loss": 0.6223, "step": 7798 }, { "epoch": 0.523472366699104, "grad_norm": 1.0911345481872559, "learning_rate": 8.65651730806571e-05, "loss": 0.6661, "step": 7800 }, { "epoch": 0.523606590382873, "grad_norm": 1.163367748260498, "learning_rate": 8.655775946424895e-05, "loss": 0.5699, "step": 7802 }, { "epoch": 0.5237408140666421, "grad_norm": 1.1395347118377686, "learning_rate": 8.655034412055856e-05, "loss": 0.6649, "step": 7804 }, { "epoch": 0.5238750377504111, "grad_norm": 1.1319963932037354, "learning_rate": 8.654292704993627e-05, "loss": 0.6253, "step": 7806 }, { "epoch": 0.52400926143418, "grad_norm": 0.9931795597076416, "learning_rate": 8.653550825273253e-05, "loss": 0.6329, "step": 7808 }, { "epoch": 0.5241434851179491, "grad_norm": 1.2011792659759521, "learning_rate": 8.652808772929788e-05, "loss": 0.7392, "step": 7810 }, { "epoch": 0.5242777088017181, "grad_norm": 0.9433580040931702, "learning_rate": 8.652066547998292e-05, "loss": 0.6397, "step": 7812 }, { "epoch": 0.524411932485487, "grad_norm": 1.039955973625183, "learning_rate": 8.651324150513833e-05, "loss": 0.6344, "step": 7814 }, { "epoch": 0.524546156169256, "grad_norm": 1.0503945350646973, "learning_rate": 8.650581580511487e-05, "loss": 0.6448, "step": 7816 }, { "epoch": 0.5246803798530251, "grad_norm": 1.0410972833633423, "learning_rate": 8.64983883802634e-05, "loss": 0.6067, "step": 7818 }, { "epoch": 0.5248146035367941, "grad_norm": 1.023305892944336, "learning_rate": 8.649095923093484e-05, "loss": 0.5264, "step": 7820 }, { "epoch": 0.524948827220563, "grad_norm": 1.0543869733810425, "learning_rate": 8.648352835748024e-05, "loss": 0.635, "step": 7822 }, { "epoch": 0.525083050904332, "grad_norm": 1.1276301145553589, "learning_rate": 8.647609576025064e-05, "loss": 0.5973, "step": 7824 }, { "epoch": 0.5252172745881011, "grad_norm": 1.0825186967849731, "learning_rate": 8.646866143959725e-05, "loss": 0.6765, "step": 7826 }, { "epoch": 0.5253514982718701, "grad_norm": 0.993621826171875, "learning_rate": 8.646122539587133e-05, "loss": 0.5949, "step": 7828 }, { "epoch": 0.525485721955639, "grad_norm": 1.1419330835342407, "learning_rate": 8.64537876294242e-05, "loss": 0.6804, "step": 7830 }, { "epoch": 0.5256199456394081, "grad_norm": 1.189340591430664, "learning_rate": 8.644634814060728e-05, "loss": 0.5882, "step": 7832 }, { "epoch": 0.5257541693231771, "grad_norm": 1.0714603662490845, "learning_rate": 8.643890692977209e-05, "loss": 0.6831, "step": 7834 }, { "epoch": 0.5258883930069461, "grad_norm": 1.0567216873168945, "learning_rate": 8.64314639972702e-05, "loss": 0.644, "step": 7836 }, { "epoch": 0.526022616690715, "grad_norm": 0.9821720719337463, "learning_rate": 8.642401934345328e-05, "loss": 0.6922, "step": 7838 }, { "epoch": 0.5261568403744841, "grad_norm": 0.9419900178909302, "learning_rate": 8.641657296867306e-05, "loss": 0.6089, "step": 7840 }, { "epoch": 0.5262910640582531, "grad_norm": 1.1534830331802368, "learning_rate": 8.640912487328139e-05, "loss": 0.6305, "step": 7842 }, { "epoch": 0.5264252877420221, "grad_norm": 1.1612067222595215, "learning_rate": 8.640167505763014e-05, "loss": 0.6018, "step": 7844 }, { "epoch": 0.526559511425791, "grad_norm": 0.9795418977737427, "learning_rate": 8.639422352207136e-05, "loss": 0.7098, "step": 7846 }, { "epoch": 0.5266937351095601, "grad_norm": 1.0749777555465698, "learning_rate": 8.638677026695707e-05, "loss": 0.6147, "step": 7848 }, { "epoch": 0.5268279587933291, "grad_norm": 1.1168949604034424, "learning_rate": 8.637931529263943e-05, "loss": 0.7036, "step": 7850 }, { "epoch": 0.526962182477098, "grad_norm": 0.9598569869995117, "learning_rate": 8.637185859947067e-05, "loss": 0.5887, "step": 7852 }, { "epoch": 0.5270964061608671, "grad_norm": 0.9963730573654175, "learning_rate": 8.636440018780314e-05, "loss": 0.6447, "step": 7854 }, { "epoch": 0.5272306298446361, "grad_norm": 0.9982380270957947, "learning_rate": 8.635694005798919e-05, "loss": 0.6494, "step": 7856 }, { "epoch": 0.5273648535284051, "grad_norm": 1.157914400100708, "learning_rate": 8.634947821038132e-05, "loss": 0.757, "step": 7858 }, { "epoch": 0.527499077212174, "grad_norm": 1.0347570180892944, "learning_rate": 8.634201464533207e-05, "loss": 0.6752, "step": 7860 }, { "epoch": 0.5276333008959431, "grad_norm": 1.0469486713409424, "learning_rate": 8.633454936319411e-05, "loss": 0.6475, "step": 7862 }, { "epoch": 0.5277675245797121, "grad_norm": 1.2268787622451782, "learning_rate": 8.632708236432015e-05, "loss": 0.6541, "step": 7864 }, { "epoch": 0.5279017482634811, "grad_norm": 1.1413471698760986, "learning_rate": 8.631961364906296e-05, "loss": 0.6259, "step": 7866 }, { "epoch": 0.52803597194725, "grad_norm": 1.1279798746109009, "learning_rate": 8.631214321777546e-05, "loss": 0.7006, "step": 7868 }, { "epoch": 0.5281701956310191, "grad_norm": 1.0315818786621094, "learning_rate": 8.630467107081057e-05, "loss": 0.6442, "step": 7870 }, { "epoch": 0.5283044193147881, "grad_norm": 1.1192677021026611, "learning_rate": 8.629719720852138e-05, "loss": 0.6034, "step": 7872 }, { "epoch": 0.5284386429985571, "grad_norm": 1.1972932815551758, "learning_rate": 8.628972163126101e-05, "loss": 0.6259, "step": 7874 }, { "epoch": 0.5285728666823261, "grad_norm": 1.0784581899642944, "learning_rate": 8.628224433938263e-05, "loss": 0.622, "step": 7876 }, { "epoch": 0.5287070903660951, "grad_norm": 1.0202914476394653, "learning_rate": 8.627476533323957e-05, "loss": 0.6532, "step": 7878 }, { "epoch": 0.5288413140498641, "grad_norm": 1.1538416147232056, "learning_rate": 8.626728461318518e-05, "loss": 0.6699, "step": 7880 }, { "epoch": 0.5289755377336331, "grad_norm": 1.0363773107528687, "learning_rate": 8.62598021795729e-05, "loss": 0.6393, "step": 7882 }, { "epoch": 0.5291097614174021, "grad_norm": 1.0268653631210327, "learning_rate": 8.625231803275627e-05, "loss": 0.6782, "step": 7884 }, { "epoch": 0.5292439851011711, "grad_norm": 0.8181347846984863, "learning_rate": 8.62448321730889e-05, "loss": 0.6044, "step": 7886 }, { "epoch": 0.5293782087849401, "grad_norm": 1.0739057064056396, "learning_rate": 8.623734460092449e-05, "loss": 0.6223, "step": 7888 }, { "epoch": 0.529512432468709, "grad_norm": 1.0722723007202148, "learning_rate": 8.62298553166168e-05, "loss": 0.6551, "step": 7890 }, { "epoch": 0.5296466561524781, "grad_norm": 1.0251004695892334, "learning_rate": 8.622236432051969e-05, "loss": 0.6408, "step": 7892 }, { "epoch": 0.5297808798362471, "grad_norm": 1.0956529378890991, "learning_rate": 8.621487161298709e-05, "loss": 0.6263, "step": 7894 }, { "epoch": 0.5299151035200161, "grad_norm": 1.1363166570663452, "learning_rate": 8.620737719437302e-05, "loss": 0.6748, "step": 7896 }, { "epoch": 0.5300493272037851, "grad_norm": 1.026837706565857, "learning_rate": 8.619988106503157e-05, "loss": 0.529, "step": 7898 }, { "epoch": 0.5301835508875541, "grad_norm": 0.9920128583908081, "learning_rate": 8.619238322531694e-05, "loss": 0.64, "step": 7900 }, { "epoch": 0.5303177745713231, "grad_norm": 1.0423779487609863, "learning_rate": 8.618488367558335e-05, "loss": 0.6391, "step": 7902 }, { "epoch": 0.5304519982550921, "grad_norm": 1.183112382888794, "learning_rate": 8.617738241618518e-05, "loss": 0.6261, "step": 7904 }, { "epoch": 0.5305862219388611, "grad_norm": 1.0162816047668457, "learning_rate": 8.616987944747683e-05, "loss": 0.6331, "step": 7906 }, { "epoch": 0.5307204456226301, "grad_norm": 1.2763468027114868, "learning_rate": 8.616237476981279e-05, "loss": 0.6126, "step": 7908 }, { "epoch": 0.5308546693063991, "grad_norm": 1.084007740020752, "learning_rate": 8.615486838354765e-05, "loss": 0.5957, "step": 7910 }, { "epoch": 0.5309888929901682, "grad_norm": 1.0765655040740967, "learning_rate": 8.61473602890361e-05, "loss": 0.6723, "step": 7912 }, { "epoch": 0.5311231166739371, "grad_norm": 1.0184767246246338, "learning_rate": 8.613985048663282e-05, "loss": 0.6472, "step": 7914 }, { "epoch": 0.5312573403577061, "grad_norm": 3.213615655899048, "learning_rate": 8.613233897669271e-05, "loss": 0.5945, "step": 7916 }, { "epoch": 0.5313915640414751, "grad_norm": 1.017124891281128, "learning_rate": 8.612482575957062e-05, "loss": 0.6631, "step": 7918 }, { "epoch": 0.5315257877252442, "grad_norm": 1.061355710029602, "learning_rate": 8.611731083562154e-05, "loss": 0.5845, "step": 7920 }, { "epoch": 0.5316600114090131, "grad_norm": 1.0305843353271484, "learning_rate": 8.610979420520053e-05, "loss": 0.6712, "step": 7922 }, { "epoch": 0.5317942350927821, "grad_norm": 0.9948982000350952, "learning_rate": 8.610227586866278e-05, "loss": 0.5888, "step": 7924 }, { "epoch": 0.5319284587765511, "grad_norm": 0.9479734301567078, "learning_rate": 8.609475582636348e-05, "loss": 0.5764, "step": 7926 }, { "epoch": 0.5320626824603201, "grad_norm": 1.1093517541885376, "learning_rate": 8.608723407865792e-05, "loss": 0.6771, "step": 7928 }, { "epoch": 0.5321969061440891, "grad_norm": 1.0552197694778442, "learning_rate": 8.607971062590155e-05, "loss": 0.6272, "step": 7930 }, { "epoch": 0.5323311298278581, "grad_norm": 1.1534368991851807, "learning_rate": 8.607218546844979e-05, "loss": 0.6534, "step": 7932 }, { "epoch": 0.5324653535116272, "grad_norm": 1.058334231376648, "learning_rate": 8.606465860665819e-05, "loss": 0.6327, "step": 7934 }, { "epoch": 0.5325995771953961, "grad_norm": 1.1343920230865479, "learning_rate": 8.605713004088238e-05, "loss": 0.6907, "step": 7936 }, { "epoch": 0.5327338008791651, "grad_norm": 1.3429607152938843, "learning_rate": 8.60495997714781e-05, "loss": 0.6221, "step": 7938 }, { "epoch": 0.5328680245629341, "grad_norm": 1.3038303852081299, "learning_rate": 8.60420677988011e-05, "loss": 0.6281, "step": 7940 }, { "epoch": 0.5330022482467032, "grad_norm": 1.0231696367263794, "learning_rate": 8.60345341232073e-05, "loss": 0.6278, "step": 7942 }, { "epoch": 0.5331364719304721, "grad_norm": 1.0313324928283691, "learning_rate": 8.60269987450526e-05, "loss": 0.6253, "step": 7944 }, { "epoch": 0.5332706956142411, "grad_norm": 1.0684877634048462, "learning_rate": 8.601946166469305e-05, "loss": 0.6228, "step": 7946 }, { "epoch": 0.5334049192980101, "grad_norm": 1.0405035018920898, "learning_rate": 8.601192288248478e-05, "loss": 0.6356, "step": 7948 }, { "epoch": 0.5335391429817792, "grad_norm": 1.089769721031189, "learning_rate": 8.600438239878394e-05, "loss": 0.6284, "step": 7950 }, { "epoch": 0.5336733666655481, "grad_norm": 1.114942193031311, "learning_rate": 8.599684021394686e-05, "loss": 0.748, "step": 7952 }, { "epoch": 0.5338075903493171, "grad_norm": 0.9413202404975891, "learning_rate": 8.598929632832985e-05, "loss": 0.6514, "step": 7954 }, { "epoch": 0.5339418140330862, "grad_norm": 0.9680725336074829, "learning_rate": 8.598175074228938e-05, "loss": 0.5765, "step": 7956 }, { "epoch": 0.5340760377168552, "grad_norm": 1.4321138858795166, "learning_rate": 8.597420345618193e-05, "loss": 0.6071, "step": 7958 }, { "epoch": 0.5342102614006241, "grad_norm": 1.0866897106170654, "learning_rate": 8.596665447036409e-05, "loss": 0.6463, "step": 7960 }, { "epoch": 0.5343444850843931, "grad_norm": 1.0365897417068481, "learning_rate": 8.595910378519257e-05, "loss": 0.6773, "step": 7962 }, { "epoch": 0.5344787087681622, "grad_norm": 0.9632202982902527, "learning_rate": 8.59515514010241e-05, "loss": 0.6185, "step": 7964 }, { "epoch": 0.5346129324519311, "grad_norm": 1.2617647647857666, "learning_rate": 8.594399731821552e-05, "loss": 0.6479, "step": 7966 }, { "epoch": 0.5347471561357001, "grad_norm": 0.9706656336784363, "learning_rate": 8.593644153712374e-05, "loss": 0.6248, "step": 7968 }, { "epoch": 0.5348813798194691, "grad_norm": 1.5335643291473389, "learning_rate": 8.592888405810578e-05, "loss": 0.6378, "step": 7970 }, { "epoch": 0.5350156035032382, "grad_norm": 1.1157900094985962, "learning_rate": 8.59213248815187e-05, "loss": 0.6764, "step": 7972 }, { "epoch": 0.5351498271870071, "grad_norm": 1.339160680770874, "learning_rate": 8.591376400771964e-05, "loss": 0.5938, "step": 7974 }, { "epoch": 0.5352840508707761, "grad_norm": 1.19435715675354, "learning_rate": 8.590620143706587e-05, "loss": 0.7154, "step": 7976 }, { "epoch": 0.5354182745545452, "grad_norm": 1.1539045572280884, "learning_rate": 8.589863716991469e-05, "loss": 0.7104, "step": 7978 }, { "epoch": 0.5355524982383142, "grad_norm": 1.0074177980422974, "learning_rate": 8.589107120662348e-05, "loss": 0.6027, "step": 7980 }, { "epoch": 0.5356867219220831, "grad_norm": 0.969353199005127, "learning_rate": 8.588350354754973e-05, "loss": 0.6428, "step": 7982 }, { "epoch": 0.5358209456058521, "grad_norm": 1.1570994853973389, "learning_rate": 8.587593419305101e-05, "loss": 0.6547, "step": 7984 }, { "epoch": 0.5359551692896212, "grad_norm": 0.9162207841873169, "learning_rate": 8.586836314348494e-05, "loss": 0.675, "step": 7986 }, { "epoch": 0.5360893929733902, "grad_norm": 0.9713578820228577, "learning_rate": 8.586079039920924e-05, "loss": 0.6232, "step": 7988 }, { "epoch": 0.5362236166571591, "grad_norm": 1.3750872611999512, "learning_rate": 8.585321596058174e-05, "loss": 0.7061, "step": 7990 }, { "epoch": 0.5363578403409281, "grad_norm": 0.9677409529685974, "learning_rate": 8.584563982796026e-05, "loss": 0.6361, "step": 7992 }, { "epoch": 0.5364920640246972, "grad_norm": 1.0356523990631104, "learning_rate": 8.583806200170279e-05, "loss": 0.6689, "step": 7994 }, { "epoch": 0.5366262877084662, "grad_norm": 1.0907829999923706, "learning_rate": 8.583048248216736e-05, "loss": 0.6733, "step": 7996 }, { "epoch": 0.5367605113922351, "grad_norm": 1.0503954887390137, "learning_rate": 8.582290126971209e-05, "loss": 0.5998, "step": 7998 }, { "epoch": 0.5368947350760042, "grad_norm": 1.0227012634277344, "learning_rate": 8.581531836469518e-05, "loss": 0.6032, "step": 8000 }, { "epoch": 0.5370289587597732, "grad_norm": 1.0412101745605469, "learning_rate": 8.580773376747492e-05, "loss": 0.6144, "step": 8002 }, { "epoch": 0.5371631824435421, "grad_norm": 1.048636794090271, "learning_rate": 8.580014747840964e-05, "loss": 0.6312, "step": 8004 }, { "epoch": 0.5372974061273111, "grad_norm": 1.1250464916229248, "learning_rate": 8.579255949785779e-05, "loss": 0.6273, "step": 8006 }, { "epoch": 0.5374316298110802, "grad_norm": 1.1802406311035156, "learning_rate": 8.578496982617788e-05, "loss": 0.6626, "step": 8008 }, { "epoch": 0.5375658534948492, "grad_norm": 0.9356011152267456, "learning_rate": 8.577737846372853e-05, "loss": 0.6837, "step": 8010 }, { "epoch": 0.5377000771786181, "grad_norm": 1.0496151447296143, "learning_rate": 8.576978541086838e-05, "loss": 0.6203, "step": 8012 }, { "epoch": 0.5378343008623871, "grad_norm": 1.1250532865524292, "learning_rate": 8.576219066795622e-05, "loss": 0.5904, "step": 8014 }, { "epoch": 0.5379685245461562, "grad_norm": 0.9540520906448364, "learning_rate": 8.575459423535089e-05, "loss": 0.6346, "step": 8016 }, { "epoch": 0.5381027482299252, "grad_norm": 1.5916898250579834, "learning_rate": 8.574699611341125e-05, "loss": 0.6193, "step": 8018 }, { "epoch": 0.5382369719136941, "grad_norm": 0.9837737083435059, "learning_rate": 8.573939630249638e-05, "loss": 0.686, "step": 8020 }, { "epoch": 0.5383711955974632, "grad_norm": 1.1416816711425781, "learning_rate": 8.573179480296529e-05, "loss": 0.7032, "step": 8022 }, { "epoch": 0.5385054192812322, "grad_norm": 1.125101089477539, "learning_rate": 8.572419161517716e-05, "loss": 0.6531, "step": 8024 }, { "epoch": 0.5386396429650012, "grad_norm": 1.0765750408172607, "learning_rate": 8.571658673949124e-05, "loss": 0.6607, "step": 8026 }, { "epoch": 0.5387738666487701, "grad_norm": 0.9669967889785767, "learning_rate": 8.570898017626681e-05, "loss": 0.6227, "step": 8028 }, { "epoch": 0.5389080903325392, "grad_norm": 1.1211695671081543, "learning_rate": 8.570137192586329e-05, "loss": 0.6798, "step": 8030 }, { "epoch": 0.5390423140163082, "grad_norm": 0.96068274974823, "learning_rate": 8.569376198864015e-05, "loss": 0.6164, "step": 8032 }, { "epoch": 0.5391765377000772, "grad_norm": 1.313602089881897, "learning_rate": 8.568615036495696e-05, "loss": 0.646, "step": 8034 }, { "epoch": 0.5393107613838461, "grad_norm": 0.9256817698478699, "learning_rate": 8.567853705517332e-05, "loss": 0.5842, "step": 8036 }, { "epoch": 0.5394449850676152, "grad_norm": 1.083078145980835, "learning_rate": 8.567092205964897e-05, "loss": 0.6093, "step": 8038 }, { "epoch": 0.5395792087513842, "grad_norm": 1.7828437089920044, "learning_rate": 8.566330537874369e-05, "loss": 0.6242, "step": 8040 }, { "epoch": 0.5397134324351531, "grad_norm": 1.140256404876709, "learning_rate": 8.565568701281738e-05, "loss": 0.5567, "step": 8042 }, { "epoch": 0.5398476561189222, "grad_norm": 0.9843252301216125, "learning_rate": 8.564806696222995e-05, "loss": 0.5182, "step": 8044 }, { "epoch": 0.5399818798026912, "grad_norm": 1.1507277488708496, "learning_rate": 8.564044522734147e-05, "loss": 0.63, "step": 8046 }, { "epoch": 0.5401161034864602, "grad_norm": 1.0563549995422363, "learning_rate": 8.563282180851203e-05, "loss": 0.6801, "step": 8048 }, { "epoch": 0.5402503271702291, "grad_norm": 1.0808017253875732, "learning_rate": 8.562519670610183e-05, "loss": 0.6887, "step": 8050 }, { "epoch": 0.5403845508539982, "grad_norm": 1.0849084854125977, "learning_rate": 8.561756992047114e-05, "loss": 0.6625, "step": 8052 }, { "epoch": 0.5405187745377672, "grad_norm": 0.9871025085449219, "learning_rate": 8.560994145198031e-05, "loss": 0.6241, "step": 8054 }, { "epoch": 0.5406529982215362, "grad_norm": 1.465746521949768, "learning_rate": 8.560231130098977e-05, "loss": 0.6822, "step": 8056 }, { "epoch": 0.5407872219053051, "grad_norm": 1.1308164596557617, "learning_rate": 8.559467946786002e-05, "loss": 0.6862, "step": 8058 }, { "epoch": 0.5409214455890742, "grad_norm": 1.1542991399765015, "learning_rate": 8.558704595295168e-05, "loss": 0.6147, "step": 8060 }, { "epoch": 0.5410556692728432, "grad_norm": 1.1769368648529053, "learning_rate": 8.557941075662538e-05, "loss": 0.6499, "step": 8062 }, { "epoch": 0.5411898929566122, "grad_norm": 0.9702188372612, "learning_rate": 8.55717738792419e-05, "loss": 0.643, "step": 8064 }, { "epoch": 0.5413241166403812, "grad_norm": 1.063549518585205, "learning_rate": 8.556413532116204e-05, "loss": 0.5977, "step": 8066 }, { "epoch": 0.5414583403241502, "grad_norm": 1.269137978553772, "learning_rate": 8.555649508274672e-05, "loss": 0.6633, "step": 8068 }, { "epoch": 0.5415925640079192, "grad_norm": 1.0866457223892212, "learning_rate": 8.554885316435694e-05, "loss": 0.6889, "step": 8070 }, { "epoch": 0.5417267876916882, "grad_norm": 0.9888244867324829, "learning_rate": 8.554120956635375e-05, "loss": 0.6395, "step": 8072 }, { "epoch": 0.5418610113754572, "grad_norm": 1.6463063955307007, "learning_rate": 8.553356428909827e-05, "loss": 0.647, "step": 8074 }, { "epoch": 0.5419952350592262, "grad_norm": 1.0329785346984863, "learning_rate": 8.552591733295179e-05, "loss": 0.6518, "step": 8076 }, { "epoch": 0.5421294587429952, "grad_norm": 1.084775447845459, "learning_rate": 8.551826869827555e-05, "loss": 0.714, "step": 8078 }, { "epoch": 0.5422636824267641, "grad_norm": 1.1005643606185913, "learning_rate": 8.551061838543097e-05, "loss": 0.6846, "step": 8080 }, { "epoch": 0.5423979061105332, "grad_norm": 1.1735957860946655, "learning_rate": 8.550296639477948e-05, "loss": 0.6822, "step": 8082 }, { "epoch": 0.5425321297943022, "grad_norm": 1.0340960025787354, "learning_rate": 8.549531272668266e-05, "loss": 0.6334, "step": 8084 }, { "epoch": 0.5426663534780712, "grad_norm": 1.2648706436157227, "learning_rate": 8.54876573815021e-05, "loss": 0.6676, "step": 8086 }, { "epoch": 0.5428005771618402, "grad_norm": 1.3927974700927734, "learning_rate": 8.548000035959953e-05, "loss": 0.6405, "step": 8088 }, { "epoch": 0.5429348008456092, "grad_norm": 1.0736503601074219, "learning_rate": 8.547234166133671e-05, "loss": 0.6455, "step": 8090 }, { "epoch": 0.5430690245293782, "grad_norm": 1.0800048112869263, "learning_rate": 8.546468128707548e-05, "loss": 0.5956, "step": 8092 }, { "epoch": 0.5432032482131472, "grad_norm": 1.027530312538147, "learning_rate": 8.545701923717781e-05, "loss": 0.7324, "step": 8094 }, { "epoch": 0.5433374718969162, "grad_norm": 0.9228115081787109, "learning_rate": 8.54493555120057e-05, "loss": 0.5722, "step": 8096 }, { "epoch": 0.5434716955806852, "grad_norm": 1.122521162033081, "learning_rate": 8.544169011192125e-05, "loss": 0.7119, "step": 8098 }, { "epoch": 0.5436059192644542, "grad_norm": 1.0947986841201782, "learning_rate": 8.543402303728665e-05, "loss": 0.6815, "step": 8100 }, { "epoch": 0.5437401429482233, "grad_norm": 1.00077486038208, "learning_rate": 8.542635428846413e-05, "loss": 0.6045, "step": 8102 }, { "epoch": 0.5438743666319922, "grad_norm": 1.0414117574691772, "learning_rate": 8.541868386581601e-05, "loss": 0.6676, "step": 8104 }, { "epoch": 0.5440085903157612, "grad_norm": 1.5793548822402954, "learning_rate": 8.541101176970476e-05, "loss": 0.7247, "step": 8106 }, { "epoch": 0.5441428139995302, "grad_norm": 0.999428927898407, "learning_rate": 8.540333800049282e-05, "loss": 0.6897, "step": 8108 }, { "epoch": 0.5442770376832993, "grad_norm": 1.000823974609375, "learning_rate": 8.539566255854277e-05, "loss": 0.6522, "step": 8110 }, { "epoch": 0.5444112613670682, "grad_norm": 0.9637446999549866, "learning_rate": 8.538798544421729e-05, "loss": 0.6125, "step": 8112 }, { "epoch": 0.5445454850508372, "grad_norm": 1.002723217010498, "learning_rate": 8.538030665787906e-05, "loss": 0.5744, "step": 8114 }, { "epoch": 0.5446797087346062, "grad_norm": 1.1904152631759644, "learning_rate": 8.537262619989093e-05, "loss": 0.7247, "step": 8116 }, { "epoch": 0.5448139324183752, "grad_norm": 1.1540770530700684, "learning_rate": 8.536494407061575e-05, "loss": 0.6186, "step": 8118 }, { "epoch": 0.5449481561021442, "grad_norm": 0.9250359535217285, "learning_rate": 8.535726027041652e-05, "loss": 0.7193, "step": 8120 }, { "epoch": 0.5450823797859132, "grad_norm": 0.9992945790290833, "learning_rate": 8.534957479965629e-05, "loss": 0.6211, "step": 8122 }, { "epoch": 0.5452166034696823, "grad_norm": 0.9270564317703247, "learning_rate": 8.534188765869813e-05, "loss": 0.6254, "step": 8124 }, { "epoch": 0.5453508271534512, "grad_norm": 1.0162103176116943, "learning_rate": 8.533419884790528e-05, "loss": 0.6142, "step": 8126 }, { "epoch": 0.5454850508372202, "grad_norm": 1.0450429916381836, "learning_rate": 8.532650836764102e-05, "loss": 0.6941, "step": 8128 }, { "epoch": 0.5456192745209892, "grad_norm": 1.0318299531936646, "learning_rate": 8.531881621826871e-05, "loss": 0.6611, "step": 8130 }, { "epoch": 0.5457534982047583, "grad_norm": 1.0086703300476074, "learning_rate": 8.531112240015178e-05, "loss": 0.677, "step": 8132 }, { "epoch": 0.5458877218885272, "grad_norm": 0.9467586278915405, "learning_rate": 8.530342691365377e-05, "loss": 0.6441, "step": 8134 }, { "epoch": 0.5460219455722962, "grad_norm": 0.8377612233161926, "learning_rate": 8.529572975913824e-05, "loss": 0.565, "step": 8136 }, { "epoch": 0.5461561692560652, "grad_norm": 0.8942846655845642, "learning_rate": 8.528803093696889e-05, "loss": 0.5783, "step": 8138 }, { "epoch": 0.5462903929398343, "grad_norm": 1.1009197235107422, "learning_rate": 8.528033044750947e-05, "loss": 0.6599, "step": 8140 }, { "epoch": 0.5464246166236032, "grad_norm": 1.0929917097091675, "learning_rate": 8.527262829112382e-05, "loss": 0.7262, "step": 8142 }, { "epoch": 0.5465588403073722, "grad_norm": 1.159877896308899, "learning_rate": 8.526492446817583e-05, "loss": 0.6566, "step": 8144 }, { "epoch": 0.5466930639911413, "grad_norm": 1.178993821144104, "learning_rate": 8.52572189790295e-05, "loss": 0.5907, "step": 8146 }, { "epoch": 0.5468272876749103, "grad_norm": 1.305436372756958, "learning_rate": 8.524951182404892e-05, "loss": 0.593, "step": 8148 }, { "epoch": 0.5469615113586792, "grad_norm": 1.1507524251937866, "learning_rate": 8.52418030035982e-05, "loss": 0.6921, "step": 8150 }, { "epoch": 0.5470957350424482, "grad_norm": 1.0701565742492676, "learning_rate": 8.523409251804163e-05, "loss": 0.6252, "step": 8152 }, { "epoch": 0.5472299587262173, "grad_norm": 1.0704950094223022, "learning_rate": 8.522638036774345e-05, "loss": 0.6535, "step": 8154 }, { "epoch": 0.5473641824099862, "grad_norm": 0.9062939882278442, "learning_rate": 8.521866655306805e-05, "loss": 0.6828, "step": 8156 }, { "epoch": 0.5474984060937552, "grad_norm": 1.1163350343704224, "learning_rate": 8.521095107437993e-05, "loss": 0.6318, "step": 8158 }, { "epoch": 0.5476326297775242, "grad_norm": 0.9973238706588745, "learning_rate": 8.52032339320436e-05, "loss": 0.616, "step": 8160 }, { "epoch": 0.5477668534612933, "grad_norm": 1.248987078666687, "learning_rate": 8.51955151264237e-05, "loss": 0.5971, "step": 8162 }, { "epoch": 0.5479010771450622, "grad_norm": 1.2610836029052734, "learning_rate": 8.518779465788491e-05, "loss": 0.7045, "step": 8164 }, { "epoch": 0.5480353008288312, "grad_norm": 0.9630738496780396, "learning_rate": 8.518007252679202e-05, "loss": 0.6211, "step": 8166 }, { "epoch": 0.5481695245126003, "grad_norm": 1.0208569765090942, "learning_rate": 8.517234873350987e-05, "loss": 0.6194, "step": 8168 }, { "epoch": 0.5483037481963693, "grad_norm": 1.011013388633728, "learning_rate": 8.516462327840343e-05, "loss": 0.5775, "step": 8170 }, { "epoch": 0.5484379718801382, "grad_norm": 1.3043915033340454, "learning_rate": 8.515689616183769e-05, "loss": 0.642, "step": 8172 }, { "epoch": 0.5485721955639072, "grad_norm": 1.1023441553115845, "learning_rate": 8.51491673841777e-05, "loss": 0.6545, "step": 8174 }, { "epoch": 0.5487064192476763, "grad_norm": 1.0701701641082764, "learning_rate": 8.51414369457887e-05, "loss": 0.6236, "step": 8176 }, { "epoch": 0.5488406429314453, "grad_norm": 1.1929540634155273, "learning_rate": 8.513370484703591e-05, "loss": 0.6672, "step": 8178 }, { "epoch": 0.5489748666152142, "grad_norm": 1.1066863536834717, "learning_rate": 8.512597108828464e-05, "loss": 0.6638, "step": 8180 }, { "epoch": 0.5491090902989832, "grad_norm": 1.0696840286254883, "learning_rate": 8.511823566990031e-05, "loss": 0.697, "step": 8182 }, { "epoch": 0.5492433139827523, "grad_norm": 1.2313902378082275, "learning_rate": 8.511049859224842e-05, "loss": 0.7155, "step": 8184 }, { "epoch": 0.5493775376665213, "grad_norm": 1.0531792640686035, "learning_rate": 8.510275985569449e-05, "loss": 0.6543, "step": 8186 }, { "epoch": 0.5495117613502902, "grad_norm": 1.1346839666366577, "learning_rate": 8.509501946060421e-05, "loss": 0.6262, "step": 8188 }, { "epoch": 0.5496459850340593, "grad_norm": 1.1444693803787231, "learning_rate": 8.508727740734324e-05, "loss": 0.64, "step": 8190 }, { "epoch": 0.5497802087178283, "grad_norm": 1.1017465591430664, "learning_rate": 8.507953369627743e-05, "loss": 0.6515, "step": 8192 }, { "epoch": 0.5499144324015972, "grad_norm": 1.0046477317810059, "learning_rate": 8.507178832777263e-05, "loss": 0.6283, "step": 8194 }, { "epoch": 0.5500486560853662, "grad_norm": 1.0455659627914429, "learning_rate": 8.506404130219479e-05, "loss": 0.6674, "step": 8196 }, { "epoch": 0.5501828797691353, "grad_norm": 1.0172463655471802, "learning_rate": 8.505629261990995e-05, "loss": 0.691, "step": 8198 }, { "epoch": 0.5503171034529043, "grad_norm": 1.0176228284835815, "learning_rate": 8.504854228128422e-05, "loss": 0.6475, "step": 8200 }, { "epoch": 0.5504513271366732, "grad_norm": 0.9945377707481384, "learning_rate": 8.504079028668379e-05, "loss": 0.6226, "step": 8202 }, { "epoch": 0.5505855508204422, "grad_norm": 1.0326147079467773, "learning_rate": 8.503303663647492e-05, "loss": 0.5734, "step": 8204 }, { "epoch": 0.5507197745042113, "grad_norm": 1.0489661693572998, "learning_rate": 8.502528133102397e-05, "loss": 0.5622, "step": 8206 }, { "epoch": 0.5508539981879803, "grad_norm": 1.0645067691802979, "learning_rate": 8.501752437069734e-05, "loss": 0.6315, "step": 8208 }, { "epoch": 0.5509882218717492, "grad_norm": 2.246250629425049, "learning_rate": 8.500976575586155e-05, "loss": 0.6108, "step": 8210 }, { "epoch": 0.5511224455555183, "grad_norm": 0.9789879322052002, "learning_rate": 8.500200548688318e-05, "loss": 0.7113, "step": 8212 }, { "epoch": 0.5512566692392873, "grad_norm": 1.8534557819366455, "learning_rate": 8.499424356412886e-05, "loss": 0.6391, "step": 8214 }, { "epoch": 0.5513908929230563, "grad_norm": 1.0547062158584595, "learning_rate": 8.498647998796534e-05, "loss": 0.6031, "step": 8216 }, { "epoch": 0.5515251166068252, "grad_norm": 1.0872576236724854, "learning_rate": 8.497871475875946e-05, "loss": 0.6412, "step": 8218 }, { "epoch": 0.5516593402905943, "grad_norm": 1.004032850265503, "learning_rate": 8.497094787687807e-05, "loss": 0.6903, "step": 8220 }, { "epoch": 0.5517935639743633, "grad_norm": 1.2047981023788452, "learning_rate": 8.496317934268818e-05, "loss": 0.6767, "step": 8222 }, { "epoch": 0.5519277876581323, "grad_norm": 1.089982271194458, "learning_rate": 8.49554091565568e-05, "loss": 0.6252, "step": 8224 }, { "epoch": 0.5520620113419012, "grad_norm": 1.1389589309692383, "learning_rate": 8.49476373188511e-05, "loss": 0.6627, "step": 8226 }, { "epoch": 0.5521962350256703, "grad_norm": 1.0517256259918213, "learning_rate": 8.493986382993823e-05, "loss": 0.6275, "step": 8228 }, { "epoch": 0.5523304587094393, "grad_norm": 0.9931901693344116, "learning_rate": 8.49320886901855e-05, "loss": 0.6115, "step": 8230 }, { "epoch": 0.5524646823932082, "grad_norm": 0.9084005951881409, "learning_rate": 8.492431189996029e-05, "loss": 0.5321, "step": 8232 }, { "epoch": 0.5525989060769773, "grad_norm": 1.0069876909255981, "learning_rate": 8.491653345963002e-05, "loss": 0.6006, "step": 8234 }, { "epoch": 0.5527331297607463, "grad_norm": 1.0372812747955322, "learning_rate": 8.490875336956219e-05, "loss": 0.6397, "step": 8236 }, { "epoch": 0.5528673534445153, "grad_norm": 1.2959024906158447, "learning_rate": 8.49009716301244e-05, "loss": 0.67, "step": 8238 }, { "epoch": 0.5530015771282842, "grad_norm": 0.9682850241661072, "learning_rate": 8.489318824168435e-05, "loss": 0.6599, "step": 8240 }, { "epoch": 0.5531358008120533, "grad_norm": 1.1090869903564453, "learning_rate": 8.488540320460975e-05, "loss": 0.6012, "step": 8242 }, { "epoch": 0.5532700244958223, "grad_norm": 1.2157872915267944, "learning_rate": 8.487761651926848e-05, "loss": 0.6433, "step": 8244 }, { "epoch": 0.5534042481795913, "grad_norm": 1.5864654779434204, "learning_rate": 8.48698281860284e-05, "loss": 0.6466, "step": 8246 }, { "epoch": 0.5535384718633602, "grad_norm": 1.0488078594207764, "learning_rate": 8.486203820525748e-05, "loss": 0.5906, "step": 8248 }, { "epoch": 0.5536726955471293, "grad_norm": 1.1140676736831665, "learning_rate": 8.485424657732384e-05, "loss": 0.6756, "step": 8250 }, { "epoch": 0.5538069192308983, "grad_norm": 1.0485445261001587, "learning_rate": 8.484645330259557e-05, "loss": 0.6702, "step": 8252 }, { "epoch": 0.5539411429146673, "grad_norm": 1.054240107536316, "learning_rate": 8.483865838144092e-05, "loss": 0.6003, "step": 8254 }, { "epoch": 0.5540753665984363, "grad_norm": 1.060673475265503, "learning_rate": 8.483086181422816e-05, "loss": 0.604, "step": 8256 }, { "epoch": 0.5542095902822053, "grad_norm": 1.134052038192749, "learning_rate": 8.482306360132567e-05, "loss": 0.6691, "step": 8258 }, { "epoch": 0.5543438139659743, "grad_norm": 0.9479762315750122, "learning_rate": 8.48152637431019e-05, "loss": 0.6215, "step": 8260 }, { "epoch": 0.5544780376497433, "grad_norm": 1.013800024986267, "learning_rate": 8.480746223992539e-05, "loss": 0.6914, "step": 8262 }, { "epoch": 0.5546122613335123, "grad_norm": 1.0438390970230103, "learning_rate": 8.479965909216471e-05, "loss": 0.6848, "step": 8264 }, { "epoch": 0.5547464850172813, "grad_norm": 1.0111439228057861, "learning_rate": 8.479185430018858e-05, "loss": 0.6174, "step": 8266 }, { "epoch": 0.5548807087010503, "grad_norm": 1.811466097831726, "learning_rate": 8.478404786436577e-05, "loss": 0.6816, "step": 8268 }, { "epoch": 0.5550149323848192, "grad_norm": 1.1030131578445435, "learning_rate": 8.477623978506508e-05, "loss": 0.6428, "step": 8270 }, { "epoch": 0.5551491560685883, "grad_norm": 1.0027936697006226, "learning_rate": 8.476843006265545e-05, "loss": 0.5804, "step": 8272 }, { "epoch": 0.5552833797523573, "grad_norm": 1.1092429161071777, "learning_rate": 8.476061869750586e-05, "loss": 0.698, "step": 8274 }, { "epoch": 0.5554176034361263, "grad_norm": 1.0696680545806885, "learning_rate": 8.475280568998541e-05, "loss": 0.6074, "step": 8276 }, { "epoch": 0.5555518271198953, "grad_norm": 2.5581133365631104, "learning_rate": 8.474499104046322e-05, "loss": 0.6166, "step": 8278 }, { "epoch": 0.5556860508036643, "grad_norm": 1.0897672176361084, "learning_rate": 8.473717474930851e-05, "loss": 0.6882, "step": 8280 }, { "epoch": 0.5558202744874333, "grad_norm": 1.124261736869812, "learning_rate": 8.472935681689063e-05, "loss": 0.5806, "step": 8282 }, { "epoch": 0.5559544981712023, "grad_norm": 1.0583689212799072, "learning_rate": 8.472153724357892e-05, "loss": 0.7294, "step": 8284 }, { "epoch": 0.5560887218549713, "grad_norm": 1.1011725664138794, "learning_rate": 8.471371602974284e-05, "loss": 0.6707, "step": 8286 }, { "epoch": 0.5562229455387403, "grad_norm": 1.0401387214660645, "learning_rate": 8.470589317575194e-05, "loss": 0.5701, "step": 8288 }, { "epoch": 0.5563571692225093, "grad_norm": 1.064677119255066, "learning_rate": 8.469806868197585e-05, "loss": 0.6673, "step": 8290 }, { "epoch": 0.5564913929062784, "grad_norm": 1.0974767208099365, "learning_rate": 8.469024254878424e-05, "loss": 0.6954, "step": 8292 }, { "epoch": 0.5566256165900473, "grad_norm": 1.0445501804351807, "learning_rate": 8.468241477654688e-05, "loss": 0.6022, "step": 8294 }, { "epoch": 0.5567598402738163, "grad_norm": 1.1602227687835693, "learning_rate": 8.467458536563362e-05, "loss": 0.6687, "step": 8296 }, { "epoch": 0.5568940639575853, "grad_norm": 0.9773215651512146, "learning_rate": 8.46667543164144e-05, "loss": 0.6445, "step": 8298 }, { "epoch": 0.5570282876413544, "grad_norm": 1.013282299041748, "learning_rate": 8.465892162925919e-05, "loss": 0.6519, "step": 8300 }, { "epoch": 0.5571625113251233, "grad_norm": 1.0231740474700928, "learning_rate": 8.46510873045381e-05, "loss": 0.643, "step": 8302 }, { "epoch": 0.5572967350088923, "grad_norm": 1.308756709098816, "learning_rate": 8.464325134262126e-05, "loss": 0.6785, "step": 8304 }, { "epoch": 0.5574309586926613, "grad_norm": 0.9831346273422241, "learning_rate": 8.463541374387894e-05, "loss": 0.7026, "step": 8306 }, { "epoch": 0.5575651823764303, "grad_norm": 1.050607442855835, "learning_rate": 8.462757450868139e-05, "loss": 0.6293, "step": 8308 }, { "epoch": 0.5576994060601993, "grad_norm": 1.0742460489273071, "learning_rate": 8.461973363739906e-05, "loss": 0.6521, "step": 8310 }, { "epoch": 0.5578336297439683, "grad_norm": 1.148586392402649, "learning_rate": 8.46118911304024e-05, "loss": 0.6586, "step": 8312 }, { "epoch": 0.5579678534277374, "grad_norm": 1.0171434879302979, "learning_rate": 8.460404698806194e-05, "loss": 0.6123, "step": 8314 }, { "epoch": 0.5581020771115063, "grad_norm": 0.8852545022964478, "learning_rate": 8.45962012107483e-05, "loss": 0.5919, "step": 8316 }, { "epoch": 0.5582363007952753, "grad_norm": 1.4615974426269531, "learning_rate": 8.458835379883219e-05, "loss": 0.6773, "step": 8318 }, { "epoch": 0.5583705244790443, "grad_norm": 1.1346313953399658, "learning_rate": 8.458050475268437e-05, "loss": 0.6652, "step": 8320 }, { "epoch": 0.5585047481628134, "grad_norm": 2.0286874771118164, "learning_rate": 8.457265407267572e-05, "loss": 0.6489, "step": 8322 }, { "epoch": 0.5586389718465823, "grad_norm": 1.1897507905960083, "learning_rate": 8.456480175917713e-05, "loss": 0.6463, "step": 8324 }, { "epoch": 0.5587731955303513, "grad_norm": 1.1002095937728882, "learning_rate": 8.455694781255963e-05, "loss": 0.6908, "step": 8326 }, { "epoch": 0.5589074192141203, "grad_norm": 1.0195788145065308, "learning_rate": 8.454909223319429e-05, "loss": 0.5828, "step": 8328 }, { "epoch": 0.5590416428978894, "grad_norm": 1.0193231105804443, "learning_rate": 8.45412350214523e-05, "loss": 0.6408, "step": 8330 }, { "epoch": 0.5591758665816583, "grad_norm": 1.0430853366851807, "learning_rate": 8.453337617770486e-05, "loss": 0.6657, "step": 8332 }, { "epoch": 0.5593100902654273, "grad_norm": 1.0074166059494019, "learning_rate": 8.452551570232331e-05, "loss": 0.6076, "step": 8334 }, { "epoch": 0.5594443139491964, "grad_norm": 0.95444256067276, "learning_rate": 8.451765359567903e-05, "loss": 0.6303, "step": 8336 }, { "epoch": 0.5595785376329654, "grad_norm": 1.0800069570541382, "learning_rate": 8.45097898581435e-05, "loss": 0.652, "step": 8338 }, { "epoch": 0.5597127613167343, "grad_norm": 1.1131151914596558, "learning_rate": 8.450192449008825e-05, "loss": 0.6429, "step": 8340 }, { "epoch": 0.5598469850005033, "grad_norm": 1.0435813665390015, "learning_rate": 8.449405749188493e-05, "loss": 0.5579, "step": 8342 }, { "epoch": 0.5599812086842724, "grad_norm": 1.211788535118103, "learning_rate": 8.448618886390522e-05, "loss": 0.6314, "step": 8344 }, { "epoch": 0.5601154323680413, "grad_norm": 1.8600342273712158, "learning_rate": 8.447831860652089e-05, "loss": 0.6245, "step": 8346 }, { "epoch": 0.5602496560518103, "grad_norm": 1.1609203815460205, "learning_rate": 8.447044672010382e-05, "loss": 0.652, "step": 8348 }, { "epoch": 0.5603838797355793, "grad_norm": 1.0970147848129272, "learning_rate": 8.44625732050259e-05, "loss": 0.6329, "step": 8350 }, { "epoch": 0.5605181034193484, "grad_norm": 0.9530807733535767, "learning_rate": 8.445469806165918e-05, "loss": 0.6211, "step": 8352 }, { "epoch": 0.5606523271031173, "grad_norm": 1.526509165763855, "learning_rate": 8.444682129037574e-05, "loss": 0.6128, "step": 8354 }, { "epoch": 0.5607865507868863, "grad_norm": 1.1120033264160156, "learning_rate": 8.44389428915477e-05, "loss": 0.6429, "step": 8356 }, { "epoch": 0.5609207744706554, "grad_norm": 1.0139830112457275, "learning_rate": 8.443106286554737e-05, "loss": 0.6424, "step": 8358 }, { "epoch": 0.5610549981544244, "grad_norm": 1.029532551765442, "learning_rate": 8.4423181212747e-05, "loss": 0.6396, "step": 8360 }, { "epoch": 0.5611892218381933, "grad_norm": 1.0851560831069946, "learning_rate": 8.4415297933519e-05, "loss": 0.6841, "step": 8362 }, { "epoch": 0.5613234455219623, "grad_norm": 1.0855809450149536, "learning_rate": 8.440741302823585e-05, "loss": 0.6103, "step": 8364 }, { "epoch": 0.5614576692057314, "grad_norm": 0.9500477313995361, "learning_rate": 8.43995264972701e-05, "loss": 0.5849, "step": 8366 }, { "epoch": 0.5615918928895004, "grad_norm": 1.2251765727996826, "learning_rate": 8.439163834099437e-05, "loss": 0.6787, "step": 8368 }, { "epoch": 0.5617261165732693, "grad_norm": 1.1215698719024658, "learning_rate": 8.438374855978136e-05, "loss": 0.5894, "step": 8370 }, { "epoch": 0.5618603402570383, "grad_norm": 0.994908332824707, "learning_rate": 8.437585715400384e-05, "loss": 0.5808, "step": 8372 }, { "epoch": 0.5619945639408074, "grad_norm": 0.9743037819862366, "learning_rate": 8.436796412403466e-05, "loss": 0.5433, "step": 8374 }, { "epoch": 0.5621287876245764, "grad_norm": 0.9671458601951599, "learning_rate": 8.436006947024677e-05, "loss": 0.5827, "step": 8376 }, { "epoch": 0.5622630113083453, "grad_norm": 1.092995047569275, "learning_rate": 8.435217319301315e-05, "loss": 0.5866, "step": 8378 }, { "epoch": 0.5623972349921144, "grad_norm": 1.0426753759384155, "learning_rate": 8.43442752927069e-05, "loss": 0.6717, "step": 8380 }, { "epoch": 0.5625314586758834, "grad_norm": 0.9841560125350952, "learning_rate": 8.433637576970119e-05, "loss": 0.6382, "step": 8382 }, { "epoch": 0.5626656823596523, "grad_norm": 1.1644257307052612, "learning_rate": 8.432847462436924e-05, "loss": 0.6793, "step": 8384 }, { "epoch": 0.5627999060434213, "grad_norm": 1.0643762350082397, "learning_rate": 8.432057185708436e-05, "loss": 0.5893, "step": 8386 }, { "epoch": 0.5629341297271904, "grad_norm": 1.0374327898025513, "learning_rate": 8.431266746821995e-05, "loss": 0.63, "step": 8388 }, { "epoch": 0.5630683534109594, "grad_norm": 0.9825643301010132, "learning_rate": 8.430476145814948e-05, "loss": 0.6431, "step": 8390 }, { "epoch": 0.5632025770947283, "grad_norm": 1.111504077911377, "learning_rate": 8.429685382724648e-05, "loss": 0.6441, "step": 8392 }, { "epoch": 0.5633368007784973, "grad_norm": 1.1499359607696533, "learning_rate": 8.428894457588458e-05, "loss": 0.6429, "step": 8394 }, { "epoch": 0.5634710244622664, "grad_norm": 1.018369197845459, "learning_rate": 8.428103370443748e-05, "loss": 0.6267, "step": 8396 }, { "epoch": 0.5636052481460354, "grad_norm": 0.9640548229217529, "learning_rate": 8.427312121327895e-05, "loss": 0.6659, "step": 8398 }, { "epoch": 0.5637394718298043, "grad_norm": 1.0157394409179688, "learning_rate": 8.426520710278283e-05, "loss": 0.6143, "step": 8400 }, { "epoch": 0.5638736955135734, "grad_norm": 1.0569133758544922, "learning_rate": 8.425729137332306e-05, "loss": 0.6107, "step": 8402 }, { "epoch": 0.5640079191973424, "grad_norm": 0.9883301854133606, "learning_rate": 8.424937402527363e-05, "loss": 0.6164, "step": 8404 }, { "epoch": 0.5641421428811114, "grad_norm": 1.2704092264175415, "learning_rate": 8.424145505900863e-05, "loss": 0.6932, "step": 8406 }, { "epoch": 0.5642763665648803, "grad_norm": 1.6753979921340942, "learning_rate": 8.423353447490221e-05, "loss": 0.6067, "step": 8408 }, { "epoch": 0.5644105902486494, "grad_norm": 0.9905928373336792, "learning_rate": 8.422561227332861e-05, "loss": 0.6036, "step": 8410 }, { "epoch": 0.5645448139324184, "grad_norm": 1.1601665019989014, "learning_rate": 8.421768845466213e-05, "loss": 0.6253, "step": 8412 }, { "epoch": 0.5646790376161874, "grad_norm": 0.9416701197624207, "learning_rate": 8.420976301927716e-05, "loss": 0.6642, "step": 8414 }, { "epoch": 0.5648132612999563, "grad_norm": 1.2624140977859497, "learning_rate": 8.420183596754816e-05, "loss": 0.6034, "step": 8416 }, { "epoch": 0.5649474849837254, "grad_norm": 1.0920568704605103, "learning_rate": 8.419390729984966e-05, "loss": 0.593, "step": 8418 }, { "epoch": 0.5650817086674944, "grad_norm": 0.9582425355911255, "learning_rate": 8.418597701655628e-05, "loss": 0.6524, "step": 8420 }, { "epoch": 0.5652159323512633, "grad_norm": 1.02560555934906, "learning_rate": 8.41780451180427e-05, "loss": 0.6018, "step": 8422 }, { "epoch": 0.5653501560350324, "grad_norm": 1.0681962966918945, "learning_rate": 8.417011160468371e-05, "loss": 0.5941, "step": 8424 }, { "epoch": 0.5654843797188014, "grad_norm": 0.892360270023346, "learning_rate": 8.416217647685411e-05, "loss": 0.6703, "step": 8426 }, { "epoch": 0.5656186034025704, "grad_norm": 0.9655805826187134, "learning_rate": 8.415423973492888e-05, "loss": 0.5849, "step": 8428 }, { "epoch": 0.5657528270863393, "grad_norm": 1.1360067129135132, "learning_rate": 8.414630137928298e-05, "loss": 0.6583, "step": 8430 }, { "epoch": 0.5658870507701084, "grad_norm": 0.9449290633201599, "learning_rate": 8.413836141029148e-05, "loss": 0.6052, "step": 8432 }, { "epoch": 0.5660212744538774, "grad_norm": 0.9781023263931274, "learning_rate": 8.413041982832954e-05, "loss": 0.6945, "step": 8434 }, { "epoch": 0.5661554981376464, "grad_norm": 1.038053035736084, "learning_rate": 8.412247663377237e-05, "loss": 0.6892, "step": 8436 }, { "epoch": 0.5662897218214153, "grad_norm": 1.099488615989685, "learning_rate": 8.411453182699529e-05, "loss": 0.6034, "step": 8438 }, { "epoch": 0.5664239455051844, "grad_norm": 1.1301482915878296, "learning_rate": 8.410658540837364e-05, "loss": 0.6724, "step": 8440 }, { "epoch": 0.5665581691889534, "grad_norm": 1.0925681591033936, "learning_rate": 8.409863737828292e-05, "loss": 0.6736, "step": 8442 }, { "epoch": 0.5666923928727224, "grad_norm": 1.0710530281066895, "learning_rate": 8.409068773709862e-05, "loss": 0.7052, "step": 8444 }, { "epoch": 0.5668266165564914, "grad_norm": 1.0949926376342773, "learning_rate": 8.408273648519638e-05, "loss": 0.697, "step": 8446 }, { "epoch": 0.5669608402402604, "grad_norm": 1.0935205221176147, "learning_rate": 8.407478362295184e-05, "loss": 0.656, "step": 8448 }, { "epoch": 0.5670950639240294, "grad_norm": 1.2679986953735352, "learning_rate": 8.40668291507408e-05, "loss": 0.6642, "step": 8450 }, { "epoch": 0.5672292876077984, "grad_norm": 1.0844709873199463, "learning_rate": 8.405887306893906e-05, "loss": 0.6574, "step": 8452 }, { "epoch": 0.5673635112915674, "grad_norm": 1.0213552713394165, "learning_rate": 8.405091537792253e-05, "loss": 0.6433, "step": 8454 }, { "epoch": 0.5674977349753364, "grad_norm": 0.9217315912246704, "learning_rate": 8.404295607806723e-05, "loss": 0.6107, "step": 8456 }, { "epoch": 0.5676319586591054, "grad_norm": 1.058948278427124, "learning_rate": 8.40349951697492e-05, "loss": 0.7232, "step": 8458 }, { "epoch": 0.5677661823428743, "grad_norm": 1.2261768579483032, "learning_rate": 8.402703265334455e-05, "loss": 0.658, "step": 8460 }, { "epoch": 0.5679004060266434, "grad_norm": 1.1236143112182617, "learning_rate": 8.401906852922954e-05, "loss": 0.6335, "step": 8462 }, { "epoch": 0.5680346297104124, "grad_norm": 1.1284643411636353, "learning_rate": 8.401110279778043e-05, "loss": 0.625, "step": 8464 }, { "epoch": 0.5681688533941814, "grad_norm": 1.1031262874603271, "learning_rate": 8.400313545937359e-05, "loss": 0.663, "step": 8466 }, { "epoch": 0.5683030770779504, "grad_norm": 1.0434975624084473, "learning_rate": 8.399516651438546e-05, "loss": 0.6443, "step": 8468 }, { "epoch": 0.5684373007617194, "grad_norm": 1.0460388660430908, "learning_rate": 8.398719596319258e-05, "loss": 0.6978, "step": 8470 }, { "epoch": 0.5685715244454884, "grad_norm": 1.0447012186050415, "learning_rate": 8.39792238061715e-05, "loss": 0.6135, "step": 8472 }, { "epoch": 0.5687057481292574, "grad_norm": 1.3942762613296509, "learning_rate": 8.397125004369892e-05, "loss": 0.5721, "step": 8474 }, { "epoch": 0.5688399718130264, "grad_norm": 1.027471661567688, "learning_rate": 8.396327467615158e-05, "loss": 0.62, "step": 8476 }, { "epoch": 0.5689741954967954, "grad_norm": 0.9808257222175598, "learning_rate": 8.39552977039063e-05, "loss": 0.6354, "step": 8478 }, { "epoch": 0.5691084191805644, "grad_norm": 1.0226244926452637, "learning_rate": 8.394731912733998e-05, "loss": 0.6141, "step": 8480 }, { "epoch": 0.5692426428643335, "grad_norm": 1.0700534582138062, "learning_rate": 8.393933894682957e-05, "loss": 0.6044, "step": 8482 }, { "epoch": 0.5693768665481024, "grad_norm": 0.8504033088684082, "learning_rate": 8.393135716275215e-05, "loss": 0.5722, "step": 8484 }, { "epoch": 0.5695110902318714, "grad_norm": 1.060200810432434, "learning_rate": 8.392337377548481e-05, "loss": 0.6351, "step": 8486 }, { "epoch": 0.5696453139156404, "grad_norm": 1.0870810747146606, "learning_rate": 8.391538878540477e-05, "loss": 0.622, "step": 8488 }, { "epoch": 0.5697795375994095, "grad_norm": 1.059228777885437, "learning_rate": 8.390740219288931e-05, "loss": 0.662, "step": 8490 }, { "epoch": 0.5699137612831784, "grad_norm": 1.3571457862854004, "learning_rate": 8.389941399831578e-05, "loss": 0.6109, "step": 8492 }, { "epoch": 0.5700479849669474, "grad_norm": 1.2545329332351685, "learning_rate": 8.389142420206158e-05, "loss": 0.729, "step": 8494 }, { "epoch": 0.5701822086507164, "grad_norm": 1.0990122556686401, "learning_rate": 8.388343280450424e-05, "loss": 0.6821, "step": 8496 }, { "epoch": 0.5703164323344854, "grad_norm": 1.2406752109527588, "learning_rate": 8.387543980602133e-05, "loss": 0.5928, "step": 8498 }, { "epoch": 0.5704506560182544, "grad_norm": 1.21001398563385, "learning_rate": 8.38674452069905e-05, "loss": 0.6895, "step": 8500 }, { "epoch": 0.5705848797020234, "grad_norm": 1.1252810955047607, "learning_rate": 8.385944900778948e-05, "loss": 0.6108, "step": 8502 }, { "epoch": 0.5707191033857925, "grad_norm": 1.0943095684051514, "learning_rate": 8.385145120879607e-05, "loss": 0.6347, "step": 8504 }, { "epoch": 0.5708533270695614, "grad_norm": 1.1311140060424805, "learning_rate": 8.384345181038818e-05, "loss": 0.6565, "step": 8506 }, { "epoch": 0.5709875507533304, "grad_norm": 1.0514534711837769, "learning_rate": 8.383545081294371e-05, "loss": 0.607, "step": 8508 }, { "epoch": 0.5711217744370994, "grad_norm": 0.8207479119300842, "learning_rate": 8.382744821684074e-05, "loss": 0.5246, "step": 8510 }, { "epoch": 0.5712559981208685, "grad_norm": 1.771498680114746, "learning_rate": 8.381944402245736e-05, "loss": 0.5949, "step": 8512 }, { "epoch": 0.5713902218046374, "grad_norm": 1.1286104917526245, "learning_rate": 8.381143823017176e-05, "loss": 0.5539, "step": 8514 }, { "epoch": 0.5715244454884064, "grad_norm": 1.0424946546554565, "learning_rate": 8.380343084036218e-05, "loss": 0.6896, "step": 8516 }, { "epoch": 0.5716586691721754, "grad_norm": 1.213830590248108, "learning_rate": 8.379542185340696e-05, "loss": 0.6418, "step": 8518 }, { "epoch": 0.5717928928559445, "grad_norm": 1.050347924232483, "learning_rate": 8.378741126968453e-05, "loss": 0.6655, "step": 8520 }, { "epoch": 0.5719271165397134, "grad_norm": 1.3137097358703613, "learning_rate": 8.377939908957336e-05, "loss": 0.6701, "step": 8522 }, { "epoch": 0.5720613402234824, "grad_norm": 1.074523687362671, "learning_rate": 8.377138531345198e-05, "loss": 0.6345, "step": 8524 }, { "epoch": 0.5721955639072515, "grad_norm": 1.134978175163269, "learning_rate": 8.376336994169908e-05, "loss": 0.6571, "step": 8526 }, { "epoch": 0.5723297875910205, "grad_norm": 1.0185121297836304, "learning_rate": 8.375535297469332e-05, "loss": 0.6489, "step": 8528 }, { "epoch": 0.5724640112747894, "grad_norm": 1.0336315631866455, "learning_rate": 8.374733441281353e-05, "loss": 0.6149, "step": 8530 }, { "epoch": 0.5725982349585584, "grad_norm": 1.0753029584884644, "learning_rate": 8.373931425643855e-05, "loss": 0.6491, "step": 8532 }, { "epoch": 0.5727324586423275, "grad_norm": 1.0328943729400635, "learning_rate": 8.37312925059473e-05, "loss": 0.6224, "step": 8534 }, { "epoch": 0.5728666823260964, "grad_norm": 1.2001701593399048, "learning_rate": 8.372326916171882e-05, "loss": 0.6266, "step": 8536 }, { "epoch": 0.5730009060098654, "grad_norm": 1.1070367097854614, "learning_rate": 8.371524422413218e-05, "loss": 0.6204, "step": 8538 }, { "epoch": 0.5731351296936344, "grad_norm": 1.0270267724990845, "learning_rate": 8.370721769356656e-05, "loss": 0.6376, "step": 8540 }, { "epoch": 0.5732693533774035, "grad_norm": 1.089913249015808, "learning_rate": 8.369918957040116e-05, "loss": 0.6562, "step": 8542 }, { "epoch": 0.5734035770611724, "grad_norm": 0.9418351054191589, "learning_rate": 8.369115985501534e-05, "loss": 0.5989, "step": 8544 }, { "epoch": 0.5735378007449414, "grad_norm": 1.1004632711410522, "learning_rate": 8.368312854778848e-05, "loss": 0.604, "step": 8546 }, { "epoch": 0.5736720244287105, "grad_norm": 1.0900779962539673, "learning_rate": 8.367509564910001e-05, "loss": 0.6368, "step": 8548 }, { "epoch": 0.5738062481124795, "grad_norm": 1.1029690504074097, "learning_rate": 8.366706115932951e-05, "loss": 0.6536, "step": 8550 }, { "epoch": 0.5739404717962484, "grad_norm": 1.3655116558074951, "learning_rate": 8.365902507885656e-05, "loss": 0.699, "step": 8552 }, { "epoch": 0.5740746954800174, "grad_norm": 0.9136799573898315, "learning_rate": 8.365098740806086e-05, "loss": 0.6259, "step": 8554 }, { "epoch": 0.5742089191637865, "grad_norm": 1.1151015758514404, "learning_rate": 8.364294814732218e-05, "loss": 0.6007, "step": 8556 }, { "epoch": 0.5743431428475555, "grad_norm": 0.9604775309562683, "learning_rate": 8.363490729702034e-05, "loss": 0.5917, "step": 8558 }, { "epoch": 0.5744773665313244, "grad_norm": 1.218360424041748, "learning_rate": 8.362686485753531e-05, "loss": 0.6404, "step": 8560 }, { "epoch": 0.5746115902150934, "grad_norm": 0.926568329334259, "learning_rate": 8.361882082924702e-05, "loss": 0.5643, "step": 8562 }, { "epoch": 0.5747458138988625, "grad_norm": 1.341833233833313, "learning_rate": 8.361077521253553e-05, "loss": 0.6566, "step": 8564 }, { "epoch": 0.5748800375826315, "grad_norm": 1.016913890838623, "learning_rate": 8.360272800778103e-05, "loss": 0.6195, "step": 8566 }, { "epoch": 0.5750142612664004, "grad_norm": 1.1553874015808105, "learning_rate": 8.359467921536371e-05, "loss": 0.6431, "step": 8568 }, { "epoch": 0.5751484849501695, "grad_norm": 1.1712065935134888, "learning_rate": 8.358662883566383e-05, "loss": 0.6416, "step": 8570 }, { "epoch": 0.5752827086339385, "grad_norm": 1.0247400999069214, "learning_rate": 8.357857686906182e-05, "loss": 0.692, "step": 8572 }, { "epoch": 0.5754169323177074, "grad_norm": 0.9636111855506897, "learning_rate": 8.357052331593807e-05, "loss": 0.6063, "step": 8574 }, { "epoch": 0.5755511560014764, "grad_norm": 1.0941756963729858, "learning_rate": 8.35624681766731e-05, "loss": 0.5797, "step": 8576 }, { "epoch": 0.5756853796852455, "grad_norm": 1.0538661479949951, "learning_rate": 8.355441145164751e-05, "loss": 0.6479, "step": 8578 }, { "epoch": 0.5758196033690145, "grad_norm": 1.0697072744369507, "learning_rate": 8.354635314124195e-05, "loss": 0.5541, "step": 8580 }, { "epoch": 0.5759538270527834, "grad_norm": 1.0581369400024414, "learning_rate": 8.35382932458372e-05, "loss": 0.6651, "step": 8582 }, { "epoch": 0.5760880507365524, "grad_norm": 1.2765485048294067, "learning_rate": 8.353023176581401e-05, "loss": 0.6432, "step": 8584 }, { "epoch": 0.5762222744203215, "grad_norm": 1.1463816165924072, "learning_rate": 8.352216870155331e-05, "loss": 0.7451, "step": 8586 }, { "epoch": 0.5763564981040905, "grad_norm": 0.9963604211807251, "learning_rate": 8.351410405343607e-05, "loss": 0.6129, "step": 8588 }, { "epoch": 0.5764907217878594, "grad_norm": 0.9272325038909912, "learning_rate": 8.350603782184332e-05, "loss": 0.6512, "step": 8590 }, { "epoch": 0.5766249454716285, "grad_norm": 0.9915642738342285, "learning_rate": 8.349797000715616e-05, "loss": 0.6561, "step": 8592 }, { "epoch": 0.5767591691553975, "grad_norm": 1.0291296243667603, "learning_rate": 8.348990060975578e-05, "loss": 0.6414, "step": 8594 }, { "epoch": 0.5768933928391665, "grad_norm": 0.9996158480644226, "learning_rate": 8.348182963002346e-05, "loss": 0.6249, "step": 8596 }, { "epoch": 0.5770276165229354, "grad_norm": 1.1353191137313843, "learning_rate": 8.347375706834055e-05, "loss": 0.6113, "step": 8598 }, { "epoch": 0.5771618402067045, "grad_norm": 0.9508293271064758, "learning_rate": 8.34656829250884e-05, "loss": 0.6033, "step": 8600 }, { "epoch": 0.5772960638904735, "grad_norm": 1.0945531129837036, "learning_rate": 8.345760720064856e-05, "loss": 0.7605, "step": 8602 }, { "epoch": 0.5774302875742425, "grad_norm": 1.0342299938201904, "learning_rate": 8.344952989540259e-05, "loss": 0.5899, "step": 8604 }, { "epoch": 0.5775645112580114, "grad_norm": 0.990175724029541, "learning_rate": 8.344145100973209e-05, "loss": 0.5604, "step": 8606 }, { "epoch": 0.5776987349417805, "grad_norm": 1.0384178161621094, "learning_rate": 8.343337054401878e-05, "loss": 0.6544, "step": 8608 }, { "epoch": 0.5778329586255495, "grad_norm": 1.055012583732605, "learning_rate": 8.342528849864447e-05, "loss": 0.624, "step": 8610 }, { "epoch": 0.5779671823093184, "grad_norm": 1.0311509370803833, "learning_rate": 8.3417204873991e-05, "loss": 0.6355, "step": 8612 }, { "epoch": 0.5781014059930875, "grad_norm": 0.9782349467277527, "learning_rate": 8.340911967044032e-05, "loss": 0.6929, "step": 8614 }, { "epoch": 0.5782356296768565, "grad_norm": 1.315628170967102, "learning_rate": 8.340103288837443e-05, "loss": 0.6516, "step": 8616 }, { "epoch": 0.5783698533606255, "grad_norm": 1.2841339111328125, "learning_rate": 8.33929445281754e-05, "loss": 0.6612, "step": 8618 }, { "epoch": 0.5785040770443944, "grad_norm": 1.027443766593933, "learning_rate": 8.338485459022544e-05, "loss": 0.5905, "step": 8620 }, { "epoch": 0.5786383007281635, "grad_norm": 1.0931084156036377, "learning_rate": 8.337676307490671e-05, "loss": 0.6221, "step": 8622 }, { "epoch": 0.5787725244119325, "grad_norm": 1.0320110321044922, "learning_rate": 8.336866998260159e-05, "loss": 0.6046, "step": 8624 }, { "epoch": 0.5789067480957015, "grad_norm": 1.0718752145767212, "learning_rate": 8.336057531369241e-05, "loss": 0.5888, "step": 8626 }, { "epoch": 0.5790409717794704, "grad_norm": 1.1324846744537354, "learning_rate": 8.335247906856165e-05, "loss": 0.6907, "step": 8628 }, { "epoch": 0.5791751954632395, "grad_norm": 1.034125804901123, "learning_rate": 8.334438124759184e-05, "loss": 0.6191, "step": 8630 }, { "epoch": 0.5793094191470085, "grad_norm": 1.0781384706497192, "learning_rate": 8.333628185116559e-05, "loss": 0.6483, "step": 8632 }, { "epoch": 0.5794436428307775, "grad_norm": 1.0112380981445312, "learning_rate": 8.332818087966556e-05, "loss": 0.6759, "step": 8634 }, { "epoch": 0.5795778665145465, "grad_norm": 1.0589327812194824, "learning_rate": 8.332007833347454e-05, "loss": 0.6546, "step": 8636 }, { "epoch": 0.5797120901983155, "grad_norm": 1.160819411277771, "learning_rate": 8.331197421297534e-05, "loss": 0.5642, "step": 8638 }, { "epoch": 0.5798463138820845, "grad_norm": 1.0162273645401, "learning_rate": 8.330386851855086e-05, "loss": 0.5833, "step": 8640 }, { "epoch": 0.5799805375658535, "grad_norm": 0.9922532439231873, "learning_rate": 8.329576125058406e-05, "loss": 0.672, "step": 8642 }, { "epoch": 0.5801147612496225, "grad_norm": 1.1330982446670532, "learning_rate": 8.328765240945803e-05, "loss": 0.6178, "step": 8644 }, { "epoch": 0.5802489849333915, "grad_norm": 1.0933043956756592, "learning_rate": 8.327954199555587e-05, "loss": 0.6216, "step": 8646 }, { "epoch": 0.5803832086171605, "grad_norm": 0.939629852771759, "learning_rate": 8.327143000926082e-05, "loss": 0.5978, "step": 8648 }, { "epoch": 0.5805174323009294, "grad_norm": 1.3238186836242676, "learning_rate": 8.32633164509561e-05, "loss": 0.6486, "step": 8650 }, { "epoch": 0.5806516559846985, "grad_norm": 0.9212239980697632, "learning_rate": 8.32552013210251e-05, "loss": 0.5924, "step": 8652 }, { "epoch": 0.5807858796684675, "grad_norm": 1.0109730958938599, "learning_rate": 8.324708461985124e-05, "loss": 0.6334, "step": 8654 }, { "epoch": 0.5809201033522365, "grad_norm": 1.1129043102264404, "learning_rate": 8.323896634781799e-05, "loss": 0.614, "step": 8656 }, { "epoch": 0.5810543270360055, "grad_norm": 0.9045149087905884, "learning_rate": 8.323084650530895e-05, "loss": 0.6003, "step": 8658 }, { "epoch": 0.5811885507197745, "grad_norm": 1.002319097518921, "learning_rate": 8.322272509270777e-05, "loss": 0.5334, "step": 8660 }, { "epoch": 0.5813227744035435, "grad_norm": 1.0560046434402466, "learning_rate": 8.321460211039814e-05, "loss": 0.6468, "step": 8662 }, { "epoch": 0.5814569980873125, "grad_norm": 1.061021089553833, "learning_rate": 8.320647755876389e-05, "loss": 0.6378, "step": 8664 }, { "epoch": 0.5815912217710815, "grad_norm": 1.386268973350525, "learning_rate": 8.319835143818887e-05, "loss": 0.5953, "step": 8666 }, { "epoch": 0.5817254454548505, "grad_norm": 1.3192063570022583, "learning_rate": 8.319022374905701e-05, "loss": 0.6294, "step": 8668 }, { "epoch": 0.5818596691386195, "grad_norm": 1.0040688514709473, "learning_rate": 8.318209449175236e-05, "loss": 0.5872, "step": 8670 }, { "epoch": 0.5819938928223886, "grad_norm": 1.0160542726516724, "learning_rate": 8.317396366665899e-05, "loss": 0.589, "step": 8672 }, { "epoch": 0.5821281165061575, "grad_norm": 1.5633481740951538, "learning_rate": 8.316583127416107e-05, "loss": 0.631, "step": 8674 }, { "epoch": 0.5822623401899265, "grad_norm": 1.3659327030181885, "learning_rate": 8.315769731464284e-05, "loss": 0.6798, "step": 8676 }, { "epoch": 0.5823965638736955, "grad_norm": 1.0467239618301392, "learning_rate": 8.314956178848861e-05, "loss": 0.6086, "step": 8678 }, { "epoch": 0.5825307875574646, "grad_norm": 1.1177936792373657, "learning_rate": 8.314142469608278e-05, "loss": 0.7187, "step": 8680 }, { "epoch": 0.5826650112412335, "grad_norm": 1.061246633529663, "learning_rate": 8.313328603780979e-05, "loss": 0.6271, "step": 8682 }, { "epoch": 0.5827992349250025, "grad_norm": 0.9757559895515442, "learning_rate": 8.31251458140542e-05, "loss": 0.6199, "step": 8684 }, { "epoch": 0.5829334586087715, "grad_norm": 1.4294708967208862, "learning_rate": 8.311700402520062e-05, "loss": 0.6139, "step": 8686 }, { "epoch": 0.5830676822925405, "grad_norm": 0.9727804660797119, "learning_rate": 8.31088606716337e-05, "loss": 0.637, "step": 8688 }, { "epoch": 0.5832019059763095, "grad_norm": 1.0098894834518433, "learning_rate": 8.310071575373823e-05, "loss": 0.6181, "step": 8690 }, { "epoch": 0.5833361296600785, "grad_norm": 1.1134440898895264, "learning_rate": 8.309256927189903e-05, "loss": 0.5976, "step": 8692 }, { "epoch": 0.5834703533438476, "grad_norm": 1.059842586517334, "learning_rate": 8.308442122650099e-05, "loss": 0.591, "step": 8694 }, { "epoch": 0.5836045770276165, "grad_norm": 1.0251708030700684, "learning_rate": 8.307627161792912e-05, "loss": 0.616, "step": 8696 }, { "epoch": 0.5837388007113855, "grad_norm": 1.0120950937271118, "learning_rate": 8.306812044656846e-05, "loss": 0.6321, "step": 8698 }, { "epoch": 0.5838730243951545, "grad_norm": 1.1307423114776611, "learning_rate": 8.305996771280413e-05, "loss": 0.6375, "step": 8700 }, { "epoch": 0.5840072480789236, "grad_norm": 1.4316738843917847, "learning_rate": 8.305181341702134e-05, "loss": 0.6423, "step": 8702 }, { "epoch": 0.5841414717626925, "grad_norm": 1.0380795001983643, "learning_rate": 8.304365755960534e-05, "loss": 0.7367, "step": 8704 }, { "epoch": 0.5842756954464615, "grad_norm": 1.0263220071792603, "learning_rate": 8.303550014094153e-05, "loss": 0.6002, "step": 8706 }, { "epoch": 0.5844099191302305, "grad_norm": 0.9541037678718567, "learning_rate": 8.302734116141527e-05, "loss": 0.5746, "step": 8708 }, { "epoch": 0.5845441428139996, "grad_norm": 1.034319519996643, "learning_rate": 8.30191806214121e-05, "loss": 0.6303, "step": 8710 }, { "epoch": 0.5846783664977685, "grad_norm": 1.0026488304138184, "learning_rate": 8.301101852131757e-05, "loss": 0.7003, "step": 8712 }, { "epoch": 0.5848125901815375, "grad_norm": 1.1234585046768188, "learning_rate": 8.300285486151734e-05, "loss": 0.5807, "step": 8714 }, { "epoch": 0.5849468138653066, "grad_norm": 1.171585202217102, "learning_rate": 8.299468964239709e-05, "loss": 0.643, "step": 8716 }, { "epoch": 0.5850810375490756, "grad_norm": 1.1702193021774292, "learning_rate": 8.298652286434265e-05, "loss": 0.6512, "step": 8718 }, { "epoch": 0.5852152612328445, "grad_norm": 0.9223757982254028, "learning_rate": 8.297835452773988e-05, "loss": 0.5944, "step": 8720 }, { "epoch": 0.5853494849166135, "grad_norm": 1.0043474435806274, "learning_rate": 8.297018463297471e-05, "loss": 0.6106, "step": 8722 }, { "epoch": 0.5854837086003826, "grad_norm": 1.0168185234069824, "learning_rate": 8.296201318043313e-05, "loss": 0.6452, "step": 8724 }, { "epoch": 0.5856179322841515, "grad_norm": 0.9247617721557617, "learning_rate": 8.295384017050125e-05, "loss": 0.5894, "step": 8726 }, { "epoch": 0.5857521559679205, "grad_norm": 1.0373064279556274, "learning_rate": 8.294566560356522e-05, "loss": 0.6591, "step": 8728 }, { "epoch": 0.5858863796516895, "grad_norm": 1.0301313400268555, "learning_rate": 8.293748948001129e-05, "loss": 0.6364, "step": 8730 }, { "epoch": 0.5860206033354586, "grad_norm": 1.2865626811981201, "learning_rate": 8.292931180022574e-05, "loss": 0.6155, "step": 8732 }, { "epoch": 0.5861548270192275, "grad_norm": 1.0519150495529175, "learning_rate": 8.292113256459495e-05, "loss": 0.6381, "step": 8734 }, { "epoch": 0.5862890507029965, "grad_norm": 1.0075963735580444, "learning_rate": 8.29129517735054e-05, "loss": 0.6134, "step": 8736 }, { "epoch": 0.5864232743867656, "grad_norm": 0.9536563754081726, "learning_rate": 8.29047694273436e-05, "loss": 0.6308, "step": 8738 }, { "epoch": 0.5865574980705346, "grad_norm": 1.0983659029006958, "learning_rate": 8.289658552649614e-05, "loss": 0.6543, "step": 8740 }, { "epoch": 0.5866917217543035, "grad_norm": 1.0358927249908447, "learning_rate": 8.28884000713497e-05, "loss": 0.6165, "step": 8742 }, { "epoch": 0.5868259454380725, "grad_norm": 1.001076102256775, "learning_rate": 8.288021306229103e-05, "loss": 0.6404, "step": 8744 }, { "epoch": 0.5869601691218416, "grad_norm": 1.0301929712295532, "learning_rate": 8.287202449970695e-05, "loss": 0.633, "step": 8746 }, { "epoch": 0.5870943928056106, "grad_norm": 1.6269079446792603, "learning_rate": 8.286383438398437e-05, "loss": 0.6431, "step": 8748 }, { "epoch": 0.5872286164893795, "grad_norm": 1.1193408966064453, "learning_rate": 8.285564271551023e-05, "loss": 0.7092, "step": 8750 }, { "epoch": 0.5873628401731485, "grad_norm": 1.1632122993469238, "learning_rate": 8.284744949467156e-05, "loss": 0.5789, "step": 8752 }, { "epoch": 0.5874970638569176, "grad_norm": 1.1115137338638306, "learning_rate": 8.283925472185552e-05, "loss": 0.6834, "step": 8754 }, { "epoch": 0.5876312875406866, "grad_norm": 1.1421805620193481, "learning_rate": 8.283105839744925e-05, "loss": 0.5362, "step": 8756 }, { "epoch": 0.5877655112244555, "grad_norm": 1.0075037479400635, "learning_rate": 8.282286052184005e-05, "loss": 0.5989, "step": 8758 }, { "epoch": 0.5878997349082246, "grad_norm": 1.2430760860443115, "learning_rate": 8.281466109541521e-05, "loss": 0.624, "step": 8760 }, { "epoch": 0.5880339585919936, "grad_norm": 1.0672948360443115, "learning_rate": 8.280646011856218e-05, "loss": 0.6325, "step": 8762 }, { "epoch": 0.5881681822757625, "grad_norm": 1.056333065032959, "learning_rate": 8.279825759166842e-05, "loss": 0.6255, "step": 8764 }, { "epoch": 0.5883024059595315, "grad_norm": 1.1188087463378906, "learning_rate": 8.279005351512146e-05, "loss": 0.6541, "step": 8766 }, { "epoch": 0.5884366296433006, "grad_norm": 1.011813998222351, "learning_rate": 8.278184788930897e-05, "loss": 0.6631, "step": 8768 }, { "epoch": 0.5885708533270696, "grad_norm": 1.5705772638320923, "learning_rate": 8.277364071461862e-05, "loss": 0.6036, "step": 8770 }, { "epoch": 0.5887050770108385, "grad_norm": 0.9317624568939209, "learning_rate": 8.27654319914382e-05, "loss": 0.6173, "step": 8772 }, { "epoch": 0.5888393006946075, "grad_norm": 1.0767476558685303, "learning_rate": 8.275722172015555e-05, "loss": 0.6394, "step": 8774 }, { "epoch": 0.5889735243783766, "grad_norm": 1.0390230417251587, "learning_rate": 8.274900990115859e-05, "loss": 0.6507, "step": 8776 }, { "epoch": 0.5891077480621456, "grad_norm": 1.225508451461792, "learning_rate": 8.274079653483531e-05, "loss": 0.6498, "step": 8778 }, { "epoch": 0.5892419717459145, "grad_norm": 1.0217746496200562, "learning_rate": 8.273258162157377e-05, "loss": 0.6289, "step": 8780 }, { "epoch": 0.5893761954296836, "grad_norm": 1.1209293603897095, "learning_rate": 8.272436516176212e-05, "loss": 0.6353, "step": 8782 }, { "epoch": 0.5895104191134526, "grad_norm": 1.0880781412124634, "learning_rate": 8.271614715578856e-05, "loss": 0.6749, "step": 8784 }, { "epoch": 0.5896446427972216, "grad_norm": 1.2088608741760254, "learning_rate": 8.270792760404138e-05, "loss": 0.6776, "step": 8786 }, { "epoch": 0.5897788664809905, "grad_norm": 1.173379898071289, "learning_rate": 8.269970650690894e-05, "loss": 0.6267, "step": 8788 }, { "epoch": 0.5899130901647596, "grad_norm": 1.0878911018371582, "learning_rate": 8.269148386477968e-05, "loss": 0.6153, "step": 8790 }, { "epoch": 0.5900473138485286, "grad_norm": 0.9524178504943848, "learning_rate": 8.268325967804207e-05, "loss": 0.6521, "step": 8792 }, { "epoch": 0.5901815375322976, "grad_norm": 0.9811989665031433, "learning_rate": 8.267503394708472e-05, "loss": 0.6251, "step": 8794 }, { "epoch": 0.5903157612160665, "grad_norm": 1.040177822113037, "learning_rate": 8.266680667229627e-05, "loss": 0.6039, "step": 8796 }, { "epoch": 0.5904499848998356, "grad_norm": 1.0749226808547974, "learning_rate": 8.265857785406544e-05, "loss": 0.5825, "step": 8798 }, { "epoch": 0.5905842085836046, "grad_norm": 1.1496376991271973, "learning_rate": 8.265034749278103e-05, "loss": 0.7057, "step": 8800 }, { "epoch": 0.5907184322673735, "grad_norm": 1.0202300548553467, "learning_rate": 8.264211558883191e-05, "loss": 0.6245, "step": 8802 }, { "epoch": 0.5908526559511426, "grad_norm": 1.3150079250335693, "learning_rate": 8.263388214260702e-05, "loss": 0.6337, "step": 8804 }, { "epoch": 0.5909868796349116, "grad_norm": 0.9684760570526123, "learning_rate": 8.262564715449535e-05, "loss": 0.637, "step": 8806 }, { "epoch": 0.5911211033186806, "grad_norm": 1.236946940422058, "learning_rate": 8.261741062488602e-05, "loss": 0.6555, "step": 8808 }, { "epoch": 0.5912553270024495, "grad_norm": 0.9326079487800598, "learning_rate": 8.260917255416817e-05, "loss": 0.6187, "step": 8810 }, { "epoch": 0.5913895506862186, "grad_norm": 1.0765881538391113, "learning_rate": 8.260093294273103e-05, "loss": 0.6183, "step": 8812 }, { "epoch": 0.5915237743699876, "grad_norm": 1.1676933765411377, "learning_rate": 8.259269179096393e-05, "loss": 0.6559, "step": 8814 }, { "epoch": 0.5916579980537566, "grad_norm": 1.091220736503601, "learning_rate": 8.258444909925624e-05, "loss": 0.6776, "step": 8816 }, { "epoch": 0.5917922217375255, "grad_norm": 1.1847777366638184, "learning_rate": 8.25762048679974e-05, "loss": 0.5775, "step": 8818 }, { "epoch": 0.5919264454212946, "grad_norm": 1.190104603767395, "learning_rate": 8.256795909757694e-05, "loss": 0.6238, "step": 8820 }, { "epoch": 0.5920606691050636, "grad_norm": 1.1804845333099365, "learning_rate": 8.255971178838445e-05, "loss": 0.6297, "step": 8822 }, { "epoch": 0.5921948927888326, "grad_norm": 1.1453303098678589, "learning_rate": 8.25514629408096e-05, "loss": 0.6756, "step": 8824 }, { "epoch": 0.5923291164726016, "grad_norm": 1.0372350215911865, "learning_rate": 8.254321255524215e-05, "loss": 0.6554, "step": 8826 }, { "epoch": 0.5924633401563706, "grad_norm": 1.1543527841567993, "learning_rate": 8.25349606320719e-05, "loss": 0.5751, "step": 8828 }, { "epoch": 0.5925975638401396, "grad_norm": 1.1377294063568115, "learning_rate": 8.252670717168872e-05, "loss": 0.6465, "step": 8830 }, { "epoch": 0.5927317875239086, "grad_norm": 1.1026825904846191, "learning_rate": 8.25184521744826e-05, "loss": 0.5892, "step": 8832 }, { "epoch": 0.5928660112076776, "grad_norm": 1.012364387512207, "learning_rate": 8.251019564084355e-05, "loss": 0.6018, "step": 8834 }, { "epoch": 0.5930002348914466, "grad_norm": 0.9433966875076294, "learning_rate": 8.250193757116169e-05, "loss": 0.6373, "step": 8836 }, { "epoch": 0.5931344585752156, "grad_norm": 1.0152767896652222, "learning_rate": 8.24936779658272e-05, "loss": 0.6141, "step": 8838 }, { "epoch": 0.5932686822589845, "grad_norm": 1.0709465742111206, "learning_rate": 8.248541682523032e-05, "loss": 0.6747, "step": 8840 }, { "epoch": 0.5934029059427536, "grad_norm": 1.8665101528167725, "learning_rate": 8.247715414976136e-05, "loss": 0.6451, "step": 8842 }, { "epoch": 0.5935371296265226, "grad_norm": 1.182285189628601, "learning_rate": 8.246888993981074e-05, "loss": 0.6019, "step": 8844 }, { "epoch": 0.5936713533102916, "grad_norm": 0.9702306389808655, "learning_rate": 8.246062419576892e-05, "loss": 0.6399, "step": 8846 }, { "epoch": 0.5938055769940606, "grad_norm": 1.0066275596618652, "learning_rate": 8.245235691802644e-05, "loss": 0.6147, "step": 8848 }, { "epoch": 0.5939398006778296, "grad_norm": 1.2227391004562378, "learning_rate": 8.24440881069739e-05, "loss": 0.5987, "step": 8850 }, { "epoch": 0.5940740243615986, "grad_norm": 0.9837579131126404, "learning_rate": 8.2435817763002e-05, "loss": 0.6061, "step": 8852 }, { "epoch": 0.5942082480453676, "grad_norm": 1.8223912715911865, "learning_rate": 8.24275458865015e-05, "loss": 0.6253, "step": 8854 }, { "epoch": 0.5943424717291366, "grad_norm": 0.9962928891181946, "learning_rate": 8.24192724778632e-05, "loss": 0.6692, "step": 8856 }, { "epoch": 0.5944766954129056, "grad_norm": 0.9767764210700989, "learning_rate": 8.241099753747805e-05, "loss": 0.6543, "step": 8858 }, { "epoch": 0.5946109190966746, "grad_norm": 1.0351556539535522, "learning_rate": 8.240272106573699e-05, "loss": 0.5912, "step": 8860 }, { "epoch": 0.5947451427804437, "grad_norm": 1.0033786296844482, "learning_rate": 8.239444306303109e-05, "loss": 0.6525, "step": 8862 }, { "epoch": 0.5948793664642126, "grad_norm": 0.9843260645866394, "learning_rate": 8.238616352975143e-05, "loss": 0.6849, "step": 8864 }, { "epoch": 0.5950135901479816, "grad_norm": 1.065222144126892, "learning_rate": 8.237788246628925e-05, "loss": 0.5955, "step": 8866 }, { "epoch": 0.5951478138317506, "grad_norm": 1.0120952129364014, "learning_rate": 8.236959987303579e-05, "loss": 0.6793, "step": 8868 }, { "epoch": 0.5952820375155197, "grad_norm": 1.0915205478668213, "learning_rate": 8.236131575038238e-05, "loss": 0.6915, "step": 8870 }, { "epoch": 0.5954162611992886, "grad_norm": 0.9849734902381897, "learning_rate": 8.235303009872043e-05, "loss": 0.6771, "step": 8872 }, { "epoch": 0.5955504848830576, "grad_norm": 1.1612014770507812, "learning_rate": 8.234474291844144e-05, "loss": 0.6502, "step": 8874 }, { "epoch": 0.5956847085668266, "grad_norm": 1.1154402494430542, "learning_rate": 8.233645420993695e-05, "loss": 0.6228, "step": 8876 }, { "epoch": 0.5958189322505956, "grad_norm": 0.9973312020301819, "learning_rate": 8.232816397359858e-05, "loss": 0.6254, "step": 8878 }, { "epoch": 0.5959531559343646, "grad_norm": 1.0583056211471558, "learning_rate": 8.231987220981804e-05, "loss": 0.621, "step": 8880 }, { "epoch": 0.5960873796181336, "grad_norm": 1.1791260242462158, "learning_rate": 8.231157891898708e-05, "loss": 0.6521, "step": 8882 }, { "epoch": 0.5962216033019027, "grad_norm": 1.0589373111724854, "learning_rate": 8.230328410149756e-05, "loss": 0.6304, "step": 8884 }, { "epoch": 0.5963558269856716, "grad_norm": 0.9476355910301208, "learning_rate": 8.22949877577414e-05, "loss": 0.6518, "step": 8886 }, { "epoch": 0.5964900506694406, "grad_norm": 1.0454285144805908, "learning_rate": 8.228668988811055e-05, "loss": 0.6237, "step": 8888 }, { "epoch": 0.5966242743532096, "grad_norm": 1.1258161067962646, "learning_rate": 8.227839049299711e-05, "loss": 0.632, "step": 8890 }, { "epoch": 0.5967584980369787, "grad_norm": 0.9451509714126587, "learning_rate": 8.227008957279319e-05, "loss": 0.6004, "step": 8892 }, { "epoch": 0.5968927217207476, "grad_norm": 1.1254085302352905, "learning_rate": 8.2261787127891e-05, "loss": 0.5896, "step": 8894 }, { "epoch": 0.5970269454045166, "grad_norm": 0.9593334197998047, "learning_rate": 8.22534831586828e-05, "loss": 0.5775, "step": 8896 }, { "epoch": 0.5971611690882856, "grad_norm": 1.0605380535125732, "learning_rate": 8.224517766556094e-05, "loss": 0.5977, "step": 8898 }, { "epoch": 0.5972953927720547, "grad_norm": 1.4234702587127686, "learning_rate": 8.223687064891785e-05, "loss": 0.6361, "step": 8900 }, { "epoch": 0.5974296164558236, "grad_norm": 1.0671988725662231, "learning_rate": 8.222856210914601e-05, "loss": 0.643, "step": 8902 }, { "epoch": 0.5975638401395926, "grad_norm": 1.0523698329925537, "learning_rate": 8.222025204663799e-05, "loss": 0.6824, "step": 8904 }, { "epoch": 0.5976980638233617, "grad_norm": 1.166785717010498, "learning_rate": 8.221194046178641e-05, "loss": 0.6208, "step": 8906 }, { "epoch": 0.5978322875071307, "grad_norm": 1.129096508026123, "learning_rate": 8.220362735498399e-05, "loss": 0.5974, "step": 8908 }, { "epoch": 0.5979665111908996, "grad_norm": 1.0599874258041382, "learning_rate": 8.21953127266235e-05, "loss": 0.671, "step": 8910 }, { "epoch": 0.5981007348746686, "grad_norm": 1.0998214483261108, "learning_rate": 8.21869965770978e-05, "loss": 0.638, "step": 8912 }, { "epoch": 0.5982349585584377, "grad_norm": 1.0187451839447021, "learning_rate": 8.21786789067998e-05, "loss": 0.6521, "step": 8914 }, { "epoch": 0.5983691822422066, "grad_norm": 1.2205415964126587, "learning_rate": 8.21703597161225e-05, "loss": 0.6373, "step": 8916 }, { "epoch": 0.5985034059259756, "grad_norm": 1.3533973693847656, "learning_rate": 8.216203900545895e-05, "loss": 0.6429, "step": 8918 }, { "epoch": 0.5986376296097446, "grad_norm": 0.7839868068695068, "learning_rate": 8.215371677520231e-05, "loss": 0.5764, "step": 8920 }, { "epoch": 0.5987718532935137, "grad_norm": 1.0047537088394165, "learning_rate": 8.21453930257458e-05, "loss": 0.5767, "step": 8922 }, { "epoch": 0.5989060769772826, "grad_norm": 0.9989202618598938, "learning_rate": 8.213706775748265e-05, "loss": 0.5818, "step": 8924 }, { "epoch": 0.5990403006610516, "grad_norm": 1.103752613067627, "learning_rate": 8.212874097080626e-05, "loss": 0.7014, "step": 8926 }, { "epoch": 0.5991745243448207, "grad_norm": 1.228872299194336, "learning_rate": 8.212041266611003e-05, "loss": 0.6309, "step": 8928 }, { "epoch": 0.5993087480285897, "grad_norm": 1.09929621219635, "learning_rate": 8.211208284378747e-05, "loss": 0.6905, "step": 8930 }, { "epoch": 0.5994429717123586, "grad_norm": 1.056538462638855, "learning_rate": 8.210375150423214e-05, "loss": 0.6044, "step": 8932 }, { "epoch": 0.5995771953961276, "grad_norm": 0.9706535935401917, "learning_rate": 8.209541864783769e-05, "loss": 0.6148, "step": 8934 }, { "epoch": 0.5997114190798967, "grad_norm": 1.103705883026123, "learning_rate": 8.20870842749978e-05, "loss": 0.675, "step": 8936 }, { "epoch": 0.5998456427636657, "grad_norm": 1.0811362266540527, "learning_rate": 8.20787483861063e-05, "loss": 0.6167, "step": 8938 }, { "epoch": 0.5999798664474346, "grad_norm": 1.8568896055221558, "learning_rate": 8.2070410981557e-05, "loss": 0.6104, "step": 8940 }, { "epoch": 0.6001140901312036, "grad_norm": 1.0807304382324219, "learning_rate": 8.206207206174386e-05, "loss": 0.6114, "step": 8942 }, { "epoch": 0.6002483138149727, "grad_norm": 1.0534753799438477, "learning_rate": 8.205373162706085e-05, "loss": 0.6841, "step": 8944 }, { "epoch": 0.6003825374987416, "grad_norm": 1.3214335441589355, "learning_rate": 8.204538967790205e-05, "loss": 0.6906, "step": 8946 }, { "epoch": 0.6005167611825106, "grad_norm": 1.0081099271774292, "learning_rate": 8.20370462146616e-05, "loss": 0.6297, "step": 8948 }, { "epoch": 0.6006509848662797, "grad_norm": 1.1878976821899414, "learning_rate": 8.202870123773371e-05, "loss": 0.7054, "step": 8950 }, { "epoch": 0.6007852085500487, "grad_norm": 1.0634784698486328, "learning_rate": 8.20203547475127e-05, "loss": 0.6073, "step": 8952 }, { "epoch": 0.6009194322338176, "grad_norm": 1.013996958732605, "learning_rate": 8.201200674439287e-05, "loss": 0.5917, "step": 8954 }, { "epoch": 0.6010536559175866, "grad_norm": 1.01338529586792, "learning_rate": 8.200365722876868e-05, "loss": 0.6146, "step": 8956 }, { "epoch": 0.6011878796013557, "grad_norm": 1.0896825790405273, "learning_rate": 8.199530620103461e-05, "loss": 0.5908, "step": 8958 }, { "epoch": 0.6013221032851247, "grad_norm": 1.0631287097930908, "learning_rate": 8.198695366158523e-05, "loss": 0.5892, "step": 8960 }, { "epoch": 0.6014563269688936, "grad_norm": 1.2561473846435547, "learning_rate": 8.197859961081522e-05, "loss": 0.6779, "step": 8962 }, { "epoch": 0.6015905506526626, "grad_norm": 1.100650429725647, "learning_rate": 8.197024404911924e-05, "loss": 0.6193, "step": 8964 }, { "epoch": 0.6017247743364317, "grad_norm": 1.1069337129592896, "learning_rate": 8.196188697689209e-05, "loss": 0.7133, "step": 8966 }, { "epoch": 0.6018589980202007, "grad_norm": 1.0815422534942627, "learning_rate": 8.195352839452863e-05, "loss": 0.6056, "step": 8968 }, { "epoch": 0.6019932217039696, "grad_norm": 0.9499128460884094, "learning_rate": 8.19451683024238e-05, "loss": 0.6629, "step": 8970 }, { "epoch": 0.6021274453877387, "grad_norm": 1.1803507804870605, "learning_rate": 8.193680670097257e-05, "loss": 0.6733, "step": 8972 }, { "epoch": 0.6022616690715077, "grad_norm": 1.0044008493423462, "learning_rate": 8.192844359057003e-05, "loss": 0.5965, "step": 8974 }, { "epoch": 0.6023958927552767, "grad_norm": 0.9643215537071228, "learning_rate": 8.192007897161133e-05, "loss": 0.5732, "step": 8976 }, { "epoch": 0.6025301164390456, "grad_norm": 0.9642957448959351, "learning_rate": 8.191171284449166e-05, "loss": 0.596, "step": 8978 }, { "epoch": 0.6026643401228147, "grad_norm": 1.2209882736206055, "learning_rate": 8.19033452096063e-05, "loss": 0.6444, "step": 8980 }, { "epoch": 0.6027985638065837, "grad_norm": 0.9373955726623535, "learning_rate": 8.189497606735061e-05, "loss": 0.6267, "step": 8982 }, { "epoch": 0.6029327874903526, "grad_norm": 1.0083324909210205, "learning_rate": 8.188660541812004e-05, "loss": 0.6713, "step": 8984 }, { "epoch": 0.6030670111741216, "grad_norm": 1.2278261184692383, "learning_rate": 8.187823326231005e-05, "loss": 0.6651, "step": 8986 }, { "epoch": 0.6032012348578907, "grad_norm": 1.4669445753097534, "learning_rate": 8.186985960031623e-05, "loss": 0.7042, "step": 8988 }, { "epoch": 0.6033354585416597, "grad_norm": 1.0924478769302368, "learning_rate": 8.18614844325342e-05, "loss": 0.6319, "step": 8990 }, { "epoch": 0.6034696822254286, "grad_norm": 0.9888027906417847, "learning_rate": 8.185310775935971e-05, "loss": 0.6269, "step": 8992 }, { "epoch": 0.6036039059091977, "grad_norm": 0.9969624876976013, "learning_rate": 8.184472958118851e-05, "loss": 0.6402, "step": 8994 }, { "epoch": 0.6037381295929667, "grad_norm": 1.018033742904663, "learning_rate": 8.183634989841644e-05, "loss": 0.6584, "step": 8996 }, { "epoch": 0.6038723532767357, "grad_norm": 0.8649704456329346, "learning_rate": 8.182796871143945e-05, "loss": 0.5438, "step": 8998 }, { "epoch": 0.6040065769605046, "grad_norm": 1.122671365737915, "learning_rate": 8.181958602065351e-05, "loss": 0.6088, "step": 9000 }, { "epoch": 0.6041408006442737, "grad_norm": 1.2516348361968994, "learning_rate": 8.181120182645473e-05, "loss": 0.6303, "step": 9002 }, { "epoch": 0.6042750243280427, "grad_norm": 1.1585074663162231, "learning_rate": 8.18028161292392e-05, "loss": 0.6331, "step": 9004 }, { "epoch": 0.6044092480118117, "grad_norm": 0.9821868538856506, "learning_rate": 8.179442892940314e-05, "loss": 0.6229, "step": 9006 }, { "epoch": 0.6045434716955806, "grad_norm": 1.0640473365783691, "learning_rate": 8.178604022734287e-05, "loss": 0.7202, "step": 9008 }, { "epoch": 0.6046776953793497, "grad_norm": 1.0368740558624268, "learning_rate": 8.177765002345466e-05, "loss": 0.6664, "step": 9010 }, { "epoch": 0.6048119190631187, "grad_norm": 0.9344690442085266, "learning_rate": 8.176925831813499e-05, "loss": 0.6318, "step": 9012 }, { "epoch": 0.6049461427468877, "grad_norm": 0.9644981026649475, "learning_rate": 8.176086511178034e-05, "loss": 0.6323, "step": 9014 }, { "epoch": 0.6050803664306567, "grad_norm": 1.0489426851272583, "learning_rate": 8.175247040478727e-05, "loss": 0.6223, "step": 9016 }, { "epoch": 0.6052145901144257, "grad_norm": 1.0508383512496948, "learning_rate": 8.17440741975524e-05, "loss": 0.6097, "step": 9018 }, { "epoch": 0.6053488137981947, "grad_norm": 1.0071632862091064, "learning_rate": 8.173567649047246e-05, "loss": 0.6127, "step": 9020 }, { "epoch": 0.6054830374819636, "grad_norm": 1.4123948812484741, "learning_rate": 8.17272772839442e-05, "loss": 0.6371, "step": 9022 }, { "epoch": 0.6056172611657327, "grad_norm": 1.0234935283660889, "learning_rate": 8.171887657836448e-05, "loss": 0.6502, "step": 9024 }, { "epoch": 0.6057514848495017, "grad_norm": 1.1688114404678345, "learning_rate": 8.171047437413022e-05, "loss": 0.6422, "step": 9026 }, { "epoch": 0.6058857085332707, "grad_norm": 0.9891804456710815, "learning_rate": 8.17020706716384e-05, "loss": 0.6012, "step": 9028 }, { "epoch": 0.6060199322170396, "grad_norm": 1.1169568300247192, "learning_rate": 8.169366547128607e-05, "loss": 0.5939, "step": 9030 }, { "epoch": 0.6061541559008087, "grad_norm": 1.1597013473510742, "learning_rate": 8.16852587734704e-05, "loss": 0.5878, "step": 9032 }, { "epoch": 0.6062883795845777, "grad_norm": 1.082456111907959, "learning_rate": 8.167685057858853e-05, "loss": 0.6445, "step": 9034 }, { "epoch": 0.6064226032683467, "grad_norm": 1.0745278596878052, "learning_rate": 8.166844088703777e-05, "loss": 0.5711, "step": 9036 }, { "epoch": 0.6065568269521157, "grad_norm": 1.0846468210220337, "learning_rate": 8.166002969921545e-05, "loss": 0.5778, "step": 9038 }, { "epoch": 0.6066910506358847, "grad_norm": 1.240317463874817, "learning_rate": 8.165161701551898e-05, "loss": 0.6438, "step": 9040 }, { "epoch": 0.6068252743196537, "grad_norm": 1.1379884481430054, "learning_rate": 8.164320283634585e-05, "loss": 0.6407, "step": 9042 }, { "epoch": 0.6069594980034227, "grad_norm": 0.9429364800453186, "learning_rate": 8.16347871620936e-05, "loss": 0.6703, "step": 9044 }, { "epoch": 0.6070937216871917, "grad_norm": 1.0277690887451172, "learning_rate": 8.162636999315987e-05, "loss": 0.6008, "step": 9046 }, { "epoch": 0.6072279453709607, "grad_norm": 1.5023620128631592, "learning_rate": 8.161795132994235e-05, "loss": 0.605, "step": 9048 }, { "epoch": 0.6073621690547297, "grad_norm": 1.0926510095596313, "learning_rate": 8.160953117283881e-05, "loss": 0.5932, "step": 9050 }, { "epoch": 0.6074963927384988, "grad_norm": 1.0439032316207886, "learning_rate": 8.160110952224707e-05, "loss": 0.621, "step": 9052 }, { "epoch": 0.6076306164222677, "grad_norm": 0.9344532489776611, "learning_rate": 8.159268637856506e-05, "loss": 0.631, "step": 9054 }, { "epoch": 0.6077648401060367, "grad_norm": 1.1640667915344238, "learning_rate": 8.158426174219074e-05, "loss": 0.6677, "step": 9056 }, { "epoch": 0.6078990637898057, "grad_norm": 1.0521429777145386, "learning_rate": 8.157583561352213e-05, "loss": 0.631, "step": 9058 }, { "epoch": 0.6080332874735747, "grad_norm": 1.0165327787399292, "learning_rate": 8.156740799295741e-05, "loss": 0.6065, "step": 9060 }, { "epoch": 0.6081675111573437, "grad_norm": 0.997218906879425, "learning_rate": 8.155897888089473e-05, "loss": 0.5984, "step": 9062 }, { "epoch": 0.6083017348411127, "grad_norm": 1.0150151252746582, "learning_rate": 8.155054827773237e-05, "loss": 0.612, "step": 9064 }, { "epoch": 0.6084359585248817, "grad_norm": 1.089469313621521, "learning_rate": 8.154211618386862e-05, "loss": 0.6098, "step": 9066 }, { "epoch": 0.6085701822086507, "grad_norm": 1.1109071969985962, "learning_rate": 8.153368259970191e-05, "loss": 0.7168, "step": 9068 }, { "epoch": 0.6087044058924197, "grad_norm": 0.9952682852745056, "learning_rate": 8.15252475256307e-05, "loss": 0.594, "step": 9070 }, { "epoch": 0.6088386295761887, "grad_norm": 1.1361262798309326, "learning_rate": 8.151681096205356e-05, "loss": 0.673, "step": 9072 }, { "epoch": 0.6089728532599578, "grad_norm": 1.187704086303711, "learning_rate": 8.150837290936905e-05, "loss": 0.6023, "step": 9074 }, { "epoch": 0.6091070769437267, "grad_norm": 1.0512447357177734, "learning_rate": 8.14999333679759e-05, "loss": 0.6197, "step": 9076 }, { "epoch": 0.6092413006274957, "grad_norm": 1.0476611852645874, "learning_rate": 8.149149233827285e-05, "loss": 0.574, "step": 9078 }, { "epoch": 0.6093755243112647, "grad_norm": 1.0389647483825684, "learning_rate": 8.148304982065869e-05, "loss": 0.6468, "step": 9080 }, { "epoch": 0.6095097479950338, "grad_norm": 1.1799911260604858, "learning_rate": 8.147460581553233e-05, "loss": 0.601, "step": 9082 }, { "epoch": 0.6096439716788027, "grad_norm": 1.0153534412384033, "learning_rate": 8.146616032329275e-05, "loss": 0.6727, "step": 9084 }, { "epoch": 0.6097781953625717, "grad_norm": 1.1318472623825073, "learning_rate": 8.145771334433896e-05, "loss": 0.6342, "step": 9086 }, { "epoch": 0.6099124190463407, "grad_norm": 1.0327775478363037, "learning_rate": 8.144926487907009e-05, "loss": 0.6956, "step": 9088 }, { "epoch": 0.6100466427301098, "grad_norm": 0.922252357006073, "learning_rate": 8.144081492788528e-05, "loss": 0.636, "step": 9090 }, { "epoch": 0.6101808664138787, "grad_norm": 1.1907691955566406, "learning_rate": 8.143236349118381e-05, "loss": 0.6395, "step": 9092 }, { "epoch": 0.6103150900976477, "grad_norm": 1.0901716947555542, "learning_rate": 8.142391056936495e-05, "loss": 0.7745, "step": 9094 }, { "epoch": 0.6104493137814168, "grad_norm": 0.9876933097839355, "learning_rate": 8.141545616282812e-05, "loss": 0.6439, "step": 9096 }, { "epoch": 0.6105835374651857, "grad_norm": 0.9889509677886963, "learning_rate": 8.140700027197277e-05, "loss": 0.6059, "step": 9098 }, { "epoch": 0.6107177611489547, "grad_norm": 1.2002860307693481, "learning_rate": 8.139854289719841e-05, "loss": 0.6139, "step": 9100 }, { "epoch": 0.6108519848327237, "grad_norm": 1.1429078578948975, "learning_rate": 8.139008403890465e-05, "loss": 0.6401, "step": 9102 }, { "epoch": 0.6109862085164928, "grad_norm": 1.1443973779678345, "learning_rate": 8.138162369749114e-05, "loss": 0.6632, "step": 9104 }, { "epoch": 0.6111204322002617, "grad_norm": 1.1087549924850464, "learning_rate": 8.137316187335761e-05, "loss": 0.653, "step": 9106 }, { "epoch": 0.6112546558840307, "grad_norm": 1.0414327383041382, "learning_rate": 8.136469856690387e-05, "loss": 0.6801, "step": 9108 }, { "epoch": 0.6113888795677997, "grad_norm": 1.0863112211227417, "learning_rate": 8.135623377852982e-05, "loss": 0.6728, "step": 9110 }, { "epoch": 0.6115231032515688, "grad_norm": 1.082608938217163, "learning_rate": 8.134776750863535e-05, "loss": 0.6256, "step": 9112 }, { "epoch": 0.6116573269353377, "grad_norm": 1.0916283130645752, "learning_rate": 8.133929975762053e-05, "loss": 0.552, "step": 9114 }, { "epoch": 0.6117915506191067, "grad_norm": 1.1448688507080078, "learning_rate": 8.133083052588543e-05, "loss": 0.652, "step": 9116 }, { "epoch": 0.6119257743028758, "grad_norm": 1.1834670305252075, "learning_rate": 8.132235981383018e-05, "loss": 0.6892, "step": 9118 }, { "epoch": 0.6120599979866448, "grad_norm": 1.07806396484375, "learning_rate": 8.131388762185503e-05, "loss": 0.6864, "step": 9120 }, { "epoch": 0.6121942216704137, "grad_norm": 0.9594216346740723, "learning_rate": 8.130541395036027e-05, "loss": 0.5829, "step": 9122 }, { "epoch": 0.6123284453541827, "grad_norm": 1.0681747198104858, "learning_rate": 8.129693879974626e-05, "loss": 0.6485, "step": 9124 }, { "epoch": 0.6124626690379518, "grad_norm": 1.1206474304199219, "learning_rate": 8.128846217041344e-05, "loss": 0.5984, "step": 9126 }, { "epoch": 0.6125968927217208, "grad_norm": 1.0924535989761353, "learning_rate": 8.12799840627623e-05, "loss": 0.5855, "step": 9128 }, { "epoch": 0.6127311164054897, "grad_norm": 1.2383434772491455, "learning_rate": 8.127150447719342e-05, "loss": 0.5934, "step": 9130 }, { "epoch": 0.6128653400892587, "grad_norm": 1.0007473230361938, "learning_rate": 8.126302341410744e-05, "loss": 0.5637, "step": 9132 }, { "epoch": 0.6129995637730278, "grad_norm": 1.0839866399765015, "learning_rate": 8.125454087390509e-05, "loss": 0.6521, "step": 9134 }, { "epoch": 0.6131337874567967, "grad_norm": 1.111741542816162, "learning_rate": 8.124605685698714e-05, "loss": 0.6516, "step": 9136 }, { "epoch": 0.6132680111405657, "grad_norm": 0.9247532486915588, "learning_rate": 8.123757136375445e-05, "loss": 0.6133, "step": 9138 }, { "epoch": 0.6134022348243348, "grad_norm": 0.9069095849990845, "learning_rate": 8.122908439460794e-05, "loss": 0.6424, "step": 9140 }, { "epoch": 0.6135364585081038, "grad_norm": 1.0073060989379883, "learning_rate": 8.12205959499486e-05, "loss": 0.6146, "step": 9142 }, { "epoch": 0.6136706821918727, "grad_norm": 1.0860568284988403, "learning_rate": 8.121210603017748e-05, "loss": 0.5838, "step": 9144 }, { "epoch": 0.6138049058756417, "grad_norm": 1.023048996925354, "learning_rate": 8.120361463569575e-05, "loss": 0.6019, "step": 9146 }, { "epoch": 0.6139391295594108, "grad_norm": 1.0431309938430786, "learning_rate": 8.119512176690455e-05, "loss": 0.671, "step": 9148 }, { "epoch": 0.6140733532431798, "grad_norm": 1.0847612619400024, "learning_rate": 8.118662742420523e-05, "loss": 0.6467, "step": 9150 }, { "epoch": 0.6142075769269487, "grad_norm": 0.9644896388053894, "learning_rate": 8.117813160799908e-05, "loss": 0.5775, "step": 9152 }, { "epoch": 0.6143418006107177, "grad_norm": 1.1092700958251953, "learning_rate": 8.116963431868751e-05, "loss": 0.5999, "step": 9154 }, { "epoch": 0.6144760242944868, "grad_norm": 1.053810954093933, "learning_rate": 8.116113555667203e-05, "loss": 0.7222, "step": 9156 }, { "epoch": 0.6146102479782558, "grad_norm": 0.9161946177482605, "learning_rate": 8.115263532235416e-05, "loss": 0.6246, "step": 9158 }, { "epoch": 0.6147444716620247, "grad_norm": 1.719303011894226, "learning_rate": 8.114413361613551e-05, "loss": 0.5393, "step": 9160 }, { "epoch": 0.6148786953457938, "grad_norm": 1.1035606861114502, "learning_rate": 8.113563043841781e-05, "loss": 0.5999, "step": 9162 }, { "epoch": 0.6150129190295628, "grad_norm": 0.940985918045044, "learning_rate": 8.11271257896028e-05, "loss": 0.5872, "step": 9164 }, { "epoch": 0.6151471427133318, "grad_norm": 1.0824211835861206, "learning_rate": 8.111861967009229e-05, "loss": 0.6177, "step": 9166 }, { "epoch": 0.6152813663971007, "grad_norm": 1.059979796409607, "learning_rate": 8.111011208028821e-05, "loss": 0.6636, "step": 9168 }, { "epoch": 0.6154155900808698, "grad_norm": 1.0575546026229858, "learning_rate": 8.11016030205925e-05, "loss": 0.6336, "step": 9170 }, { "epoch": 0.6155498137646388, "grad_norm": 1.0367028713226318, "learning_rate": 8.109309249140721e-05, "loss": 0.7206, "step": 9172 }, { "epoch": 0.6156840374484077, "grad_norm": 1.1719602346420288, "learning_rate": 8.108458049313443e-05, "loss": 0.6034, "step": 9174 }, { "epoch": 0.6158182611321767, "grad_norm": 1.026654839515686, "learning_rate": 8.107606702617636e-05, "loss": 0.6252, "step": 9176 }, { "epoch": 0.6159524848159458, "grad_norm": 0.9755574464797974, "learning_rate": 8.106755209093523e-05, "loss": 0.6031, "step": 9178 }, { "epoch": 0.6160867084997148, "grad_norm": 1.0690412521362305, "learning_rate": 8.105903568781335e-05, "loss": 0.6279, "step": 9180 }, { "epoch": 0.6162209321834837, "grad_norm": 1.1642770767211914, "learning_rate": 8.105051781721312e-05, "loss": 0.6367, "step": 9182 }, { "epoch": 0.6163551558672528, "grad_norm": 1.4145997762680054, "learning_rate": 8.104199847953696e-05, "loss": 0.65, "step": 9184 }, { "epoch": 0.6164893795510218, "grad_norm": 1.0357612371444702, "learning_rate": 8.103347767518743e-05, "loss": 0.6423, "step": 9186 }, { "epoch": 0.6166236032347908, "grad_norm": 0.9822550415992737, "learning_rate": 8.102495540456711e-05, "loss": 0.6013, "step": 9188 }, { "epoch": 0.6167578269185597, "grad_norm": 1.0211374759674072, "learning_rate": 8.101643166807864e-05, "loss": 0.5946, "step": 9190 }, { "epoch": 0.6168920506023288, "grad_norm": 1.137711524963379, "learning_rate": 8.100790646612477e-05, "loss": 0.5964, "step": 9192 }, { "epoch": 0.6170262742860978, "grad_norm": 1.0474289655685425, "learning_rate": 8.09993797991083e-05, "loss": 0.626, "step": 9194 }, { "epoch": 0.6171604979698668, "grad_norm": 1.2463290691375732, "learning_rate": 8.099085166743208e-05, "loss": 0.6328, "step": 9196 }, { "epoch": 0.6172947216536357, "grad_norm": 1.0076327323913574, "learning_rate": 8.098232207149907e-05, "loss": 0.6683, "step": 9198 }, { "epoch": 0.6174289453374048, "grad_norm": 1.053296446800232, "learning_rate": 8.097379101171225e-05, "loss": 0.6389, "step": 9200 }, { "epoch": 0.6175631690211738, "grad_norm": 1.063968539237976, "learning_rate": 8.096525848847473e-05, "loss": 0.6451, "step": 9202 }, { "epoch": 0.6176973927049428, "grad_norm": 0.9891154170036316, "learning_rate": 8.095672450218964e-05, "loss": 0.5218, "step": 9204 }, { "epoch": 0.6178316163887118, "grad_norm": 1.092598795890808, "learning_rate": 8.094818905326019e-05, "loss": 0.6316, "step": 9206 }, { "epoch": 0.6179658400724808, "grad_norm": 1.0701557397842407, "learning_rate": 8.093965214208964e-05, "loss": 0.6268, "step": 9208 }, { "epoch": 0.6181000637562498, "grad_norm": 1.4913824796676636, "learning_rate": 8.09311137690814e-05, "loss": 0.6582, "step": 9210 }, { "epoch": 0.6182342874400187, "grad_norm": 0.962090790271759, "learning_rate": 8.092257393463882e-05, "loss": 0.6144, "step": 9212 }, { "epoch": 0.6183685111237878, "grad_norm": 1.2915080785751343, "learning_rate": 8.091403263916546e-05, "loss": 0.6584, "step": 9214 }, { "epoch": 0.6185027348075568, "grad_norm": 0.9153003096580505, "learning_rate": 8.090548988306483e-05, "loss": 0.5892, "step": 9216 }, { "epoch": 0.6186369584913258, "grad_norm": 1.1019619703292847, "learning_rate": 8.08969456667406e-05, "loss": 0.661, "step": 9218 }, { "epoch": 0.6187711821750947, "grad_norm": 1.0655659437179565, "learning_rate": 8.088839999059642e-05, "loss": 0.705, "step": 9220 }, { "epoch": 0.6189054058588638, "grad_norm": 1.0463535785675049, "learning_rate": 8.087985285503609e-05, "loss": 0.5788, "step": 9222 }, { "epoch": 0.6190396295426328, "grad_norm": 0.9526081681251526, "learning_rate": 8.087130426046343e-05, "loss": 0.5708, "step": 9224 }, { "epoch": 0.6191738532264018, "grad_norm": 1.0919690132141113, "learning_rate": 8.086275420728235e-05, "loss": 0.604, "step": 9226 }, { "epoch": 0.6193080769101708, "grad_norm": 0.8936063647270203, "learning_rate": 8.085420269589682e-05, "loss": 0.5557, "step": 9228 }, { "epoch": 0.6194423005939398, "grad_norm": 1.1682130098342896, "learning_rate": 8.08456497267109e-05, "loss": 0.6856, "step": 9230 }, { "epoch": 0.6195765242777088, "grad_norm": 1.5496854782104492, "learning_rate": 8.083709530012867e-05, "loss": 0.6227, "step": 9232 }, { "epoch": 0.6197107479614778, "grad_norm": 1.0385692119598389, "learning_rate": 8.082853941655433e-05, "loss": 0.6537, "step": 9234 }, { "epoch": 0.6198449716452468, "grad_norm": 0.9992638826370239, "learning_rate": 8.081998207639212e-05, "loss": 0.6415, "step": 9236 }, { "epoch": 0.6199791953290158, "grad_norm": 1.149176836013794, "learning_rate": 8.081142328004637e-05, "loss": 0.6335, "step": 9238 }, { "epoch": 0.6201134190127848, "grad_norm": 1.0221976041793823, "learning_rate": 8.080286302792146e-05, "loss": 0.687, "step": 9240 }, { "epoch": 0.6202476426965539, "grad_norm": 1.1299258470535278, "learning_rate": 8.079430132042183e-05, "loss": 0.5745, "step": 9242 }, { "epoch": 0.6203818663803228, "grad_norm": 0.9322912096977234, "learning_rate": 8.078573815795203e-05, "loss": 0.6768, "step": 9244 }, { "epoch": 0.6205160900640918, "grad_norm": 0.8978147506713867, "learning_rate": 8.077717354091663e-05, "loss": 0.5811, "step": 9246 }, { "epoch": 0.6206503137478608, "grad_norm": 0.9364485144615173, "learning_rate": 8.07686074697203e-05, "loss": 0.5774, "step": 9248 }, { "epoch": 0.6207845374316298, "grad_norm": 1.059453010559082, "learning_rate": 8.076003994476778e-05, "loss": 0.6521, "step": 9250 }, { "epoch": 0.6209187611153988, "grad_norm": 1.4061264991760254, "learning_rate": 8.075147096646387e-05, "loss": 0.6372, "step": 9252 }, { "epoch": 0.6210529847991678, "grad_norm": 1.0206187963485718, "learning_rate": 8.074290053521341e-05, "loss": 0.5841, "step": 9254 }, { "epoch": 0.6211872084829368, "grad_norm": 1.3406577110290527, "learning_rate": 8.073432865142135e-05, "loss": 0.7033, "step": 9256 }, { "epoch": 0.6213214321667058, "grad_norm": 1.1175163984298706, "learning_rate": 8.07257553154927e-05, "loss": 0.6678, "step": 9258 }, { "epoch": 0.6214556558504748, "grad_norm": 1.0556981563568115, "learning_rate": 8.071718052783253e-05, "loss": 0.6165, "step": 9260 }, { "epoch": 0.6215898795342438, "grad_norm": 1.007433295249939, "learning_rate": 8.070860428884599e-05, "loss": 0.588, "step": 9262 }, { "epoch": 0.6217241032180129, "grad_norm": 1.1114963293075562, "learning_rate": 8.070002659893829e-05, "loss": 0.5717, "step": 9264 }, { "epoch": 0.6218583269017818, "grad_norm": 1.250169277191162, "learning_rate": 8.069144745851469e-05, "loss": 0.5943, "step": 9266 }, { "epoch": 0.6219925505855508, "grad_norm": 1.0131369829177856, "learning_rate": 8.068286686798054e-05, "loss": 0.5999, "step": 9268 }, { "epoch": 0.6221267742693198, "grad_norm": 0.9430072903633118, "learning_rate": 8.067428482774129e-05, "loss": 0.6898, "step": 9270 }, { "epoch": 0.6222609979530889, "grad_norm": 1.0309466123580933, "learning_rate": 8.06657013382024e-05, "loss": 0.6147, "step": 9272 }, { "epoch": 0.6223952216368578, "grad_norm": 1.0565474033355713, "learning_rate": 8.065711639976939e-05, "loss": 0.6038, "step": 9274 }, { "epoch": 0.6225294453206268, "grad_norm": 1.2949970960617065, "learning_rate": 8.064853001284793e-05, "loss": 0.7469, "step": 9276 }, { "epoch": 0.6226636690043958, "grad_norm": 1.0799611806869507, "learning_rate": 8.063994217784372e-05, "loss": 0.6407, "step": 9278 }, { "epoch": 0.6227978926881649, "grad_norm": 1.3754076957702637, "learning_rate": 8.063135289516245e-05, "loss": 0.6355, "step": 9280 }, { "epoch": 0.6229321163719338, "grad_norm": 1.0370794534683228, "learning_rate": 8.062276216521003e-05, "loss": 0.623, "step": 9282 }, { "epoch": 0.6230663400557028, "grad_norm": 1.0482323169708252, "learning_rate": 8.061416998839231e-05, "loss": 0.6153, "step": 9284 }, { "epoch": 0.6232005637394719, "grad_norm": 0.9800313711166382, "learning_rate": 8.060557636511523e-05, "loss": 0.5863, "step": 9286 }, { "epoch": 0.6233347874232408, "grad_norm": 1.1380650997161865, "learning_rate": 8.059698129578486e-05, "loss": 0.6261, "step": 9288 }, { "epoch": 0.6234690111070098, "grad_norm": 1.0158618688583374, "learning_rate": 8.058838478080731e-05, "loss": 0.6202, "step": 9290 }, { "epoch": 0.6236032347907788, "grad_norm": 1.075456976890564, "learning_rate": 8.05797868205887e-05, "loss": 0.6662, "step": 9292 }, { "epoch": 0.6237374584745479, "grad_norm": 0.8624019622802734, "learning_rate": 8.057118741553533e-05, "loss": 0.5554, "step": 9294 }, { "epoch": 0.6238716821583168, "grad_norm": 1.1782561540603638, "learning_rate": 8.056258656605344e-05, "loss": 0.6102, "step": 9296 }, { "epoch": 0.6240059058420858, "grad_norm": 1.1685196161270142, "learning_rate": 8.055398427254945e-05, "loss": 0.6215, "step": 9298 }, { "epoch": 0.6241401295258548, "grad_norm": 1.2361077070236206, "learning_rate": 8.054538053542978e-05, "loss": 0.6379, "step": 9300 }, { "epoch": 0.6242743532096239, "grad_norm": 1.1202350854873657, "learning_rate": 8.053677535510094e-05, "loss": 0.6236, "step": 9302 }, { "epoch": 0.6244085768933928, "grad_norm": 0.9103413820266724, "learning_rate": 8.052816873196952e-05, "loss": 0.584, "step": 9304 }, { "epoch": 0.6245428005771618, "grad_norm": 1.1568377017974854, "learning_rate": 8.051956066644217e-05, "loss": 0.6205, "step": 9306 }, { "epoch": 0.6246770242609309, "grad_norm": 0.978415846824646, "learning_rate": 8.051095115892557e-05, "loss": 0.675, "step": 9308 }, { "epoch": 0.6248112479446999, "grad_norm": 0.939058244228363, "learning_rate": 8.050234020982653e-05, "loss": 0.6457, "step": 9310 }, { "epoch": 0.6249454716284688, "grad_norm": 0.9567255973815918, "learning_rate": 8.04937278195519e-05, "loss": 0.6117, "step": 9312 }, { "epoch": 0.6250796953122378, "grad_norm": 0.9711152911186218, "learning_rate": 8.04851139885086e-05, "loss": 0.5895, "step": 9314 }, { "epoch": 0.6252139189960069, "grad_norm": 0.9670515060424805, "learning_rate": 8.047649871710359e-05, "loss": 0.6798, "step": 9316 }, { "epoch": 0.6253481426797759, "grad_norm": 1.1726408004760742, "learning_rate": 8.046788200574395e-05, "loss": 0.6938, "step": 9318 }, { "epoch": 0.6254823663635448, "grad_norm": 0.9561095237731934, "learning_rate": 8.045926385483682e-05, "loss": 0.6577, "step": 9320 }, { "epoch": 0.6256165900473138, "grad_norm": 0.926373302936554, "learning_rate": 8.045064426478935e-05, "loss": 0.6572, "step": 9322 }, { "epoch": 0.6257508137310829, "grad_norm": 1.00580894947052, "learning_rate": 8.044202323600882e-05, "loss": 0.5591, "step": 9324 }, { "epoch": 0.6258850374148518, "grad_norm": 1.028042197227478, "learning_rate": 8.043340076890256e-05, "loss": 0.6075, "step": 9326 }, { "epoch": 0.6260192610986208, "grad_norm": 1.2293074131011963, "learning_rate": 8.042477686387794e-05, "loss": 0.5969, "step": 9328 }, { "epoch": 0.6261534847823899, "grad_norm": 0.9517163038253784, "learning_rate": 8.041615152134247e-05, "loss": 0.605, "step": 9330 }, { "epoch": 0.6262877084661589, "grad_norm": 1.2889013290405273, "learning_rate": 8.040752474170364e-05, "loss": 0.5905, "step": 9332 }, { "epoch": 0.6264219321499278, "grad_norm": 0.9943846464157104, "learning_rate": 8.039889652536905e-05, "loss": 0.5722, "step": 9334 }, { "epoch": 0.6265561558336968, "grad_norm": 1.0777925252914429, "learning_rate": 8.039026687274638e-05, "loss": 0.6051, "step": 9336 }, { "epoch": 0.6266903795174659, "grad_norm": 2.0752015113830566, "learning_rate": 8.038163578424336e-05, "loss": 0.6598, "step": 9338 }, { "epoch": 0.6268246032012349, "grad_norm": 1.0468372106552124, "learning_rate": 8.037300326026779e-05, "loss": 0.6887, "step": 9340 }, { "epoch": 0.6269588268850038, "grad_norm": 1.1117974519729614, "learning_rate": 8.036436930122754e-05, "loss": 0.6429, "step": 9342 }, { "epoch": 0.6270930505687728, "grad_norm": 1.2182691097259521, "learning_rate": 8.035573390753054e-05, "loss": 0.6092, "step": 9344 }, { "epoch": 0.6272272742525419, "grad_norm": 1.0793417692184448, "learning_rate": 8.034709707958483e-05, "loss": 0.7057, "step": 9346 }, { "epoch": 0.6273614979363109, "grad_norm": 1.2882823944091797, "learning_rate": 8.033845881779845e-05, "loss": 0.6182, "step": 9348 }, { "epoch": 0.6274957216200798, "grad_norm": 1.1499978303909302, "learning_rate": 8.032981912257955e-05, "loss": 0.6296, "step": 9350 }, { "epoch": 0.6276299453038489, "grad_norm": 0.9525328278541565, "learning_rate": 8.032117799433634e-05, "loss": 0.594, "step": 9352 }, { "epoch": 0.6277641689876179, "grad_norm": 1.1425232887268066, "learning_rate": 8.031253543347708e-05, "loss": 0.7363, "step": 9354 }, { "epoch": 0.6278983926713869, "grad_norm": 0.9608234167098999, "learning_rate": 8.030389144041014e-05, "loss": 0.5717, "step": 9356 }, { "epoch": 0.6280326163551558, "grad_norm": 1.0381462574005127, "learning_rate": 8.029524601554392e-05, "loss": 0.588, "step": 9358 }, { "epoch": 0.6281668400389249, "grad_norm": 1.111899495124817, "learning_rate": 8.028659915928689e-05, "loss": 0.6468, "step": 9360 }, { "epoch": 0.6283010637226939, "grad_norm": 0.9303768277168274, "learning_rate": 8.027795087204761e-05, "loss": 0.599, "step": 9362 }, { "epoch": 0.6284352874064628, "grad_norm": 0.9777799844741821, "learning_rate": 8.026930115423469e-05, "loss": 0.6209, "step": 9364 }, { "epoch": 0.6285695110902318, "grad_norm": 1.0423401594161987, "learning_rate": 8.026065000625681e-05, "loss": 0.5481, "step": 9366 }, { "epoch": 0.6287037347740009, "grad_norm": 1.0492119789123535, "learning_rate": 8.025199742852272e-05, "loss": 0.6571, "step": 9368 }, { "epoch": 0.6288379584577699, "grad_norm": 0.9871809482574463, "learning_rate": 8.024334342144124e-05, "loss": 0.6433, "step": 9370 }, { "epoch": 0.6289721821415388, "grad_norm": 0.9131199717521667, "learning_rate": 8.023468798542127e-05, "loss": 0.5715, "step": 9372 }, { "epoch": 0.6291064058253079, "grad_norm": 0.976167619228363, "learning_rate": 8.022603112087174e-05, "loss": 0.5816, "step": 9374 }, { "epoch": 0.6292406295090769, "grad_norm": 1.010567545890808, "learning_rate": 8.021737282820167e-05, "loss": 0.6013, "step": 9376 }, { "epoch": 0.6293748531928459, "grad_norm": 1.1628448963165283, "learning_rate": 8.020871310782015e-05, "loss": 0.667, "step": 9378 }, { "epoch": 0.6295090768766148, "grad_norm": 1.045589804649353, "learning_rate": 8.020005196013636e-05, "loss": 0.6611, "step": 9380 }, { "epoch": 0.6296433005603839, "grad_norm": 0.9522994160652161, "learning_rate": 8.019138938555947e-05, "loss": 0.6149, "step": 9382 }, { "epoch": 0.6297775242441529, "grad_norm": 2.1528701782226562, "learning_rate": 8.018272538449882e-05, "loss": 0.616, "step": 9384 }, { "epoch": 0.6299117479279219, "grad_norm": 1.1953459978103638, "learning_rate": 8.017405995736374e-05, "loss": 0.6366, "step": 9386 }, { "epoch": 0.6300459716116908, "grad_norm": 1.0215733051300049, "learning_rate": 8.016539310456367e-05, "loss": 0.5684, "step": 9388 }, { "epoch": 0.6301801952954599, "grad_norm": 0.9657465815544128, "learning_rate": 8.01567248265081e-05, "loss": 0.6225, "step": 9390 }, { "epoch": 0.6303144189792289, "grad_norm": 0.8911698460578918, "learning_rate": 8.014805512360655e-05, "loss": 0.5509, "step": 9392 }, { "epoch": 0.6304486426629979, "grad_norm": 1.2820117473602295, "learning_rate": 8.01393839962687e-05, "loss": 0.6528, "step": 9394 }, { "epoch": 0.6305828663467669, "grad_norm": 1.0842069387435913, "learning_rate": 8.013071144490423e-05, "loss": 0.5732, "step": 9396 }, { "epoch": 0.6307170900305359, "grad_norm": 1.0435683727264404, "learning_rate": 8.012203746992288e-05, "loss": 0.6356, "step": 9398 }, { "epoch": 0.6308513137143049, "grad_norm": 1.0976462364196777, "learning_rate": 8.01133620717345e-05, "loss": 0.5647, "step": 9400 }, { "epoch": 0.6309855373980738, "grad_norm": 1.0342777967453003, "learning_rate": 8.010468525074897e-05, "loss": 0.6827, "step": 9402 }, { "epoch": 0.6311197610818429, "grad_norm": 1.0499624013900757, "learning_rate": 8.009600700737627e-05, "loss": 0.5681, "step": 9404 }, { "epoch": 0.6312539847656119, "grad_norm": 1.211155652999878, "learning_rate": 8.008732734202642e-05, "loss": 0.6753, "step": 9406 }, { "epoch": 0.6313882084493809, "grad_norm": 1.012489676475525, "learning_rate": 8.007864625510951e-05, "loss": 0.6366, "step": 9408 }, { "epoch": 0.6315224321331498, "grad_norm": 1.0493266582489014, "learning_rate": 8.006996374703572e-05, "loss": 0.5444, "step": 9410 }, { "epoch": 0.6316566558169189, "grad_norm": 1.1480631828308105, "learning_rate": 8.006127981821526e-05, "loss": 0.5618, "step": 9412 }, { "epoch": 0.6317908795006879, "grad_norm": 1.0542213916778564, "learning_rate": 8.005259446905845e-05, "loss": 0.5866, "step": 9414 }, { "epoch": 0.6319251031844569, "grad_norm": 1.1040873527526855, "learning_rate": 8.004390769997565e-05, "loss": 0.742, "step": 9416 }, { "epoch": 0.6320593268682259, "grad_norm": 1.099540114402771, "learning_rate": 8.003521951137728e-05, "loss": 0.6227, "step": 9418 }, { "epoch": 0.6321935505519949, "grad_norm": 1.1449235677719116, "learning_rate": 8.002652990367385e-05, "loss": 0.6253, "step": 9420 }, { "epoch": 0.6323277742357639, "grad_norm": 1.0806974172592163, "learning_rate": 8.001783887727594e-05, "loss": 0.6733, "step": 9422 }, { "epoch": 0.632461997919533, "grad_norm": 1.1577404737472534, "learning_rate": 8.000914643259416e-05, "loss": 0.6744, "step": 9424 }, { "epoch": 0.6325962216033019, "grad_norm": 1.5830479860305786, "learning_rate": 8.000045257003923e-05, "loss": 0.6348, "step": 9426 }, { "epoch": 0.6327304452870709, "grad_norm": 1.038620948791504, "learning_rate": 7.999175729002191e-05, "loss": 0.6083, "step": 9428 }, { "epoch": 0.6328646689708399, "grad_norm": 0.9739518165588379, "learning_rate": 7.998306059295301e-05, "loss": 0.6105, "step": 9430 }, { "epoch": 0.632998892654609, "grad_norm": 0.9545291662216187, "learning_rate": 7.997436247924348e-05, "loss": 0.5806, "step": 9432 }, { "epoch": 0.6331331163383779, "grad_norm": 0.987770676612854, "learning_rate": 7.996566294930428e-05, "loss": 0.6121, "step": 9434 }, { "epoch": 0.6332673400221469, "grad_norm": 1.0466232299804688, "learning_rate": 7.99569620035464e-05, "loss": 0.6446, "step": 9436 }, { "epoch": 0.6334015637059159, "grad_norm": 0.9641134142875671, "learning_rate": 7.994825964238099e-05, "loss": 0.7188, "step": 9438 }, { "epoch": 0.6335357873896849, "grad_norm": 1.0057556629180908, "learning_rate": 7.993955586621918e-05, "loss": 0.6589, "step": 9440 }, { "epoch": 0.6336700110734539, "grad_norm": 1.01612389087677, "learning_rate": 7.993085067547226e-05, "loss": 0.6157, "step": 9442 }, { "epoch": 0.6338042347572229, "grad_norm": 1.1171088218688965, "learning_rate": 7.992214407055148e-05, "loss": 0.6277, "step": 9444 }, { "epoch": 0.633938458440992, "grad_norm": 1.028316617012024, "learning_rate": 7.991343605186826e-05, "loss": 0.6152, "step": 9446 }, { "epoch": 0.6340726821247609, "grad_norm": 0.9967698454856873, "learning_rate": 7.990472661983398e-05, "loss": 0.6577, "step": 9448 }, { "epoch": 0.6342069058085299, "grad_norm": 1.0741626024246216, "learning_rate": 7.989601577486017e-05, "loss": 0.5831, "step": 9450 }, { "epoch": 0.6343411294922989, "grad_norm": 1.047202706336975, "learning_rate": 7.988730351735843e-05, "loss": 0.6453, "step": 9452 }, { "epoch": 0.634475353176068, "grad_norm": 0.9847677946090698, "learning_rate": 7.987858984774035e-05, "loss": 0.6171, "step": 9454 }, { "epoch": 0.6346095768598369, "grad_norm": 1.1441807746887207, "learning_rate": 7.986987476641764e-05, "loss": 0.752, "step": 9456 }, { "epoch": 0.6347438005436059, "grad_norm": 1.0286108255386353, "learning_rate": 7.98611582738021e-05, "loss": 0.6004, "step": 9458 }, { "epoch": 0.6348780242273749, "grad_norm": 1.0483300685882568, "learning_rate": 7.985244037030556e-05, "loss": 0.5558, "step": 9460 }, { "epoch": 0.635012247911144, "grad_norm": 0.9127335548400879, "learning_rate": 7.984372105633991e-05, "loss": 0.6242, "step": 9462 }, { "epoch": 0.6351464715949129, "grad_norm": 0.9870041012763977, "learning_rate": 7.983500033231711e-05, "loss": 0.5861, "step": 9464 }, { "epoch": 0.6352806952786819, "grad_norm": 1.08979070186615, "learning_rate": 7.982627819864923e-05, "loss": 0.6495, "step": 9466 }, { "epoch": 0.635414918962451, "grad_norm": 1.0896207094192505, "learning_rate": 7.981755465574834e-05, "loss": 0.5842, "step": 9468 }, { "epoch": 0.63554914264622, "grad_norm": 0.995100736618042, "learning_rate": 7.980882970402664e-05, "loss": 0.5988, "step": 9470 }, { "epoch": 0.6356833663299889, "grad_norm": 1.0793224573135376, "learning_rate": 7.980010334389636e-05, "loss": 0.646, "step": 9472 }, { "epoch": 0.6358175900137579, "grad_norm": 1.2096102237701416, "learning_rate": 7.979137557576978e-05, "loss": 0.6993, "step": 9474 }, { "epoch": 0.635951813697527, "grad_norm": 0.9157567024230957, "learning_rate": 7.978264640005928e-05, "loss": 0.5822, "step": 9476 }, { "epoch": 0.6360860373812959, "grad_norm": 1.0237129926681519, "learning_rate": 7.977391581717733e-05, "loss": 0.5526, "step": 9478 }, { "epoch": 0.6362202610650649, "grad_norm": 1.0919779539108276, "learning_rate": 7.976518382753637e-05, "loss": 0.5939, "step": 9480 }, { "epoch": 0.6363544847488339, "grad_norm": 0.9949460029602051, "learning_rate": 7.975645043154903e-05, "loss": 0.6182, "step": 9482 }, { "epoch": 0.636488708432603, "grad_norm": 1.152925968170166, "learning_rate": 7.974771562962791e-05, "loss": 0.6723, "step": 9484 }, { "epoch": 0.6366229321163719, "grad_norm": 0.9883755445480347, "learning_rate": 7.973897942218573e-05, "loss": 0.6788, "step": 9486 }, { "epoch": 0.6367571558001409, "grad_norm": 0.9843109846115112, "learning_rate": 7.973024180963526e-05, "loss": 0.5981, "step": 9488 }, { "epoch": 0.63689137948391, "grad_norm": 0.9135860204696655, "learning_rate": 7.97215027923893e-05, "loss": 0.6213, "step": 9490 }, { "epoch": 0.637025603167679, "grad_norm": 0.9985994696617126, "learning_rate": 7.97127623708608e-05, "loss": 0.5922, "step": 9492 }, { "epoch": 0.6371598268514479, "grad_norm": 1.1184334754943848, "learning_rate": 7.97040205454627e-05, "loss": 0.6606, "step": 9494 }, { "epoch": 0.6372940505352169, "grad_norm": 0.9954715371131897, "learning_rate": 7.969527731660805e-05, "loss": 0.6347, "step": 9496 }, { "epoch": 0.637428274218986, "grad_norm": 1.0252074003219604, "learning_rate": 7.968653268470992e-05, "loss": 0.6833, "step": 9498 }, { "epoch": 0.637562497902755, "grad_norm": 1.1799784898757935, "learning_rate": 7.967778665018151e-05, "loss": 0.5944, "step": 9500 }, { "epoch": 0.6376967215865239, "grad_norm": 0.8477250337600708, "learning_rate": 7.966903921343603e-05, "loss": 0.588, "step": 9502 }, { "epoch": 0.6378309452702929, "grad_norm": 1.1228663921356201, "learning_rate": 7.966029037488681e-05, "loss": 0.6696, "step": 9504 }, { "epoch": 0.637965168954062, "grad_norm": 0.9697222709655762, "learning_rate": 7.965154013494717e-05, "loss": 0.629, "step": 9506 }, { "epoch": 0.638099392637831, "grad_norm": 1.421544075012207, "learning_rate": 7.964278849403057e-05, "loss": 0.6485, "step": 9508 }, { "epoch": 0.6382336163215999, "grad_norm": 1.0229979753494263, "learning_rate": 7.96340354525505e-05, "loss": 0.585, "step": 9510 }, { "epoch": 0.638367840005369, "grad_norm": 1.0671172142028809, "learning_rate": 7.962528101092054e-05, "loss": 0.6059, "step": 9512 }, { "epoch": 0.638502063689138, "grad_norm": 0.9892696738243103, "learning_rate": 7.96165251695543e-05, "loss": 0.6253, "step": 9514 }, { "epoch": 0.6386362873729069, "grad_norm": 1.5110344886779785, "learning_rate": 7.96077679288655e-05, "loss": 0.5844, "step": 9516 }, { "epoch": 0.6387705110566759, "grad_norm": 1.0938777923583984, "learning_rate": 7.959900928926788e-05, "loss": 0.6188, "step": 9518 }, { "epoch": 0.638904734740445, "grad_norm": 1.3613741397857666, "learning_rate": 7.959024925117527e-05, "loss": 0.6533, "step": 9520 }, { "epoch": 0.639038958424214, "grad_norm": 1.0754303932189941, "learning_rate": 7.958148781500156e-05, "loss": 0.7255, "step": 9522 }, { "epoch": 0.6391731821079829, "grad_norm": 0.9454665184020996, "learning_rate": 7.957272498116073e-05, "loss": 0.6007, "step": 9524 }, { "epoch": 0.6393074057917519, "grad_norm": 0.860373318195343, "learning_rate": 7.956396075006681e-05, "loss": 0.646, "step": 9526 }, { "epoch": 0.639441629475521, "grad_norm": 0.944023847579956, "learning_rate": 7.955519512213386e-05, "loss": 0.6317, "step": 9528 }, { "epoch": 0.63957585315929, "grad_norm": 1.1715887784957886, "learning_rate": 7.954642809777606e-05, "loss": 0.6406, "step": 9530 }, { "epoch": 0.6397100768430589, "grad_norm": 1.045246958732605, "learning_rate": 7.953765967740764e-05, "loss": 0.5636, "step": 9532 }, { "epoch": 0.639844300526828, "grad_norm": 1.015788197517395, "learning_rate": 7.952888986144288e-05, "loss": 0.6354, "step": 9534 }, { "epoch": 0.639978524210597, "grad_norm": 0.9352192878723145, "learning_rate": 7.952011865029614e-05, "loss": 0.6232, "step": 9536 }, { "epoch": 0.640112747894366, "grad_norm": 1.4465974569320679, "learning_rate": 7.951134604438183e-05, "loss": 0.585, "step": 9538 }, { "epoch": 0.6402469715781349, "grad_norm": 1.1227067708969116, "learning_rate": 7.950257204411448e-05, "loss": 0.6315, "step": 9540 }, { "epoch": 0.640381195261904, "grad_norm": 0.9985955357551575, "learning_rate": 7.949379664990859e-05, "loss": 0.5605, "step": 9542 }, { "epoch": 0.640515418945673, "grad_norm": 0.9917994141578674, "learning_rate": 7.948501986217883e-05, "loss": 0.5681, "step": 9544 }, { "epoch": 0.640649642629442, "grad_norm": 1.1306357383728027, "learning_rate": 7.947624168133984e-05, "loss": 0.695, "step": 9546 }, { "epoch": 0.6407838663132109, "grad_norm": 1.1166436672210693, "learning_rate": 7.94674621078064e-05, "loss": 0.623, "step": 9548 }, { "epoch": 0.64091808999698, "grad_norm": 1.310808539390564, "learning_rate": 7.945868114199332e-05, "loss": 0.6501, "step": 9550 }, { "epoch": 0.641052313680749, "grad_norm": 1.0382336378097534, "learning_rate": 7.944989878431548e-05, "loss": 0.5603, "step": 9552 }, { "epoch": 0.6411865373645179, "grad_norm": 1.0245572328567505, "learning_rate": 7.944111503518782e-05, "loss": 0.552, "step": 9554 }, { "epoch": 0.641320761048287, "grad_norm": 1.0605592727661133, "learning_rate": 7.943232989502541e-05, "loss": 0.616, "step": 9556 }, { "epoch": 0.641454984732056, "grad_norm": 1.1174709796905518, "learning_rate": 7.942354336424325e-05, "loss": 0.6506, "step": 9558 }, { "epoch": 0.641589208415825, "grad_norm": 1.1075295209884644, "learning_rate": 7.941475544325654e-05, "loss": 0.5926, "step": 9560 }, { "epoch": 0.6417234320995939, "grad_norm": 1.1021950244903564, "learning_rate": 7.940596613248048e-05, "loss": 0.6594, "step": 9562 }, { "epoch": 0.641857655783363, "grad_norm": 1.1112298965454102, "learning_rate": 7.939717543233034e-05, "loss": 0.6499, "step": 9564 }, { "epoch": 0.641991879467132, "grad_norm": 0.9838439226150513, "learning_rate": 7.938838334322147e-05, "loss": 0.6146, "step": 9566 }, { "epoch": 0.642126103150901, "grad_norm": 1.048783779144287, "learning_rate": 7.937958986556927e-05, "loss": 0.6297, "step": 9568 }, { "epoch": 0.6422603268346699, "grad_norm": 1.0209814310073853, "learning_rate": 7.937079499978923e-05, "loss": 0.5854, "step": 9570 }, { "epoch": 0.642394550518439, "grad_norm": 0.9504020810127258, "learning_rate": 7.936199874629689e-05, "loss": 0.645, "step": 9572 }, { "epoch": 0.642528774202208, "grad_norm": 0.9745621085166931, "learning_rate": 7.935320110550783e-05, "loss": 0.5738, "step": 9574 }, { "epoch": 0.642662997885977, "grad_norm": 1.0484421253204346, "learning_rate": 7.934440207783773e-05, "loss": 0.6287, "step": 9576 }, { "epoch": 0.642797221569746, "grad_norm": 0.9482985138893127, "learning_rate": 7.933560166370235e-05, "loss": 0.5943, "step": 9578 }, { "epoch": 0.642931445253515, "grad_norm": 1.031992793083191, "learning_rate": 7.932679986351748e-05, "loss": 0.6229, "step": 9580 }, { "epoch": 0.643065668937284, "grad_norm": 1.068016767501831, "learning_rate": 7.931799667769899e-05, "loss": 0.6183, "step": 9582 }, { "epoch": 0.643199892621053, "grad_norm": 0.9357966780662537, "learning_rate": 7.93091921066628e-05, "loss": 0.635, "step": 9584 }, { "epoch": 0.643334116304822, "grad_norm": 0.9679009914398193, "learning_rate": 7.930038615082491e-05, "loss": 0.5797, "step": 9586 }, { "epoch": 0.643468339988591, "grad_norm": 1.1679974794387817, "learning_rate": 7.92915788106014e-05, "loss": 0.6083, "step": 9588 }, { "epoch": 0.64360256367236, "grad_norm": 1.0740517377853394, "learning_rate": 7.928277008640838e-05, "loss": 0.562, "step": 9590 }, { "epoch": 0.6437367873561289, "grad_norm": 1.0679657459259033, "learning_rate": 7.927395997866207e-05, "loss": 0.5675, "step": 9592 }, { "epoch": 0.643871011039898, "grad_norm": 1.0845129489898682, "learning_rate": 7.926514848777871e-05, "loss": 0.6104, "step": 9594 }, { "epoch": 0.644005234723667, "grad_norm": 1.0254936218261719, "learning_rate": 7.925633561417462e-05, "loss": 0.6037, "step": 9596 }, { "epoch": 0.644139458407436, "grad_norm": 1.9534566402435303, "learning_rate": 7.924752135826623e-05, "loss": 0.6097, "step": 9598 }, { "epoch": 0.644273682091205, "grad_norm": 1.1238148212432861, "learning_rate": 7.923870572046994e-05, "loss": 0.5883, "step": 9600 }, { "epoch": 0.644407905774974, "grad_norm": 1.180912733078003, "learning_rate": 7.922988870120232e-05, "loss": 0.6708, "step": 9602 }, { "epoch": 0.644542129458743, "grad_norm": 1.112041711807251, "learning_rate": 7.922107030087992e-05, "loss": 0.5837, "step": 9604 }, { "epoch": 0.644676353142512, "grad_norm": 1.07122802734375, "learning_rate": 7.921225051991942e-05, "loss": 0.616, "step": 9606 }, { "epoch": 0.644810576826281, "grad_norm": 1.001197338104248, "learning_rate": 7.920342935873752e-05, "loss": 0.6973, "step": 9608 }, { "epoch": 0.64494480051005, "grad_norm": 1.050616979598999, "learning_rate": 7.919460681775101e-05, "loss": 0.5911, "step": 9610 }, { "epoch": 0.645079024193819, "grad_norm": 0.9228583574295044, "learning_rate": 7.918578289737675e-05, "loss": 0.6037, "step": 9612 }, { "epoch": 0.645213247877588, "grad_norm": 1.0434296131134033, "learning_rate": 7.917695759803163e-05, "loss": 0.6193, "step": 9614 }, { "epoch": 0.645347471561357, "grad_norm": 0.98051518201828, "learning_rate": 7.916813092013264e-05, "loss": 0.5831, "step": 9616 }, { "epoch": 0.645481695245126, "grad_norm": 0.9835851192474365, "learning_rate": 7.91593028640968e-05, "loss": 0.6135, "step": 9618 }, { "epoch": 0.645615918928895, "grad_norm": 1.0886749029159546, "learning_rate": 7.915047343034128e-05, "loss": 0.625, "step": 9620 }, { "epoch": 0.6457501426126641, "grad_norm": 1.1845823526382446, "learning_rate": 7.914164261928318e-05, "loss": 0.5843, "step": 9622 }, { "epoch": 0.645884366296433, "grad_norm": 0.809241533279419, "learning_rate": 7.913281043133978e-05, "loss": 0.5533, "step": 9624 }, { "epoch": 0.646018589980202, "grad_norm": 0.9766884446144104, "learning_rate": 7.912397686692837e-05, "loss": 0.6157, "step": 9626 }, { "epoch": 0.646152813663971, "grad_norm": 1.074600338935852, "learning_rate": 7.911514192646632e-05, "loss": 0.598, "step": 9628 }, { "epoch": 0.64628703734774, "grad_norm": 1.175805687904358, "learning_rate": 7.910630561037109e-05, "loss": 0.6646, "step": 9630 }, { "epoch": 0.646421261031509, "grad_norm": 1.9594348669052124, "learning_rate": 7.909746791906013e-05, "loss": 0.621, "step": 9632 }, { "epoch": 0.646555484715278, "grad_norm": 1.1910063028335571, "learning_rate": 7.908862885295103e-05, "loss": 0.6405, "step": 9634 }, { "epoch": 0.646689708399047, "grad_norm": 0.9968884587287903, "learning_rate": 7.907978841246144e-05, "loss": 0.5714, "step": 9636 }, { "epoch": 0.646823932082816, "grad_norm": 1.0536750555038452, "learning_rate": 7.907094659800902e-05, "loss": 0.6174, "step": 9638 }, { "epoch": 0.646958155766585, "grad_norm": 2.651026725769043, "learning_rate": 7.906210341001152e-05, "loss": 0.5597, "step": 9640 }, { "epoch": 0.647092379450354, "grad_norm": 1.0783993005752563, "learning_rate": 7.905325884888679e-05, "loss": 0.6645, "step": 9642 }, { "epoch": 0.6472266031341231, "grad_norm": 0.9372936487197876, "learning_rate": 7.90444129150527e-05, "loss": 0.5837, "step": 9644 }, { "epoch": 0.647360826817892, "grad_norm": 1.0757248401641846, "learning_rate": 7.903556560892724e-05, "loss": 0.6846, "step": 9646 }, { "epoch": 0.647495050501661, "grad_norm": 1.0966620445251465, "learning_rate": 7.90267169309284e-05, "loss": 0.6526, "step": 9648 }, { "epoch": 0.64762927418543, "grad_norm": 0.9838707447052002, "learning_rate": 7.901786688147426e-05, "loss": 0.6245, "step": 9650 }, { "epoch": 0.6477634978691991, "grad_norm": 1.0680439472198486, "learning_rate": 7.900901546098296e-05, "loss": 0.6486, "step": 9652 }, { "epoch": 0.647897721552968, "grad_norm": 1.0925967693328857, "learning_rate": 7.900016266987274e-05, "loss": 0.6563, "step": 9654 }, { "epoch": 0.648031945236737, "grad_norm": 1.2490516901016235, "learning_rate": 7.899130850856184e-05, "loss": 0.6367, "step": 9656 }, { "epoch": 0.648166168920506, "grad_norm": 1.0596709251403809, "learning_rate": 7.898245297746863e-05, "loss": 0.6069, "step": 9658 }, { "epoch": 0.6483003926042751, "grad_norm": 1.0604771375656128, "learning_rate": 7.897359607701151e-05, "loss": 0.6104, "step": 9660 }, { "epoch": 0.648434616288044, "grad_norm": 1.1038539409637451, "learning_rate": 7.896473780760895e-05, "loss": 0.5665, "step": 9662 }, { "epoch": 0.648568839971813, "grad_norm": 0.9951624870300293, "learning_rate": 7.895587816967948e-05, "loss": 0.6164, "step": 9664 }, { "epoch": 0.6487030636555821, "grad_norm": 1.1739658117294312, "learning_rate": 7.89470171636417e-05, "loss": 0.6248, "step": 9666 }, { "epoch": 0.648837287339351, "grad_norm": 1.0021990537643433, "learning_rate": 7.89381547899143e-05, "loss": 0.5915, "step": 9668 }, { "epoch": 0.64897151102312, "grad_norm": 1.0784162282943726, "learning_rate": 7.892929104891598e-05, "loss": 0.5992, "step": 9670 }, { "epoch": 0.649105734706889, "grad_norm": 0.9180759191513062, "learning_rate": 7.892042594106555e-05, "loss": 0.5745, "step": 9672 }, { "epoch": 0.6492399583906581, "grad_norm": 1.7134888172149658, "learning_rate": 7.891155946678185e-05, "loss": 0.6183, "step": 9674 }, { "epoch": 0.649374182074427, "grad_norm": 1.005355715751648, "learning_rate": 7.890269162648382e-05, "loss": 0.6206, "step": 9676 }, { "epoch": 0.649508405758196, "grad_norm": 1.0124843120574951, "learning_rate": 7.889382242059044e-05, "loss": 0.5833, "step": 9678 }, { "epoch": 0.649642629441965, "grad_norm": 1.091232180595398, "learning_rate": 7.888495184952079e-05, "loss": 0.6312, "step": 9680 }, { "epoch": 0.6497768531257341, "grad_norm": 0.9182711243629456, "learning_rate": 7.887607991369396e-05, "loss": 0.6144, "step": 9682 }, { "epoch": 0.649911076809503, "grad_norm": 1.022800326347351, "learning_rate": 7.886720661352913e-05, "loss": 0.6135, "step": 9684 }, { "epoch": 0.650045300493272, "grad_norm": 0.9301436543464661, "learning_rate": 7.885833194944555e-05, "loss": 0.5886, "step": 9686 }, { "epoch": 0.6501795241770411, "grad_norm": 1.4711514711380005, "learning_rate": 7.884945592186255e-05, "loss": 0.6559, "step": 9688 }, { "epoch": 0.6503137478608101, "grad_norm": 1.037255883216858, "learning_rate": 7.884057853119947e-05, "loss": 0.5976, "step": 9690 }, { "epoch": 0.650447971544579, "grad_norm": 1.082667350769043, "learning_rate": 7.883169977787577e-05, "loss": 0.639, "step": 9692 }, { "epoch": 0.650582195228348, "grad_norm": 1.0646504163742065, "learning_rate": 7.882281966231094e-05, "loss": 0.6241, "step": 9694 }, { "epoch": 0.6507164189121171, "grad_norm": 0.9921514391899109, "learning_rate": 7.881393818492457e-05, "loss": 0.554, "step": 9696 }, { "epoch": 0.6508506425958861, "grad_norm": 1.0990134477615356, "learning_rate": 7.880505534613629e-05, "loss": 0.6075, "step": 9698 }, { "epoch": 0.650984866279655, "grad_norm": 1.0666207075119019, "learning_rate": 7.879617114636577e-05, "loss": 0.6823, "step": 9700 }, { "epoch": 0.651119089963424, "grad_norm": 0.94024658203125, "learning_rate": 7.87872855860328e-05, "loss": 0.5955, "step": 9702 }, { "epoch": 0.6512533136471931, "grad_norm": 1.038899302482605, "learning_rate": 7.877839866555719e-05, "loss": 0.6509, "step": 9704 }, { "epoch": 0.651387537330962, "grad_norm": 1.012138843536377, "learning_rate": 7.876951038535883e-05, "loss": 0.6723, "step": 9706 }, { "epoch": 0.651521761014731, "grad_norm": 1.1637332439422607, "learning_rate": 7.876062074585768e-05, "loss": 0.6284, "step": 9708 }, { "epoch": 0.6516559846985001, "grad_norm": 1.057187557220459, "learning_rate": 7.875172974747376e-05, "loss": 0.5865, "step": 9710 }, { "epoch": 0.6517902083822691, "grad_norm": 0.8246267437934875, "learning_rate": 7.874283739062715e-05, "loss": 0.5282, "step": 9712 }, { "epoch": 0.651924432066038, "grad_norm": 1.2830623388290405, "learning_rate": 7.873394367573798e-05, "loss": 0.5927, "step": 9714 }, { "epoch": 0.652058655749807, "grad_norm": 1.0558857917785645, "learning_rate": 7.87250486032265e-05, "loss": 0.5818, "step": 9716 }, { "epoch": 0.6521928794335761, "grad_norm": 1.1389917135238647, "learning_rate": 7.871615217351294e-05, "loss": 0.6734, "step": 9718 }, { "epoch": 0.6523271031173451, "grad_norm": 1.0388612747192383, "learning_rate": 7.870725438701765e-05, "loss": 0.6553, "step": 9720 }, { "epoch": 0.652461326801114, "grad_norm": 0.9338822364807129, "learning_rate": 7.869835524416104e-05, "loss": 0.6292, "step": 9722 }, { "epoch": 0.652595550484883, "grad_norm": 1.1659862995147705, "learning_rate": 7.86894547453636e-05, "loss": 0.6133, "step": 9724 }, { "epoch": 0.6527297741686521, "grad_norm": 0.9935160279273987, "learning_rate": 7.868055289104581e-05, "loss": 0.5499, "step": 9726 }, { "epoch": 0.6528639978524211, "grad_norm": 1.248638391494751, "learning_rate": 7.867164968162833e-05, "loss": 0.6316, "step": 9728 }, { "epoch": 0.65299822153619, "grad_norm": 1.0811436176300049, "learning_rate": 7.866274511753175e-05, "loss": 0.63, "step": 9730 }, { "epoch": 0.6531324452199591, "grad_norm": 1.0859920978546143, "learning_rate": 7.865383919917683e-05, "loss": 0.572, "step": 9732 }, { "epoch": 0.6532666689037281, "grad_norm": 1.0905475616455078, "learning_rate": 7.864493192698437e-05, "loss": 0.5775, "step": 9734 }, { "epoch": 0.6534008925874971, "grad_norm": 1.0847246646881104, "learning_rate": 7.863602330137519e-05, "loss": 0.6127, "step": 9736 }, { "epoch": 0.653535116271266, "grad_norm": 0.9611513614654541, "learning_rate": 7.862711332277023e-05, "loss": 0.5731, "step": 9738 }, { "epoch": 0.6536693399550351, "grad_norm": 0.94939786195755, "learning_rate": 7.861820199159044e-05, "loss": 0.6028, "step": 9740 }, { "epoch": 0.6538035636388041, "grad_norm": 0.9335368275642395, "learning_rate": 7.860928930825691e-05, "loss": 0.5852, "step": 9742 }, { "epoch": 0.653937787322573, "grad_norm": 1.0535403490066528, "learning_rate": 7.86003752731907e-05, "loss": 0.6592, "step": 9744 }, { "epoch": 0.654072011006342, "grad_norm": 1.0651839971542358, "learning_rate": 7.8591459886813e-05, "loss": 0.6449, "step": 9746 }, { "epoch": 0.6542062346901111, "grad_norm": 1.0187623500823975, "learning_rate": 7.858254314954505e-05, "loss": 0.6007, "step": 9748 }, { "epoch": 0.6543404583738801, "grad_norm": 1.020383596420288, "learning_rate": 7.857362506180813e-05, "loss": 0.6603, "step": 9750 }, { "epoch": 0.654474682057649, "grad_norm": 1.2457799911499023, "learning_rate": 7.856470562402362e-05, "loss": 0.5842, "step": 9752 }, { "epoch": 0.6546089057414181, "grad_norm": 1.057938814163208, "learning_rate": 7.855578483661293e-05, "loss": 0.6104, "step": 9754 }, { "epoch": 0.6547431294251871, "grad_norm": 1.1072943210601807, "learning_rate": 7.854686269999757e-05, "loss": 0.5934, "step": 9756 }, { "epoch": 0.6548773531089561, "grad_norm": 1.2148404121398926, "learning_rate": 7.853793921459909e-05, "loss": 0.6323, "step": 9758 }, { "epoch": 0.655011576792725, "grad_norm": 1.1100873947143555, "learning_rate": 7.852901438083908e-05, "loss": 0.6788, "step": 9760 }, { "epoch": 0.6551458004764941, "grad_norm": 0.9000804424285889, "learning_rate": 7.852008819913925e-05, "loss": 0.6299, "step": 9762 }, { "epoch": 0.6552800241602631, "grad_norm": 1.2299797534942627, "learning_rate": 7.851116066992133e-05, "loss": 0.603, "step": 9764 }, { "epoch": 0.6554142478440321, "grad_norm": 1.0158828496932983, "learning_rate": 7.850223179360714e-05, "loss": 0.6332, "step": 9766 }, { "epoch": 0.655548471527801, "grad_norm": 1.0191811323165894, "learning_rate": 7.849330157061854e-05, "loss": 0.5966, "step": 9768 }, { "epoch": 0.6556826952115701, "grad_norm": 1.0282412767410278, "learning_rate": 7.848437000137747e-05, "loss": 0.6104, "step": 9770 }, { "epoch": 0.6558169188953391, "grad_norm": 1.0161830186843872, "learning_rate": 7.847543708630593e-05, "loss": 0.6041, "step": 9772 }, { "epoch": 0.6559511425791081, "grad_norm": 1.0597655773162842, "learning_rate": 7.8466502825826e-05, "loss": 0.5792, "step": 9774 }, { "epoch": 0.6560853662628771, "grad_norm": 1.132139801979065, "learning_rate": 7.845756722035978e-05, "loss": 0.5909, "step": 9776 }, { "epoch": 0.6562195899466461, "grad_norm": 0.9694353342056274, "learning_rate": 7.844863027032945e-05, "loss": 0.5742, "step": 9778 }, { "epoch": 0.6563538136304151, "grad_norm": 1.0666712522506714, "learning_rate": 7.84396919761573e-05, "loss": 0.6605, "step": 9780 }, { "epoch": 0.656488037314184, "grad_norm": 0.9630760550498962, "learning_rate": 7.843075233826561e-05, "loss": 0.5949, "step": 9782 }, { "epoch": 0.6566222609979531, "grad_norm": 1.3532354831695557, "learning_rate": 7.84218113570768e-05, "loss": 0.5712, "step": 9784 }, { "epoch": 0.6567564846817221, "grad_norm": 1.3067315816879272, "learning_rate": 7.841286903301328e-05, "loss": 0.5465, "step": 9786 }, { "epoch": 0.6568907083654911, "grad_norm": 0.9235031604766846, "learning_rate": 7.840392536649758e-05, "loss": 0.5865, "step": 9788 }, { "epoch": 0.65702493204926, "grad_norm": 1.137033224105835, "learning_rate": 7.839498035795224e-05, "loss": 0.6354, "step": 9790 }, { "epoch": 0.6571591557330291, "grad_norm": 1.0483646392822266, "learning_rate": 7.838603400779993e-05, "loss": 0.6268, "step": 9792 }, { "epoch": 0.6572933794167981, "grad_norm": 1.0721083879470825, "learning_rate": 7.837708631646333e-05, "loss": 0.6168, "step": 9794 }, { "epoch": 0.6574276031005671, "grad_norm": 1.1923506259918213, "learning_rate": 7.836813728436521e-05, "loss": 0.625, "step": 9796 }, { "epoch": 0.6575618267843361, "grad_norm": 1.0472887754440308, "learning_rate": 7.835918691192837e-05, "loss": 0.6315, "step": 9798 }, { "epoch": 0.6576960504681051, "grad_norm": 1.0348756313323975, "learning_rate": 7.835023519957571e-05, "loss": 0.6201, "step": 9800 }, { "epoch": 0.6578302741518741, "grad_norm": 1.2074779272079468, "learning_rate": 7.834128214773022e-05, "loss": 0.5988, "step": 9802 }, { "epoch": 0.6579644978356431, "grad_norm": 1.0870145559310913, "learning_rate": 7.833232775681485e-05, "loss": 0.6503, "step": 9804 }, { "epoch": 0.6580987215194121, "grad_norm": 0.9929474592208862, "learning_rate": 7.83233720272527e-05, "loss": 0.6137, "step": 9806 }, { "epoch": 0.6582329452031811, "grad_norm": 1.4966610670089722, "learning_rate": 7.831441495946694e-05, "loss": 0.6278, "step": 9808 }, { "epoch": 0.6583671688869501, "grad_norm": 1.0693825483322144, "learning_rate": 7.830545655388075e-05, "loss": 0.5809, "step": 9810 }, { "epoch": 0.6585013925707192, "grad_norm": 1.047507405281067, "learning_rate": 7.829649681091738e-05, "loss": 0.5508, "step": 9812 }, { "epoch": 0.6586356162544881, "grad_norm": 1.130644679069519, "learning_rate": 7.828753573100019e-05, "loss": 0.6143, "step": 9814 }, { "epoch": 0.6587698399382571, "grad_norm": 0.9263031482696533, "learning_rate": 7.827857331455256e-05, "loss": 0.5803, "step": 9816 }, { "epoch": 0.6589040636220261, "grad_norm": 1.7844195365905762, "learning_rate": 7.826960956199794e-05, "loss": 0.5558, "step": 9818 }, { "epoch": 0.6590382873057951, "grad_norm": 1.1864254474639893, "learning_rate": 7.826064447375987e-05, "loss": 0.5985, "step": 9820 }, { "epoch": 0.6591725109895641, "grad_norm": 1.0454565286636353, "learning_rate": 7.825167805026193e-05, "loss": 0.6035, "step": 9822 }, { "epoch": 0.6593067346733331, "grad_norm": 0.9805462956428528, "learning_rate": 7.824271029192773e-05, "loss": 0.6726, "step": 9824 }, { "epoch": 0.6594409583571021, "grad_norm": 0.9167178869247437, "learning_rate": 7.823374119918103e-05, "loss": 0.5407, "step": 9826 }, { "epoch": 0.6595751820408711, "grad_norm": 0.9714899659156799, "learning_rate": 7.822477077244557e-05, "loss": 0.6605, "step": 9828 }, { "epoch": 0.6597094057246401, "grad_norm": 1.0378906726837158, "learning_rate": 7.821579901214518e-05, "loss": 0.5913, "step": 9830 }, { "epoch": 0.6598436294084091, "grad_norm": 1.2218577861785889, "learning_rate": 7.820682591870378e-05, "loss": 0.6122, "step": 9832 }, { "epoch": 0.6599778530921782, "grad_norm": 1.1689800024032593, "learning_rate": 7.819785149254532e-05, "loss": 0.6229, "step": 9834 }, { "epoch": 0.6601120767759471, "grad_norm": 1.148080587387085, "learning_rate": 7.818887573409383e-05, "loss": 0.6555, "step": 9836 }, { "epoch": 0.6602463004597161, "grad_norm": 1.0432515144348145, "learning_rate": 7.817989864377339e-05, "loss": 0.5595, "step": 9838 }, { "epoch": 0.6603805241434851, "grad_norm": 0.9731805324554443, "learning_rate": 7.817092022200816e-05, "loss": 0.5145, "step": 9840 }, { "epoch": 0.6605147478272542, "grad_norm": 0.9755486249923706, "learning_rate": 7.816194046922234e-05, "loss": 0.5965, "step": 9842 }, { "epoch": 0.6606489715110231, "grad_norm": 1.0782371759414673, "learning_rate": 7.815295938584021e-05, "loss": 0.6653, "step": 9844 }, { "epoch": 0.6607831951947921, "grad_norm": 1.184570550918579, "learning_rate": 7.81439769722861e-05, "loss": 0.6332, "step": 9846 }, { "epoch": 0.6609174188785611, "grad_norm": 0.9063990116119385, "learning_rate": 7.813499322898443e-05, "loss": 0.5187, "step": 9848 }, { "epoch": 0.6610516425623302, "grad_norm": 0.9837430715560913, "learning_rate": 7.812600815635967e-05, "loss": 0.6289, "step": 9850 }, { "epoch": 0.6611858662460991, "grad_norm": 0.9125910401344299, "learning_rate": 7.81170217548363e-05, "loss": 0.6033, "step": 9852 }, { "epoch": 0.6613200899298681, "grad_norm": 1.0609534978866577, "learning_rate": 7.810803402483897e-05, "loss": 0.5779, "step": 9854 }, { "epoch": 0.6614543136136372, "grad_norm": 0.9907367825508118, "learning_rate": 7.80990449667923e-05, "loss": 0.734, "step": 9856 }, { "epoch": 0.6615885372974061, "grad_norm": 0.9950106143951416, "learning_rate": 7.8090054581121e-05, "loss": 0.6049, "step": 9858 }, { "epoch": 0.6617227609811751, "grad_norm": 1.0695754289627075, "learning_rate": 7.808106286824986e-05, "loss": 0.6191, "step": 9860 }, { "epoch": 0.6618569846649441, "grad_norm": 1.0091043710708618, "learning_rate": 7.807206982860371e-05, "loss": 0.5934, "step": 9862 }, { "epoch": 0.6619912083487132, "grad_norm": 1.0125815868377686, "learning_rate": 7.806307546260748e-05, "loss": 0.6075, "step": 9864 }, { "epoch": 0.6621254320324821, "grad_norm": 1.187417984008789, "learning_rate": 7.80540797706861e-05, "loss": 0.6647, "step": 9866 }, { "epoch": 0.6622596557162511, "grad_norm": 1.3254411220550537, "learning_rate": 7.804508275326462e-05, "loss": 0.7018, "step": 9868 }, { "epoch": 0.6623938794000201, "grad_norm": 0.9366840124130249, "learning_rate": 7.803608441076815e-05, "loss": 0.551, "step": 9870 }, { "epoch": 0.6625281030837892, "grad_norm": 1.0737332105636597, "learning_rate": 7.80270847436218e-05, "loss": 0.6484, "step": 9872 }, { "epoch": 0.6626623267675581, "grad_norm": 0.9313533902168274, "learning_rate": 7.801808375225082e-05, "loss": 0.6453, "step": 9874 }, { "epoch": 0.6627965504513271, "grad_norm": 1.0625557899475098, "learning_rate": 7.800908143708047e-05, "loss": 0.5741, "step": 9876 }, { "epoch": 0.6629307741350962, "grad_norm": 1.0991019010543823, "learning_rate": 7.800007779853611e-05, "loss": 0.5746, "step": 9878 }, { "epoch": 0.6630649978188652, "grad_norm": 0.9642413258552551, "learning_rate": 7.799107283704312e-05, "loss": 0.6528, "step": 9880 }, { "epoch": 0.6631992215026341, "grad_norm": 1.091901421546936, "learning_rate": 7.798206655302698e-05, "loss": 0.5899, "step": 9882 }, { "epoch": 0.6633334451864031, "grad_norm": 1.035330891609192, "learning_rate": 7.797305894691322e-05, "loss": 0.6699, "step": 9884 }, { "epoch": 0.6634676688701722, "grad_norm": 1.0007174015045166, "learning_rate": 7.796405001912746e-05, "loss": 0.5949, "step": 9886 }, { "epoch": 0.6636018925539412, "grad_norm": 0.9840034246444702, "learning_rate": 7.795503977009528e-05, "loss": 0.6358, "step": 9888 }, { "epoch": 0.6637361162377101, "grad_norm": 0.9621015787124634, "learning_rate": 7.794602820024248e-05, "loss": 0.5859, "step": 9890 }, { "epoch": 0.6638703399214791, "grad_norm": 0.9338459372520447, "learning_rate": 7.793701530999478e-05, "loss": 0.58, "step": 9892 }, { "epoch": 0.6640045636052482, "grad_norm": 1.088112473487854, "learning_rate": 7.792800109977804e-05, "loss": 0.5988, "step": 9894 }, { "epoch": 0.6641387872890171, "grad_norm": 1.1450368165969849, "learning_rate": 7.791898557001818e-05, "loss": 0.664, "step": 9896 }, { "epoch": 0.6642730109727861, "grad_norm": 0.8303860425949097, "learning_rate": 7.790996872114116e-05, "loss": 0.6325, "step": 9898 }, { "epoch": 0.6644072346565552, "grad_norm": 1.0726555585861206, "learning_rate": 7.790095055357298e-05, "loss": 0.6394, "step": 9900 }, { "epoch": 0.6645414583403242, "grad_norm": 1.5649341344833374, "learning_rate": 7.789193106773976e-05, "loss": 0.6188, "step": 9902 }, { "epoch": 0.6646756820240931, "grad_norm": 1.0068936347961426, "learning_rate": 7.788291026406764e-05, "loss": 0.5768, "step": 9904 }, { "epoch": 0.6648099057078621, "grad_norm": 1.008538007736206, "learning_rate": 7.787388814298284e-05, "loss": 0.6108, "step": 9906 }, { "epoch": 0.6649441293916312, "grad_norm": 1.0401108264923096, "learning_rate": 7.786486470491165e-05, "loss": 0.6475, "step": 9908 }, { "epoch": 0.6650783530754002, "grad_norm": 1.1032782793045044, "learning_rate": 7.785583995028038e-05, "loss": 0.6025, "step": 9910 }, { "epoch": 0.6652125767591691, "grad_norm": 1.1373237371444702, "learning_rate": 7.784681387951546e-05, "loss": 0.6146, "step": 9912 }, { "epoch": 0.6653468004429381, "grad_norm": 0.8877438306808472, "learning_rate": 7.783778649304333e-05, "loss": 0.5687, "step": 9914 }, { "epoch": 0.6654810241267072, "grad_norm": 1.1277129650115967, "learning_rate": 7.782875779129055e-05, "loss": 0.5518, "step": 9916 }, { "epoch": 0.6656152478104762, "grad_norm": 1.2100125551223755, "learning_rate": 7.781972777468367e-05, "loss": 0.6513, "step": 9918 }, { "epoch": 0.6657494714942451, "grad_norm": 1.1718670129776, "learning_rate": 7.781069644364936e-05, "loss": 0.6516, "step": 9920 }, { "epoch": 0.6658836951780142, "grad_norm": 0.9917896389961243, "learning_rate": 7.780166379861432e-05, "loss": 0.604, "step": 9922 }, { "epoch": 0.6660179188617832, "grad_norm": 0.9580020308494568, "learning_rate": 7.779262984000536e-05, "loss": 0.6172, "step": 9924 }, { "epoch": 0.6661521425455522, "grad_norm": 1.0788674354553223, "learning_rate": 7.778359456824929e-05, "loss": 0.6439, "step": 9926 }, { "epoch": 0.6662863662293211, "grad_norm": 1.4100440740585327, "learning_rate": 7.777455798377297e-05, "loss": 0.6557, "step": 9928 }, { "epoch": 0.6664205899130902, "grad_norm": 1.071541428565979, "learning_rate": 7.776552008700344e-05, "loss": 0.6275, "step": 9930 }, { "epoch": 0.6665548135968592, "grad_norm": 0.9982123374938965, "learning_rate": 7.775648087836768e-05, "loss": 0.5679, "step": 9932 }, { "epoch": 0.6666890372806281, "grad_norm": 1.0546722412109375, "learning_rate": 7.774744035829277e-05, "loss": 0.7266, "step": 9934 }, { "epoch": 0.6668232609643971, "grad_norm": 1.2762454748153687, "learning_rate": 7.773839852720589e-05, "loss": 0.6516, "step": 9936 }, { "epoch": 0.6669574846481662, "grad_norm": 1.0322846174240112, "learning_rate": 7.772935538553418e-05, "loss": 0.6761, "step": 9938 }, { "epoch": 0.6670917083319352, "grad_norm": 1.0097932815551758, "learning_rate": 7.772031093370499e-05, "loss": 0.6756, "step": 9940 }, { "epoch": 0.6672259320157041, "grad_norm": 0.9951656460762024, "learning_rate": 7.771126517214561e-05, "loss": 0.6605, "step": 9942 }, { "epoch": 0.6673601556994732, "grad_norm": 0.9250411987304688, "learning_rate": 7.770221810128343e-05, "loss": 0.6245, "step": 9944 }, { "epoch": 0.6674943793832422, "grad_norm": 0.9195786714553833, "learning_rate": 7.769316972154594e-05, "loss": 0.6032, "step": 9946 }, { "epoch": 0.6676286030670112, "grad_norm": 1.2926996946334839, "learning_rate": 7.768412003336064e-05, "loss": 0.6591, "step": 9948 }, { "epoch": 0.6677628267507801, "grad_norm": 0.9514656066894531, "learning_rate": 7.767506903715509e-05, "loss": 0.6087, "step": 9950 }, { "epoch": 0.6678970504345492, "grad_norm": 1.1330329179763794, "learning_rate": 7.766601673335694e-05, "loss": 0.6591, "step": 9952 }, { "epoch": 0.6680312741183182, "grad_norm": 0.8967975974082947, "learning_rate": 7.765696312239392e-05, "loss": 0.5153, "step": 9954 }, { "epoch": 0.6681654978020872, "grad_norm": 1.05215322971344, "learning_rate": 7.764790820469377e-05, "loss": 0.5942, "step": 9956 }, { "epoch": 0.6682997214858561, "grad_norm": 1.0084857940673828, "learning_rate": 7.763885198068433e-05, "loss": 0.552, "step": 9958 }, { "epoch": 0.6684339451696252, "grad_norm": 1.0705190896987915, "learning_rate": 7.762979445079348e-05, "loss": 0.6433, "step": 9960 }, { "epoch": 0.6685681688533942, "grad_norm": 1.0716350078582764, "learning_rate": 7.762073561544918e-05, "loss": 0.6164, "step": 9962 }, { "epoch": 0.6687023925371632, "grad_norm": 0.9401054382324219, "learning_rate": 7.761167547507942e-05, "loss": 0.6359, "step": 9964 }, { "epoch": 0.6688366162209322, "grad_norm": 0.818011999130249, "learning_rate": 7.760261403011231e-05, "loss": 0.5992, "step": 9966 }, { "epoch": 0.6689708399047012, "grad_norm": 0.9985383152961731, "learning_rate": 7.759355128097595e-05, "loss": 0.6641, "step": 9968 }, { "epoch": 0.6691050635884702, "grad_norm": 1.016554832458496, "learning_rate": 7.758448722809856e-05, "loss": 0.6243, "step": 9970 }, { "epoch": 0.6692392872722391, "grad_norm": 1.0722575187683105, "learning_rate": 7.757542187190838e-05, "loss": 0.6227, "step": 9972 }, { "epoch": 0.6693735109560082, "grad_norm": 1.0455728769302368, "learning_rate": 7.756635521283375e-05, "loss": 0.63, "step": 9974 }, { "epoch": 0.6695077346397772, "grad_norm": 1.0678956508636475, "learning_rate": 7.755728725130304e-05, "loss": 0.5656, "step": 9976 }, { "epoch": 0.6696419583235462, "grad_norm": 0.9857407808303833, "learning_rate": 7.754821798774471e-05, "loss": 0.6717, "step": 9978 }, { "epoch": 0.6697761820073151, "grad_norm": 0.9332596063613892, "learning_rate": 7.753914742258724e-05, "loss": 0.5908, "step": 9980 }, { "epoch": 0.6699104056910842, "grad_norm": 1.1080443859100342, "learning_rate": 7.75300755562592e-05, "loss": 0.67, "step": 9982 }, { "epoch": 0.6700446293748532, "grad_norm": 1.1003410816192627, "learning_rate": 7.752100238918925e-05, "loss": 0.7029, "step": 9984 }, { "epoch": 0.6701788530586222, "grad_norm": 0.9595111608505249, "learning_rate": 7.751192792180604e-05, "loss": 0.5828, "step": 9986 }, { "epoch": 0.6703130767423912, "grad_norm": 1.0013678073883057, "learning_rate": 7.750285215453833e-05, "loss": 0.5596, "step": 9988 }, { "epoch": 0.6704473004261602, "grad_norm": 1.0782396793365479, "learning_rate": 7.749377508781495e-05, "loss": 0.6067, "step": 9990 }, { "epoch": 0.6705815241099292, "grad_norm": 1.0426450967788696, "learning_rate": 7.748469672206476e-05, "loss": 0.5675, "step": 9992 }, { "epoch": 0.6707157477936982, "grad_norm": 1.0867704153060913, "learning_rate": 7.747561705771669e-05, "loss": 0.6423, "step": 9994 }, { "epoch": 0.6708499714774672, "grad_norm": 0.949182391166687, "learning_rate": 7.746653609519973e-05, "loss": 0.6095, "step": 9996 }, { "epoch": 0.6709841951612362, "grad_norm": 0.9397948384284973, "learning_rate": 7.745745383494296e-05, "loss": 0.5811, "step": 9998 }, { "epoch": 0.6711184188450052, "grad_norm": 0.9894161820411682, "learning_rate": 7.74483702773755e-05, "loss": 0.6259, "step": 10000 }, { "epoch": 0.6712526425287743, "grad_norm": 1.345678448677063, "learning_rate": 7.74392854229265e-05, "loss": 0.5753, "step": 10002 }, { "epoch": 0.6713868662125432, "grad_norm": 1.0500710010528564, "learning_rate": 7.743019927202524e-05, "loss": 0.6433, "step": 10004 }, { "epoch": 0.6715210898963122, "grad_norm": 0.8496576547622681, "learning_rate": 7.7421111825101e-05, "loss": 0.5807, "step": 10006 }, { "epoch": 0.6716553135800812, "grad_norm": 1.0088499784469604, "learning_rate": 7.741202308258314e-05, "loss": 0.5823, "step": 10008 }, { "epoch": 0.6717895372638502, "grad_norm": 1.0959690809249878, "learning_rate": 7.74029330449011e-05, "loss": 0.5742, "step": 10010 }, { "epoch": 0.6719237609476192, "grad_norm": 1.1204447746276855, "learning_rate": 7.739384171248435e-05, "loss": 0.665, "step": 10012 }, { "epoch": 0.6720579846313882, "grad_norm": 1.0192700624465942, "learning_rate": 7.738474908576246e-05, "loss": 0.6114, "step": 10014 }, { "epoch": 0.6721922083151572, "grad_norm": 1.0640549659729004, "learning_rate": 7.737565516516501e-05, "loss": 0.6024, "step": 10016 }, { "epoch": 0.6723264319989262, "grad_norm": 0.9268406629562378, "learning_rate": 7.73665599511217e-05, "loss": 0.6366, "step": 10018 }, { "epoch": 0.6724606556826952, "grad_norm": 1.0345485210418701, "learning_rate": 7.735746344406223e-05, "loss": 0.5904, "step": 10020 }, { "epoch": 0.6725948793664642, "grad_norm": 1.0905174016952515, "learning_rate": 7.734836564441642e-05, "loss": 0.6618, "step": 10022 }, { "epoch": 0.6727291030502333, "grad_norm": 2.176734447479248, "learning_rate": 7.733926655261411e-05, "loss": 0.6098, "step": 10024 }, { "epoch": 0.6728633267340022, "grad_norm": 1.0297119617462158, "learning_rate": 7.733016616908521e-05, "loss": 0.624, "step": 10026 }, { "epoch": 0.6729975504177712, "grad_norm": 1.0259052515029907, "learning_rate": 7.73210644942597e-05, "loss": 0.6121, "step": 10028 }, { "epoch": 0.6731317741015402, "grad_norm": 1.0044418573379517, "learning_rate": 7.731196152856763e-05, "loss": 0.5884, "step": 10030 }, { "epoch": 0.6732659977853093, "grad_norm": 1.0181063413619995, "learning_rate": 7.730285727243907e-05, "loss": 0.6566, "step": 10032 }, { "epoch": 0.6734002214690782, "grad_norm": 1.5579719543457031, "learning_rate": 7.72937517263042e-05, "loss": 0.5857, "step": 10034 }, { "epoch": 0.6735344451528472, "grad_norm": 1.3661072254180908, "learning_rate": 7.728464489059324e-05, "loss": 0.6911, "step": 10036 }, { "epoch": 0.6736686688366162, "grad_norm": 1.1502236127853394, "learning_rate": 7.727553676573644e-05, "loss": 0.6244, "step": 10038 }, { "epoch": 0.6738028925203853, "grad_norm": 1.0469461679458618, "learning_rate": 7.726642735216418e-05, "loss": 0.5851, "step": 10040 }, { "epoch": 0.6739371162041542, "grad_norm": 1.1454188823699951, "learning_rate": 7.725731665030684e-05, "loss": 0.626, "step": 10042 }, { "epoch": 0.6740713398879232, "grad_norm": 1.0856674909591675, "learning_rate": 7.72482046605949e-05, "loss": 0.6999, "step": 10044 }, { "epoch": 0.6742055635716923, "grad_norm": 1.0814210176467896, "learning_rate": 7.723909138345886e-05, "loss": 0.6791, "step": 10046 }, { "epoch": 0.6743397872554612, "grad_norm": 1.538653016090393, "learning_rate": 7.722997681932931e-05, "loss": 0.6329, "step": 10048 }, { "epoch": 0.6744740109392302, "grad_norm": 0.9768089056015015, "learning_rate": 7.722086096863692e-05, "loss": 0.6018, "step": 10050 }, { "epoch": 0.6746082346229992, "grad_norm": 1.0232763290405273, "learning_rate": 7.721174383181235e-05, "loss": 0.6223, "step": 10052 }, { "epoch": 0.6747424583067683, "grad_norm": 1.0074292421340942, "learning_rate": 7.720262540928641e-05, "loss": 0.613, "step": 10054 }, { "epoch": 0.6748766819905372, "grad_norm": 1.0369144678115845, "learning_rate": 7.719350570148991e-05, "loss": 0.5817, "step": 10056 }, { "epoch": 0.6750109056743062, "grad_norm": 1.017956256866455, "learning_rate": 7.718438470885375e-05, "loss": 0.5936, "step": 10058 }, { "epoch": 0.6751451293580752, "grad_norm": 1.0069831609725952, "learning_rate": 7.717526243180883e-05, "loss": 0.5982, "step": 10060 }, { "epoch": 0.6752793530418443, "grad_norm": 0.9372869729995728, "learning_rate": 7.716613887078623e-05, "loss": 0.6255, "step": 10062 }, { "epoch": 0.6754135767256132, "grad_norm": 1.2310115098953247, "learning_rate": 7.715701402621699e-05, "loss": 0.6804, "step": 10064 }, { "epoch": 0.6755478004093822, "grad_norm": 1.140617847442627, "learning_rate": 7.714788789853223e-05, "loss": 0.5764, "step": 10066 }, { "epoch": 0.6756820240931513, "grad_norm": 1.0454936027526855, "learning_rate": 7.713876048816316e-05, "loss": 0.6441, "step": 10068 }, { "epoch": 0.6758162477769203, "grad_norm": 1.1380059719085693, "learning_rate": 7.7129631795541e-05, "loss": 0.6409, "step": 10070 }, { "epoch": 0.6759504714606892, "grad_norm": 1.0036143064498901, "learning_rate": 7.712050182109711e-05, "loss": 0.5859, "step": 10072 }, { "epoch": 0.6760846951444582, "grad_norm": 1.0466136932373047, "learning_rate": 7.711137056526283e-05, "loss": 0.6363, "step": 10074 }, { "epoch": 0.6762189188282273, "grad_norm": 1.055659294128418, "learning_rate": 7.71022380284696e-05, "loss": 0.6195, "step": 10076 }, { "epoch": 0.6763531425119963, "grad_norm": 1.0443370342254639, "learning_rate": 7.709310421114892e-05, "loss": 0.6196, "step": 10078 }, { "epoch": 0.6764873661957652, "grad_norm": 1.1877312660217285, "learning_rate": 7.708396911373233e-05, "loss": 0.5761, "step": 10080 }, { "epoch": 0.6766215898795342, "grad_norm": 1.0187946557998657, "learning_rate": 7.707483273665149e-05, "loss": 0.6427, "step": 10082 }, { "epoch": 0.6767558135633033, "grad_norm": 1.349446177482605, "learning_rate": 7.706569508033801e-05, "loss": 0.637, "step": 10084 }, { "epoch": 0.6768900372470722, "grad_norm": 1.0880353450775146, "learning_rate": 7.705655614522367e-05, "loss": 0.6479, "step": 10086 }, { "epoch": 0.6770242609308412, "grad_norm": 0.7679103016853333, "learning_rate": 7.704741593174026e-05, "loss": 0.5232, "step": 10088 }, { "epoch": 0.6771584846146103, "grad_norm": 1.0523754358291626, "learning_rate": 7.703827444031963e-05, "loss": 0.5853, "step": 10090 }, { "epoch": 0.6772927082983793, "grad_norm": 1.0461854934692383, "learning_rate": 7.702913167139372e-05, "loss": 0.6162, "step": 10092 }, { "epoch": 0.6774269319821482, "grad_norm": 0.9754141569137573, "learning_rate": 7.701998762539446e-05, "loss": 0.621, "step": 10094 }, { "epoch": 0.6775611556659172, "grad_norm": 1.0695887804031372, "learning_rate": 7.701084230275392e-05, "loss": 0.6168, "step": 10096 }, { "epoch": 0.6776953793496863, "grad_norm": 1.1557365655899048, "learning_rate": 7.700169570390422e-05, "loss": 0.6171, "step": 10098 }, { "epoch": 0.6778296030334553, "grad_norm": 1.1904844045639038, "learning_rate": 7.699254782927749e-05, "loss": 0.6529, "step": 10100 }, { "epoch": 0.6779638267172242, "grad_norm": 1.0055497884750366, "learning_rate": 7.698339867930592e-05, "loss": 0.6299, "step": 10102 }, { "epoch": 0.6780980504009932, "grad_norm": 1.1452466249465942, "learning_rate": 7.697424825442187e-05, "loss": 0.5837, "step": 10104 }, { "epoch": 0.6782322740847623, "grad_norm": 1.1297428607940674, "learning_rate": 7.69650965550576e-05, "loss": 0.6165, "step": 10106 }, { "epoch": 0.6783664977685313, "grad_norm": 1.0672310590744019, "learning_rate": 7.695594358164557e-05, "loss": 0.6501, "step": 10108 }, { "epoch": 0.6785007214523002, "grad_norm": 1.1842536926269531, "learning_rate": 7.694678933461818e-05, "loss": 0.595, "step": 10110 }, { "epoch": 0.6786349451360693, "grad_norm": 1.150343418121338, "learning_rate": 7.6937633814408e-05, "loss": 0.6444, "step": 10112 }, { "epoch": 0.6787691688198383, "grad_norm": 0.9357632398605347, "learning_rate": 7.692847702144759e-05, "loss": 0.6438, "step": 10114 }, { "epoch": 0.6789033925036073, "grad_norm": 1.2471938133239746, "learning_rate": 7.69193189561696e-05, "loss": 0.6013, "step": 10116 }, { "epoch": 0.6790376161873762, "grad_norm": 1.1025973558425903, "learning_rate": 7.691015961900671e-05, "loss": 0.5623, "step": 10118 }, { "epoch": 0.6791718398711453, "grad_norm": 1.0556445121765137, "learning_rate": 7.69009990103917e-05, "loss": 0.5833, "step": 10120 }, { "epoch": 0.6793060635549143, "grad_norm": 1.170667052268982, "learning_rate": 7.689183713075741e-05, "loss": 0.6205, "step": 10122 }, { "epoch": 0.6794402872386832, "grad_norm": 1.0111286640167236, "learning_rate": 7.688267398053665e-05, "loss": 0.626, "step": 10124 }, { "epoch": 0.6795745109224522, "grad_norm": 1.1592979431152344, "learning_rate": 7.687350956016244e-05, "loss": 0.5957, "step": 10126 }, { "epoch": 0.6797087346062213, "grad_norm": 1.0017783641815186, "learning_rate": 7.686434387006773e-05, "loss": 0.6221, "step": 10128 }, { "epoch": 0.6798429582899903, "grad_norm": 1.0687118768692017, "learning_rate": 7.685517691068563e-05, "loss": 0.6079, "step": 10130 }, { "epoch": 0.6799771819737592, "grad_norm": 1.0000358819961548, "learning_rate": 7.68460086824492e-05, "loss": 0.6036, "step": 10132 }, { "epoch": 0.6801114056575283, "grad_norm": 1.069441318511963, "learning_rate": 7.683683918579165e-05, "loss": 0.5702, "step": 10134 }, { "epoch": 0.6802456293412973, "grad_norm": 1.5402475595474243, "learning_rate": 7.682766842114622e-05, "loss": 0.6325, "step": 10136 }, { "epoch": 0.6803798530250663, "grad_norm": 0.9602324366569519, "learning_rate": 7.681849638894623e-05, "loss": 0.5932, "step": 10138 }, { "epoch": 0.6805140767088352, "grad_norm": 1.008661150932312, "learning_rate": 7.6809323089625e-05, "loss": 0.63, "step": 10140 }, { "epoch": 0.6806483003926043, "grad_norm": 1.1476720571517944, "learning_rate": 7.680014852361598e-05, "loss": 0.5783, "step": 10142 }, { "epoch": 0.6807825240763733, "grad_norm": 1.0170818567276, "learning_rate": 7.679097269135266e-05, "loss": 0.6574, "step": 10144 }, { "epoch": 0.6809167477601423, "grad_norm": 1.764005184173584, "learning_rate": 7.678179559326855e-05, "loss": 0.6232, "step": 10146 }, { "epoch": 0.6810509714439112, "grad_norm": 1.8771541118621826, "learning_rate": 7.677261722979725e-05, "loss": 0.5922, "step": 10148 }, { "epoch": 0.6811851951276803, "grad_norm": 1.4425020217895508, "learning_rate": 7.676343760137244e-05, "loss": 0.62, "step": 10150 }, { "epoch": 0.6813194188114493, "grad_norm": 0.840770423412323, "learning_rate": 7.675425670842783e-05, "loss": 0.5946, "step": 10152 }, { "epoch": 0.6814536424952182, "grad_norm": 1.0578558444976807, "learning_rate": 7.674507455139721e-05, "loss": 0.6644, "step": 10154 }, { "epoch": 0.6815878661789873, "grad_norm": 1.0931966304779053, "learning_rate": 7.673589113071442e-05, "loss": 0.6275, "step": 10156 }, { "epoch": 0.6817220898627563, "grad_norm": 1.1474859714508057, "learning_rate": 7.672670644681332e-05, "loss": 0.5998, "step": 10158 }, { "epoch": 0.6818563135465253, "grad_norm": 0.9686486124992371, "learning_rate": 7.671752050012792e-05, "loss": 0.5811, "step": 10160 }, { "epoch": 0.6819905372302942, "grad_norm": 1.1089202165603638, "learning_rate": 7.670833329109219e-05, "loss": 0.6239, "step": 10162 }, { "epoch": 0.6821247609140633, "grad_norm": 1.8264477252960205, "learning_rate": 7.669914482014025e-05, "loss": 0.6387, "step": 10164 }, { "epoch": 0.6822589845978323, "grad_norm": 1.0332576036453247, "learning_rate": 7.668995508770621e-05, "loss": 0.6105, "step": 10166 }, { "epoch": 0.6823932082816013, "grad_norm": 1.0819076299667358, "learning_rate": 7.668076409422427e-05, "loss": 0.6164, "step": 10168 }, { "epoch": 0.6825274319653702, "grad_norm": 1.1721596717834473, "learning_rate": 7.667157184012871e-05, "loss": 0.6715, "step": 10170 }, { "epoch": 0.6826616556491393, "grad_norm": 1.121983528137207, "learning_rate": 7.666237832585382e-05, "loss": 0.5908, "step": 10172 }, { "epoch": 0.6827958793329083, "grad_norm": 0.9286911487579346, "learning_rate": 7.665318355183398e-05, "loss": 0.5408, "step": 10174 }, { "epoch": 0.6829301030166773, "grad_norm": 1.180193305015564, "learning_rate": 7.664398751850363e-05, "loss": 0.6882, "step": 10176 }, { "epoch": 0.6830643267004463, "grad_norm": 1.121951699256897, "learning_rate": 7.663479022629727e-05, "loss": 0.6304, "step": 10178 }, { "epoch": 0.6831985503842153, "grad_norm": 1.0333795547485352, "learning_rate": 7.662559167564944e-05, "loss": 0.6446, "step": 10180 }, { "epoch": 0.6833327740679843, "grad_norm": 0.9968959093093872, "learning_rate": 7.661639186699474e-05, "loss": 0.5881, "step": 10182 }, { "epoch": 0.6834669977517533, "grad_norm": 0.9462453126907349, "learning_rate": 7.66071908007679e-05, "loss": 0.5612, "step": 10184 }, { "epoch": 0.6836012214355223, "grad_norm": 1.2085262537002563, "learning_rate": 7.659798847740359e-05, "loss": 0.6888, "step": 10186 }, { "epoch": 0.6837354451192913, "grad_norm": 1.0166971683502197, "learning_rate": 7.658878489733664e-05, "loss": 0.6123, "step": 10188 }, { "epoch": 0.6838696688030603, "grad_norm": 1.029884934425354, "learning_rate": 7.657958006100188e-05, "loss": 0.6025, "step": 10190 }, { "epoch": 0.6840038924868292, "grad_norm": 0.9909621477127075, "learning_rate": 7.657037396883423e-05, "loss": 0.6118, "step": 10192 }, { "epoch": 0.6841381161705983, "grad_norm": 1.0653865337371826, "learning_rate": 7.656116662126866e-05, "loss": 0.6363, "step": 10194 }, { "epoch": 0.6842723398543673, "grad_norm": 1.0455743074417114, "learning_rate": 7.65519580187402e-05, "loss": 0.5604, "step": 10196 }, { "epoch": 0.6844065635381363, "grad_norm": 0.9630902409553528, "learning_rate": 7.654274816168396e-05, "loss": 0.5414, "step": 10198 }, { "epoch": 0.6845407872219053, "grad_norm": 1.1408867835998535, "learning_rate": 7.653353705053503e-05, "loss": 0.5913, "step": 10200 }, { "epoch": 0.6846750109056743, "grad_norm": 1.0438798666000366, "learning_rate": 7.652432468572865e-05, "loss": 0.5727, "step": 10202 }, { "epoch": 0.6848092345894433, "grad_norm": 2.1974589824676514, "learning_rate": 7.65151110677001e-05, "loss": 0.5832, "step": 10204 }, { "epoch": 0.6849434582732123, "grad_norm": 0.9814490079879761, "learning_rate": 7.650589619688469e-05, "loss": 0.6232, "step": 10206 }, { "epoch": 0.6850776819569813, "grad_norm": 0.9011107683181763, "learning_rate": 7.64966800737178e-05, "loss": 0.5893, "step": 10208 }, { "epoch": 0.6852119056407503, "grad_norm": 1.0110008716583252, "learning_rate": 7.648746269863487e-05, "loss": 0.583, "step": 10210 }, { "epoch": 0.6853461293245193, "grad_norm": 1.6427772045135498, "learning_rate": 7.647824407207144e-05, "loss": 0.6113, "step": 10212 }, { "epoch": 0.6854803530082884, "grad_norm": 0.9514898657798767, "learning_rate": 7.646902419446302e-05, "loss": 0.5717, "step": 10214 }, { "epoch": 0.6856145766920573, "grad_norm": 0.8733771443367004, "learning_rate": 7.645980306624528e-05, "loss": 0.6078, "step": 10216 }, { "epoch": 0.6857488003758263, "grad_norm": 0.9661299586296082, "learning_rate": 7.645058068785386e-05, "loss": 0.6116, "step": 10218 }, { "epoch": 0.6858830240595953, "grad_norm": 1.0476787090301514, "learning_rate": 7.64413570597245e-05, "loss": 0.6379, "step": 10220 }, { "epoch": 0.6860172477433644, "grad_norm": 1.0753364562988281, "learning_rate": 7.643213218229305e-05, "loss": 0.6188, "step": 10222 }, { "epoch": 0.6861514714271333, "grad_norm": 1.123155951499939, "learning_rate": 7.642290605599531e-05, "loss": 0.6564, "step": 10224 }, { "epoch": 0.6862856951109023, "grad_norm": 1.0590447187423706, "learning_rate": 7.64136786812672e-05, "loss": 0.5836, "step": 10226 }, { "epoch": 0.6864199187946713, "grad_norm": 0.9150590896606445, "learning_rate": 7.640445005854475e-05, "loss": 0.627, "step": 10228 }, { "epoch": 0.6865541424784403, "grad_norm": 0.9836304783821106, "learning_rate": 7.639522018826393e-05, "loss": 0.6949, "step": 10230 }, { "epoch": 0.6866883661622093, "grad_norm": 1.3256171941757202, "learning_rate": 7.638598907086088e-05, "loss": 0.6433, "step": 10232 }, { "epoch": 0.6868225898459783, "grad_norm": 1.0061068534851074, "learning_rate": 7.637675670677171e-05, "loss": 0.6485, "step": 10234 }, { "epoch": 0.6869568135297474, "grad_norm": 0.9491173624992371, "learning_rate": 7.636752309643267e-05, "loss": 0.6128, "step": 10236 }, { "epoch": 0.6870910372135163, "grad_norm": 0.9607385993003845, "learning_rate": 7.635828824028001e-05, "loss": 0.6076, "step": 10238 }, { "epoch": 0.6872252608972853, "grad_norm": 1.398555040359497, "learning_rate": 7.634905213875006e-05, "loss": 0.5525, "step": 10240 }, { "epoch": 0.6873594845810543, "grad_norm": 1.0753309726715088, "learning_rate": 7.633981479227921e-05, "loss": 0.6379, "step": 10242 }, { "epoch": 0.6874937082648234, "grad_norm": 1.0636236667633057, "learning_rate": 7.633057620130391e-05, "loss": 0.6578, "step": 10244 }, { "epoch": 0.6876279319485923, "grad_norm": 0.9041749238967896, "learning_rate": 7.632133636626064e-05, "loss": 0.5808, "step": 10246 }, { "epoch": 0.6877621556323613, "grad_norm": 0.9568422436714172, "learning_rate": 7.6312095287586e-05, "loss": 0.6274, "step": 10248 }, { "epoch": 0.6878963793161303, "grad_norm": 1.0265642404556274, "learning_rate": 7.630285296571661e-05, "loss": 0.6536, "step": 10250 }, { "epoch": 0.6880306029998994, "grad_norm": 0.9597381353378296, "learning_rate": 7.629360940108913e-05, "loss": 0.5919, "step": 10252 }, { "epoch": 0.6881648266836683, "grad_norm": 1.0445297956466675, "learning_rate": 7.62843645941403e-05, "loss": 0.5497, "step": 10254 }, { "epoch": 0.6882990503674373, "grad_norm": 1.0023874044418335, "learning_rate": 7.627511854530695e-05, "loss": 0.6252, "step": 10256 }, { "epoch": 0.6884332740512064, "grad_norm": 1.0493043661117554, "learning_rate": 7.62658712550259e-05, "loss": 0.5709, "step": 10258 }, { "epoch": 0.6885674977349754, "grad_norm": 0.9132547378540039, "learning_rate": 7.625662272373409e-05, "loss": 0.5709, "step": 10260 }, { "epoch": 0.6887017214187443, "grad_norm": 1.17551851272583, "learning_rate": 7.624737295186849e-05, "loss": 0.5738, "step": 10262 }, { "epoch": 0.6888359451025133, "grad_norm": 1.090474009513855, "learning_rate": 7.623812193986612e-05, "loss": 0.647, "step": 10264 }, { "epoch": 0.6889701687862824, "grad_norm": 0.9255596995353699, "learning_rate": 7.62288696881641e-05, "loss": 0.5732, "step": 10266 }, { "epoch": 0.6891043924700513, "grad_norm": 1.0065983533859253, "learning_rate": 7.621961619719954e-05, "loss": 0.615, "step": 10268 }, { "epoch": 0.6892386161538203, "grad_norm": 2.6161906719207764, "learning_rate": 7.621036146740972e-05, "loss": 0.5857, "step": 10270 }, { "epoch": 0.6893728398375893, "grad_norm": 1.048517107963562, "learning_rate": 7.620110549923181e-05, "loss": 0.6099, "step": 10272 }, { "epoch": 0.6895070635213584, "grad_norm": 0.9987454414367676, "learning_rate": 7.619184829310322e-05, "loss": 0.6543, "step": 10274 }, { "epoch": 0.6896412872051273, "grad_norm": 0.9380282163619995, "learning_rate": 7.618258984946128e-05, "loss": 0.5941, "step": 10276 }, { "epoch": 0.6897755108888963, "grad_norm": 1.1077454090118408, "learning_rate": 7.617333016874346e-05, "loss": 0.6435, "step": 10278 }, { "epoch": 0.6899097345726654, "grad_norm": 1.1259806156158447, "learning_rate": 7.616406925138727e-05, "loss": 0.6018, "step": 10280 }, { "epoch": 0.6900439582564344, "grad_norm": 0.9674026966094971, "learning_rate": 7.615480709783025e-05, "loss": 0.5838, "step": 10282 }, { "epoch": 0.6901781819402033, "grad_norm": 1.1969586610794067, "learning_rate": 7.614554370851003e-05, "loss": 0.6811, "step": 10284 }, { "epoch": 0.6903124056239723, "grad_norm": 0.9744465947151184, "learning_rate": 7.613627908386427e-05, "loss": 0.6712, "step": 10286 }, { "epoch": 0.6904466293077414, "grad_norm": 0.9148718118667603, "learning_rate": 7.612701322433074e-05, "loss": 0.5742, "step": 10288 }, { "epoch": 0.6905808529915104, "grad_norm": 0.9658867716789246, "learning_rate": 7.61177461303472e-05, "loss": 0.627, "step": 10290 }, { "epoch": 0.6907150766752793, "grad_norm": 1.0178569555282593, "learning_rate": 7.610847780235149e-05, "loss": 0.5996, "step": 10292 }, { "epoch": 0.6908493003590483, "grad_norm": 1.414971947669983, "learning_rate": 7.609920824078157e-05, "loss": 0.5695, "step": 10294 }, { "epoch": 0.6909835240428174, "grad_norm": 1.0085248947143555, "learning_rate": 7.608993744607538e-05, "loss": 0.6539, "step": 10296 }, { "epoch": 0.6911177477265864, "grad_norm": 1.0939044952392578, "learning_rate": 7.608066541867093e-05, "loss": 0.5884, "step": 10298 }, { "epoch": 0.6912519714103553, "grad_norm": 1.0116087198257446, "learning_rate": 7.607139215900636e-05, "loss": 0.6157, "step": 10300 }, { "epoch": 0.6913861950941244, "grad_norm": 0.9540835022926331, "learning_rate": 7.606211766751976e-05, "loss": 0.6214, "step": 10302 }, { "epoch": 0.6915204187778934, "grad_norm": 1.0693719387054443, "learning_rate": 7.605284194464934e-05, "loss": 0.6726, "step": 10304 }, { "epoch": 0.6916546424616623, "grad_norm": 1.0814954042434692, "learning_rate": 7.604356499083338e-05, "loss": 0.6213, "step": 10306 }, { "epoch": 0.6917888661454313, "grad_norm": 1.1147676706314087, "learning_rate": 7.603428680651019e-05, "loss": 0.6544, "step": 10308 }, { "epoch": 0.6919230898292004, "grad_norm": 1.0701748132705688, "learning_rate": 7.602500739211813e-05, "loss": 0.6196, "step": 10310 }, { "epoch": 0.6920573135129694, "grad_norm": 1.0478525161743164, "learning_rate": 7.601572674809565e-05, "loss": 0.57, "step": 10312 }, { "epoch": 0.6921915371967383, "grad_norm": 1.0839003324508667, "learning_rate": 7.600644487488124e-05, "loss": 0.6296, "step": 10314 }, { "epoch": 0.6923257608805073, "grad_norm": 1.0307375192642212, "learning_rate": 7.599716177291345e-05, "loss": 0.5879, "step": 10316 }, { "epoch": 0.6924599845642764, "grad_norm": 1.3292783498764038, "learning_rate": 7.59878774426309e-05, "loss": 0.5892, "step": 10318 }, { "epoch": 0.6925942082480454, "grad_norm": 1.1004843711853027, "learning_rate": 7.597859188447223e-05, "loss": 0.6271, "step": 10320 }, { "epoch": 0.6927284319318143, "grad_norm": 0.9349532723426819, "learning_rate": 7.596930509887618e-05, "loss": 0.5415, "step": 10322 }, { "epoch": 0.6928626556155834, "grad_norm": 1.0800589323043823, "learning_rate": 7.596001708628152e-05, "loss": 0.645, "step": 10324 }, { "epoch": 0.6929968792993524, "grad_norm": 1.141237497329712, "learning_rate": 7.595072784712711e-05, "loss": 0.6522, "step": 10326 }, { "epoch": 0.6931311029831214, "grad_norm": 1.1993459463119507, "learning_rate": 7.594143738185184e-05, "loss": 0.6175, "step": 10328 }, { "epoch": 0.6932653266668903, "grad_norm": 1.0714892148971558, "learning_rate": 7.593214569089467e-05, "loss": 0.6167, "step": 10330 }, { "epoch": 0.6933995503506594, "grad_norm": 1.1327598094940186, "learning_rate": 7.59228527746946e-05, "loss": 0.6299, "step": 10332 }, { "epoch": 0.6935337740344284, "grad_norm": 1.010339379310608, "learning_rate": 7.591355863369071e-05, "loss": 0.6034, "step": 10334 }, { "epoch": 0.6936679977181974, "grad_norm": 1.156844139099121, "learning_rate": 7.590426326832217e-05, "loss": 0.6784, "step": 10336 }, { "epoch": 0.6938022214019663, "grad_norm": 0.9312037229537964, "learning_rate": 7.589496667902809e-05, "loss": 0.561, "step": 10338 }, { "epoch": 0.6939364450857354, "grad_norm": 1.0371742248535156, "learning_rate": 7.588566886624774e-05, "loss": 0.6169, "step": 10340 }, { "epoch": 0.6940706687695044, "grad_norm": 1.2180014848709106, "learning_rate": 7.587636983042048e-05, "loss": 0.6081, "step": 10342 }, { "epoch": 0.6942048924532733, "grad_norm": 1.029098391532898, "learning_rate": 7.586706957198562e-05, "loss": 0.6158, "step": 10344 }, { "epoch": 0.6943391161370424, "grad_norm": 1.1107252836227417, "learning_rate": 7.585776809138257e-05, "loss": 0.6046, "step": 10346 }, { "epoch": 0.6944733398208114, "grad_norm": 1.2475100755691528, "learning_rate": 7.584846538905083e-05, "loss": 0.6487, "step": 10348 }, { "epoch": 0.6946075635045804, "grad_norm": 1.0322431325912476, "learning_rate": 7.583916146542992e-05, "loss": 0.649, "step": 10350 }, { "epoch": 0.6947417871883493, "grad_norm": 0.919420599937439, "learning_rate": 7.582985632095946e-05, "loss": 0.5707, "step": 10352 }, { "epoch": 0.6948760108721184, "grad_norm": 1.2439117431640625, "learning_rate": 7.582054995607908e-05, "loss": 0.6862, "step": 10354 }, { "epoch": 0.6950102345558874, "grad_norm": 1.0442447662353516, "learning_rate": 7.581124237122848e-05, "loss": 0.5455, "step": 10356 }, { "epoch": 0.6951444582396564, "grad_norm": 0.994797945022583, "learning_rate": 7.580193356684743e-05, "loss": 0.5781, "step": 10358 }, { "epoch": 0.6952786819234253, "grad_norm": 1.0670169591903687, "learning_rate": 7.579262354337577e-05, "loss": 0.5641, "step": 10360 }, { "epoch": 0.6954129056071944, "grad_norm": 2.9009644985198975, "learning_rate": 7.578331230125336e-05, "loss": 0.6412, "step": 10362 }, { "epoch": 0.6955471292909634, "grad_norm": 0.9840713739395142, "learning_rate": 7.577399984092015e-05, "loss": 0.5508, "step": 10364 }, { "epoch": 0.6956813529747324, "grad_norm": 1.0158319473266602, "learning_rate": 7.576468616281612e-05, "loss": 0.5954, "step": 10366 }, { "epoch": 0.6958155766585014, "grad_norm": 0.9923294186592102, "learning_rate": 7.575537126738132e-05, "loss": 0.5792, "step": 10368 }, { "epoch": 0.6959498003422704, "grad_norm": 0.9123964905738831, "learning_rate": 7.57460551550559e-05, "loss": 0.5705, "step": 10370 }, { "epoch": 0.6960840240260394, "grad_norm": 1.0065155029296875, "learning_rate": 7.573673782628e-05, "loss": 0.5463, "step": 10372 }, { "epoch": 0.6962182477098084, "grad_norm": 1.0273391008377075, "learning_rate": 7.572741928149384e-05, "loss": 0.5893, "step": 10374 }, { "epoch": 0.6963524713935774, "grad_norm": 1.121673583984375, "learning_rate": 7.57180995211377e-05, "loss": 0.5428, "step": 10376 }, { "epoch": 0.6964866950773464, "grad_norm": 1.0862985849380493, "learning_rate": 7.570877854565195e-05, "loss": 0.5782, "step": 10378 }, { "epoch": 0.6966209187611154, "grad_norm": 1.1917251348495483, "learning_rate": 7.569945635547695e-05, "loss": 0.6539, "step": 10380 }, { "epoch": 0.6967551424448843, "grad_norm": 0.9577513933181763, "learning_rate": 7.569013295105318e-05, "loss": 0.5804, "step": 10382 }, { "epoch": 0.6968893661286534, "grad_norm": 1.0312405824661255, "learning_rate": 7.568080833282114e-05, "loss": 0.6378, "step": 10384 }, { "epoch": 0.6970235898124224, "grad_norm": 2.1493964195251465, "learning_rate": 7.56714825012214e-05, "loss": 0.543, "step": 10386 }, { "epoch": 0.6971578134961914, "grad_norm": 1.163265347480774, "learning_rate": 7.566215545669462e-05, "loss": 0.6308, "step": 10388 }, { "epoch": 0.6972920371799604, "grad_norm": 0.9730448722839355, "learning_rate": 7.565282719968143e-05, "loss": 0.5708, "step": 10390 }, { "epoch": 0.6974262608637294, "grad_norm": 0.9758681058883667, "learning_rate": 7.564349773062262e-05, "loss": 0.5878, "step": 10392 }, { "epoch": 0.6975604845474984, "grad_norm": 1.1344770193099976, "learning_rate": 7.563416704995894e-05, "loss": 0.6225, "step": 10394 }, { "epoch": 0.6976947082312674, "grad_norm": 1.0707207918167114, "learning_rate": 7.56248351581313e-05, "loss": 0.5813, "step": 10396 }, { "epoch": 0.6978289319150364, "grad_norm": 0.972823977470398, "learning_rate": 7.561550205558058e-05, "loss": 0.5609, "step": 10398 }, { "epoch": 0.6979631555988054, "grad_norm": 0.9138381481170654, "learning_rate": 7.560616774274775e-05, "loss": 0.5109, "step": 10400 }, { "epoch": 0.6980973792825744, "grad_norm": 1.0473132133483887, "learning_rate": 7.559683222007386e-05, "loss": 0.6346, "step": 10402 }, { "epoch": 0.6982316029663435, "grad_norm": 1.0219554901123047, "learning_rate": 7.558749548799997e-05, "loss": 0.5836, "step": 10404 }, { "epoch": 0.6983658266501124, "grad_norm": 1.2376022338867188, "learning_rate": 7.557815754696724e-05, "loss": 0.6474, "step": 10406 }, { "epoch": 0.6985000503338814, "grad_norm": 1.0176584720611572, "learning_rate": 7.556881839741687e-05, "loss": 0.5756, "step": 10408 }, { "epoch": 0.6986342740176504, "grad_norm": 1.1369335651397705, "learning_rate": 7.55594780397901e-05, "loss": 0.6256, "step": 10410 }, { "epoch": 0.6987684977014195, "grad_norm": 0.9506736397743225, "learning_rate": 7.555013647452826e-05, "loss": 0.5925, "step": 10412 }, { "epoch": 0.6989027213851884, "grad_norm": 0.9491580724716187, "learning_rate": 7.554079370207269e-05, "loss": 0.5816, "step": 10414 }, { "epoch": 0.6990369450689574, "grad_norm": 1.0840520858764648, "learning_rate": 7.553144972286488e-05, "loss": 0.5691, "step": 10416 }, { "epoch": 0.6991711687527264, "grad_norm": 1.1048444509506226, "learning_rate": 7.552210453734625e-05, "loss": 0.5801, "step": 10418 }, { "epoch": 0.6993053924364954, "grad_norm": 1.0872331857681274, "learning_rate": 7.551275814595837e-05, "loss": 0.6006, "step": 10420 }, { "epoch": 0.6994396161202644, "grad_norm": 1.0584237575531006, "learning_rate": 7.550341054914284e-05, "loss": 0.6611, "step": 10422 }, { "epoch": 0.6995738398040334, "grad_norm": 1.0477406978607178, "learning_rate": 7.549406174734132e-05, "loss": 0.5806, "step": 10424 }, { "epoch": 0.6997080634878025, "grad_norm": 1.157261848449707, "learning_rate": 7.548471174099551e-05, "loss": 0.5282, "step": 10426 }, { "epoch": 0.6998422871715714, "grad_norm": 1.0036718845367432, "learning_rate": 7.547536053054718e-05, "loss": 0.606, "step": 10428 }, { "epoch": 0.6999765108553404, "grad_norm": 1.0837408304214478, "learning_rate": 7.546600811643816e-05, "loss": 0.5896, "step": 10430 }, { "epoch": 0.7001107345391094, "grad_norm": 1.0489901304244995, "learning_rate": 7.545665449911032e-05, "loss": 0.6062, "step": 10432 }, { "epoch": 0.7002449582228785, "grad_norm": 1.1422386169433594, "learning_rate": 7.544729967900563e-05, "loss": 0.5789, "step": 10434 }, { "epoch": 0.7003791819066474, "grad_norm": 1.064909815788269, "learning_rate": 7.543794365656609e-05, "loss": 0.6285, "step": 10436 }, { "epoch": 0.7005134055904164, "grad_norm": 1.1166802644729614, "learning_rate": 7.542858643223369e-05, "loss": 0.5959, "step": 10438 }, { "epoch": 0.7006476292741854, "grad_norm": 1.0952495336532593, "learning_rate": 7.541922800645061e-05, "loss": 0.6267, "step": 10440 }, { "epoch": 0.7007818529579545, "grad_norm": 1.1388980150222778, "learning_rate": 7.540986837965899e-05, "loss": 0.5997, "step": 10442 }, { "epoch": 0.7009160766417234, "grad_norm": 1.0324063301086426, "learning_rate": 7.540050755230104e-05, "loss": 0.5926, "step": 10444 }, { "epoch": 0.7010503003254924, "grad_norm": 1.14800226688385, "learning_rate": 7.539114552481908e-05, "loss": 0.5649, "step": 10446 }, { "epoch": 0.7011845240092615, "grad_norm": 1.0952218770980835, "learning_rate": 7.53817822976554e-05, "loss": 0.6183, "step": 10448 }, { "epoch": 0.7013187476930305, "grad_norm": 1.2546967267990112, "learning_rate": 7.537241787125245e-05, "loss": 0.6727, "step": 10450 }, { "epoch": 0.7014529713767994, "grad_norm": 1.0081515312194824, "learning_rate": 7.536305224605261e-05, "loss": 0.6433, "step": 10452 }, { "epoch": 0.7015871950605684, "grad_norm": 1.042733907699585, "learning_rate": 7.535368542249846e-05, "loss": 0.6368, "step": 10454 }, { "epoch": 0.7017214187443375, "grad_norm": 1.0268605947494507, "learning_rate": 7.534431740103249e-05, "loss": 0.5982, "step": 10456 }, { "epoch": 0.7018556424281064, "grad_norm": 0.9910513162612915, "learning_rate": 7.53349481820974e-05, "loss": 0.5713, "step": 10458 }, { "epoch": 0.7019898661118754, "grad_norm": 1.038655161857605, "learning_rate": 7.53255777661358e-05, "loss": 0.5926, "step": 10460 }, { "epoch": 0.7021240897956444, "grad_norm": 1.022022008895874, "learning_rate": 7.531620615359046e-05, "loss": 0.6858, "step": 10462 }, { "epoch": 0.7022583134794135, "grad_norm": 1.5238577127456665, "learning_rate": 7.530683334490416e-05, "loss": 0.5713, "step": 10464 }, { "epoch": 0.7023925371631824, "grad_norm": 1.0431960821151733, "learning_rate": 7.529745934051976e-05, "loss": 0.6103, "step": 10466 }, { "epoch": 0.7025267608469514, "grad_norm": 1.0651239156723022, "learning_rate": 7.528808414088015e-05, "loss": 0.5874, "step": 10468 }, { "epoch": 0.7026609845307205, "grad_norm": 1.1000802516937256, "learning_rate": 7.527870774642828e-05, "loss": 0.5662, "step": 10470 }, { "epoch": 0.7027952082144895, "grad_norm": 1.0806845426559448, "learning_rate": 7.526933015760717e-05, "loss": 0.5968, "step": 10472 }, { "epoch": 0.7029294318982584, "grad_norm": 1.2063080072402954, "learning_rate": 7.525995137485993e-05, "loss": 0.6168, "step": 10474 }, { "epoch": 0.7030636555820274, "grad_norm": 1.074565052986145, "learning_rate": 7.525057139862964e-05, "loss": 0.6239, "step": 10476 }, { "epoch": 0.7031978792657965, "grad_norm": 1.032949686050415, "learning_rate": 7.524119022935949e-05, "loss": 0.525, "step": 10478 }, { "epoch": 0.7033321029495655, "grad_norm": 1.0446885824203491, "learning_rate": 7.523180786749276e-05, "loss": 0.5857, "step": 10480 }, { "epoch": 0.7034663266333344, "grad_norm": 1.0869170427322388, "learning_rate": 7.522242431347272e-05, "loss": 0.5863, "step": 10482 }, { "epoch": 0.7036005503171034, "grad_norm": 1.000910758972168, "learning_rate": 7.521303956774273e-05, "loss": 0.5666, "step": 10484 }, { "epoch": 0.7037347740008725, "grad_norm": 1.1870709657669067, "learning_rate": 7.52036536307462e-05, "loss": 0.6281, "step": 10486 }, { "epoch": 0.7038689976846415, "grad_norm": 1.1357227563858032, "learning_rate": 7.519426650292658e-05, "loss": 0.645, "step": 10488 }, { "epoch": 0.7040032213684104, "grad_norm": 0.9073771834373474, "learning_rate": 7.518487818472743e-05, "loss": 0.6408, "step": 10490 }, { "epoch": 0.7041374450521795, "grad_norm": 1.2567945718765259, "learning_rate": 7.51754886765923e-05, "loss": 0.6292, "step": 10492 }, { "epoch": 0.7042716687359485, "grad_norm": 1.2820225954055786, "learning_rate": 7.516609797896483e-05, "loss": 0.6182, "step": 10494 }, { "epoch": 0.7044058924197174, "grad_norm": 1.0083980560302734, "learning_rate": 7.515670609228873e-05, "loss": 0.5833, "step": 10496 }, { "epoch": 0.7045401161034864, "grad_norm": 0.9857763648033142, "learning_rate": 7.514731301700773e-05, "loss": 0.6576, "step": 10498 }, { "epoch": 0.7046743397872555, "grad_norm": 1.4885666370391846, "learning_rate": 7.513791875356564e-05, "loss": 0.6317, "step": 10500 }, { "epoch": 0.7048085634710245, "grad_norm": 1.116719365119934, "learning_rate": 7.512852330240632e-05, "loss": 0.6104, "step": 10502 }, { "epoch": 0.7049427871547934, "grad_norm": 0.9923556447029114, "learning_rate": 7.51191266639737e-05, "loss": 0.6866, "step": 10504 }, { "epoch": 0.7050770108385624, "grad_norm": 1.264246940612793, "learning_rate": 7.51097288387117e-05, "loss": 0.564, "step": 10506 }, { "epoch": 0.7052112345223315, "grad_norm": 1.0330630540847778, "learning_rate": 7.510032982706443e-05, "loss": 0.6315, "step": 10508 }, { "epoch": 0.7053454582061005, "grad_norm": 1.1161117553710938, "learning_rate": 7.509092962947591e-05, "loss": 0.6257, "step": 10510 }, { "epoch": 0.7054796818898694, "grad_norm": 1.8677253723144531, "learning_rate": 7.508152824639032e-05, "loss": 0.6676, "step": 10512 }, { "epoch": 0.7056139055736385, "grad_norm": 1.144211769104004, "learning_rate": 7.507212567825184e-05, "loss": 0.6177, "step": 10514 }, { "epoch": 0.7057481292574075, "grad_norm": 0.9658868312835693, "learning_rate": 7.506272192550472e-05, "loss": 0.6054, "step": 10516 }, { "epoch": 0.7058823529411765, "grad_norm": 4.2473978996276855, "learning_rate": 7.505331698859325e-05, "loss": 0.5925, "step": 10518 }, { "epoch": 0.7060165766249454, "grad_norm": 1.01564621925354, "learning_rate": 7.504391086796186e-05, "loss": 0.6168, "step": 10520 }, { "epoch": 0.7061508003087145, "grad_norm": 1.3537542819976807, "learning_rate": 7.50345035640549e-05, "loss": 0.6144, "step": 10522 }, { "epoch": 0.7062850239924835, "grad_norm": 0.8762102723121643, "learning_rate": 7.502509507731688e-05, "loss": 0.5816, "step": 10524 }, { "epoch": 0.7064192476762525, "grad_norm": 0.9736351370811462, "learning_rate": 7.501568540819233e-05, "loss": 0.6564, "step": 10526 }, { "epoch": 0.7065534713600214, "grad_norm": 1.4718120098114014, "learning_rate": 7.500627455712583e-05, "loss": 0.6406, "step": 10528 }, { "epoch": 0.7066876950437905, "grad_norm": 0.9880357384681702, "learning_rate": 7.499686252456205e-05, "loss": 0.5654, "step": 10530 }, { "epoch": 0.7068219187275595, "grad_norm": 1.083616018295288, "learning_rate": 7.498744931094565e-05, "loss": 0.6184, "step": 10532 }, { "epoch": 0.7069561424113284, "grad_norm": 1.0512688159942627, "learning_rate": 7.497803491672141e-05, "loss": 0.6053, "step": 10534 }, { "epoch": 0.7070903660950975, "grad_norm": 1.0200892686843872, "learning_rate": 7.496861934233414e-05, "loss": 0.625, "step": 10536 }, { "epoch": 0.7072245897788665, "grad_norm": 1.6291522979736328, "learning_rate": 7.495920258822869e-05, "loss": 0.6425, "step": 10538 }, { "epoch": 0.7073588134626355, "grad_norm": 1.2425224781036377, "learning_rate": 7.494978465485002e-05, "loss": 0.6697, "step": 10540 }, { "epoch": 0.7074930371464044, "grad_norm": 0.95583575963974, "learning_rate": 7.494036554264308e-05, "loss": 0.5765, "step": 10542 }, { "epoch": 0.7076272608301735, "grad_norm": 0.9750218987464905, "learning_rate": 7.49309452520529e-05, "loss": 0.593, "step": 10544 }, { "epoch": 0.7077614845139425, "grad_norm": 1.0550893545150757, "learning_rate": 7.49215237835246e-05, "loss": 0.6519, "step": 10546 }, { "epoch": 0.7078957081977115, "grad_norm": 0.9797727465629578, "learning_rate": 7.49121011375033e-05, "loss": 0.5769, "step": 10548 }, { "epoch": 0.7080299318814804, "grad_norm": 3.812112808227539, "learning_rate": 7.490267731443422e-05, "loss": 0.5869, "step": 10550 }, { "epoch": 0.7081641555652495, "grad_norm": 0.9804056286811829, "learning_rate": 7.489325231476258e-05, "loss": 0.5874, "step": 10552 }, { "epoch": 0.7082983792490185, "grad_norm": 1.2232451438903809, "learning_rate": 7.488382613893372e-05, "loss": 0.6239, "step": 10554 }, { "epoch": 0.7084326029327875, "grad_norm": 0.9250823259353638, "learning_rate": 7.487439878739303e-05, "loss": 0.6, "step": 10556 }, { "epoch": 0.7085668266165565, "grad_norm": 0.9323468804359436, "learning_rate": 7.486497026058588e-05, "loss": 0.6055, "step": 10558 }, { "epoch": 0.7087010503003255, "grad_norm": 0.9436549544334412, "learning_rate": 7.485554055895781e-05, "loss": 0.5185, "step": 10560 }, { "epoch": 0.7088352739840945, "grad_norm": 1.084033489227295, "learning_rate": 7.484610968295431e-05, "loss": 0.6145, "step": 10562 }, { "epoch": 0.7089694976678635, "grad_norm": 1.198091745376587, "learning_rate": 7.4836677633021e-05, "loss": 0.5858, "step": 10564 }, { "epoch": 0.7091037213516325, "grad_norm": 1.241812825202942, "learning_rate": 7.48272444096035e-05, "loss": 0.5898, "step": 10566 }, { "epoch": 0.7092379450354015, "grad_norm": 0.9129748344421387, "learning_rate": 7.481781001314751e-05, "loss": 0.6265, "step": 10568 }, { "epoch": 0.7093721687191705, "grad_norm": 0.901710033416748, "learning_rate": 7.480837444409882e-05, "loss": 0.6173, "step": 10570 }, { "epoch": 0.7095063924029394, "grad_norm": 1.212648630142212, "learning_rate": 7.479893770290321e-05, "loss": 0.6168, "step": 10572 }, { "epoch": 0.7096406160867085, "grad_norm": 0.9216916561126709, "learning_rate": 7.478949979000656e-05, "loss": 0.5529, "step": 10574 }, { "epoch": 0.7097748397704775, "grad_norm": 1.0843908786773682, "learning_rate": 7.47800607058548e-05, "loss": 0.5992, "step": 10576 }, { "epoch": 0.7099090634542465, "grad_norm": 1.1257072687149048, "learning_rate": 7.477062045089389e-05, "loss": 0.639, "step": 10578 }, { "epoch": 0.7100432871380155, "grad_norm": 1.2055573463439941, "learning_rate": 7.476117902556987e-05, "loss": 0.6759, "step": 10580 }, { "epoch": 0.7101775108217845, "grad_norm": 1.157829999923706, "learning_rate": 7.475173643032882e-05, "loss": 0.6496, "step": 10582 }, { "epoch": 0.7103117345055535, "grad_norm": 1.0404984951019287, "learning_rate": 7.474229266561692e-05, "loss": 0.6716, "step": 10584 }, { "epoch": 0.7104459581893225, "grad_norm": 1.039047360420227, "learning_rate": 7.473284773188034e-05, "loss": 0.641, "step": 10586 }, { "epoch": 0.7105801818730915, "grad_norm": 0.9575285315513611, "learning_rate": 7.472340162956534e-05, "loss": 0.6802, "step": 10588 }, { "epoch": 0.7107144055568605, "grad_norm": 1.1343613862991333, "learning_rate": 7.471395435911822e-05, "loss": 0.5732, "step": 10590 }, { "epoch": 0.7108486292406295, "grad_norm": 1.0703816413879395, "learning_rate": 7.470450592098537e-05, "loss": 0.5905, "step": 10592 }, { "epoch": 0.7109828529243986, "grad_norm": 1.05869460105896, "learning_rate": 7.469505631561317e-05, "loss": 0.6367, "step": 10594 }, { "epoch": 0.7111170766081675, "grad_norm": 1.1756197214126587, "learning_rate": 7.468560554344814e-05, "loss": 0.5681, "step": 10596 }, { "epoch": 0.7112513002919365, "grad_norm": 0.9795808792114258, "learning_rate": 7.46761536049368e-05, "loss": 0.58, "step": 10598 }, { "epoch": 0.7113855239757055, "grad_norm": 0.9587852358818054, "learning_rate": 7.46667005005257e-05, "loss": 0.6089, "step": 10600 }, { "epoch": 0.7115197476594746, "grad_norm": 1.1246082782745361, "learning_rate": 7.465724623066153e-05, "loss": 0.5236, "step": 10602 }, { "epoch": 0.7116539713432435, "grad_norm": 1.0588661432266235, "learning_rate": 7.464779079579094e-05, "loss": 0.5631, "step": 10604 }, { "epoch": 0.7117881950270125, "grad_norm": 0.8241974711418152, "learning_rate": 7.463833419636072e-05, "loss": 0.5622, "step": 10606 }, { "epoch": 0.7119224187107815, "grad_norm": 0.9836342930793762, "learning_rate": 7.462887643281764e-05, "loss": 0.5949, "step": 10608 }, { "epoch": 0.7120566423945505, "grad_norm": 1.0671802759170532, "learning_rate": 7.46194175056086e-05, "loss": 0.5492, "step": 10610 }, { "epoch": 0.7121908660783195, "grad_norm": 0.9698143601417542, "learning_rate": 7.460995741518049e-05, "loss": 0.6173, "step": 10612 }, { "epoch": 0.7123250897620885, "grad_norm": 1.0343726873397827, "learning_rate": 7.460049616198027e-05, "loss": 0.5927, "step": 10614 }, { "epoch": 0.7124593134458576, "grad_norm": 1.0122747421264648, "learning_rate": 7.4591033746455e-05, "loss": 0.5812, "step": 10616 }, { "epoch": 0.7125935371296265, "grad_norm": 0.9365114569664001, "learning_rate": 7.458157016905173e-05, "loss": 0.593, "step": 10618 }, { "epoch": 0.7127277608133955, "grad_norm": 1.0280989408493042, "learning_rate": 7.45721054302176e-05, "loss": 0.5717, "step": 10620 }, { "epoch": 0.7128619844971645, "grad_norm": 1.0961730480194092, "learning_rate": 7.456263953039984e-05, "loss": 0.645, "step": 10622 }, { "epoch": 0.7129962081809336, "grad_norm": 1.0248903036117554, "learning_rate": 7.455317247004563e-05, "loss": 0.5717, "step": 10624 }, { "epoch": 0.7131304318647025, "grad_norm": 0.9453458189964294, "learning_rate": 7.454370424960231e-05, "loss": 0.6218, "step": 10626 }, { "epoch": 0.7132646555484715, "grad_norm": 1.2950961589813232, "learning_rate": 7.453423486951723e-05, "loss": 0.6676, "step": 10628 }, { "epoch": 0.7133988792322405, "grad_norm": 1.004745364189148, "learning_rate": 7.45247643302378e-05, "loss": 0.6064, "step": 10630 }, { "epoch": 0.7135331029160096, "grad_norm": 1.0576225519180298, "learning_rate": 7.451529263221147e-05, "loss": 0.647, "step": 10632 }, { "epoch": 0.7136673265997785, "grad_norm": 0.8936069011688232, "learning_rate": 7.450581977588577e-05, "loss": 0.6206, "step": 10634 }, { "epoch": 0.7138015502835475, "grad_norm": 1.0472456216812134, "learning_rate": 7.44963457617083e-05, "loss": 0.6276, "step": 10636 }, { "epoch": 0.7139357739673166, "grad_norm": 1.0001219511032104, "learning_rate": 7.448687059012665e-05, "loss": 0.6069, "step": 10638 }, { "epoch": 0.7140699976510856, "grad_norm": 0.9868367314338684, "learning_rate": 7.44773942615885e-05, "loss": 0.548, "step": 10640 }, { "epoch": 0.7142042213348545, "grad_norm": 1.2530124187469482, "learning_rate": 7.446791677654162e-05, "loss": 0.6735, "step": 10642 }, { "epoch": 0.7143384450186235, "grad_norm": 2.3150460720062256, "learning_rate": 7.445843813543379e-05, "loss": 0.5418, "step": 10644 }, { "epoch": 0.7144726687023926, "grad_norm": 1.0615333318710327, "learning_rate": 7.444895833871283e-05, "loss": 0.7008, "step": 10646 }, { "epoch": 0.7146068923861615, "grad_norm": 1.098844289779663, "learning_rate": 7.443947738682668e-05, "loss": 0.7514, "step": 10648 }, { "epoch": 0.7147411160699305, "grad_norm": 1.056339979171753, "learning_rate": 7.44299952802233e-05, "loss": 0.6633, "step": 10650 }, { "epoch": 0.7148753397536995, "grad_norm": 0.9609205722808838, "learning_rate": 7.442051201935065e-05, "loss": 0.6155, "step": 10652 }, { "epoch": 0.7150095634374686, "grad_norm": 1.1464987993240356, "learning_rate": 7.441102760465686e-05, "loss": 0.614, "step": 10654 }, { "epoch": 0.7151437871212375, "grad_norm": 1.0660523176193237, "learning_rate": 7.440154203658999e-05, "loss": 0.6386, "step": 10656 }, { "epoch": 0.7152780108050065, "grad_norm": 1.0465962886810303, "learning_rate": 7.439205531559825e-05, "loss": 0.6342, "step": 10658 }, { "epoch": 0.7154122344887756, "grad_norm": 0.9891012907028198, "learning_rate": 7.438256744212987e-05, "loss": 0.6225, "step": 10660 }, { "epoch": 0.7155464581725446, "grad_norm": 0.9910521507263184, "learning_rate": 7.437307841663312e-05, "loss": 0.6156, "step": 10662 }, { "epoch": 0.7156806818563135, "grad_norm": 0.9643241167068481, "learning_rate": 7.436358823955634e-05, "loss": 0.586, "step": 10664 }, { "epoch": 0.7158149055400825, "grad_norm": 1.088691234588623, "learning_rate": 7.435409691134792e-05, "loss": 0.5797, "step": 10666 }, { "epoch": 0.7159491292238516, "grad_norm": 1.0931916236877441, "learning_rate": 7.434460443245632e-05, "loss": 0.6551, "step": 10668 }, { "epoch": 0.7160833529076206, "grad_norm": 1.0558147430419922, "learning_rate": 7.433511080333004e-05, "loss": 0.6085, "step": 10670 }, { "epoch": 0.7162175765913895, "grad_norm": 0.9660263657569885, "learning_rate": 7.43256160244176e-05, "loss": 0.5556, "step": 10672 }, { "epoch": 0.7163518002751585, "grad_norm": 0.9382823705673218, "learning_rate": 7.431612009616767e-05, "loss": 0.5926, "step": 10674 }, { "epoch": 0.7164860239589276, "grad_norm": 1.1861258745193481, "learning_rate": 7.430662301902885e-05, "loss": 0.6212, "step": 10676 }, { "epoch": 0.7166202476426966, "grad_norm": 1.098279356956482, "learning_rate": 7.429712479344992e-05, "loss": 0.5871, "step": 10678 }, { "epoch": 0.7167544713264655, "grad_norm": 1.1667335033416748, "learning_rate": 7.42876254198796e-05, "loss": 0.676, "step": 10680 }, { "epoch": 0.7168886950102346, "grad_norm": 0.9760928153991699, "learning_rate": 7.427812489876674e-05, "loss": 0.5523, "step": 10682 }, { "epoch": 0.7170229186940036, "grad_norm": 1.0477968454360962, "learning_rate": 7.426862323056023e-05, "loss": 0.6248, "step": 10684 }, { "epoch": 0.7171571423777725, "grad_norm": 0.933026909828186, "learning_rate": 7.425912041570899e-05, "loss": 0.5802, "step": 10686 }, { "epoch": 0.7172913660615415, "grad_norm": 0.999974250793457, "learning_rate": 7.424961645466202e-05, "loss": 0.6241, "step": 10688 }, { "epoch": 0.7174255897453106, "grad_norm": 1.1083967685699463, "learning_rate": 7.424011134786835e-05, "loss": 0.5953, "step": 10690 }, { "epoch": 0.7175598134290796, "grad_norm": 1.1214714050292969, "learning_rate": 7.423060509577707e-05, "loss": 0.6042, "step": 10692 }, { "epoch": 0.7176940371128485, "grad_norm": 0.9524616599082947, "learning_rate": 7.422109769883738e-05, "loss": 0.553, "step": 10694 }, { "epoch": 0.7178282607966175, "grad_norm": 1.2512476444244385, "learning_rate": 7.421158915749842e-05, "loss": 0.5933, "step": 10696 }, { "epoch": 0.7179624844803866, "grad_norm": 0.9280952215194702, "learning_rate": 7.42020794722095e-05, "loss": 0.56, "step": 10698 }, { "epoch": 0.7180967081641556, "grad_norm": 1.0245829820632935, "learning_rate": 7.419256864341992e-05, "loss": 0.6062, "step": 10700 }, { "epoch": 0.7182309318479245, "grad_norm": 0.9840590953826904, "learning_rate": 7.418305667157903e-05, "loss": 0.5775, "step": 10702 }, { "epoch": 0.7183651555316936, "grad_norm": 1.1951439380645752, "learning_rate": 7.417354355713627e-05, "loss": 0.6861, "step": 10704 }, { "epoch": 0.7184993792154626, "grad_norm": 1.0898561477661133, "learning_rate": 7.416402930054111e-05, "loss": 0.6109, "step": 10706 }, { "epoch": 0.7186336028992316, "grad_norm": 1.06401789188385, "learning_rate": 7.415451390224309e-05, "loss": 0.5883, "step": 10708 }, { "epoch": 0.7187678265830005, "grad_norm": 0.910780131816864, "learning_rate": 7.414499736269178e-05, "loss": 0.5871, "step": 10710 }, { "epoch": 0.7189020502667696, "grad_norm": 1.069695234298706, "learning_rate": 7.413547968233684e-05, "loss": 0.5824, "step": 10712 }, { "epoch": 0.7190362739505386, "grad_norm": 1.0132920742034912, "learning_rate": 7.412596086162793e-05, "loss": 0.6835, "step": 10714 }, { "epoch": 0.7191704976343076, "grad_norm": 0.9329550862312317, "learning_rate": 7.411644090101481e-05, "loss": 0.5653, "step": 10716 }, { "epoch": 0.7193047213180765, "grad_norm": 0.8903213739395142, "learning_rate": 7.410691980094728e-05, "loss": 0.6081, "step": 10718 }, { "epoch": 0.7194389450018456, "grad_norm": 1.0895676612854004, "learning_rate": 7.409739756187519e-05, "loss": 0.628, "step": 10720 }, { "epoch": 0.7195731686856146, "grad_norm": 1.0903286933898926, "learning_rate": 7.408787418424848e-05, "loss": 0.6389, "step": 10722 }, { "epoch": 0.7197073923693835, "grad_norm": 1.4827929735183716, "learning_rate": 7.407834966851705e-05, "loss": 0.5801, "step": 10724 }, { "epoch": 0.7198416160531526, "grad_norm": 0.9832586050033569, "learning_rate": 7.406882401513096e-05, "loss": 0.6168, "step": 10726 }, { "epoch": 0.7199758397369216, "grad_norm": 1.0722204446792603, "learning_rate": 7.405929722454026e-05, "loss": 0.6553, "step": 10728 }, { "epoch": 0.7201100634206906, "grad_norm": 1.5461186170578003, "learning_rate": 7.404976929719507e-05, "loss": 0.5757, "step": 10730 }, { "epoch": 0.7202442871044595, "grad_norm": 0.9481789469718933, "learning_rate": 7.404024023354558e-05, "loss": 0.626, "step": 10732 }, { "epoch": 0.7203785107882286, "grad_norm": 1.0034617185592651, "learning_rate": 7.4030710034042e-05, "loss": 0.5735, "step": 10734 }, { "epoch": 0.7205127344719976, "grad_norm": 1.0112169981002808, "learning_rate": 7.402117869913465e-05, "loss": 0.6503, "step": 10736 }, { "epoch": 0.7206469581557666, "grad_norm": 1.1357429027557373, "learning_rate": 7.401164622927382e-05, "loss": 0.6094, "step": 10738 }, { "epoch": 0.7207811818395355, "grad_norm": 0.9947372078895569, "learning_rate": 7.400211262490994e-05, "loss": 0.6288, "step": 10740 }, { "epoch": 0.7209154055233046, "grad_norm": 1.0178544521331787, "learning_rate": 7.399257788649342e-05, "loss": 0.6021, "step": 10742 }, { "epoch": 0.7210496292070736, "grad_norm": 1.1283224821090698, "learning_rate": 7.398304201447478e-05, "loss": 0.5801, "step": 10744 }, { "epoch": 0.7211838528908426, "grad_norm": 0.999299943447113, "learning_rate": 7.397350500930458e-05, "loss": 0.5839, "step": 10746 }, { "epoch": 0.7213180765746116, "grad_norm": 1.0841357707977295, "learning_rate": 7.39639668714334e-05, "loss": 0.602, "step": 10748 }, { "epoch": 0.7214523002583806, "grad_norm": 1.0852906703948975, "learning_rate": 7.395442760131192e-05, "loss": 0.6459, "step": 10750 }, { "epoch": 0.7215865239421496, "grad_norm": 1.0680116415023804, "learning_rate": 7.394488719939081e-05, "loss": 0.5973, "step": 10752 }, { "epoch": 0.7217207476259186, "grad_norm": 1.0845413208007812, "learning_rate": 7.393534566612089e-05, "loss": 0.633, "step": 10754 }, { "epoch": 0.7218549713096876, "grad_norm": 1.012047290802002, "learning_rate": 7.392580300195296e-05, "loss": 0.5649, "step": 10756 }, { "epoch": 0.7219891949934566, "grad_norm": 1.0492441654205322, "learning_rate": 7.391625920733786e-05, "loss": 0.6044, "step": 10758 }, { "epoch": 0.7221234186772256, "grad_norm": 1.1300857067108154, "learning_rate": 7.390671428272655e-05, "loss": 0.6455, "step": 10760 }, { "epoch": 0.7222576423609945, "grad_norm": 0.9608523845672607, "learning_rate": 7.389716822857e-05, "loss": 0.5733, "step": 10762 }, { "epoch": 0.7223918660447636, "grad_norm": 1.1491624116897583, "learning_rate": 7.388762104531925e-05, "loss": 0.6238, "step": 10764 }, { "epoch": 0.7225260897285326, "grad_norm": 0.9276455640792847, "learning_rate": 7.387807273342539e-05, "loss": 0.6136, "step": 10766 }, { "epoch": 0.7226603134123016, "grad_norm": 1.0152052640914917, "learning_rate": 7.386852329333953e-05, "loss": 0.6227, "step": 10768 }, { "epoch": 0.7227945370960706, "grad_norm": 1.027574062347412, "learning_rate": 7.385897272551287e-05, "loss": 0.5818, "step": 10770 }, { "epoch": 0.7229287607798396, "grad_norm": 1.1471387147903442, "learning_rate": 7.38494210303967e-05, "loss": 0.6054, "step": 10772 }, { "epoch": 0.7230629844636086, "grad_norm": 1.2042551040649414, "learning_rate": 7.383986820844226e-05, "loss": 0.6077, "step": 10774 }, { "epoch": 0.7231972081473776, "grad_norm": 1.2099021673202515, "learning_rate": 7.383031426010092e-05, "loss": 0.6161, "step": 10776 }, { "epoch": 0.7233314318311466, "grad_norm": 1.0751736164093018, "learning_rate": 7.38207591858241e-05, "loss": 0.5854, "step": 10778 }, { "epoch": 0.7234656555149156, "grad_norm": 1.0264619588851929, "learning_rate": 7.381120298606325e-05, "loss": 0.5747, "step": 10780 }, { "epoch": 0.7235998791986846, "grad_norm": 1.0129170417785645, "learning_rate": 7.380164566126989e-05, "loss": 0.5775, "step": 10782 }, { "epoch": 0.7237341028824537, "grad_norm": 1.0577455759048462, "learning_rate": 7.379208721189557e-05, "loss": 0.6174, "step": 10784 }, { "epoch": 0.7238683265662226, "grad_norm": 1.015270471572876, "learning_rate": 7.37825276383919e-05, "loss": 0.597, "step": 10786 }, { "epoch": 0.7240025502499916, "grad_norm": 1.2538281679153442, "learning_rate": 7.377296694121058e-05, "loss": 0.5995, "step": 10788 }, { "epoch": 0.7241367739337606, "grad_norm": 1.055049180984497, "learning_rate": 7.376340512080334e-05, "loss": 0.6603, "step": 10790 }, { "epoch": 0.7242709976175297, "grad_norm": 1.0343923568725586, "learning_rate": 7.375384217762191e-05, "loss": 0.5742, "step": 10792 }, { "epoch": 0.7244052213012986, "grad_norm": 1.1283427476882935, "learning_rate": 7.374427811211815e-05, "loss": 0.6378, "step": 10794 }, { "epoch": 0.7245394449850676, "grad_norm": 1.2214810848236084, "learning_rate": 7.373471292474393e-05, "loss": 0.6573, "step": 10796 }, { "epoch": 0.7246736686688366, "grad_norm": 4.867005348205566, "learning_rate": 7.372514661595122e-05, "loss": 0.5434, "step": 10798 }, { "epoch": 0.7248078923526056, "grad_norm": 1.8141260147094727, "learning_rate": 7.371557918619198e-05, "loss": 0.5713, "step": 10800 }, { "epoch": 0.7249421160363746, "grad_norm": 1.0678434371948242, "learning_rate": 7.370601063591824e-05, "loss": 0.6123, "step": 10802 }, { "epoch": 0.7250763397201436, "grad_norm": 1.0353124141693115, "learning_rate": 7.369644096558213e-05, "loss": 0.5903, "step": 10804 }, { "epoch": 0.7252105634039127, "grad_norm": 0.9907625317573547, "learning_rate": 7.368687017563578e-05, "loss": 0.6341, "step": 10806 }, { "epoch": 0.7253447870876816, "grad_norm": 1.0260570049285889, "learning_rate": 7.36772982665314e-05, "loss": 0.6229, "step": 10808 }, { "epoch": 0.7254790107714506, "grad_norm": 1.0138272047042847, "learning_rate": 7.366772523872122e-05, "loss": 0.6119, "step": 10810 }, { "epoch": 0.7256132344552196, "grad_norm": 2.370683431625366, "learning_rate": 7.365815109265757e-05, "loss": 0.6508, "step": 10812 }, { "epoch": 0.7257474581389887, "grad_norm": 0.9320681691169739, "learning_rate": 7.36485758287928e-05, "loss": 0.5675, "step": 10814 }, { "epoch": 0.7258816818227576, "grad_norm": 1.097596526145935, "learning_rate": 7.363899944757935e-05, "loss": 0.5803, "step": 10816 }, { "epoch": 0.7260159055065266, "grad_norm": 1.05289888381958, "learning_rate": 7.362942194946962e-05, "loss": 0.6129, "step": 10818 }, { "epoch": 0.7261501291902956, "grad_norm": 0.9760967493057251, "learning_rate": 7.361984333491618e-05, "loss": 0.5784, "step": 10820 }, { "epoch": 0.7262843528740647, "grad_norm": 0.9339340925216675, "learning_rate": 7.361026360437162e-05, "loss": 0.5642, "step": 10822 }, { "epoch": 0.7264185765578336, "grad_norm": 0.8658645153045654, "learning_rate": 7.360068275828849e-05, "loss": 0.5677, "step": 10824 }, { "epoch": 0.7265528002416026, "grad_norm": 1.546994686126709, "learning_rate": 7.359110079711953e-05, "loss": 0.642, "step": 10826 }, { "epoch": 0.7266870239253717, "grad_norm": 0.8931393027305603, "learning_rate": 7.358151772131744e-05, "loss": 0.6122, "step": 10828 }, { "epoch": 0.7268212476091407, "grad_norm": 1.133793830871582, "learning_rate": 7.357193353133503e-05, "loss": 0.6548, "step": 10830 }, { "epoch": 0.7269554712929096, "grad_norm": 1.0960601568222046, "learning_rate": 7.356234822762508e-05, "loss": 0.5956, "step": 10832 }, { "epoch": 0.7270896949766786, "grad_norm": 1.1904242038726807, "learning_rate": 7.355276181064052e-05, "loss": 0.5764, "step": 10834 }, { "epoch": 0.7272239186604477, "grad_norm": 1.222369909286499, "learning_rate": 7.354317428083428e-05, "loss": 0.6833, "step": 10836 }, { "epoch": 0.7273581423442166, "grad_norm": 1.057033658027649, "learning_rate": 7.353358563865935e-05, "loss": 0.5845, "step": 10838 }, { "epoch": 0.7274923660279856, "grad_norm": 1.0513136386871338, "learning_rate": 7.35239958845688e-05, "loss": 0.574, "step": 10840 }, { "epoch": 0.7276265897117546, "grad_norm": 1.9896886348724365, "learning_rate": 7.351440501901567e-05, "loss": 0.5445, "step": 10842 }, { "epoch": 0.7277608133955237, "grad_norm": 1.1040643453598022, "learning_rate": 7.350481304245315e-05, "loss": 0.6422, "step": 10844 }, { "epoch": 0.7278950370792926, "grad_norm": 0.9568721055984497, "learning_rate": 7.349521995533444e-05, "loss": 0.5501, "step": 10846 }, { "epoch": 0.7280292607630616, "grad_norm": 1.1906222105026245, "learning_rate": 7.348562575811279e-05, "loss": 0.6471, "step": 10848 }, { "epoch": 0.7281634844468307, "grad_norm": 0.911615788936615, "learning_rate": 7.347603045124149e-05, "loss": 0.5898, "step": 10850 }, { "epoch": 0.7282977081305997, "grad_norm": 1.226242184638977, "learning_rate": 7.346643403517394e-05, "loss": 0.6071, "step": 10852 }, { "epoch": 0.7284319318143686, "grad_norm": 1.0330523252487183, "learning_rate": 7.345683651036351e-05, "loss": 0.6134, "step": 10854 }, { "epoch": 0.7285661554981376, "grad_norm": 0.9957926869392395, "learning_rate": 7.344723787726368e-05, "loss": 0.5669, "step": 10856 }, { "epoch": 0.7287003791819067, "grad_norm": 1.0053691864013672, "learning_rate": 7.343763813632798e-05, "loss": 0.5931, "step": 10858 }, { "epoch": 0.7288346028656757, "grad_norm": 1.0712882280349731, "learning_rate": 7.342803728800995e-05, "loss": 0.6325, "step": 10860 }, { "epoch": 0.7289688265494446, "grad_norm": 0.9607504606246948, "learning_rate": 7.341843533276323e-05, "loss": 0.5539, "step": 10862 }, { "epoch": 0.7291030502332136, "grad_norm": 0.9818548560142517, "learning_rate": 7.340883227104151e-05, "loss": 0.5815, "step": 10864 }, { "epoch": 0.7292372739169827, "grad_norm": 0.9800499677658081, "learning_rate": 7.339922810329846e-05, "loss": 0.5471, "step": 10866 }, { "epoch": 0.7293714976007517, "grad_norm": 1.1134047508239746, "learning_rate": 7.338962282998794e-05, "loss": 0.584, "step": 10868 }, { "epoch": 0.7295057212845206, "grad_norm": 1.0640432834625244, "learning_rate": 7.33800164515637e-05, "loss": 0.5663, "step": 10870 }, { "epoch": 0.7296399449682897, "grad_norm": 1.0774834156036377, "learning_rate": 7.337040896847967e-05, "loss": 0.5586, "step": 10872 }, { "epoch": 0.7297741686520587, "grad_norm": 0.9616897702217102, "learning_rate": 7.336080038118978e-05, "loss": 0.5472, "step": 10874 }, { "epoch": 0.7299083923358276, "grad_norm": 1.0113489627838135, "learning_rate": 7.335119069014798e-05, "loss": 0.5974, "step": 10876 }, { "epoch": 0.7300426160195966, "grad_norm": 1.2057607173919678, "learning_rate": 7.334157989580838e-05, "loss": 0.589, "step": 10878 }, { "epoch": 0.7301768397033657, "grad_norm": 0.9961106181144714, "learning_rate": 7.333196799862499e-05, "loss": 0.6091, "step": 10880 }, { "epoch": 0.7303110633871347, "grad_norm": 0.883374810218811, "learning_rate": 7.332235499905202e-05, "loss": 0.6059, "step": 10882 }, { "epoch": 0.7304452870709036, "grad_norm": 0.9781058430671692, "learning_rate": 7.331274089754363e-05, "loss": 0.5797, "step": 10884 }, { "epoch": 0.7305795107546726, "grad_norm": 1.0016698837280273, "learning_rate": 7.330312569455408e-05, "loss": 0.5686, "step": 10886 }, { "epoch": 0.7307137344384417, "grad_norm": 1.010748267173767, "learning_rate": 7.329350939053766e-05, "loss": 0.533, "step": 10888 }, { "epoch": 0.7308479581222107, "grad_norm": 1.062753438949585, "learning_rate": 7.328389198594872e-05, "loss": 0.6047, "step": 10890 }, { "epoch": 0.7309821818059796, "grad_norm": 1.06117844581604, "learning_rate": 7.327427348124167e-05, "loss": 0.6129, "step": 10892 }, { "epoch": 0.7311164054897487, "grad_norm": 1.3572216033935547, "learning_rate": 7.326465387687097e-05, "loss": 0.5429, "step": 10894 }, { "epoch": 0.7312506291735177, "grad_norm": 1.013454794883728, "learning_rate": 7.325503317329112e-05, "loss": 0.5456, "step": 10896 }, { "epoch": 0.7313848528572867, "grad_norm": 1.1997941732406616, "learning_rate": 7.324541137095669e-05, "loss": 0.6264, "step": 10898 }, { "epoch": 0.7315190765410556, "grad_norm": 1.0833473205566406, "learning_rate": 7.323578847032226e-05, "loss": 0.6353, "step": 10900 }, { "epoch": 0.7316533002248247, "grad_norm": 1.4310764074325562, "learning_rate": 7.322616447184254e-05, "loss": 0.6689, "step": 10902 }, { "epoch": 0.7317875239085937, "grad_norm": 1.1005669832229614, "learning_rate": 7.321653937597222e-05, "loss": 0.557, "step": 10904 }, { "epoch": 0.7319217475923627, "grad_norm": 1.1346960067749023, "learning_rate": 7.320691318316606e-05, "loss": 0.6179, "step": 10906 }, { "epoch": 0.7320559712761316, "grad_norm": 1.0411349534988403, "learning_rate": 7.319728589387888e-05, "loss": 0.6131, "step": 10908 }, { "epoch": 0.7321901949599007, "grad_norm": 0.9274797439575195, "learning_rate": 7.318765750856555e-05, "loss": 0.6152, "step": 10910 }, { "epoch": 0.7323244186436697, "grad_norm": 1.1459671258926392, "learning_rate": 7.317802802768102e-05, "loss": 0.6046, "step": 10912 }, { "epoch": 0.7324586423274386, "grad_norm": 0.9764459133148193, "learning_rate": 7.316839745168024e-05, "loss": 0.5685, "step": 10914 }, { "epoch": 0.7325928660112077, "grad_norm": 1.1787185668945312, "learning_rate": 7.315876578101823e-05, "loss": 0.61, "step": 10916 }, { "epoch": 0.7327270896949767, "grad_norm": 1.0656797885894775, "learning_rate": 7.314913301615008e-05, "loss": 0.5772, "step": 10918 }, { "epoch": 0.7328613133787457, "grad_norm": 0.9993340969085693, "learning_rate": 7.313949915753093e-05, "loss": 0.5461, "step": 10920 }, { "epoch": 0.7329955370625146, "grad_norm": 1.1809332370758057, "learning_rate": 7.312986420561593e-05, "loss": 0.5383, "step": 10922 }, { "epoch": 0.7331297607462837, "grad_norm": 1.0336933135986328, "learning_rate": 7.312022816086033e-05, "loss": 0.6016, "step": 10924 }, { "epoch": 0.7332639844300527, "grad_norm": 0.8597950339317322, "learning_rate": 7.311059102371942e-05, "loss": 0.5386, "step": 10926 }, { "epoch": 0.7333982081138217, "grad_norm": 1.0131632089614868, "learning_rate": 7.310095279464852e-05, "loss": 0.6432, "step": 10928 }, { "epoch": 0.7335324317975906, "grad_norm": 1.0859897136688232, "learning_rate": 7.309131347410303e-05, "loss": 0.6246, "step": 10930 }, { "epoch": 0.7336666554813597, "grad_norm": 0.9647706747055054, "learning_rate": 7.308167306253839e-05, "loss": 0.6463, "step": 10932 }, { "epoch": 0.7338008791651287, "grad_norm": 1.065216302871704, "learning_rate": 7.30720315604101e-05, "loss": 0.6372, "step": 10934 }, { "epoch": 0.7339351028488977, "grad_norm": 0.9903578162193298, "learning_rate": 7.306238896817366e-05, "loss": 0.5638, "step": 10936 }, { "epoch": 0.7340693265326667, "grad_norm": 1.1092201471328735, "learning_rate": 7.305274528628469e-05, "loss": 0.5985, "step": 10938 }, { "epoch": 0.7342035502164357, "grad_norm": 1.0268052816390991, "learning_rate": 7.304310051519886e-05, "loss": 0.6446, "step": 10940 }, { "epoch": 0.7343377739002047, "grad_norm": 1.010176181793213, "learning_rate": 7.303345465537184e-05, "loss": 0.5817, "step": 10942 }, { "epoch": 0.7344719975839737, "grad_norm": 0.9174301624298096, "learning_rate": 7.302380770725937e-05, "loss": 0.6188, "step": 10944 }, { "epoch": 0.7346062212677427, "grad_norm": 0.9151535034179688, "learning_rate": 7.301415967131727e-05, "loss": 0.6744, "step": 10946 }, { "epoch": 0.7347404449515117, "grad_norm": 1.102332353591919, "learning_rate": 7.300451054800137e-05, "loss": 0.6855, "step": 10948 }, { "epoch": 0.7348746686352807, "grad_norm": 1.0601832866668701, "learning_rate": 7.29948603377676e-05, "loss": 0.6343, "step": 10950 }, { "epoch": 0.7350088923190496, "grad_norm": 0.9875639081001282, "learning_rate": 7.298520904107187e-05, "loss": 0.5326, "step": 10952 }, { "epoch": 0.7351431160028187, "grad_norm": 1.0342354774475098, "learning_rate": 7.297555665837024e-05, "loss": 0.6317, "step": 10954 }, { "epoch": 0.7352773396865877, "grad_norm": 1.122740626335144, "learning_rate": 7.296590319011871e-05, "loss": 0.603, "step": 10956 }, { "epoch": 0.7354115633703567, "grad_norm": 1.1834608316421509, "learning_rate": 7.295624863677343e-05, "loss": 0.6336, "step": 10958 }, { "epoch": 0.7355457870541257, "grad_norm": 1.2283573150634766, "learning_rate": 7.294659299879054e-05, "loss": 0.569, "step": 10960 }, { "epoch": 0.7356800107378947, "grad_norm": 0.9388507604598999, "learning_rate": 7.293693627662625e-05, "loss": 0.5259, "step": 10962 }, { "epoch": 0.7358142344216637, "grad_norm": 1.1806296110153198, "learning_rate": 7.292727847073684e-05, "loss": 0.5792, "step": 10964 }, { "epoch": 0.7359484581054327, "grad_norm": 1.0787748098373413, "learning_rate": 7.29176195815786e-05, "loss": 0.5581, "step": 10966 }, { "epoch": 0.7360826817892017, "grad_norm": 1.2471569776535034, "learning_rate": 7.29079596096079e-05, "loss": 0.6241, "step": 10968 }, { "epoch": 0.7362169054729707, "grad_norm": 0.9744227528572083, "learning_rate": 7.289829855528114e-05, "loss": 0.6011, "step": 10970 }, { "epoch": 0.7363511291567397, "grad_norm": 0.9760153889656067, "learning_rate": 7.288863641905481e-05, "loss": 0.5663, "step": 10972 }, { "epoch": 0.7364853528405088, "grad_norm": 0.8509034514427185, "learning_rate": 7.287897320138542e-05, "loss": 0.5625, "step": 10974 }, { "epoch": 0.7366195765242777, "grad_norm": 1.7126649618148804, "learning_rate": 7.286930890272954e-05, "loss": 0.6058, "step": 10976 }, { "epoch": 0.7367538002080467, "grad_norm": 1.046785593032837, "learning_rate": 7.285964352354378e-05, "loss": 0.5684, "step": 10978 }, { "epoch": 0.7368880238918157, "grad_norm": 1.021963119506836, "learning_rate": 7.284997706428482e-05, "loss": 0.6269, "step": 10980 }, { "epoch": 0.7370222475755848, "grad_norm": 1.2418005466461182, "learning_rate": 7.284030952540937e-05, "loss": 0.6135, "step": 10982 }, { "epoch": 0.7371564712593537, "grad_norm": 1.1099112033843994, "learning_rate": 7.28306409073742e-05, "loss": 0.6034, "step": 10984 }, { "epoch": 0.7372906949431227, "grad_norm": 1.0667752027511597, "learning_rate": 7.282097121063616e-05, "loss": 0.6169, "step": 10986 }, { "epoch": 0.7374249186268917, "grad_norm": 1.0118855237960815, "learning_rate": 7.28113004356521e-05, "loss": 0.5993, "step": 10988 }, { "epoch": 0.7375591423106607, "grad_norm": 1.100252628326416, "learning_rate": 7.280162858287894e-05, "loss": 0.6263, "step": 10990 }, { "epoch": 0.7376933659944297, "grad_norm": 1.1078400611877441, "learning_rate": 7.279195565277369e-05, "loss": 0.6074, "step": 10992 }, { "epoch": 0.7378275896781987, "grad_norm": 1.0341601371765137, "learning_rate": 7.278228164579333e-05, "loss": 0.6238, "step": 10994 }, { "epoch": 0.7379618133619678, "grad_norm": 1.0098555088043213, "learning_rate": 7.277260656239497e-05, "loss": 0.6187, "step": 10996 }, { "epoch": 0.7380960370457367, "grad_norm": 1.0140324831008911, "learning_rate": 7.276293040303573e-05, "loss": 0.6115, "step": 10998 }, { "epoch": 0.7382302607295057, "grad_norm": 1.0767360925674438, "learning_rate": 7.275325316817279e-05, "loss": 0.6177, "step": 11000 }, { "epoch": 0.7383644844132747, "grad_norm": 1.2175441980361938, "learning_rate": 7.274357485826339e-05, "loss": 0.6105, "step": 11002 }, { "epoch": 0.7384987080970438, "grad_norm": 1.2842063903808594, "learning_rate": 7.27338954737648e-05, "loss": 0.6284, "step": 11004 }, { "epoch": 0.7386329317808127, "grad_norm": 1.1585752964019775, "learning_rate": 7.272421501513434e-05, "loss": 0.6264, "step": 11006 }, { "epoch": 0.7387671554645817, "grad_norm": 1.0015320777893066, "learning_rate": 7.27145334828294e-05, "loss": 0.6096, "step": 11008 }, { "epoch": 0.7389013791483507, "grad_norm": 1.0719194412231445, "learning_rate": 7.270485087730744e-05, "loss": 0.584, "step": 11010 }, { "epoch": 0.7390356028321198, "grad_norm": 1.157904863357544, "learning_rate": 7.269516719902591e-05, "loss": 0.5593, "step": 11012 }, { "epoch": 0.7391698265158887, "grad_norm": 0.820466935634613, "learning_rate": 7.268548244844236e-05, "loss": 0.6092, "step": 11014 }, { "epoch": 0.7393040501996577, "grad_norm": 0.9571653604507446, "learning_rate": 7.267579662601439e-05, "loss": 0.5796, "step": 11016 }, { "epoch": 0.7394382738834268, "grad_norm": 0.9789788126945496, "learning_rate": 7.266610973219959e-05, "loss": 0.5665, "step": 11018 }, { "epoch": 0.7395724975671958, "grad_norm": 1.0947332382202148, "learning_rate": 7.265642176745571e-05, "loss": 0.6398, "step": 11020 }, { "epoch": 0.7397067212509647, "grad_norm": 1.1176904439926147, "learning_rate": 7.264673273224042e-05, "loss": 0.6227, "step": 11022 }, { "epoch": 0.7398409449347337, "grad_norm": 1.037615180015564, "learning_rate": 7.263704262701157e-05, "loss": 0.6312, "step": 11024 }, { "epoch": 0.7399751686185028, "grad_norm": 1.0402809381484985, "learning_rate": 7.262735145222696e-05, "loss": 0.5933, "step": 11026 }, { "epoch": 0.7401093923022717, "grad_norm": 0.8275668025016785, "learning_rate": 7.261765920834447e-05, "loss": 0.5473, "step": 11028 }, { "epoch": 0.7402436159860407, "grad_norm": 0.9896917939186096, "learning_rate": 7.260796589582208e-05, "loss": 0.5697, "step": 11030 }, { "epoch": 0.7403778396698097, "grad_norm": 1.028495192527771, "learning_rate": 7.259827151511775e-05, "loss": 0.629, "step": 11032 }, { "epoch": 0.7405120633535788, "grad_norm": 1.251771330833435, "learning_rate": 7.258857606668951e-05, "loss": 0.6175, "step": 11034 }, { "epoch": 0.7406462870373477, "grad_norm": 1.0744801759719849, "learning_rate": 7.257887955099551e-05, "loss": 0.6593, "step": 11036 }, { "epoch": 0.7407805107211167, "grad_norm": 1.0244027376174927, "learning_rate": 7.256918196849382e-05, "loss": 0.5309, "step": 11038 }, { "epoch": 0.7409147344048858, "grad_norm": 1.1488616466522217, "learning_rate": 7.255948331964268e-05, "loss": 0.6284, "step": 11040 }, { "epoch": 0.7410489580886548, "grad_norm": 0.962162971496582, "learning_rate": 7.25497836049003e-05, "loss": 0.6222, "step": 11042 }, { "epoch": 0.7411831817724237, "grad_norm": 0.9596481323242188, "learning_rate": 7.2540082824725e-05, "loss": 0.6159, "step": 11044 }, { "epoch": 0.7413174054561927, "grad_norm": 0.9856430292129517, "learning_rate": 7.25303809795751e-05, "loss": 0.5966, "step": 11046 }, { "epoch": 0.7414516291399618, "grad_norm": 1.4479644298553467, "learning_rate": 7.2520678069909e-05, "loss": 0.5527, "step": 11048 }, { "epoch": 0.7415858528237308, "grad_norm": 1.0082017183303833, "learning_rate": 7.251097409618515e-05, "loss": 0.6666, "step": 11050 }, { "epoch": 0.7417200765074997, "grad_norm": 1.1670246124267578, "learning_rate": 7.250126905886204e-05, "loss": 0.575, "step": 11052 }, { "epoch": 0.7418543001912687, "grad_norm": 0.9044466018676758, "learning_rate": 7.249156295839824e-05, "loss": 0.5328, "step": 11054 }, { "epoch": 0.7419885238750378, "grad_norm": 1.0731639862060547, "learning_rate": 7.248185579525228e-05, "loss": 0.5571, "step": 11056 }, { "epoch": 0.7421227475588068, "grad_norm": 1.0607680082321167, "learning_rate": 7.247214756988285e-05, "loss": 0.5882, "step": 11058 }, { "epoch": 0.7422569712425757, "grad_norm": 0.9615009427070618, "learning_rate": 7.246243828274863e-05, "loss": 0.5327, "step": 11060 }, { "epoch": 0.7423911949263448, "grad_norm": 1.143050193786621, "learning_rate": 7.24527279343084e-05, "loss": 0.5794, "step": 11062 }, { "epoch": 0.7425254186101138, "grad_norm": 1.4304579496383667, "learning_rate": 7.24430165250209e-05, "loss": 0.5818, "step": 11064 }, { "epoch": 0.7426596422938827, "grad_norm": 0.991141140460968, "learning_rate": 7.2433304055345e-05, "loss": 0.6599, "step": 11066 }, { "epoch": 0.7427938659776517, "grad_norm": 1.0762946605682373, "learning_rate": 7.242359052573963e-05, "loss": 0.6425, "step": 11068 }, { "epoch": 0.7429280896614208, "grad_norm": 1.0666600465774536, "learning_rate": 7.241387593666368e-05, "loss": 0.5718, "step": 11070 }, { "epoch": 0.7430623133451898, "grad_norm": 1.2328325510025024, "learning_rate": 7.240416028857617e-05, "loss": 0.5796, "step": 11072 }, { "epoch": 0.7431965370289587, "grad_norm": 0.9430737495422363, "learning_rate": 7.239444358193613e-05, "loss": 0.6491, "step": 11074 }, { "epoch": 0.7433307607127277, "grad_norm": 1.041743278503418, "learning_rate": 7.238472581720268e-05, "loss": 0.64, "step": 11076 }, { "epoch": 0.7434649843964968, "grad_norm": 1.0096365213394165, "learning_rate": 7.237500699483495e-05, "loss": 0.5913, "step": 11078 }, { "epoch": 0.7435992080802658, "grad_norm": 1.6502892971038818, "learning_rate": 7.236528711529213e-05, "loss": 0.5335, "step": 11080 }, { "epoch": 0.7437334317640347, "grad_norm": 1.1586986780166626, "learning_rate": 7.23555661790335e-05, "loss": 0.5467, "step": 11082 }, { "epoch": 0.7438676554478038, "grad_norm": 0.7920651435852051, "learning_rate": 7.23458441865183e-05, "loss": 0.4931, "step": 11084 }, { "epoch": 0.7440018791315728, "grad_norm": 1.4115488529205322, "learning_rate": 7.233612113820592e-05, "loss": 0.6777, "step": 11086 }, { "epoch": 0.7441361028153418, "grad_norm": 1.0248574018478394, "learning_rate": 7.232639703455573e-05, "loss": 0.5823, "step": 11088 }, { "epoch": 0.7442703264991107, "grad_norm": 1.38546621799469, "learning_rate": 7.231667187602718e-05, "loss": 0.6076, "step": 11090 }, { "epoch": 0.7444045501828798, "grad_norm": 1.055503487586975, "learning_rate": 7.230694566307978e-05, "loss": 0.5366, "step": 11092 }, { "epoch": 0.7445387738666488, "grad_norm": 1.0996710062026978, "learning_rate": 7.229721839617306e-05, "loss": 0.6016, "step": 11094 }, { "epoch": 0.7446729975504178, "grad_norm": 1.1728242635726929, "learning_rate": 7.228749007576661e-05, "loss": 0.6528, "step": 11096 }, { "epoch": 0.7448072212341867, "grad_norm": 1.067726492881775, "learning_rate": 7.227776070232008e-05, "loss": 0.7174, "step": 11098 }, { "epoch": 0.7449414449179558, "grad_norm": 1.052566647529602, "learning_rate": 7.226803027629316e-05, "loss": 0.5095, "step": 11100 }, { "epoch": 0.7450756686017248, "grad_norm": 0.9860169887542725, "learning_rate": 7.225829879814561e-05, "loss": 0.5785, "step": 11102 }, { "epoch": 0.7452098922854937, "grad_norm": 0.9966432452201843, "learning_rate": 7.22485662683372e-05, "loss": 0.5927, "step": 11104 }, { "epoch": 0.7453441159692628, "grad_norm": 0.8664472699165344, "learning_rate": 7.223883268732779e-05, "loss": 0.5297, "step": 11106 }, { "epoch": 0.7454783396530318, "grad_norm": 1.3534940481185913, "learning_rate": 7.222909805557726e-05, "loss": 0.6371, "step": 11108 }, { "epoch": 0.7456125633368008, "grad_norm": 1.2141648530960083, "learning_rate": 7.221936237354557e-05, "loss": 0.5975, "step": 11110 }, { "epoch": 0.7457467870205697, "grad_norm": 1.1479947566986084, "learning_rate": 7.22096256416927e-05, "loss": 0.5909, "step": 11112 }, { "epoch": 0.7458810107043388, "grad_norm": 1.2423744201660156, "learning_rate": 7.219988786047866e-05, "loss": 0.5182, "step": 11114 }, { "epoch": 0.7460152343881078, "grad_norm": 0.9990307092666626, "learning_rate": 7.219014903036361e-05, "loss": 0.608, "step": 11116 }, { "epoch": 0.7461494580718768, "grad_norm": 0.9691771268844604, "learning_rate": 7.218040915180764e-05, "loss": 0.5972, "step": 11118 }, { "epoch": 0.7462836817556457, "grad_norm": 1.0182071924209595, "learning_rate": 7.217066822527096e-05, "loss": 0.5767, "step": 11120 }, { "epoch": 0.7464179054394148, "grad_norm": 1.1678303480148315, "learning_rate": 7.216092625121379e-05, "loss": 0.5906, "step": 11122 }, { "epoch": 0.7465521291231838, "grad_norm": 1.0928617715835571, "learning_rate": 7.215118323009643e-05, "loss": 0.5542, "step": 11124 }, { "epoch": 0.7466863528069528, "grad_norm": 1.1073976755142212, "learning_rate": 7.214143916237925e-05, "loss": 0.592, "step": 11126 }, { "epoch": 0.7468205764907218, "grad_norm": 1.024496078491211, "learning_rate": 7.213169404852258e-05, "loss": 0.6111, "step": 11128 }, { "epoch": 0.7469548001744908, "grad_norm": 0.96971195936203, "learning_rate": 7.21219478889869e-05, "loss": 0.6242, "step": 11130 }, { "epoch": 0.7470890238582598, "grad_norm": 1.0988749265670776, "learning_rate": 7.211220068423266e-05, "loss": 0.6015, "step": 11132 }, { "epoch": 0.7472232475420288, "grad_norm": 1.0498028993606567, "learning_rate": 7.210245243472046e-05, "loss": 0.5652, "step": 11134 }, { "epoch": 0.7473574712257978, "grad_norm": 0.9467136859893799, "learning_rate": 7.209270314091081e-05, "loss": 0.6386, "step": 11136 }, { "epoch": 0.7474916949095668, "grad_norm": 1.2103420495986938, "learning_rate": 7.208295280326439e-05, "loss": 0.5894, "step": 11138 }, { "epoch": 0.7476259185933358, "grad_norm": 1.0159958600997925, "learning_rate": 7.207320142224188e-05, "loss": 0.6403, "step": 11140 }, { "epoch": 0.7477601422771047, "grad_norm": 0.9736151695251465, "learning_rate": 7.206344899830401e-05, "loss": 0.6065, "step": 11142 }, { "epoch": 0.7478943659608738, "grad_norm": 1.1738616228103638, "learning_rate": 7.205369553191156e-05, "loss": 0.5054, "step": 11144 }, { "epoch": 0.7480285896446428, "grad_norm": 0.9719931483268738, "learning_rate": 7.204394102352535e-05, "loss": 0.6456, "step": 11146 }, { "epoch": 0.7481628133284118, "grad_norm": 0.9128687381744385, "learning_rate": 7.20341854736063e-05, "loss": 0.5473, "step": 11148 }, { "epoch": 0.7482970370121808, "grad_norm": 1.0695174932479858, "learning_rate": 7.20244288826153e-05, "loss": 0.5721, "step": 11150 }, { "epoch": 0.7484312606959498, "grad_norm": 1.0116002559661865, "learning_rate": 7.201467125101332e-05, "loss": 0.5793, "step": 11152 }, { "epoch": 0.7485654843797188, "grad_norm": 1.1777184009552002, "learning_rate": 7.200491257926145e-05, "loss": 0.6363, "step": 11154 }, { "epoch": 0.7486997080634878, "grad_norm": 1.085649847984314, "learning_rate": 7.19951528678207e-05, "loss": 0.6127, "step": 11156 }, { "epoch": 0.7488339317472568, "grad_norm": 1.094965934753418, "learning_rate": 7.198539211715226e-05, "loss": 0.6476, "step": 11158 }, { "epoch": 0.7489681554310258, "grad_norm": 0.9376707673072815, "learning_rate": 7.197563032771727e-05, "loss": 0.5898, "step": 11160 }, { "epoch": 0.7491023791147948, "grad_norm": 1.0657490491867065, "learning_rate": 7.196586749997694e-05, "loss": 0.6939, "step": 11162 }, { "epoch": 0.7492366027985639, "grad_norm": 0.9906362891197205, "learning_rate": 7.195610363439259e-05, "loss": 0.5848, "step": 11164 }, { "epoch": 0.7493708264823328, "grad_norm": 1.4110857248306274, "learning_rate": 7.194633873142548e-05, "loss": 0.6378, "step": 11166 }, { "epoch": 0.7495050501661018, "grad_norm": 1.144875168800354, "learning_rate": 7.193657279153706e-05, "loss": 0.6205, "step": 11168 }, { "epoch": 0.7496392738498708, "grad_norm": 1.0834269523620605, "learning_rate": 7.19268058151887e-05, "loss": 0.5946, "step": 11170 }, { "epoch": 0.7497734975336399, "grad_norm": 1.1376134157180786, "learning_rate": 7.191703780284187e-05, "loss": 0.6001, "step": 11172 }, { "epoch": 0.7499077212174088, "grad_norm": 0.9572040438652039, "learning_rate": 7.190726875495812e-05, "loss": 0.6463, "step": 11174 }, { "epoch": 0.7500419449011778, "grad_norm": 0.9035720229148865, "learning_rate": 7.189749867199899e-05, "loss": 0.5924, "step": 11176 }, { "epoch": 0.7501761685849468, "grad_norm": 1.004518747329712, "learning_rate": 7.188772755442611e-05, "loss": 0.5729, "step": 11178 }, { "epoch": 0.7503103922687158, "grad_norm": 1.8473553657531738, "learning_rate": 7.187795540270114e-05, "loss": 0.612, "step": 11180 }, { "epoch": 0.7504446159524848, "grad_norm": 5.95490026473999, "learning_rate": 7.18681822172858e-05, "loss": 0.5695, "step": 11182 }, { "epoch": 0.7505788396362538, "grad_norm": 1.2424906492233276, "learning_rate": 7.185840799864186e-05, "loss": 0.6821, "step": 11184 }, { "epoch": 0.7507130633200229, "grad_norm": 1.002289056777954, "learning_rate": 7.184863274723111e-05, "loss": 0.6464, "step": 11186 }, { "epoch": 0.7508472870037918, "grad_norm": 1.2320494651794434, "learning_rate": 7.183885646351542e-05, "loss": 0.6134, "step": 11188 }, { "epoch": 0.7509815106875608, "grad_norm": 1.0252602100372314, "learning_rate": 7.182907914795672e-05, "loss": 0.6894, "step": 11190 }, { "epoch": 0.7511157343713298, "grad_norm": 1.0166563987731934, "learning_rate": 7.181930080101696e-05, "loss": 0.5954, "step": 11192 }, { "epoch": 0.7512499580550989, "grad_norm": 1.0519530773162842, "learning_rate": 7.180952142315813e-05, "loss": 0.6021, "step": 11194 }, { "epoch": 0.7513841817388678, "grad_norm": 1.084174633026123, "learning_rate": 7.179974101484232e-05, "loss": 0.5695, "step": 11196 }, { "epoch": 0.7515184054226368, "grad_norm": 0.9310356974601746, "learning_rate": 7.17899595765316e-05, "loss": 0.4996, "step": 11198 }, { "epoch": 0.7516526291064058, "grad_norm": 0.9874021410942078, "learning_rate": 7.178017710868814e-05, "loss": 0.5587, "step": 11200 }, { "epoch": 0.7517868527901749, "grad_norm": 2.154062032699585, "learning_rate": 7.177039361177413e-05, "loss": 0.6281, "step": 11202 }, { "epoch": 0.7519210764739438, "grad_norm": 1.0503273010253906, "learning_rate": 7.176060908625184e-05, "loss": 0.5806, "step": 11204 }, { "epoch": 0.7520553001577128, "grad_norm": 1.2201011180877686, "learning_rate": 7.175082353258358e-05, "loss": 0.6996, "step": 11206 }, { "epoch": 0.7521895238414819, "grad_norm": 0.8703465461730957, "learning_rate": 7.174103695123166e-05, "loss": 0.4985, "step": 11208 }, { "epoch": 0.7523237475252509, "grad_norm": 1.0070104598999023, "learning_rate": 7.17312493426585e-05, "loss": 0.535, "step": 11210 }, { "epoch": 0.7524579712090198, "grad_norm": 1.2064918279647827, "learning_rate": 7.172146070732652e-05, "loss": 0.6514, "step": 11212 }, { "epoch": 0.7525921948927888, "grad_norm": 1.5008562803268433, "learning_rate": 7.171167104569826e-05, "loss": 0.5332, "step": 11214 }, { "epoch": 0.7527264185765579, "grad_norm": 1.0957249402999878, "learning_rate": 7.170188035823624e-05, "loss": 0.6015, "step": 11216 }, { "epoch": 0.7528606422603268, "grad_norm": 1.0030272006988525, "learning_rate": 7.169208864540303e-05, "loss": 0.5409, "step": 11218 }, { "epoch": 0.7529948659440958, "grad_norm": 0.9358177185058594, "learning_rate": 7.16822959076613e-05, "loss": 0.57, "step": 11220 }, { "epoch": 0.7531290896278648, "grad_norm": 1.0475953817367554, "learning_rate": 7.167250214547372e-05, "loss": 0.6187, "step": 11222 }, { "epoch": 0.7532633133116339, "grad_norm": 0.8726040124893188, "learning_rate": 7.166270735930304e-05, "loss": 0.5462, "step": 11224 }, { "epoch": 0.7533975369954028, "grad_norm": 0.8723013401031494, "learning_rate": 7.165291154961202e-05, "loss": 0.5357, "step": 11226 }, { "epoch": 0.7535317606791718, "grad_norm": 0.9909194111824036, "learning_rate": 7.164311471686352e-05, "loss": 0.6319, "step": 11228 }, { "epoch": 0.7536659843629409, "grad_norm": 1.0541589260101318, "learning_rate": 7.163331686152042e-05, "loss": 0.6679, "step": 11230 }, { "epoch": 0.7538002080467099, "grad_norm": 0.9891774654388428, "learning_rate": 7.16235179840456e-05, "loss": 0.6496, "step": 11232 }, { "epoch": 0.7539344317304788, "grad_norm": 1.0104056596755981, "learning_rate": 7.161371808490212e-05, "loss": 0.5731, "step": 11234 }, { "epoch": 0.7540686554142478, "grad_norm": 1.0035972595214844, "learning_rate": 7.160391716455292e-05, "loss": 0.6075, "step": 11236 }, { "epoch": 0.7542028790980169, "grad_norm": 0.9934561252593994, "learning_rate": 7.159411522346115e-05, "loss": 0.5643, "step": 11238 }, { "epoch": 0.7543371027817859, "grad_norm": 1.009917140007019, "learning_rate": 7.158431226208988e-05, "loss": 0.5825, "step": 11240 }, { "epoch": 0.7544713264655548, "grad_norm": 0.9430350661277771, "learning_rate": 7.157450828090231e-05, "loss": 0.5723, "step": 11242 }, { "epoch": 0.7546055501493238, "grad_norm": 1.1066055297851562, "learning_rate": 7.156470328036165e-05, "loss": 0.6028, "step": 11244 }, { "epoch": 0.7547397738330929, "grad_norm": 0.9261332154273987, "learning_rate": 7.155489726093114e-05, "loss": 0.5163, "step": 11246 }, { "epoch": 0.7548739975168619, "grad_norm": 1.0471699237823486, "learning_rate": 7.154509022307415e-05, "loss": 0.6726, "step": 11248 }, { "epoch": 0.7550082212006308, "grad_norm": 1.1183794736862183, "learning_rate": 7.1535282167254e-05, "loss": 0.602, "step": 11250 }, { "epoch": 0.7551424448843999, "grad_norm": 1.6070631742477417, "learning_rate": 7.152547309393411e-05, "loss": 0.6271, "step": 11252 }, { "epoch": 0.7552766685681689, "grad_norm": 2.8954949378967285, "learning_rate": 7.151566300357796e-05, "loss": 0.5473, "step": 11254 }, { "epoch": 0.7554108922519378, "grad_norm": 1.2352590560913086, "learning_rate": 7.150585189664902e-05, "loss": 0.5778, "step": 11256 }, { "epoch": 0.7555451159357068, "grad_norm": 1.0116827487945557, "learning_rate": 7.14960397736109e-05, "loss": 0.617, "step": 11258 }, { "epoch": 0.7556793396194759, "grad_norm": 1.0050992965698242, "learning_rate": 7.148622663492715e-05, "loss": 0.5584, "step": 11260 }, { "epoch": 0.7558135633032449, "grad_norm": 1.0844353437423706, "learning_rate": 7.147641248106142e-05, "loss": 0.5867, "step": 11262 }, { "epoch": 0.7559477869870138, "grad_norm": 1.0055772066116333, "learning_rate": 7.146659731247747e-05, "loss": 0.6329, "step": 11264 }, { "epoch": 0.7560820106707828, "grad_norm": 1.0988073348999023, "learning_rate": 7.1456781129639e-05, "loss": 0.5738, "step": 11266 }, { "epoch": 0.7562162343545519, "grad_norm": 1.0617798566818237, "learning_rate": 7.144696393300981e-05, "loss": 0.5328, "step": 11268 }, { "epoch": 0.7563504580383209, "grad_norm": 1.1145955324172974, "learning_rate": 7.143714572305374e-05, "loss": 0.5807, "step": 11270 }, { "epoch": 0.7564846817220898, "grad_norm": 1.2228416204452515, "learning_rate": 7.14273265002347e-05, "loss": 0.6712, "step": 11272 }, { "epoch": 0.7566189054058589, "grad_norm": 1.2184854745864868, "learning_rate": 7.141750626501661e-05, "loss": 0.6264, "step": 11274 }, { "epoch": 0.7567531290896279, "grad_norm": 0.8387179374694824, "learning_rate": 7.140768501786347e-05, "loss": 0.5983, "step": 11276 }, { "epoch": 0.7568873527733969, "grad_norm": 1.1214351654052734, "learning_rate": 7.13978627592393e-05, "loss": 0.576, "step": 11278 }, { "epoch": 0.7570215764571658, "grad_norm": 1.0166372060775757, "learning_rate": 7.13880394896082e-05, "loss": 0.5419, "step": 11280 }, { "epoch": 0.7571558001409349, "grad_norm": 1.0215328931808472, "learning_rate": 7.13782152094343e-05, "loss": 0.58, "step": 11282 }, { "epoch": 0.7572900238247039, "grad_norm": 2.02290678024292, "learning_rate": 7.136838991918175e-05, "loss": 0.7162, "step": 11284 }, { "epoch": 0.7574242475084729, "grad_norm": 1.0733305215835571, "learning_rate": 7.135856361931482e-05, "loss": 0.6071, "step": 11286 }, { "epoch": 0.7575584711922418, "grad_norm": 1.0772255659103394, "learning_rate": 7.134873631029775e-05, "loss": 0.5756, "step": 11288 }, { "epoch": 0.7576926948760109, "grad_norm": 1.0512280464172363, "learning_rate": 7.133890799259486e-05, "loss": 0.6125, "step": 11290 }, { "epoch": 0.7578269185597799, "grad_norm": 1.0169273614883423, "learning_rate": 7.132907866667053e-05, "loss": 0.5617, "step": 11292 }, { "epoch": 0.7579611422435488, "grad_norm": 1.1695070266723633, "learning_rate": 7.131924833298918e-05, "loss": 0.6074, "step": 11294 }, { "epoch": 0.7580953659273179, "grad_norm": 1.4435280561447144, "learning_rate": 7.130941699201528e-05, "loss": 0.619, "step": 11296 }, { "epoch": 0.7582295896110869, "grad_norm": 1.000217080116272, "learning_rate": 7.129958464421331e-05, "loss": 0.5763, "step": 11298 }, { "epoch": 0.7583638132948559, "grad_norm": 1.026188611984253, "learning_rate": 7.128975129004786e-05, "loss": 0.5677, "step": 11300 }, { "epoch": 0.7584980369786248, "grad_norm": 1.1340270042419434, "learning_rate": 7.127991692998353e-05, "loss": 0.621, "step": 11302 }, { "epoch": 0.7586322606623939, "grad_norm": 1.1129652261734009, "learning_rate": 7.127008156448496e-05, "loss": 0.7227, "step": 11304 }, { "epoch": 0.7587664843461629, "grad_norm": 0.9840350151062012, "learning_rate": 7.126024519401687e-05, "loss": 0.6042, "step": 11306 }, { "epoch": 0.7589007080299319, "grad_norm": 1.0387786626815796, "learning_rate": 7.1250407819044e-05, "loss": 0.6167, "step": 11308 }, { "epoch": 0.7590349317137008, "grad_norm": 0.9231281876564026, "learning_rate": 7.124056944003114e-05, "loss": 0.5275, "step": 11310 }, { "epoch": 0.7591691553974699, "grad_norm": 1.3023412227630615, "learning_rate": 7.123073005744314e-05, "loss": 0.5872, "step": 11312 }, { "epoch": 0.7593033790812389, "grad_norm": 3.0256383419036865, "learning_rate": 7.12208896717449e-05, "loss": 0.5885, "step": 11314 }, { "epoch": 0.7594376027650079, "grad_norm": 1.1872811317443848, "learning_rate": 7.121104828340136e-05, "loss": 0.5464, "step": 11316 }, { "epoch": 0.7595718264487769, "grad_norm": 0.9879008531570435, "learning_rate": 7.120120589287749e-05, "loss": 0.5875, "step": 11318 }, { "epoch": 0.7597060501325459, "grad_norm": 1.045357346534729, "learning_rate": 7.119136250063833e-05, "loss": 0.5428, "step": 11320 }, { "epoch": 0.7598402738163149, "grad_norm": 1.0363407135009766, "learning_rate": 7.118151810714896e-05, "loss": 0.5652, "step": 11322 }, { "epoch": 0.759974497500084, "grad_norm": 1.0396844148635864, "learning_rate": 7.117167271287453e-05, "loss": 0.661, "step": 11324 }, { "epoch": 0.7601087211838529, "grad_norm": 1.00739586353302, "learning_rate": 7.116182631828016e-05, "loss": 0.6256, "step": 11326 }, { "epoch": 0.7602429448676219, "grad_norm": 1.515653133392334, "learning_rate": 7.115197892383114e-05, "loss": 0.6617, "step": 11328 }, { "epoch": 0.7603771685513909, "grad_norm": 1.025126338005066, "learning_rate": 7.11421305299927e-05, "loss": 0.6959, "step": 11330 }, { "epoch": 0.7605113922351598, "grad_norm": 1.0322550535202026, "learning_rate": 7.113228113723016e-05, "loss": 0.5964, "step": 11332 }, { "epoch": 0.7606456159189289, "grad_norm": 1.1067694425582886, "learning_rate": 7.11224307460089e-05, "loss": 0.4972, "step": 11334 }, { "epoch": 0.7607798396026979, "grad_norm": 1.0172574520111084, "learning_rate": 7.111257935679433e-05, "loss": 0.5196, "step": 11336 }, { "epoch": 0.7609140632864669, "grad_norm": 0.9377800822257996, "learning_rate": 7.110272697005189e-05, "loss": 0.5904, "step": 11338 }, { "epoch": 0.7610482869702359, "grad_norm": 0.9999615550041199, "learning_rate": 7.10928735862471e-05, "loss": 0.6526, "step": 11340 }, { "epoch": 0.7611825106540049, "grad_norm": 1.1272873878479004, "learning_rate": 7.108301920584552e-05, "loss": 0.5765, "step": 11342 }, { "epoch": 0.7613167343377739, "grad_norm": 1.0125153064727783, "learning_rate": 7.107316382931272e-05, "loss": 0.6562, "step": 11344 }, { "epoch": 0.761450958021543, "grad_norm": 1.16325044631958, "learning_rate": 7.106330745711438e-05, "loss": 0.61, "step": 11346 }, { "epoch": 0.7615851817053119, "grad_norm": 1.0269536972045898, "learning_rate": 7.105345008971619e-05, "loss": 0.6078, "step": 11348 }, { "epoch": 0.7617194053890809, "grad_norm": 1.2129024267196655, "learning_rate": 7.104359172758387e-05, "loss": 0.6532, "step": 11350 }, { "epoch": 0.7618536290728499, "grad_norm": 1.1761541366577148, "learning_rate": 7.103373237118321e-05, "loss": 0.5589, "step": 11352 }, { "epoch": 0.761987852756619, "grad_norm": 1.0910080671310425, "learning_rate": 7.102387202098008e-05, "loss": 0.5613, "step": 11354 }, { "epoch": 0.7621220764403879, "grad_norm": 1.0277620553970337, "learning_rate": 7.101401067744033e-05, "loss": 0.6106, "step": 11356 }, { "epoch": 0.7622563001241569, "grad_norm": 0.9003410339355469, "learning_rate": 7.10041483410299e-05, "loss": 0.5574, "step": 11358 }, { "epoch": 0.7623905238079259, "grad_norm": 1.14950692653656, "learning_rate": 7.099428501221476e-05, "loss": 0.5946, "step": 11360 }, { "epoch": 0.7625247474916949, "grad_norm": 0.9178588390350342, "learning_rate": 7.098442069146095e-05, "loss": 0.5615, "step": 11362 }, { "epoch": 0.7626589711754639, "grad_norm": 1.1091738939285278, "learning_rate": 7.09745553792345e-05, "loss": 0.6495, "step": 11364 }, { "epoch": 0.7627931948592329, "grad_norm": 1.4295737743377686, "learning_rate": 7.096468907600157e-05, "loss": 0.6089, "step": 11366 }, { "epoch": 0.762927418543002, "grad_norm": 0.9904151558876038, "learning_rate": 7.095482178222832e-05, "loss": 0.6365, "step": 11368 }, { "epoch": 0.7630616422267709, "grad_norm": 0.9750891923904419, "learning_rate": 7.094495349838092e-05, "loss": 0.5672, "step": 11370 }, { "epoch": 0.7631958659105399, "grad_norm": 1.06587815284729, "learning_rate": 7.093508422492568e-05, "loss": 0.5864, "step": 11372 }, { "epoch": 0.7633300895943089, "grad_norm": 0.9094560742378235, "learning_rate": 7.092521396232887e-05, "loss": 0.5021, "step": 11374 }, { "epoch": 0.763464313278078, "grad_norm": 1.104093074798584, "learning_rate": 7.091534271105682e-05, "loss": 0.6175, "step": 11376 }, { "epoch": 0.7635985369618469, "grad_norm": 0.9446908831596375, "learning_rate": 7.090547047157599e-05, "loss": 0.5092, "step": 11378 }, { "epoch": 0.7637327606456159, "grad_norm": 1.130378007888794, "learning_rate": 7.089559724435277e-05, "loss": 0.645, "step": 11380 }, { "epoch": 0.7638669843293849, "grad_norm": 0.971934974193573, "learning_rate": 7.088572302985368e-05, "loss": 0.5984, "step": 11382 }, { "epoch": 0.764001208013154, "grad_norm": 1.1072789430618286, "learning_rate": 7.087584782854525e-05, "loss": 0.5966, "step": 11384 }, { "epoch": 0.7641354316969229, "grad_norm": 0.9358595013618469, "learning_rate": 7.086597164089404e-05, "loss": 0.5732, "step": 11386 }, { "epoch": 0.7642696553806919, "grad_norm": 0.9158409833908081, "learning_rate": 7.085609446736671e-05, "loss": 0.6165, "step": 11388 }, { "epoch": 0.764403879064461, "grad_norm": 1.1295926570892334, "learning_rate": 7.084621630842993e-05, "loss": 0.6221, "step": 11390 }, { "epoch": 0.76453810274823, "grad_norm": 1.6016974449157715, "learning_rate": 7.083633716455043e-05, "loss": 0.5171, "step": 11392 }, { "epoch": 0.7646723264319989, "grad_norm": 1.0827736854553223, "learning_rate": 7.082645703619496e-05, "loss": 0.6444, "step": 11394 }, { "epoch": 0.7648065501157679, "grad_norm": 1.1585004329681396, "learning_rate": 7.081657592383035e-05, "loss": 0.5721, "step": 11396 }, { "epoch": 0.764940773799537, "grad_norm": 1.180953025817871, "learning_rate": 7.080669382792346e-05, "loss": 0.5513, "step": 11398 }, { "epoch": 0.7650749974833059, "grad_norm": 0.9680315256118774, "learning_rate": 7.079681074894123e-05, "loss": 0.5934, "step": 11400 }, { "epoch": 0.7652092211670749, "grad_norm": 1.234734058380127, "learning_rate": 7.078692668735054e-05, "loss": 0.5883, "step": 11402 }, { "epoch": 0.7653434448508439, "grad_norm": 1.1248189210891724, "learning_rate": 7.077704164361848e-05, "loss": 0.5944, "step": 11404 }, { "epoch": 0.765477668534613, "grad_norm": 1.1900737285614014, "learning_rate": 7.076715561821204e-05, "loss": 0.5976, "step": 11406 }, { "epoch": 0.7656118922183819, "grad_norm": 1.0213483572006226, "learning_rate": 7.075726861159832e-05, "loss": 0.5714, "step": 11408 }, { "epoch": 0.7657461159021509, "grad_norm": 1.0772080421447754, "learning_rate": 7.07473806242445e-05, "loss": 0.5593, "step": 11410 }, { "epoch": 0.76588033958592, "grad_norm": 1.0889390707015991, "learning_rate": 7.073749165661773e-05, "loss": 0.6026, "step": 11412 }, { "epoch": 0.766014563269689, "grad_norm": 1.0070146322250366, "learning_rate": 7.072760170918526e-05, "loss": 0.6509, "step": 11414 }, { "epoch": 0.7661487869534579, "grad_norm": 0.9670523405075073, "learning_rate": 7.071771078241438e-05, "loss": 0.5696, "step": 11416 }, { "epoch": 0.7662830106372269, "grad_norm": 1.8057606220245361, "learning_rate": 7.070781887677239e-05, "loss": 0.5914, "step": 11418 }, { "epoch": 0.766417234320996, "grad_norm": 0.9283233880996704, "learning_rate": 7.069792599272669e-05, "loss": 0.606, "step": 11420 }, { "epoch": 0.766551458004765, "grad_norm": 1.1039934158325195, "learning_rate": 7.068803213074468e-05, "loss": 0.6239, "step": 11422 }, { "epoch": 0.7666856816885339, "grad_norm": 1.1731226444244385, "learning_rate": 7.067813729129384e-05, "loss": 0.5942, "step": 11424 }, { "epoch": 0.7668199053723029, "grad_norm": 1.0220401287078857, "learning_rate": 7.066824147484165e-05, "loss": 0.554, "step": 11426 }, { "epoch": 0.766954129056072, "grad_norm": 1.0181552171707153, "learning_rate": 7.065834468185573e-05, "loss": 0.6014, "step": 11428 }, { "epoch": 0.767088352739841, "grad_norm": 0.9061861038208008, "learning_rate": 7.064844691280362e-05, "loss": 0.5985, "step": 11430 }, { "epoch": 0.7672225764236099, "grad_norm": 0.8315479159355164, "learning_rate": 7.063854816815301e-05, "loss": 0.5446, "step": 11432 }, { "epoch": 0.767356800107379, "grad_norm": 1.047680377960205, "learning_rate": 7.062864844837159e-05, "loss": 0.6219, "step": 11434 }, { "epoch": 0.767491023791148, "grad_norm": 1.0420489311218262, "learning_rate": 7.061874775392709e-05, "loss": 0.5858, "step": 11436 }, { "epoch": 0.7676252474749169, "grad_norm": 1.1388205289840698, "learning_rate": 7.060884608528728e-05, "loss": 0.6973, "step": 11438 }, { "epoch": 0.7677594711586859, "grad_norm": 1.8225839138031006, "learning_rate": 7.059894344292004e-05, "loss": 0.573, "step": 11440 }, { "epoch": 0.767893694842455, "grad_norm": 1.0094853639602661, "learning_rate": 7.058903982729322e-05, "loss": 0.6169, "step": 11442 }, { "epoch": 0.768027918526224, "grad_norm": 0.9954559803009033, "learning_rate": 7.057913523887478e-05, "loss": 0.5874, "step": 11444 }, { "epoch": 0.7681621422099929, "grad_norm": 1.2659293413162231, "learning_rate": 7.056922967813263e-05, "loss": 0.6087, "step": 11446 }, { "epoch": 0.7682963658937619, "grad_norm": 0.9722047448158264, "learning_rate": 7.055932314553485e-05, "loss": 0.6462, "step": 11448 }, { "epoch": 0.768430589577531, "grad_norm": 1.0264939069747925, "learning_rate": 7.054941564154946e-05, "loss": 0.5979, "step": 11450 }, { "epoch": 0.7685648132613, "grad_norm": 1.1204912662506104, "learning_rate": 7.05395071666446e-05, "loss": 0.6051, "step": 11452 }, { "epoch": 0.7686990369450689, "grad_norm": 1.0123697519302368, "learning_rate": 7.052959772128841e-05, "loss": 0.6148, "step": 11454 }, { "epoch": 0.768833260628838, "grad_norm": 0.9574852585792542, "learning_rate": 7.051968730594911e-05, "loss": 0.573, "step": 11456 }, { "epoch": 0.768967484312607, "grad_norm": 1.3269104957580566, "learning_rate": 7.050977592109494e-05, "loss": 0.5905, "step": 11458 }, { "epoch": 0.769101707996376, "grad_norm": 1.0462757349014282, "learning_rate": 7.049986356719417e-05, "loss": 0.6053, "step": 11460 }, { "epoch": 0.7692359316801449, "grad_norm": 1.1457247734069824, "learning_rate": 7.048995024471517e-05, "loss": 0.6514, "step": 11462 }, { "epoch": 0.769370155363914, "grad_norm": 1.0068777799606323, "learning_rate": 7.048003595412632e-05, "loss": 0.6148, "step": 11464 }, { "epoch": 0.769504379047683, "grad_norm": 0.990300714969635, "learning_rate": 7.047012069589601e-05, "loss": 0.5711, "step": 11466 }, { "epoch": 0.769638602731452, "grad_norm": 1.1073365211486816, "learning_rate": 7.046020447049277e-05, "loss": 0.6263, "step": 11468 }, { "epoch": 0.7697728264152209, "grad_norm": 1.0013397932052612, "learning_rate": 7.045028727838511e-05, "loss": 0.5909, "step": 11470 }, { "epoch": 0.76990705009899, "grad_norm": 1.0649570226669312, "learning_rate": 7.044036912004159e-05, "loss": 0.6719, "step": 11472 }, { "epoch": 0.770041273782759, "grad_norm": 1.0973566770553589, "learning_rate": 7.04304499959308e-05, "loss": 0.5559, "step": 11474 }, { "epoch": 0.7701754974665279, "grad_norm": 0.9822953343391418, "learning_rate": 7.042052990652146e-05, "loss": 0.6049, "step": 11476 }, { "epoch": 0.7703097211502969, "grad_norm": 1.1304566860198975, "learning_rate": 7.041060885228222e-05, "loss": 0.6276, "step": 11478 }, { "epoch": 0.770443944834066, "grad_norm": 1.0694489479064941, "learning_rate": 7.040068683368181e-05, "loss": 0.5635, "step": 11480 }, { "epoch": 0.770578168517835, "grad_norm": 0.9529529809951782, "learning_rate": 7.039076385118911e-05, "loss": 0.5494, "step": 11482 }, { "epoch": 0.7707123922016039, "grad_norm": 1.0704907178878784, "learning_rate": 7.03808399052729e-05, "loss": 0.6072, "step": 11484 }, { "epoch": 0.770846615885373, "grad_norm": 1.017514944076538, "learning_rate": 7.03709149964021e-05, "loss": 0.5016, "step": 11486 }, { "epoch": 0.770980839569142, "grad_norm": 1.1568889617919922, "learning_rate": 7.036098912504559e-05, "loss": 0.5852, "step": 11488 }, { "epoch": 0.771115063252911, "grad_norm": 1.095329761505127, "learning_rate": 7.035106229167241e-05, "loss": 0.6255, "step": 11490 }, { "epoch": 0.7712492869366799, "grad_norm": 1.0314277410507202, "learning_rate": 7.034113449675154e-05, "loss": 0.5831, "step": 11492 }, { "epoch": 0.771383510620449, "grad_norm": 0.9504003524780273, "learning_rate": 7.033120574075206e-05, "loss": 0.6363, "step": 11494 }, { "epoch": 0.771517734304218, "grad_norm": 1.0379869937896729, "learning_rate": 7.032127602414311e-05, "loss": 0.5973, "step": 11496 }, { "epoch": 0.771651957987987, "grad_norm": 1.1097465753555298, "learning_rate": 7.031134534739381e-05, "loss": 0.6785, "step": 11498 }, { "epoch": 0.7717861816717559, "grad_norm": 0.9349481463432312, "learning_rate": 7.030141371097339e-05, "loss": 0.5758, "step": 11500 }, { "epoch": 0.771920405355525, "grad_norm": 1.1350663900375366, "learning_rate": 7.029148111535109e-05, "loss": 0.5973, "step": 11502 }, { "epoch": 0.772054629039294, "grad_norm": 1.1020756959915161, "learning_rate": 7.02815475609962e-05, "loss": 0.5947, "step": 11504 }, { "epoch": 0.772188852723063, "grad_norm": 0.9822752475738525, "learning_rate": 7.02716130483781e-05, "loss": 0.6771, "step": 11506 }, { "epoch": 0.772323076406832, "grad_norm": 0.8996891379356384, "learning_rate": 7.026167757796612e-05, "loss": 0.5805, "step": 11508 }, { "epoch": 0.772457300090601, "grad_norm": 0.9797414541244507, "learning_rate": 7.025174115022972e-05, "loss": 0.5829, "step": 11510 }, { "epoch": 0.77259152377437, "grad_norm": 1.0092127323150635, "learning_rate": 7.024180376563838e-05, "loss": 0.5577, "step": 11512 }, { "epoch": 0.7727257474581389, "grad_norm": 0.9922501444816589, "learning_rate": 7.02318654246616e-05, "loss": 0.5855, "step": 11514 }, { "epoch": 0.772859971141908, "grad_norm": 0.8650910258293152, "learning_rate": 7.022192612776899e-05, "loss": 0.6531, "step": 11516 }, { "epoch": 0.772994194825677, "grad_norm": 0.9783521890640259, "learning_rate": 7.021198587543012e-05, "loss": 0.5521, "step": 11518 }, { "epoch": 0.773128418509446, "grad_norm": 1.0634944438934326, "learning_rate": 7.020204466811467e-05, "loss": 0.6848, "step": 11520 }, { "epoch": 0.7732626421932149, "grad_norm": 1.15195631980896, "learning_rate": 7.019210250629231e-05, "loss": 0.5593, "step": 11522 }, { "epoch": 0.773396865876984, "grad_norm": 1.0139727592468262, "learning_rate": 7.018215939043285e-05, "loss": 0.653, "step": 11524 }, { "epoch": 0.773531089560753, "grad_norm": 1.0244086980819702, "learning_rate": 7.017221532100601e-05, "loss": 0.5606, "step": 11526 }, { "epoch": 0.773665313244522, "grad_norm": 1.0112557411193848, "learning_rate": 7.016227029848169e-05, "loss": 0.5724, "step": 11528 }, { "epoch": 0.773799536928291, "grad_norm": 1.0652763843536377, "learning_rate": 7.015232432332974e-05, "loss": 0.5757, "step": 11530 }, { "epoch": 0.77393376061206, "grad_norm": 1.005967617034912, "learning_rate": 7.014237739602008e-05, "loss": 0.5454, "step": 11532 }, { "epoch": 0.774067984295829, "grad_norm": 1.002595067024231, "learning_rate": 7.01324295170227e-05, "loss": 0.582, "step": 11534 }, { "epoch": 0.774202207979598, "grad_norm": 0.9543070197105408, "learning_rate": 7.012248068680762e-05, "loss": 0.5883, "step": 11536 }, { "epoch": 0.774336431663367, "grad_norm": 1.0744119882583618, "learning_rate": 7.01125309058449e-05, "loss": 0.6088, "step": 11538 }, { "epoch": 0.774470655347136, "grad_norm": 2.2598273754119873, "learning_rate": 7.010258017460463e-05, "loss": 0.601, "step": 11540 }, { "epoch": 0.774604879030905, "grad_norm": 0.9988208413124084, "learning_rate": 7.0092628493557e-05, "loss": 0.6396, "step": 11542 }, { "epoch": 0.774739102714674, "grad_norm": 1.0017659664154053, "learning_rate": 7.008267586317216e-05, "loss": 0.6025, "step": 11544 }, { "epoch": 0.774873326398443, "grad_norm": 1.0925278663635254, "learning_rate": 7.007272228392039e-05, "loss": 0.6301, "step": 11546 }, { "epoch": 0.775007550082212, "grad_norm": 1.0574859380722046, "learning_rate": 7.006276775627196e-05, "loss": 0.6225, "step": 11548 }, { "epoch": 0.775141773765981, "grad_norm": 1.3055510520935059, "learning_rate": 7.005281228069721e-05, "loss": 0.5825, "step": 11550 }, { "epoch": 0.77527599744975, "grad_norm": 1.0374795198440552, "learning_rate": 7.004285585766651e-05, "loss": 0.6058, "step": 11552 }, { "epoch": 0.775410221133519, "grad_norm": 1.1046421527862549, "learning_rate": 7.003289848765028e-05, "loss": 0.5908, "step": 11554 }, { "epoch": 0.775544444817288, "grad_norm": 1.1111980676651, "learning_rate": 7.002294017111899e-05, "loss": 0.5679, "step": 11556 }, { "epoch": 0.775678668501057, "grad_norm": 1.09193754196167, "learning_rate": 7.001298090854316e-05, "loss": 0.6047, "step": 11558 }, { "epoch": 0.775812892184826, "grad_norm": 0.9132357835769653, "learning_rate": 7.000302070039332e-05, "loss": 0.5793, "step": 11560 }, { "epoch": 0.775947115868595, "grad_norm": 1.1297905445098877, "learning_rate": 6.999305954714009e-05, "loss": 0.6307, "step": 11562 }, { "epoch": 0.776081339552364, "grad_norm": 1.0000507831573486, "learning_rate": 6.998309744925411e-05, "loss": 0.6287, "step": 11564 }, { "epoch": 0.776215563236133, "grad_norm": 1.1495134830474854, "learning_rate": 6.997313440720608e-05, "loss": 0.6487, "step": 11566 }, { "epoch": 0.776349786919902, "grad_norm": 0.9105957746505737, "learning_rate": 6.996317042146671e-05, "loss": 0.5489, "step": 11568 }, { "epoch": 0.776484010603671, "grad_norm": 0.9797573685646057, "learning_rate": 6.995320549250681e-05, "loss": 0.586, "step": 11570 }, { "epoch": 0.77661823428744, "grad_norm": 1.1910641193389893, "learning_rate": 6.99432396207972e-05, "loss": 0.5437, "step": 11572 }, { "epoch": 0.7767524579712091, "grad_norm": 1.0212557315826416, "learning_rate": 6.99332728068087e-05, "loss": 0.5356, "step": 11574 }, { "epoch": 0.776886681654978, "grad_norm": 1.0056376457214355, "learning_rate": 6.992330505101228e-05, "loss": 0.5565, "step": 11576 }, { "epoch": 0.777020905338747, "grad_norm": 0.9811734557151794, "learning_rate": 6.991333635387886e-05, "loss": 0.6, "step": 11578 }, { "epoch": 0.777155129022516, "grad_norm": 0.9935258030891418, "learning_rate": 6.990336671587946e-05, "loss": 0.6373, "step": 11580 }, { "epoch": 0.7772893527062851, "grad_norm": 1.10719895362854, "learning_rate": 6.989339613748512e-05, "loss": 0.6113, "step": 11582 }, { "epoch": 0.777423576390054, "grad_norm": 0.9676934480667114, "learning_rate": 6.988342461916693e-05, "loss": 0.589, "step": 11584 }, { "epoch": 0.777557800073823, "grad_norm": 1.0333850383758545, "learning_rate": 6.987345216139604e-05, "loss": 0.5752, "step": 11586 }, { "epoch": 0.777692023757592, "grad_norm": 1.1050474643707275, "learning_rate": 6.98634787646436e-05, "loss": 0.615, "step": 11588 }, { "epoch": 0.777826247441361, "grad_norm": 1.0079107284545898, "learning_rate": 6.985350442938084e-05, "loss": 0.5257, "step": 11590 }, { "epoch": 0.77796047112513, "grad_norm": 1.1853986978530884, "learning_rate": 6.984352915607906e-05, "loss": 0.576, "step": 11592 }, { "epoch": 0.778094694808899, "grad_norm": 0.8857834935188293, "learning_rate": 6.983355294520952e-05, "loss": 0.5174, "step": 11594 }, { "epoch": 0.7782289184926681, "grad_norm": 1.2821314334869385, "learning_rate": 6.982357579724364e-05, "loss": 0.5895, "step": 11596 }, { "epoch": 0.778363142176437, "grad_norm": 0.8643916249275208, "learning_rate": 6.981359771265276e-05, "loss": 0.5334, "step": 11598 }, { "epoch": 0.778497365860206, "grad_norm": 1.3218395709991455, "learning_rate": 6.980361869190836e-05, "loss": 0.6501, "step": 11600 }, { "epoch": 0.778631589543975, "grad_norm": 1.0390161275863647, "learning_rate": 6.97936387354819e-05, "loss": 0.531, "step": 11602 }, { "epoch": 0.7787658132277441, "grad_norm": 0.9862406253814697, "learning_rate": 6.978365784384494e-05, "loss": 0.5516, "step": 11604 }, { "epoch": 0.778900036911513, "grad_norm": 1.0138154029846191, "learning_rate": 6.977367601746907e-05, "loss": 0.6473, "step": 11606 }, { "epoch": 0.779034260595282, "grad_norm": 0.956520676612854, "learning_rate": 6.976369325682586e-05, "loss": 0.568, "step": 11608 }, { "epoch": 0.779168484279051, "grad_norm": 1.2190860509872437, "learning_rate": 6.975370956238703e-05, "loss": 0.5528, "step": 11610 }, { "epoch": 0.7793027079628201, "grad_norm": 1.0392407178878784, "learning_rate": 6.974372493462427e-05, "loss": 0.6182, "step": 11612 }, { "epoch": 0.779436931646589, "grad_norm": 1.3222901821136475, "learning_rate": 6.973373937400932e-05, "loss": 0.5887, "step": 11614 }, { "epoch": 0.779571155330358, "grad_norm": 1.0793002843856812, "learning_rate": 6.9723752881014e-05, "loss": 0.598, "step": 11616 }, { "epoch": 0.7797053790141271, "grad_norm": 1.046127200126648, "learning_rate": 6.971376545611012e-05, "loss": 0.5966, "step": 11618 }, { "epoch": 0.7798396026978961, "grad_norm": 0.9392060041427612, "learning_rate": 6.97037770997696e-05, "loss": 0.568, "step": 11620 }, { "epoch": 0.779973826381665, "grad_norm": 1.0161939859390259, "learning_rate": 6.969378781246436e-05, "loss": 0.5838, "step": 11622 }, { "epoch": 0.780108050065434, "grad_norm": 1.3568873405456543, "learning_rate": 6.968379759466638e-05, "loss": 0.6808, "step": 11624 }, { "epoch": 0.7802422737492031, "grad_norm": 1.0033005475997925, "learning_rate": 6.967380644684765e-05, "loss": 0.5981, "step": 11626 }, { "epoch": 0.780376497432972, "grad_norm": 1.0781974792480469, "learning_rate": 6.966381436948027e-05, "loss": 0.6977, "step": 11628 }, { "epoch": 0.780510721116741, "grad_norm": 1.0803115367889404, "learning_rate": 6.965382136303632e-05, "loss": 0.66, "step": 11630 }, { "epoch": 0.78064494480051, "grad_norm": 1.0441747903823853, "learning_rate": 6.964382742798797e-05, "loss": 0.587, "step": 11632 }, { "epoch": 0.7807791684842791, "grad_norm": 1.0430234670639038, "learning_rate": 6.963383256480738e-05, "loss": 0.5442, "step": 11634 }, { "epoch": 0.780913392168048, "grad_norm": 0.972874641418457, "learning_rate": 6.962383677396682e-05, "loss": 0.6236, "step": 11636 }, { "epoch": 0.781047615851817, "grad_norm": 1.1453553438186646, "learning_rate": 6.961384005593856e-05, "loss": 0.6472, "step": 11638 }, { "epoch": 0.7811818395355861, "grad_norm": 1.0501232147216797, "learning_rate": 6.960384241119494e-05, "loss": 0.5658, "step": 11640 }, { "epoch": 0.7813160632193551, "grad_norm": 1.000161051750183, "learning_rate": 6.95938438402083e-05, "loss": 0.5674, "step": 11642 }, { "epoch": 0.781450286903124, "grad_norm": 1.0791254043579102, "learning_rate": 6.958384434345107e-05, "loss": 0.6232, "step": 11644 }, { "epoch": 0.781584510586893, "grad_norm": 0.9885473847389221, "learning_rate": 6.95738439213957e-05, "loss": 0.6215, "step": 11646 }, { "epoch": 0.7817187342706621, "grad_norm": 0.9427973031997681, "learning_rate": 6.956384257451471e-05, "loss": 0.5575, "step": 11648 }, { "epoch": 0.7818529579544311, "grad_norm": 0.9807668924331665, "learning_rate": 6.955384030328063e-05, "loss": 0.6265, "step": 11650 }, { "epoch": 0.7819871816382, "grad_norm": 1.0430842638015747, "learning_rate": 6.954383710816604e-05, "loss": 0.5747, "step": 11652 }, { "epoch": 0.782121405321969, "grad_norm": 1.0898436307907104, "learning_rate": 6.953383298964357e-05, "loss": 0.617, "step": 11654 }, { "epoch": 0.7822556290057381, "grad_norm": 1.0856739282608032, "learning_rate": 6.95238279481859e-05, "loss": 0.578, "step": 11656 }, { "epoch": 0.7823898526895071, "grad_norm": 1.0635261535644531, "learning_rate": 6.951382198426577e-05, "loss": 0.6226, "step": 11658 }, { "epoch": 0.782524076373276, "grad_norm": 1.0395147800445557, "learning_rate": 6.95038150983559e-05, "loss": 0.5009, "step": 11660 }, { "epoch": 0.7826583000570451, "grad_norm": 1.0393999814987183, "learning_rate": 6.949380729092914e-05, "loss": 0.5967, "step": 11662 }, { "epoch": 0.7827925237408141, "grad_norm": 0.9801155924797058, "learning_rate": 6.948379856245832e-05, "loss": 0.5237, "step": 11664 }, { "epoch": 0.782926747424583, "grad_norm": 1.1089324951171875, "learning_rate": 6.947378891341631e-05, "loss": 0.5307, "step": 11666 }, { "epoch": 0.783060971108352, "grad_norm": 0.8968009352684021, "learning_rate": 6.946377834427608e-05, "loss": 0.5496, "step": 11668 }, { "epoch": 0.7831951947921211, "grad_norm": 1.186698317527771, "learning_rate": 6.945376685551061e-05, "loss": 0.6179, "step": 11670 }, { "epoch": 0.7833294184758901, "grad_norm": 1.0249183177947998, "learning_rate": 6.94437544475929e-05, "loss": 0.6164, "step": 11672 }, { "epoch": 0.783463642159659, "grad_norm": 0.9646841883659363, "learning_rate": 6.9433741120996e-05, "loss": 0.5585, "step": 11674 }, { "epoch": 0.783597865843428, "grad_norm": 0.9436026215553284, "learning_rate": 6.94237268761931e-05, "loss": 0.5165, "step": 11676 }, { "epoch": 0.7837320895271971, "grad_norm": 1.168487310409546, "learning_rate": 6.941371171365725e-05, "loss": 0.605, "step": 11678 }, { "epoch": 0.7838663132109661, "grad_norm": 1.008285403251648, "learning_rate": 6.940369563386172e-05, "loss": 0.5822, "step": 11680 }, { "epoch": 0.784000536894735, "grad_norm": 1.0759882926940918, "learning_rate": 6.939367863727973e-05, "loss": 0.6026, "step": 11682 }, { "epoch": 0.7841347605785041, "grad_norm": 1.0713379383087158, "learning_rate": 6.938366072438456e-05, "loss": 0.6155, "step": 11684 }, { "epoch": 0.7842689842622731, "grad_norm": 1.0884873867034912, "learning_rate": 6.937364189564954e-05, "loss": 0.6289, "step": 11686 }, { "epoch": 0.7844032079460421, "grad_norm": 0.9686011672019958, "learning_rate": 6.936362215154802e-05, "loss": 0.6258, "step": 11688 }, { "epoch": 0.784537431629811, "grad_norm": 1.0788530111312866, "learning_rate": 6.935360149255345e-05, "loss": 0.587, "step": 11690 }, { "epoch": 0.7846716553135801, "grad_norm": 0.9957976937294006, "learning_rate": 6.934357991913924e-05, "loss": 0.6245, "step": 11692 }, { "epoch": 0.7848058789973491, "grad_norm": 1.063419222831726, "learning_rate": 6.933355743177894e-05, "loss": 0.6002, "step": 11694 }, { "epoch": 0.7849401026811181, "grad_norm": 1.0071430206298828, "learning_rate": 6.932353403094605e-05, "loss": 0.593, "step": 11696 }, { "epoch": 0.785074326364887, "grad_norm": 1.0671712160110474, "learning_rate": 6.931350971711418e-05, "loss": 0.5449, "step": 11698 }, { "epoch": 0.7852085500486561, "grad_norm": 1.0007480382919312, "learning_rate": 6.930348449075699e-05, "loss": 0.6136, "step": 11700 }, { "epoch": 0.7853427737324251, "grad_norm": 1.0614603757858276, "learning_rate": 6.929345835234808e-05, "loss": 0.6491, "step": 11702 }, { "epoch": 0.785476997416194, "grad_norm": 0.9814666509628296, "learning_rate": 6.928343130236121e-05, "loss": 0.5912, "step": 11704 }, { "epoch": 0.7856112210999631, "grad_norm": 1.0447053909301758, "learning_rate": 6.927340334127013e-05, "loss": 0.6188, "step": 11706 }, { "epoch": 0.7857454447837321, "grad_norm": 1.0697200298309326, "learning_rate": 6.926337446954864e-05, "loss": 0.5908, "step": 11708 }, { "epoch": 0.7858796684675011, "grad_norm": 1.0311012268066406, "learning_rate": 6.92533446876706e-05, "loss": 0.6419, "step": 11710 }, { "epoch": 0.78601389215127, "grad_norm": 1.0517666339874268, "learning_rate": 6.924331399610986e-05, "loss": 0.6235, "step": 11712 }, { "epoch": 0.7861481158350391, "grad_norm": 1.0834307670593262, "learning_rate": 6.92332823953404e-05, "loss": 0.5855, "step": 11714 }, { "epoch": 0.7862823395188081, "grad_norm": 0.8779531717300415, "learning_rate": 6.922324988583616e-05, "loss": 0.498, "step": 11716 }, { "epoch": 0.7864165632025771, "grad_norm": 1.1895356178283691, "learning_rate": 6.921321646807113e-05, "loss": 0.5921, "step": 11718 }, { "epoch": 0.786550786886346, "grad_norm": 1.094163417816162, "learning_rate": 6.920318214251945e-05, "loss": 0.5684, "step": 11720 }, { "epoch": 0.7866850105701151, "grad_norm": 1.262536644935608, "learning_rate": 6.919314690965514e-05, "loss": 0.5357, "step": 11722 }, { "epoch": 0.7868192342538841, "grad_norm": 1.0477513074874878, "learning_rate": 6.91831107699524e-05, "loss": 0.5262, "step": 11724 }, { "epoch": 0.7869534579376531, "grad_norm": 0.9876933097839355, "learning_rate": 6.917307372388539e-05, "loss": 0.5662, "step": 11726 }, { "epoch": 0.7870876816214221, "grad_norm": 0.8962420225143433, "learning_rate": 6.916303577192835e-05, "loss": 0.5427, "step": 11728 }, { "epoch": 0.7872219053051911, "grad_norm": 0.9462629556655884, "learning_rate": 6.915299691455555e-05, "loss": 0.6347, "step": 11730 }, { "epoch": 0.7873561289889601, "grad_norm": 0.998128354549408, "learning_rate": 6.914295715224132e-05, "loss": 0.5965, "step": 11732 }, { "epoch": 0.7874903526727292, "grad_norm": 1.0247682332992554, "learning_rate": 6.913291648546001e-05, "loss": 0.611, "step": 11734 }, { "epoch": 0.7876245763564981, "grad_norm": 0.9905968308448792, "learning_rate": 6.9122874914686e-05, "loss": 0.5954, "step": 11736 }, { "epoch": 0.7877588000402671, "grad_norm": 1.0264225006103516, "learning_rate": 6.911283244039377e-05, "loss": 0.5826, "step": 11738 }, { "epoch": 0.7878930237240361, "grad_norm": 0.9632094502449036, "learning_rate": 6.910278906305778e-05, "loss": 0.5702, "step": 11740 }, { "epoch": 0.788027247407805, "grad_norm": 1.104326605796814, "learning_rate": 6.909274478315257e-05, "loss": 0.5727, "step": 11742 }, { "epoch": 0.7881614710915741, "grad_norm": 0.9858546257019043, "learning_rate": 6.908269960115273e-05, "loss": 0.6141, "step": 11744 }, { "epoch": 0.7882956947753431, "grad_norm": 0.9897134900093079, "learning_rate": 6.907265351753283e-05, "loss": 0.5913, "step": 11746 }, { "epoch": 0.7884299184591121, "grad_norm": 1.038590431213379, "learning_rate": 6.906260653276758e-05, "loss": 0.6159, "step": 11748 }, { "epoch": 0.7885641421428811, "grad_norm": 1.0222185850143433, "learning_rate": 6.905255864733164e-05, "loss": 0.5529, "step": 11750 }, { "epoch": 0.7886983658266501, "grad_norm": 1.1252750158309937, "learning_rate": 6.90425098616998e-05, "loss": 0.5685, "step": 11752 }, { "epoch": 0.7888325895104191, "grad_norm": 0.9735110998153687, "learning_rate": 6.903246017634677e-05, "loss": 0.5798, "step": 11754 }, { "epoch": 0.7889668131941882, "grad_norm": 1.099358320236206, "learning_rate": 6.902240959174745e-05, "loss": 0.6107, "step": 11756 }, { "epoch": 0.7891010368779571, "grad_norm": 1.0541526079177856, "learning_rate": 6.901235810837669e-05, "loss": 0.5444, "step": 11758 }, { "epoch": 0.7892352605617261, "grad_norm": 1.0630377531051636, "learning_rate": 6.900230572670938e-05, "loss": 0.6276, "step": 11760 }, { "epoch": 0.7893694842454951, "grad_norm": 1.0359013080596924, "learning_rate": 6.89922524472205e-05, "loss": 0.5934, "step": 11762 }, { "epoch": 0.7895037079292642, "grad_norm": 0.8728708028793335, "learning_rate": 6.898219827038503e-05, "loss": 0.5087, "step": 11764 }, { "epoch": 0.7896379316130331, "grad_norm": 0.9688023924827576, "learning_rate": 6.897214319667802e-05, "loss": 0.5695, "step": 11766 }, { "epoch": 0.7897721552968021, "grad_norm": 1.2484382390975952, "learning_rate": 6.896208722657455e-05, "loss": 0.6423, "step": 11768 }, { "epoch": 0.7899063789805711, "grad_norm": 0.9647979140281677, "learning_rate": 6.895203036054974e-05, "loss": 0.5779, "step": 11770 }, { "epoch": 0.7900406026643402, "grad_norm": 1.1346417665481567, "learning_rate": 6.894197259907879e-05, "loss": 0.5259, "step": 11772 }, { "epoch": 0.7901748263481091, "grad_norm": 1.1269654035568237, "learning_rate": 6.893191394263684e-05, "loss": 0.5567, "step": 11774 }, { "epoch": 0.7903090500318781, "grad_norm": 1.0334393978118896, "learning_rate": 6.892185439169922e-05, "loss": 0.6355, "step": 11776 }, { "epoch": 0.7904432737156472, "grad_norm": 1.088161826133728, "learning_rate": 6.891179394674119e-05, "loss": 0.5314, "step": 11778 }, { "epoch": 0.7905774973994161, "grad_norm": 1.1312185525894165, "learning_rate": 6.890173260823807e-05, "loss": 0.5725, "step": 11780 }, { "epoch": 0.7907117210831851, "grad_norm": 0.9361488819122314, "learning_rate": 6.889167037666525e-05, "loss": 0.5429, "step": 11782 }, { "epoch": 0.7908459447669541, "grad_norm": 1.098042607307434, "learning_rate": 6.888160725249816e-05, "loss": 0.6652, "step": 11784 }, { "epoch": 0.7909801684507232, "grad_norm": 1.0583075284957886, "learning_rate": 6.887154323621225e-05, "loss": 0.6155, "step": 11786 }, { "epoch": 0.7911143921344921, "grad_norm": 0.8496408462524414, "learning_rate": 6.886147832828303e-05, "loss": 0.6059, "step": 11788 }, { "epoch": 0.7912486158182611, "grad_norm": 1.0364539623260498, "learning_rate": 6.885141252918607e-05, "loss": 0.6223, "step": 11790 }, { "epoch": 0.7913828395020301, "grad_norm": 0.9641837477684021, "learning_rate": 6.884134583939692e-05, "loss": 0.6216, "step": 11792 }, { "epoch": 0.7915170631857992, "grad_norm": 1.2971800565719604, "learning_rate": 6.883127825939122e-05, "loss": 0.6559, "step": 11794 }, { "epoch": 0.7916512868695681, "grad_norm": 1.0435941219329834, "learning_rate": 6.882120978964466e-05, "loss": 0.5561, "step": 11796 }, { "epoch": 0.7917855105533371, "grad_norm": 0.9846218824386597, "learning_rate": 6.881114043063296e-05, "loss": 0.612, "step": 11798 }, { "epoch": 0.7919197342371062, "grad_norm": 1.1445139646530151, "learning_rate": 6.880107018283186e-05, "loss": 0.5746, "step": 11800 }, { "epoch": 0.7920539579208752, "grad_norm": 1.0392295122146606, "learning_rate": 6.879099904671715e-05, "loss": 0.5904, "step": 11802 }, { "epoch": 0.7921881816046441, "grad_norm": 0.9261115789413452, "learning_rate": 6.87809270227647e-05, "loss": 0.4898, "step": 11804 }, { "epoch": 0.7923224052884131, "grad_norm": 1.0234037637710571, "learning_rate": 6.877085411145038e-05, "loss": 0.5977, "step": 11806 }, { "epoch": 0.7924566289721822, "grad_norm": 1.1201796531677246, "learning_rate": 6.87607803132501e-05, "loss": 0.5648, "step": 11808 }, { "epoch": 0.7925908526559512, "grad_norm": 1.0571445226669312, "learning_rate": 6.875070562863986e-05, "loss": 0.6202, "step": 11810 }, { "epoch": 0.7927250763397201, "grad_norm": 1.1081063747406006, "learning_rate": 6.874063005809563e-05, "loss": 0.56, "step": 11812 }, { "epoch": 0.7928593000234891, "grad_norm": 1.0794700384140015, "learning_rate": 6.87305536020935e-05, "loss": 0.5982, "step": 11814 }, { "epoch": 0.7929935237072582, "grad_norm": 1.0285453796386719, "learning_rate": 6.872047626110955e-05, "loss": 0.5858, "step": 11816 }, { "epoch": 0.7931277473910271, "grad_norm": 1.135297417640686, "learning_rate": 6.87103980356199e-05, "loss": 0.5713, "step": 11818 }, { "epoch": 0.7932619710747961, "grad_norm": 1.0952262878417969, "learning_rate": 6.870031892610073e-05, "loss": 0.5671, "step": 11820 }, { "epoch": 0.7933961947585652, "grad_norm": 1.0698264837265015, "learning_rate": 6.869023893302826e-05, "loss": 0.5819, "step": 11822 }, { "epoch": 0.7935304184423342, "grad_norm": 0.984652042388916, "learning_rate": 6.868015805687877e-05, "loss": 0.5566, "step": 11824 }, { "epoch": 0.7936646421261031, "grad_norm": 1.0963594913482666, "learning_rate": 6.867007629812852e-05, "loss": 0.7261, "step": 11826 }, { "epoch": 0.7937988658098721, "grad_norm": 0.9263544678688049, "learning_rate": 6.865999365725391e-05, "loss": 0.576, "step": 11828 }, { "epoch": 0.7939330894936412, "grad_norm": 1.1272345781326294, "learning_rate": 6.864991013473125e-05, "loss": 0.5807, "step": 11830 }, { "epoch": 0.7940673131774102, "grad_norm": 0.939246416091919, "learning_rate": 6.863982573103704e-05, "loss": 0.5666, "step": 11832 }, { "epoch": 0.7942015368611791, "grad_norm": 1.0075976848602295, "learning_rate": 6.862974044664772e-05, "loss": 0.5334, "step": 11834 }, { "epoch": 0.7943357605449481, "grad_norm": 1.0745658874511719, "learning_rate": 6.861965428203978e-05, "loss": 0.603, "step": 11836 }, { "epoch": 0.7944699842287172, "grad_norm": 1.0120166540145874, "learning_rate": 6.860956723768981e-05, "loss": 0.6386, "step": 11838 }, { "epoch": 0.7946042079124862, "grad_norm": 0.9060064554214478, "learning_rate": 6.859947931407436e-05, "loss": 0.5112, "step": 11840 }, { "epoch": 0.7947384315962551, "grad_norm": 0.973538339138031, "learning_rate": 6.858939051167011e-05, "loss": 0.6096, "step": 11842 }, { "epoch": 0.7948726552800242, "grad_norm": 1.0098037719726562, "learning_rate": 6.85793008309537e-05, "loss": 0.624, "step": 11844 }, { "epoch": 0.7950068789637932, "grad_norm": 1.023622751235962, "learning_rate": 6.856921027240187e-05, "loss": 0.5849, "step": 11846 }, { "epoch": 0.7951411026475622, "grad_norm": 1.0303845405578613, "learning_rate": 6.855911883649137e-05, "loss": 0.616, "step": 11848 }, { "epoch": 0.7952753263313311, "grad_norm": 1.1904534101486206, "learning_rate": 6.854902652369898e-05, "loss": 0.7265, "step": 11850 }, { "epoch": 0.7954095500151002, "grad_norm": 0.9215459823608398, "learning_rate": 6.853893333450158e-05, "loss": 0.5722, "step": 11852 }, { "epoch": 0.7955437736988692, "grad_norm": 0.9743821620941162, "learning_rate": 6.852883926937602e-05, "loss": 0.608, "step": 11854 }, { "epoch": 0.7956779973826381, "grad_norm": 1.222190260887146, "learning_rate": 6.851874432879925e-05, "loss": 0.6626, "step": 11856 }, { "epoch": 0.7958122210664071, "grad_norm": 1.0577031373977661, "learning_rate": 6.850864851324823e-05, "loss": 0.5342, "step": 11858 }, { "epoch": 0.7959464447501762, "grad_norm": 0.9763243794441223, "learning_rate": 6.849855182319995e-05, "loss": 0.6014, "step": 11860 }, { "epoch": 0.7960806684339452, "grad_norm": 1.0097098350524902, "learning_rate": 6.848845425913149e-05, "loss": 0.6697, "step": 11862 }, { "epoch": 0.7962148921177141, "grad_norm": 1.0537142753601074, "learning_rate": 6.84783558215199e-05, "loss": 0.6646, "step": 11864 }, { "epoch": 0.7963491158014832, "grad_norm": 1.0196220874786377, "learning_rate": 6.846825651084236e-05, "loss": 0.6225, "step": 11866 }, { "epoch": 0.7964833394852522, "grad_norm": 0.8699311017990112, "learning_rate": 6.8458156327576e-05, "loss": 0.5719, "step": 11868 }, { "epoch": 0.7966175631690212, "grad_norm": 0.9252583980560303, "learning_rate": 6.844805527219804e-05, "loss": 0.5961, "step": 11870 }, { "epoch": 0.7967517868527901, "grad_norm": 1.1307587623596191, "learning_rate": 6.843795334518576e-05, "loss": 0.5956, "step": 11872 }, { "epoch": 0.7968860105365592, "grad_norm": 0.9717182517051697, "learning_rate": 6.842785054701643e-05, "loss": 0.6711, "step": 11874 }, { "epoch": 0.7970202342203282, "grad_norm": 1.0717628002166748, "learning_rate": 6.84177468781674e-05, "loss": 0.5864, "step": 11876 }, { "epoch": 0.7971544579040972, "grad_norm": 1.231685996055603, "learning_rate": 6.840764233911606e-05, "loss": 0.5761, "step": 11878 }, { "epoch": 0.7972886815878661, "grad_norm": 0.9946985840797424, "learning_rate": 6.83975369303398e-05, "loss": 0.5445, "step": 11880 }, { "epoch": 0.7974229052716352, "grad_norm": 1.0769169330596924, "learning_rate": 6.838743065231612e-05, "loss": 0.649, "step": 11882 }, { "epoch": 0.7975571289554042, "grad_norm": 0.9944827556610107, "learning_rate": 6.837732350552249e-05, "loss": 0.5963, "step": 11884 }, { "epoch": 0.7976913526391732, "grad_norm": 1.077628254890442, "learning_rate": 6.836721549043645e-05, "loss": 0.5909, "step": 11886 }, { "epoch": 0.7978255763229422, "grad_norm": 1.2306236028671265, "learning_rate": 6.835710660753561e-05, "loss": 0.5751, "step": 11888 }, { "epoch": 0.7979598000067112, "grad_norm": 0.972992479801178, "learning_rate": 6.834699685729757e-05, "loss": 0.5889, "step": 11890 }, { "epoch": 0.7980940236904802, "grad_norm": 0.9339223504066467, "learning_rate": 6.83368862402e-05, "loss": 0.5542, "step": 11892 }, { "epoch": 0.7982282473742491, "grad_norm": 1.0817885398864746, "learning_rate": 6.832677475672063e-05, "loss": 0.6131, "step": 11894 }, { "epoch": 0.7983624710580182, "grad_norm": 1.1673030853271484, "learning_rate": 6.831666240733718e-05, "loss": 0.6095, "step": 11896 }, { "epoch": 0.7984966947417872, "grad_norm": 1.21226966381073, "learning_rate": 6.830654919252745e-05, "loss": 0.6406, "step": 11898 }, { "epoch": 0.7986309184255562, "grad_norm": 1.0874611139297485, "learning_rate": 6.829643511276929e-05, "loss": 0.6048, "step": 11900 }, { "epoch": 0.7987651421093251, "grad_norm": 0.9976110458374023, "learning_rate": 6.828632016854051e-05, "loss": 0.5256, "step": 11902 }, { "epoch": 0.7988993657930942, "grad_norm": 0.9796349406242371, "learning_rate": 6.82762043603191e-05, "loss": 0.5838, "step": 11904 }, { "epoch": 0.7990335894768632, "grad_norm": 0.9170714616775513, "learning_rate": 6.826608768858294e-05, "loss": 0.5912, "step": 11906 }, { "epoch": 0.7991678131606322, "grad_norm": 1.0391581058502197, "learning_rate": 6.825597015381007e-05, "loss": 0.5676, "step": 11908 }, { "epoch": 0.7993020368444012, "grad_norm": 1.0441548824310303, "learning_rate": 6.824585175647852e-05, "loss": 0.6165, "step": 11910 }, { "epoch": 0.7994362605281702, "grad_norm": 0.9074247479438782, "learning_rate": 6.823573249706634e-05, "loss": 0.5816, "step": 11912 }, { "epoch": 0.7995704842119392, "grad_norm": 1.03797447681427, "learning_rate": 6.822561237605167e-05, "loss": 0.5578, "step": 11914 }, { "epoch": 0.7997047078957082, "grad_norm": 1.083938717842102, "learning_rate": 6.821549139391264e-05, "loss": 0.6759, "step": 11916 }, { "epoch": 0.7998389315794772, "grad_norm": 0.8666544556617737, "learning_rate": 6.820536955112747e-05, "loss": 0.5735, "step": 11918 }, { "epoch": 0.7999731552632462, "grad_norm": 1.1249221563339233, "learning_rate": 6.819524684817438e-05, "loss": 0.6596, "step": 11920 }, { "epoch": 0.8001073789470152, "grad_norm": 1.0319015979766846, "learning_rate": 6.818512328553166e-05, "loss": 0.6022, "step": 11922 }, { "epoch": 0.8002416026307843, "grad_norm": 1.0916059017181396, "learning_rate": 6.817499886367763e-05, "loss": 0.5961, "step": 11924 }, { "epoch": 0.8003758263145532, "grad_norm": 1.0699912309646606, "learning_rate": 6.816487358309064e-05, "loss": 0.6445, "step": 11926 }, { "epoch": 0.8005100499983222, "grad_norm": 1.6955249309539795, "learning_rate": 6.815474744424908e-05, "loss": 0.5892, "step": 11928 }, { "epoch": 0.8006442736820912, "grad_norm": 1.1446768045425415, "learning_rate": 6.814462044763143e-05, "loss": 0.6042, "step": 11930 }, { "epoch": 0.8007784973658602, "grad_norm": 0.95870441198349, "learning_rate": 6.813449259371611e-05, "loss": 0.6115, "step": 11932 }, { "epoch": 0.8009127210496292, "grad_norm": 1.0400450229644775, "learning_rate": 6.81243638829817e-05, "loss": 0.6142, "step": 11934 }, { "epoch": 0.8010469447333982, "grad_norm": 0.9809623956680298, "learning_rate": 6.811423431590672e-05, "loss": 0.6326, "step": 11936 }, { "epoch": 0.8011811684171672, "grad_norm": 1.0043615102767944, "learning_rate": 6.81041038929698e-05, "loss": 0.5786, "step": 11938 }, { "epoch": 0.8013153921009362, "grad_norm": 1.0907158851623535, "learning_rate": 6.809397261464957e-05, "loss": 0.5435, "step": 11940 }, { "epoch": 0.8014496157847052, "grad_norm": 1.8366481065750122, "learning_rate": 6.808384048142472e-05, "loss": 0.5893, "step": 11942 }, { "epoch": 0.8015838394684742, "grad_norm": 0.9586851596832275, "learning_rate": 6.807370749377396e-05, "loss": 0.5757, "step": 11944 }, { "epoch": 0.8017180631522433, "grad_norm": 0.9573516845703125, "learning_rate": 6.806357365217606e-05, "loss": 0.5899, "step": 11946 }, { "epoch": 0.8018522868360122, "grad_norm": 1.194173812866211, "learning_rate": 6.805343895710983e-05, "loss": 0.5861, "step": 11948 }, { "epoch": 0.8019865105197812, "grad_norm": 0.9803144931793213, "learning_rate": 6.80433034090541e-05, "loss": 0.5297, "step": 11950 }, { "epoch": 0.8021207342035502, "grad_norm": 1.0402456521987915, "learning_rate": 6.803316700848779e-05, "loss": 0.6527, "step": 11952 }, { "epoch": 0.8022549578873193, "grad_norm": 1.0439958572387695, "learning_rate": 6.802302975588976e-05, "loss": 0.5892, "step": 11954 }, { "epoch": 0.8023891815710882, "grad_norm": 1.0565516948699951, "learning_rate": 6.801289165173905e-05, "loss": 0.5179, "step": 11956 }, { "epoch": 0.8025234052548572, "grad_norm": 1.2166541814804077, "learning_rate": 6.800275269651462e-05, "loss": 0.7249, "step": 11958 }, { "epoch": 0.8026576289386262, "grad_norm": 1.0121784210205078, "learning_rate": 6.79926128906955e-05, "loss": 0.5717, "step": 11960 }, { "epoch": 0.8027918526223953, "grad_norm": 1.14600670337677, "learning_rate": 6.798247223476084e-05, "loss": 0.6073, "step": 11962 }, { "epoch": 0.8029260763061642, "grad_norm": 1.074245572090149, "learning_rate": 6.79723307291897e-05, "loss": 0.6504, "step": 11964 }, { "epoch": 0.8030602999899332, "grad_norm": 1.091021180152893, "learning_rate": 6.79621883744613e-05, "loss": 0.5784, "step": 11966 }, { "epoch": 0.8031945236737023, "grad_norm": 0.9298738837242126, "learning_rate": 6.79520451710548e-05, "loss": 0.5882, "step": 11968 }, { "epoch": 0.8033287473574712, "grad_norm": 1.0967477560043335, "learning_rate": 6.794190111944948e-05, "loss": 0.564, "step": 11970 }, { "epoch": 0.8034629710412402, "grad_norm": 0.9110896587371826, "learning_rate": 6.79317562201246e-05, "loss": 0.5958, "step": 11972 }, { "epoch": 0.8035971947250092, "grad_norm": 1.0178650617599487, "learning_rate": 6.792161047355951e-05, "loss": 0.577, "step": 11974 }, { "epoch": 0.8037314184087783, "grad_norm": 1.421694040298462, "learning_rate": 6.791146388023356e-05, "loss": 0.6112, "step": 11976 }, { "epoch": 0.8038656420925472, "grad_norm": 1.0152606964111328, "learning_rate": 6.790131644062616e-05, "loss": 0.6533, "step": 11978 }, { "epoch": 0.8039998657763162, "grad_norm": 0.9436149001121521, "learning_rate": 6.789116815521678e-05, "loss": 0.6309, "step": 11980 }, { "epoch": 0.8041340894600852, "grad_norm": 1.0372847318649292, "learning_rate": 6.788101902448486e-05, "loss": 0.519, "step": 11982 }, { "epoch": 0.8042683131438543, "grad_norm": 0.9170295000076294, "learning_rate": 6.787086904890998e-05, "loss": 0.5786, "step": 11984 }, { "epoch": 0.8044025368276232, "grad_norm": 1.0059316158294678, "learning_rate": 6.786071822897166e-05, "loss": 0.5635, "step": 11986 }, { "epoch": 0.8045367605113922, "grad_norm": 0.9609963893890381, "learning_rate": 6.785056656514953e-05, "loss": 0.584, "step": 11988 }, { "epoch": 0.8046709841951613, "grad_norm": 0.9653461575508118, "learning_rate": 6.784041405792324e-05, "loss": 0.5671, "step": 11990 }, { "epoch": 0.8048052078789303, "grad_norm": 0.8587554693222046, "learning_rate": 6.783026070777245e-05, "loss": 0.5294, "step": 11992 }, { "epoch": 0.8049394315626992, "grad_norm": 1.9127076864242554, "learning_rate": 6.782010651517691e-05, "loss": 0.5974, "step": 11994 }, { "epoch": 0.8050736552464682, "grad_norm": 1.0242037773132324, "learning_rate": 6.780995148061638e-05, "loss": 0.5632, "step": 11996 }, { "epoch": 0.8052078789302373, "grad_norm": 1.1092298030853271, "learning_rate": 6.779979560457066e-05, "loss": 0.6036, "step": 11998 }, { "epoch": 0.8053421026140063, "grad_norm": 1.0180758237838745, "learning_rate": 6.778963888751961e-05, "loss": 0.6109, "step": 12000 }, { "epoch": 0.8054763262977752, "grad_norm": 1.0010555982589722, "learning_rate": 6.77794813299431e-05, "loss": 0.5716, "step": 12002 }, { "epoch": 0.8056105499815442, "grad_norm": 1.098800539970398, "learning_rate": 6.776932293232106e-05, "loss": 0.552, "step": 12004 }, { "epoch": 0.8057447736653133, "grad_norm": 0.9850975275039673, "learning_rate": 6.775916369513344e-05, "loss": 0.629, "step": 12006 }, { "epoch": 0.8058789973490822, "grad_norm": 1.0020534992218018, "learning_rate": 6.774900361886028e-05, "loss": 0.5255, "step": 12008 }, { "epoch": 0.8060132210328512, "grad_norm": 1.052212119102478, "learning_rate": 6.773884270398158e-05, "loss": 0.6533, "step": 12010 }, { "epoch": 0.8061474447166203, "grad_norm": 1.0798852443695068, "learning_rate": 6.772868095097745e-05, "loss": 0.5716, "step": 12012 }, { "epoch": 0.8062816684003893, "grad_norm": 0.8810888528823853, "learning_rate": 6.771851836032801e-05, "loss": 0.5025, "step": 12014 }, { "epoch": 0.8064158920841582, "grad_norm": 1.1676989793777466, "learning_rate": 6.770835493251342e-05, "loss": 0.6193, "step": 12016 }, { "epoch": 0.8065501157679272, "grad_norm": 0.9496458172798157, "learning_rate": 6.769819066801388e-05, "loss": 0.5612, "step": 12018 }, { "epoch": 0.8066843394516963, "grad_norm": 0.988442063331604, "learning_rate": 6.768802556730964e-05, "loss": 0.5982, "step": 12020 }, { "epoch": 0.8068185631354653, "grad_norm": 1.0732218027114868, "learning_rate": 6.767785963088096e-05, "loss": 0.5754, "step": 12022 }, { "epoch": 0.8069527868192342, "grad_norm": 1.0232852697372437, "learning_rate": 6.766769285920819e-05, "loss": 0.605, "step": 12024 }, { "epoch": 0.8070870105030032, "grad_norm": 0.9692088961601257, "learning_rate": 6.765752525277168e-05, "loss": 0.5416, "step": 12026 }, { "epoch": 0.8072212341867723, "grad_norm": 0.9061393737792969, "learning_rate": 6.76473568120518e-05, "loss": 0.5389, "step": 12028 }, { "epoch": 0.8073554578705413, "grad_norm": 0.9491109848022461, "learning_rate": 6.763718753752901e-05, "loss": 0.5283, "step": 12030 }, { "epoch": 0.8074896815543102, "grad_norm": 1.0774158239364624, "learning_rate": 6.762701742968382e-05, "loss": 0.6744, "step": 12032 }, { "epoch": 0.8076239052380793, "grad_norm": 0.84043949842453, "learning_rate": 6.761684648899669e-05, "loss": 0.5901, "step": 12034 }, { "epoch": 0.8077581289218483, "grad_norm": 1.1024222373962402, "learning_rate": 6.760667471594821e-05, "loss": 0.6299, "step": 12036 }, { "epoch": 0.8078923526056173, "grad_norm": 1.4508450031280518, "learning_rate": 6.7596502111019e-05, "loss": 0.5974, "step": 12038 }, { "epoch": 0.8080265762893862, "grad_norm": 1.3109294176101685, "learning_rate": 6.758632867468964e-05, "loss": 0.5406, "step": 12040 }, { "epoch": 0.8081607999731553, "grad_norm": 1.1407095193862915, "learning_rate": 6.757615440744084e-05, "loss": 0.5489, "step": 12042 }, { "epoch": 0.8082950236569243, "grad_norm": 0.943513035774231, "learning_rate": 6.756597930975331e-05, "loss": 0.5673, "step": 12044 }, { "epoch": 0.8084292473406932, "grad_norm": 1.1445965766906738, "learning_rate": 6.75558033821078e-05, "loss": 0.5903, "step": 12046 }, { "epoch": 0.8085634710244622, "grad_norm": 1.021998643875122, "learning_rate": 6.754562662498509e-05, "loss": 0.5976, "step": 12048 }, { "epoch": 0.8086976947082313, "grad_norm": 1.0419927835464478, "learning_rate": 6.753544903886602e-05, "loss": 0.5785, "step": 12050 }, { "epoch": 0.8088319183920003, "grad_norm": 0.8495463728904724, "learning_rate": 6.75252706242315e-05, "loss": 0.5681, "step": 12052 }, { "epoch": 0.8089661420757692, "grad_norm": 0.9950043559074402, "learning_rate": 6.751509138156239e-05, "loss": 0.5758, "step": 12054 }, { "epoch": 0.8091003657595383, "grad_norm": 0.9346325397491455, "learning_rate": 6.750491131133962e-05, "loss": 0.6295, "step": 12056 }, { "epoch": 0.8092345894433073, "grad_norm": 1.0494962930679321, "learning_rate": 6.749473041404424e-05, "loss": 0.5382, "step": 12058 }, { "epoch": 0.8093688131270763, "grad_norm": 0.9695845246315002, "learning_rate": 6.748454869015725e-05, "loss": 0.5711, "step": 12060 }, { "epoch": 0.8095030368108452, "grad_norm": 1.1978754997253418, "learning_rate": 6.747436614015972e-05, "loss": 0.6299, "step": 12062 }, { "epoch": 0.8096372604946143, "grad_norm": 0.9886444807052612, "learning_rate": 6.746418276453275e-05, "loss": 0.5606, "step": 12064 }, { "epoch": 0.8097714841783833, "grad_norm": 1.180014729499817, "learning_rate": 6.745399856375749e-05, "loss": 0.6357, "step": 12066 }, { "epoch": 0.8099057078621523, "grad_norm": 0.8843987584114075, "learning_rate": 6.74438135383151e-05, "loss": 0.5641, "step": 12068 }, { "epoch": 0.8100399315459212, "grad_norm": 1.0571144819259644, "learning_rate": 6.743362768868682e-05, "loss": 0.607, "step": 12070 }, { "epoch": 0.8101741552296903, "grad_norm": 0.9392337799072266, "learning_rate": 6.742344101535394e-05, "loss": 0.5141, "step": 12072 }, { "epoch": 0.8103083789134593, "grad_norm": 1.0465116500854492, "learning_rate": 6.741325351879771e-05, "loss": 0.5677, "step": 12074 }, { "epoch": 0.8104426025972283, "grad_norm": 0.9749264717102051, "learning_rate": 6.740306519949952e-05, "loss": 0.5839, "step": 12076 }, { "epoch": 0.8105768262809973, "grad_norm": 1.1746549606323242, "learning_rate": 6.739287605794069e-05, "loss": 0.5538, "step": 12078 }, { "epoch": 0.8107110499647663, "grad_norm": 1.009050965309143, "learning_rate": 6.73826860946027e-05, "loss": 0.7081, "step": 12080 }, { "epoch": 0.8108452736485353, "grad_norm": 1.0723981857299805, "learning_rate": 6.737249530996694e-05, "loss": 0.5721, "step": 12082 }, { "epoch": 0.8109794973323042, "grad_norm": 0.9879979491233826, "learning_rate": 6.736230370451496e-05, "loss": 0.6328, "step": 12084 }, { "epoch": 0.8111137210160733, "grad_norm": 0.8727061152458191, "learning_rate": 6.735211127872827e-05, "loss": 0.5154, "step": 12086 }, { "epoch": 0.8112479446998423, "grad_norm": 0.9819492697715759, "learning_rate": 6.734191803308842e-05, "loss": 0.5759, "step": 12088 }, { "epoch": 0.8113821683836113, "grad_norm": 1.0325849056243896, "learning_rate": 6.733172396807708e-05, "loss": 0.5876, "step": 12090 }, { "epoch": 0.8115163920673802, "grad_norm": 0.9518097043037415, "learning_rate": 6.732152908417583e-05, "loss": 0.556, "step": 12092 }, { "epoch": 0.8116506157511493, "grad_norm": 1.0151830911636353, "learning_rate": 6.731133338186643e-05, "loss": 0.6942, "step": 12094 }, { "epoch": 0.8117848394349183, "grad_norm": 0.9402797222137451, "learning_rate": 6.730113686163055e-05, "loss": 0.635, "step": 12096 }, { "epoch": 0.8119190631186873, "grad_norm": 1.1566520929336548, "learning_rate": 6.729093952394996e-05, "loss": 0.5926, "step": 12098 }, { "epoch": 0.8120532868024563, "grad_norm": 1.03131103515625, "learning_rate": 6.72807413693065e-05, "loss": 0.635, "step": 12100 }, { "epoch": 0.8121875104862253, "grad_norm": 1.0487685203552246, "learning_rate": 6.727054239818198e-05, "loss": 0.5959, "step": 12102 }, { "epoch": 0.8123217341699943, "grad_norm": 1.0963678359985352, "learning_rate": 6.72603426110583e-05, "loss": 0.6084, "step": 12104 }, { "epoch": 0.8124559578537633, "grad_norm": 0.9777641296386719, "learning_rate": 6.725014200841738e-05, "loss": 0.5647, "step": 12106 }, { "epoch": 0.8125901815375323, "grad_norm": 0.9852986931800842, "learning_rate": 6.723994059074114e-05, "loss": 0.5841, "step": 12108 }, { "epoch": 0.8127244052213013, "grad_norm": 1.1587984561920166, "learning_rate": 6.722973835851162e-05, "loss": 0.6071, "step": 12110 }, { "epoch": 0.8128586289050703, "grad_norm": 1.3010673522949219, "learning_rate": 6.721953531221085e-05, "loss": 0.5806, "step": 12112 }, { "epoch": 0.8129928525888394, "grad_norm": 0.9649850130081177, "learning_rate": 6.720933145232091e-05, "loss": 0.5731, "step": 12114 }, { "epoch": 0.8131270762726083, "grad_norm": 1.01322603225708, "learning_rate": 6.719912677932389e-05, "loss": 0.6052, "step": 12116 }, { "epoch": 0.8132612999563773, "grad_norm": 1.1387805938720703, "learning_rate": 6.718892129370195e-05, "loss": 0.5466, "step": 12118 }, { "epoch": 0.8133955236401463, "grad_norm": 1.001020908355713, "learning_rate": 6.717871499593728e-05, "loss": 0.6387, "step": 12120 }, { "epoch": 0.8135297473239153, "grad_norm": 0.9601315259933472, "learning_rate": 6.71685078865121e-05, "loss": 0.5281, "step": 12122 }, { "epoch": 0.8136639710076843, "grad_norm": 1.089627981185913, "learning_rate": 6.71582999659087e-05, "loss": 0.6077, "step": 12124 }, { "epoch": 0.8137981946914533, "grad_norm": 1.0176795721054077, "learning_rate": 6.714809123460935e-05, "loss": 0.605, "step": 12126 }, { "epoch": 0.8139324183752223, "grad_norm": 0.9597039222717285, "learning_rate": 6.713788169309641e-05, "loss": 0.6353, "step": 12128 }, { "epoch": 0.8140666420589913, "grad_norm": 1.0546835660934448, "learning_rate": 6.712767134185228e-05, "loss": 0.5262, "step": 12130 }, { "epoch": 0.8142008657427603, "grad_norm": 1.076861023902893, "learning_rate": 6.711746018135933e-05, "loss": 0.5729, "step": 12132 }, { "epoch": 0.8143350894265293, "grad_norm": 1.1588343381881714, "learning_rate": 6.710724821210006e-05, "loss": 0.606, "step": 12134 }, { "epoch": 0.8144693131102984, "grad_norm": 1.027660608291626, "learning_rate": 6.709703543455695e-05, "loss": 0.5724, "step": 12136 }, { "epoch": 0.8146035367940673, "grad_norm": 1.0974794626235962, "learning_rate": 6.708682184921255e-05, "loss": 0.593, "step": 12138 }, { "epoch": 0.8147377604778363, "grad_norm": 1.1136621236801147, "learning_rate": 6.70766074565494e-05, "loss": 0.6177, "step": 12140 }, { "epoch": 0.8148719841616053, "grad_norm": 1.093705654144287, "learning_rate": 6.706639225705014e-05, "loss": 0.6727, "step": 12142 }, { "epoch": 0.8150062078453744, "grad_norm": 1.0256413221359253, "learning_rate": 6.705617625119738e-05, "loss": 0.6204, "step": 12144 }, { "epoch": 0.8151404315291433, "grad_norm": 1.0700013637542725, "learning_rate": 6.704595943947385e-05, "loss": 0.6268, "step": 12146 }, { "epoch": 0.8152746552129123, "grad_norm": 0.9938770532608032, "learning_rate": 6.703574182236226e-05, "loss": 0.5993, "step": 12148 }, { "epoch": 0.8154088788966813, "grad_norm": 1.084133505821228, "learning_rate": 6.702552340034535e-05, "loss": 0.688, "step": 12150 }, { "epoch": 0.8155431025804504, "grad_norm": 0.9394397139549255, "learning_rate": 6.701530417390597e-05, "loss": 0.5272, "step": 12152 }, { "epoch": 0.8156773262642193, "grad_norm": 1.0092073678970337, "learning_rate": 6.70050841435269e-05, "loss": 0.585, "step": 12154 }, { "epoch": 0.8158115499479883, "grad_norm": 1.026151418685913, "learning_rate": 6.699486330969106e-05, "loss": 0.6363, "step": 12156 }, { "epoch": 0.8159457736317574, "grad_norm": 0.9925841093063354, "learning_rate": 6.698464167288133e-05, "loss": 0.6316, "step": 12158 }, { "epoch": 0.8160799973155263, "grad_norm": 0.9733792543411255, "learning_rate": 6.697441923358068e-05, "loss": 0.5863, "step": 12160 }, { "epoch": 0.8162142209992953, "grad_norm": 0.9599014520645142, "learning_rate": 6.696419599227213e-05, "loss": 0.5733, "step": 12162 }, { "epoch": 0.8163484446830643, "grad_norm": 0.874563992023468, "learning_rate": 6.695397194943864e-05, "loss": 0.5244, "step": 12164 }, { "epoch": 0.8164826683668334, "grad_norm": 1.0366476774215698, "learning_rate": 6.694374710556335e-05, "loss": 0.6081, "step": 12166 }, { "epoch": 0.8166168920506023, "grad_norm": 1.7056056261062622, "learning_rate": 6.69335214611293e-05, "loss": 0.6429, "step": 12168 }, { "epoch": 0.8167511157343713, "grad_norm": 0.9990048408508301, "learning_rate": 6.692329501661966e-05, "loss": 0.6229, "step": 12170 }, { "epoch": 0.8168853394181403, "grad_norm": 0.9935768842697144, "learning_rate": 6.691306777251762e-05, "loss": 0.6714, "step": 12172 }, { "epoch": 0.8170195631019094, "grad_norm": 1.1473873853683472, "learning_rate": 6.690283972930639e-05, "loss": 0.6324, "step": 12174 }, { "epoch": 0.8171537867856783, "grad_norm": 1.0470446348190308, "learning_rate": 6.689261088746921e-05, "loss": 0.643, "step": 12176 }, { "epoch": 0.8172880104694473, "grad_norm": 1.36915123462677, "learning_rate": 6.688238124748939e-05, "loss": 0.5994, "step": 12178 }, { "epoch": 0.8174222341532164, "grad_norm": 1.0217487812042236, "learning_rate": 6.687215080985025e-05, "loss": 0.6359, "step": 12180 }, { "epoch": 0.8175564578369854, "grad_norm": 1.2322551012039185, "learning_rate": 6.686191957503517e-05, "loss": 0.5836, "step": 12182 }, { "epoch": 0.8176906815207543, "grad_norm": 1.0810588598251343, "learning_rate": 6.685168754352754e-05, "loss": 0.6029, "step": 12184 }, { "epoch": 0.8178249052045233, "grad_norm": 0.9595453143119812, "learning_rate": 6.684145471581081e-05, "loss": 0.5012, "step": 12186 }, { "epoch": 0.8179591288882924, "grad_norm": 1.0888211727142334, "learning_rate": 6.683122109236845e-05, "loss": 0.6424, "step": 12188 }, { "epoch": 0.8180933525720614, "grad_norm": 1.0043089389801025, "learning_rate": 6.682098667368403e-05, "loss": 0.6469, "step": 12190 }, { "epoch": 0.8182275762558303, "grad_norm": 1.1146799325942993, "learning_rate": 6.681075146024104e-05, "loss": 0.5393, "step": 12192 }, { "epoch": 0.8183617999395993, "grad_norm": 1.0782904624938965, "learning_rate": 6.68005154525231e-05, "loss": 0.6172, "step": 12194 }, { "epoch": 0.8184960236233684, "grad_norm": 0.9525636434555054, "learning_rate": 6.679027865101383e-05, "loss": 0.639, "step": 12196 }, { "epoch": 0.8186302473071373, "grad_norm": 1.0268981456756592, "learning_rate": 6.678004105619693e-05, "loss": 0.5793, "step": 12198 }, { "epoch": 0.8187644709909063, "grad_norm": 0.970932126045227, "learning_rate": 6.676980266855608e-05, "loss": 0.5664, "step": 12200 }, { "epoch": 0.8188986946746754, "grad_norm": 1.0692269802093506, "learning_rate": 6.675956348857504e-05, "loss": 0.5385, "step": 12202 }, { "epoch": 0.8190329183584444, "grad_norm": 1.116820216178894, "learning_rate": 6.674932351673758e-05, "loss": 0.674, "step": 12204 }, { "epoch": 0.8191671420422133, "grad_norm": 1.1007301807403564, "learning_rate": 6.67390827535275e-05, "loss": 0.5891, "step": 12206 }, { "epoch": 0.8193013657259823, "grad_norm": 1.049869418144226, "learning_rate": 6.672884119942868e-05, "loss": 0.6069, "step": 12208 }, { "epoch": 0.8194355894097514, "grad_norm": 1.1595649719238281, "learning_rate": 6.671859885492502e-05, "loss": 0.6151, "step": 12210 }, { "epoch": 0.8195698130935204, "grad_norm": 1.0789681673049927, "learning_rate": 6.670835572050043e-05, "loss": 0.5684, "step": 12212 }, { "epoch": 0.8197040367772893, "grad_norm": 1.0159722566604614, "learning_rate": 6.669811179663891e-05, "loss": 0.5849, "step": 12214 }, { "epoch": 0.8198382604610583, "grad_norm": 1.06174635887146, "learning_rate": 6.668786708382441e-05, "loss": 0.5139, "step": 12216 }, { "epoch": 0.8199724841448274, "grad_norm": 1.1024788618087769, "learning_rate": 6.667762158254104e-05, "loss": 0.6238, "step": 12218 }, { "epoch": 0.8201067078285964, "grad_norm": 1.0661011934280396, "learning_rate": 6.666737529327282e-05, "loss": 0.6053, "step": 12220 }, { "epoch": 0.8202409315123653, "grad_norm": 1.0743005275726318, "learning_rate": 6.66571282165039e-05, "loss": 0.6731, "step": 12222 }, { "epoch": 0.8203751551961344, "grad_norm": 1.4202250242233276, "learning_rate": 6.664688035271843e-05, "loss": 0.6163, "step": 12224 }, { "epoch": 0.8205093788799034, "grad_norm": 0.9662414789199829, "learning_rate": 6.66366317024006e-05, "loss": 0.5493, "step": 12226 }, { "epoch": 0.8206436025636724, "grad_norm": 0.9612823128700256, "learning_rate": 6.662638226603463e-05, "loss": 0.5547, "step": 12228 }, { "epoch": 0.8207778262474413, "grad_norm": 1.1255649328231812, "learning_rate": 6.661613204410479e-05, "loss": 0.6004, "step": 12230 }, { "epoch": 0.8209120499312104, "grad_norm": 1.4618395566940308, "learning_rate": 6.66058810370954e-05, "loss": 0.5452, "step": 12232 }, { "epoch": 0.8210462736149794, "grad_norm": 1.1108720302581787, "learning_rate": 6.659562924549076e-05, "loss": 0.5336, "step": 12234 }, { "epoch": 0.8211804972987483, "grad_norm": 1.011399745941162, "learning_rate": 6.658537666977529e-05, "loss": 0.6551, "step": 12236 }, { "epoch": 0.8213147209825173, "grad_norm": 1.0920634269714355, "learning_rate": 6.657512331043339e-05, "loss": 0.6252, "step": 12238 }, { "epoch": 0.8214489446662864, "grad_norm": 0.9514212608337402, "learning_rate": 6.65648691679495e-05, "loss": 0.5765, "step": 12240 }, { "epoch": 0.8215831683500554, "grad_norm": 1.2220063209533691, "learning_rate": 6.65546142428081e-05, "loss": 0.6864, "step": 12242 }, { "epoch": 0.8217173920338243, "grad_norm": 0.9927895069122314, "learning_rate": 6.654435853549375e-05, "loss": 0.5969, "step": 12244 }, { "epoch": 0.8218516157175934, "grad_norm": 1.1377277374267578, "learning_rate": 6.653410204649099e-05, "loss": 0.6233, "step": 12246 }, { "epoch": 0.8219858394013624, "grad_norm": 1.0246647596359253, "learning_rate": 6.652384477628442e-05, "loss": 0.6001, "step": 12248 }, { "epoch": 0.8221200630851314, "grad_norm": 1.1752510070800781, "learning_rate": 6.651358672535868e-05, "loss": 0.6288, "step": 12250 }, { "epoch": 0.8222542867689003, "grad_norm": 1.0195872783660889, "learning_rate": 6.650332789419844e-05, "loss": 0.6057, "step": 12252 }, { "epoch": 0.8223885104526694, "grad_norm": 0.9772490859031677, "learning_rate": 6.64930682832884e-05, "loss": 0.6264, "step": 12254 }, { "epoch": 0.8225227341364384, "grad_norm": 1.108428716659546, "learning_rate": 6.648280789311332e-05, "loss": 0.6569, "step": 12256 }, { "epoch": 0.8226569578202074, "grad_norm": 0.9418324828147888, "learning_rate": 6.6472546724158e-05, "loss": 0.5641, "step": 12258 }, { "epoch": 0.8227911815039763, "grad_norm": 1.1299644708633423, "learning_rate": 6.646228477690722e-05, "loss": 0.6731, "step": 12260 }, { "epoch": 0.8229254051877454, "grad_norm": 0.8824229836463928, "learning_rate": 6.645202205184584e-05, "loss": 0.4883, "step": 12262 }, { "epoch": 0.8230596288715144, "grad_norm": 2.0591981410980225, "learning_rate": 6.644175854945878e-05, "loss": 0.6361, "step": 12264 }, { "epoch": 0.8231938525552834, "grad_norm": 1.0393680334091187, "learning_rate": 6.643149427023097e-05, "loss": 0.6191, "step": 12266 }, { "epoch": 0.8233280762390524, "grad_norm": 1.083211064338684, "learning_rate": 6.642122921464736e-05, "loss": 0.6547, "step": 12268 }, { "epoch": 0.8234622999228214, "grad_norm": 1.21903657913208, "learning_rate": 6.641096338319297e-05, "loss": 0.5543, "step": 12270 }, { "epoch": 0.8235965236065904, "grad_norm": 1.0367486476898193, "learning_rate": 6.640069677635282e-05, "loss": 0.5979, "step": 12272 }, { "epoch": 0.8237307472903593, "grad_norm": 1.1093127727508545, "learning_rate": 6.6390429394612e-05, "loss": 0.6194, "step": 12274 }, { "epoch": 0.8238649709741284, "grad_norm": 1.100286841392517, "learning_rate": 6.638016123845562e-05, "loss": 0.6025, "step": 12276 }, { "epoch": 0.8239991946578974, "grad_norm": 1.0538876056671143, "learning_rate": 6.636989230836884e-05, "loss": 0.5919, "step": 12278 }, { "epoch": 0.8241334183416664, "grad_norm": 1.1927465200424194, "learning_rate": 6.635962260483683e-05, "loss": 0.5939, "step": 12280 }, { "epoch": 0.8242676420254353, "grad_norm": 0.8975897431373596, "learning_rate": 6.634935212834483e-05, "loss": 0.5064, "step": 12282 }, { "epoch": 0.8244018657092044, "grad_norm": 0.9954090118408203, "learning_rate": 6.63390808793781e-05, "loss": 0.602, "step": 12284 }, { "epoch": 0.8245360893929734, "grad_norm": 0.9890730977058411, "learning_rate": 6.63288088584219e-05, "loss": 0.6095, "step": 12286 }, { "epoch": 0.8246703130767424, "grad_norm": 1.0482323169708252, "learning_rate": 6.63185360659616e-05, "loss": 0.5444, "step": 12288 }, { "epoch": 0.8248045367605114, "grad_norm": 1.1700454950332642, "learning_rate": 6.630826250248256e-05, "loss": 0.5609, "step": 12290 }, { "epoch": 0.8249387604442804, "grad_norm": 0.9571924209594727, "learning_rate": 6.629798816847019e-05, "loss": 0.516, "step": 12292 }, { "epoch": 0.8250729841280494, "grad_norm": 1.028490662574768, "learning_rate": 6.628771306440994e-05, "loss": 0.5647, "step": 12294 }, { "epoch": 0.8252072078118184, "grad_norm": 1.81583571434021, "learning_rate": 6.627743719078725e-05, "loss": 0.5666, "step": 12296 }, { "epoch": 0.8253414314955874, "grad_norm": 1.122802495956421, "learning_rate": 6.626716054808768e-05, "loss": 0.6308, "step": 12298 }, { "epoch": 0.8254756551793564, "grad_norm": 1.0311051607131958, "learning_rate": 6.625688313679676e-05, "loss": 0.5847, "step": 12300 }, { "epoch": 0.8256098788631254, "grad_norm": 1.5131393671035767, "learning_rate": 6.624660495740007e-05, "loss": 0.6102, "step": 12302 }, { "epoch": 0.8257441025468945, "grad_norm": 1.0073790550231934, "learning_rate": 6.623632601038325e-05, "loss": 0.5906, "step": 12304 }, { "epoch": 0.8258783262306634, "grad_norm": 0.9229928851127625, "learning_rate": 6.622604629623196e-05, "loss": 0.6432, "step": 12306 }, { "epoch": 0.8260125499144324, "grad_norm": 0.9179010987281799, "learning_rate": 6.621576581543189e-05, "loss": 0.5733, "step": 12308 }, { "epoch": 0.8261467735982014, "grad_norm": 1.052332878112793, "learning_rate": 6.620548456846876e-05, "loss": 0.5641, "step": 12310 }, { "epoch": 0.8262809972819704, "grad_norm": 1.1295593976974487, "learning_rate": 6.619520255582834e-05, "loss": 0.6259, "step": 12312 }, { "epoch": 0.8264152209657394, "grad_norm": 0.9979819059371948, "learning_rate": 6.618491977799648e-05, "loss": 0.5714, "step": 12314 }, { "epoch": 0.8265494446495084, "grad_norm": 1.22626793384552, "learning_rate": 6.617463623545895e-05, "loss": 0.6188, "step": 12316 }, { "epoch": 0.8266836683332774, "grad_norm": 1.0156211853027344, "learning_rate": 6.61643519287017e-05, "loss": 0.6422, "step": 12318 }, { "epoch": 0.8268178920170464, "grad_norm": 1.1920782327651978, "learning_rate": 6.615406685821058e-05, "loss": 0.6056, "step": 12320 }, { "epoch": 0.8269521157008154, "grad_norm": 1.2539907693862915, "learning_rate": 6.614378102447158e-05, "loss": 0.6192, "step": 12322 }, { "epoch": 0.8270863393845844, "grad_norm": 1.0377227067947388, "learning_rate": 6.613349442797066e-05, "loss": 0.5679, "step": 12324 }, { "epoch": 0.8272205630683535, "grad_norm": 1.1849273443222046, "learning_rate": 6.612320706919387e-05, "loss": 0.5548, "step": 12326 }, { "epoch": 0.8273547867521224, "grad_norm": 1.0158774852752686, "learning_rate": 6.611291894862726e-05, "loss": 0.6035, "step": 12328 }, { "epoch": 0.8274890104358914, "grad_norm": 1.2449626922607422, "learning_rate": 6.610263006675688e-05, "loss": 0.5756, "step": 12330 }, { "epoch": 0.8276232341196604, "grad_norm": 1.0021066665649414, "learning_rate": 6.609234042406892e-05, "loss": 0.5549, "step": 12332 }, { "epoch": 0.8277574578034295, "grad_norm": 1.165596604347229, "learning_rate": 6.60820500210495e-05, "loss": 0.6329, "step": 12334 }, { "epoch": 0.8278916814871984, "grad_norm": 1.0297205448150635, "learning_rate": 6.607175885818485e-05, "loss": 0.6054, "step": 12336 }, { "epoch": 0.8280259051709674, "grad_norm": 0.9477143883705139, "learning_rate": 6.60614669359612e-05, "loss": 0.615, "step": 12338 }, { "epoch": 0.8281601288547364, "grad_norm": 1.0090079307556152, "learning_rate": 6.605117425486482e-05, "loss": 0.5564, "step": 12340 }, { "epoch": 0.8282943525385055, "grad_norm": 1.0071775913238525, "learning_rate": 6.604088081538203e-05, "loss": 0.5843, "step": 12342 }, { "epoch": 0.8284285762222744, "grad_norm": 0.9374699592590332, "learning_rate": 6.603058661799915e-05, "loss": 0.5506, "step": 12344 }, { "epoch": 0.8285627999060434, "grad_norm": 1.0525579452514648, "learning_rate": 6.602029166320258e-05, "loss": 0.6223, "step": 12346 }, { "epoch": 0.8286970235898125, "grad_norm": 0.9523213505744934, "learning_rate": 6.600999595147872e-05, "loss": 0.5676, "step": 12348 }, { "epoch": 0.8288312472735814, "grad_norm": 1.093316912651062, "learning_rate": 6.599969948331403e-05, "loss": 0.6291, "step": 12350 }, { "epoch": 0.8289654709573504, "grad_norm": 1.0761946439743042, "learning_rate": 6.598940225919504e-05, "loss": 0.5415, "step": 12352 }, { "epoch": 0.8290996946411194, "grad_norm": 1.1536874771118164, "learning_rate": 6.59791042796082e-05, "loss": 0.5226, "step": 12354 }, { "epoch": 0.8292339183248885, "grad_norm": 0.8335601091384888, "learning_rate": 6.596880554504011e-05, "loss": 0.524, "step": 12356 }, { "epoch": 0.8293681420086574, "grad_norm": 1.0278328657150269, "learning_rate": 6.595850605597736e-05, "loss": 0.5763, "step": 12358 }, { "epoch": 0.8295023656924264, "grad_norm": 1.1322448253631592, "learning_rate": 6.594820581290659e-05, "loss": 0.7102, "step": 12360 }, { "epoch": 0.8296365893761954, "grad_norm": 1.030147910118103, "learning_rate": 6.593790481631445e-05, "loss": 0.5821, "step": 12362 }, { "epoch": 0.8297708130599645, "grad_norm": 1.0838468074798584, "learning_rate": 6.592760306668763e-05, "loss": 0.5579, "step": 12364 }, { "epoch": 0.8299050367437334, "grad_norm": 1.009028434753418, "learning_rate": 6.591730056451292e-05, "loss": 0.5975, "step": 12366 }, { "epoch": 0.8300392604275024, "grad_norm": 1.03667414188385, "learning_rate": 6.590699731027703e-05, "loss": 0.5092, "step": 12368 }, { "epoch": 0.8301734841112715, "grad_norm": 0.994506299495697, "learning_rate": 6.589669330446682e-05, "loss": 0.6129, "step": 12370 }, { "epoch": 0.8303077077950405, "grad_norm": 0.9825753569602966, "learning_rate": 6.58863885475691e-05, "loss": 0.5808, "step": 12372 }, { "epoch": 0.8304419314788094, "grad_norm": 1.4217524528503418, "learning_rate": 6.587608304007075e-05, "loss": 0.5751, "step": 12374 }, { "epoch": 0.8305761551625784, "grad_norm": 1.0501091480255127, "learning_rate": 6.586577678245872e-05, "loss": 0.5963, "step": 12376 }, { "epoch": 0.8307103788463475, "grad_norm": 1.012234091758728, "learning_rate": 6.58554697752199e-05, "loss": 0.6249, "step": 12378 }, { "epoch": 0.8308446025301165, "grad_norm": 1.040315866470337, "learning_rate": 6.584516201884134e-05, "loss": 0.5786, "step": 12380 }, { "epoch": 0.8309788262138854, "grad_norm": 1.2015894651412964, "learning_rate": 6.583485351381002e-05, "loss": 0.6193, "step": 12382 }, { "epoch": 0.8311130498976544, "grad_norm": 1.040621280670166, "learning_rate": 6.582454426061302e-05, "loss": 0.5689, "step": 12384 }, { "epoch": 0.8312472735814235, "grad_norm": 1.1915639638900757, "learning_rate": 6.581423425973741e-05, "loss": 0.6585, "step": 12386 }, { "epoch": 0.8313814972651924, "grad_norm": 1.0276316404342651, "learning_rate": 6.580392351167033e-05, "loss": 0.662, "step": 12388 }, { "epoch": 0.8315157209489614, "grad_norm": 1.0397698879241943, "learning_rate": 6.579361201689895e-05, "loss": 0.6126, "step": 12390 }, { "epoch": 0.8316499446327305, "grad_norm": 1.0907901525497437, "learning_rate": 6.578329977591045e-05, "loss": 0.6089, "step": 12392 }, { "epoch": 0.8317841683164995, "grad_norm": 0.9839596152305603, "learning_rate": 6.577298678919209e-05, "loss": 0.5561, "step": 12394 }, { "epoch": 0.8319183920002684, "grad_norm": 0.9547123908996582, "learning_rate": 6.576267305723111e-05, "loss": 0.5302, "step": 12396 }, { "epoch": 0.8320526156840374, "grad_norm": 0.9410605430603027, "learning_rate": 6.575235858051481e-05, "loss": 0.6468, "step": 12398 }, { "epoch": 0.8321868393678065, "grad_norm": 0.9823176860809326, "learning_rate": 6.574204335953056e-05, "loss": 0.5582, "step": 12400 }, { "epoch": 0.8323210630515755, "grad_norm": 1.059046983718872, "learning_rate": 6.57317273947657e-05, "loss": 0.6269, "step": 12402 }, { "epoch": 0.8324552867353444, "grad_norm": 0.9903417229652405, "learning_rate": 6.572141068670768e-05, "loss": 0.5396, "step": 12404 }, { "epoch": 0.8325895104191134, "grad_norm": 0.92746502161026, "learning_rate": 6.57110932358439e-05, "loss": 0.5885, "step": 12406 }, { "epoch": 0.8327237341028825, "grad_norm": 1.1375131607055664, "learning_rate": 6.570077504266188e-05, "loss": 0.5735, "step": 12408 }, { "epoch": 0.8328579577866515, "grad_norm": 0.9820353984832764, "learning_rate": 6.569045610764909e-05, "loss": 0.5607, "step": 12410 }, { "epoch": 0.8329921814704204, "grad_norm": 1.0120148658752441, "learning_rate": 6.568013643129311e-05, "loss": 0.5833, "step": 12412 }, { "epoch": 0.8331264051541895, "grad_norm": 0.9980480074882507, "learning_rate": 6.566981601408153e-05, "loss": 0.5451, "step": 12414 }, { "epoch": 0.8332606288379585, "grad_norm": 0.9461591243743896, "learning_rate": 6.565949485650193e-05, "loss": 0.5984, "step": 12416 }, { "epoch": 0.8333948525217275, "grad_norm": 1.4781508445739746, "learning_rate": 6.564917295904201e-05, "loss": 0.5972, "step": 12418 }, { "epoch": 0.8335290762054964, "grad_norm": 1.0129133462905884, "learning_rate": 6.563885032218943e-05, "loss": 0.625, "step": 12420 }, { "epoch": 0.8336632998892655, "grad_norm": 1.0137070417404175, "learning_rate": 6.562852694643194e-05, "loss": 0.5361, "step": 12422 }, { "epoch": 0.8337975235730345, "grad_norm": 1.1191861629486084, "learning_rate": 6.561820283225726e-05, "loss": 0.5878, "step": 12424 }, { "epoch": 0.8339317472568034, "grad_norm": 0.9222754836082458, "learning_rate": 6.560787798015323e-05, "loss": 0.5672, "step": 12426 }, { "epoch": 0.8340659709405724, "grad_norm": 1.1135950088500977, "learning_rate": 6.559755239060765e-05, "loss": 0.608, "step": 12428 }, { "epoch": 0.8342001946243415, "grad_norm": 1.876238465309143, "learning_rate": 6.558722606410839e-05, "loss": 0.5573, "step": 12430 }, { "epoch": 0.8343344183081105, "grad_norm": 0.9243791699409485, "learning_rate": 6.557689900114337e-05, "loss": 0.6266, "step": 12432 }, { "epoch": 0.8344686419918794, "grad_norm": 0.975472092628479, "learning_rate": 6.556657120220046e-05, "loss": 0.5769, "step": 12434 }, { "epoch": 0.8346028656756485, "grad_norm": 1.129416584968567, "learning_rate": 6.555624266776771e-05, "loss": 0.5199, "step": 12436 }, { "epoch": 0.8347370893594175, "grad_norm": 0.9951165914535522, "learning_rate": 6.554591339833307e-05, "loss": 0.4939, "step": 12438 }, { "epoch": 0.8348713130431865, "grad_norm": 1.0705358982086182, "learning_rate": 6.553558339438457e-05, "loss": 0.5433, "step": 12440 }, { "epoch": 0.8350055367269554, "grad_norm": 1.1221299171447754, "learning_rate": 6.552525265641034e-05, "loss": 0.5944, "step": 12442 }, { "epoch": 0.8351397604107245, "grad_norm": 1.151170015335083, "learning_rate": 6.551492118489843e-05, "loss": 0.6213, "step": 12444 }, { "epoch": 0.8352739840944935, "grad_norm": 1.3591521978378296, "learning_rate": 6.550458898033701e-05, "loss": 0.5648, "step": 12446 }, { "epoch": 0.8354082077782625, "grad_norm": 0.9034987688064575, "learning_rate": 6.549425604321424e-05, "loss": 0.5845, "step": 12448 }, { "epoch": 0.8355424314620314, "grad_norm": 1.1140567064285278, "learning_rate": 6.548392237401836e-05, "loss": 0.5704, "step": 12450 }, { "epoch": 0.8356766551458005, "grad_norm": 1.1149941682815552, "learning_rate": 6.547358797323758e-05, "loss": 0.6356, "step": 12452 }, { "epoch": 0.8358108788295695, "grad_norm": 1.0062631368637085, "learning_rate": 6.546325284136019e-05, "loss": 0.6261, "step": 12454 }, { "epoch": 0.8359451025133385, "grad_norm": 1.1215084791183472, "learning_rate": 6.545291697887453e-05, "loss": 0.6334, "step": 12456 }, { "epoch": 0.8360793261971075, "grad_norm": 0.9508087038993835, "learning_rate": 6.544258038626891e-05, "loss": 0.624, "step": 12458 }, { "epoch": 0.8362135498808765, "grad_norm": 0.9110504388809204, "learning_rate": 6.543224306403174e-05, "loss": 0.5663, "step": 12460 }, { "epoch": 0.8363477735646455, "grad_norm": 0.9410451650619507, "learning_rate": 6.542190501265144e-05, "loss": 0.5636, "step": 12462 }, { "epoch": 0.8364819972484144, "grad_norm": 0.9118926525115967, "learning_rate": 6.541156623261646e-05, "loss": 0.6712, "step": 12464 }, { "epoch": 0.8366162209321835, "grad_norm": 1.238031268119812, "learning_rate": 6.540122672441526e-05, "loss": 0.6163, "step": 12466 }, { "epoch": 0.8367504446159525, "grad_norm": 0.9727116227149963, "learning_rate": 6.53908864885364e-05, "loss": 0.6714, "step": 12468 }, { "epoch": 0.8368846682997215, "grad_norm": 1.0054668188095093, "learning_rate": 6.538054552546841e-05, "loss": 0.6083, "step": 12470 }, { "epoch": 0.8370188919834904, "grad_norm": 1.0687994956970215, "learning_rate": 6.537020383569988e-05, "loss": 0.5514, "step": 12472 }, { "epoch": 0.8371531156672595, "grad_norm": 0.8260632753372192, "learning_rate": 6.535986141971946e-05, "loss": 0.514, "step": 12474 }, { "epoch": 0.8372873393510285, "grad_norm": 1.081587553024292, "learning_rate": 6.534951827801579e-05, "loss": 0.6348, "step": 12476 }, { "epoch": 0.8374215630347975, "grad_norm": 1.0609947443008423, "learning_rate": 6.533917441107755e-05, "loss": 0.5551, "step": 12478 }, { "epoch": 0.8375557867185665, "grad_norm": 1.0872089862823486, "learning_rate": 6.532882981939349e-05, "loss": 0.5786, "step": 12480 }, { "epoch": 0.8376900104023355, "grad_norm": 1.1435655355453491, "learning_rate": 6.531848450345236e-05, "loss": 0.605, "step": 12482 }, { "epoch": 0.8378242340861045, "grad_norm": 1.085896611213684, "learning_rate": 6.530813846374297e-05, "loss": 0.521, "step": 12484 }, { "epoch": 0.8379584577698735, "grad_norm": 1.0298261642456055, "learning_rate": 6.529779170075413e-05, "loss": 0.6042, "step": 12486 }, { "epoch": 0.8380926814536425, "grad_norm": 1.1587203741073608, "learning_rate": 6.528744421497471e-05, "loss": 0.6989, "step": 12488 }, { "epoch": 0.8382269051374115, "grad_norm": 0.9626758098602295, "learning_rate": 6.527709600689363e-05, "loss": 0.6008, "step": 12490 }, { "epoch": 0.8383611288211805, "grad_norm": 0.9409077763557434, "learning_rate": 6.526674707699979e-05, "loss": 0.5952, "step": 12492 }, { "epoch": 0.8384953525049496, "grad_norm": 0.9995615482330322, "learning_rate": 6.525639742578218e-05, "loss": 0.545, "step": 12494 }, { "epoch": 0.8386295761887185, "grad_norm": 0.9815777540206909, "learning_rate": 6.524604705372979e-05, "loss": 0.5895, "step": 12496 }, { "epoch": 0.8387637998724875, "grad_norm": 0.9833337664604187, "learning_rate": 6.523569596133165e-05, "loss": 0.5099, "step": 12498 }, { "epoch": 0.8388980235562565, "grad_norm": 0.9670937061309814, "learning_rate": 6.522534414907684e-05, "loss": 0.5499, "step": 12500 }, { "epoch": 0.8390322472400255, "grad_norm": 0.9951319694519043, "learning_rate": 6.521499161745445e-05, "loss": 0.5436, "step": 12502 }, { "epoch": 0.8391664709237945, "grad_norm": 0.9808186292648315, "learning_rate": 6.520463836695364e-05, "loss": 0.6691, "step": 12504 }, { "epoch": 0.8393006946075635, "grad_norm": 2.6236653327941895, "learning_rate": 6.519428439806356e-05, "loss": 0.6125, "step": 12506 }, { "epoch": 0.8394349182913325, "grad_norm": 1.3241273164749146, "learning_rate": 6.518392971127341e-05, "loss": 0.5987, "step": 12508 }, { "epoch": 0.8395691419751015, "grad_norm": 1.2755435705184937, "learning_rate": 6.517357430707245e-05, "loss": 0.6293, "step": 12510 }, { "epoch": 0.8397033656588705, "grad_norm": 1.3209245204925537, "learning_rate": 6.516321818594995e-05, "loss": 0.6042, "step": 12512 }, { "epoch": 0.8398375893426395, "grad_norm": 1.1447761058807373, "learning_rate": 6.51528613483952e-05, "loss": 0.6067, "step": 12514 }, { "epoch": 0.8399718130264086, "grad_norm": 1.0160657167434692, "learning_rate": 6.514250379489753e-05, "loss": 0.5526, "step": 12516 }, { "epoch": 0.8401060367101775, "grad_norm": 1.0563714504241943, "learning_rate": 6.513214552594637e-05, "loss": 0.5564, "step": 12518 }, { "epoch": 0.8402402603939465, "grad_norm": 1.038380742073059, "learning_rate": 6.512178654203105e-05, "loss": 0.5417, "step": 12520 }, { "epoch": 0.8403744840777155, "grad_norm": 1.062550663948059, "learning_rate": 6.511142684364109e-05, "loss": 0.5708, "step": 12522 }, { "epoch": 0.8405087077614846, "grad_norm": 0.9445386528968811, "learning_rate": 6.51010664312659e-05, "loss": 0.6243, "step": 12524 }, { "epoch": 0.8406429314452535, "grad_norm": 0.9992539882659912, "learning_rate": 6.509070530539502e-05, "loss": 0.5518, "step": 12526 }, { "epoch": 0.8407771551290225, "grad_norm": 3.426658868789673, "learning_rate": 6.508034346651798e-05, "loss": 0.6628, "step": 12528 }, { "epoch": 0.8409113788127915, "grad_norm": 1.1215256452560425, "learning_rate": 6.506998091512438e-05, "loss": 0.5626, "step": 12530 }, { "epoch": 0.8410456024965605, "grad_norm": 1.0512257814407349, "learning_rate": 6.505961765170383e-05, "loss": 0.4858, "step": 12532 }, { "epoch": 0.8411798261803295, "grad_norm": 1.0057384967803955, "learning_rate": 6.504925367674594e-05, "loss": 0.6318, "step": 12534 }, { "epoch": 0.8413140498640985, "grad_norm": 1.1570310592651367, "learning_rate": 6.503888899074041e-05, "loss": 0.5472, "step": 12536 }, { "epoch": 0.8414482735478676, "grad_norm": 0.8283010721206665, "learning_rate": 6.502852359417696e-05, "loss": 0.4961, "step": 12538 }, { "epoch": 0.8415824972316365, "grad_norm": 1.0469348430633545, "learning_rate": 6.50181574875453e-05, "loss": 0.5474, "step": 12540 }, { "epoch": 0.8417167209154055, "grad_norm": 1.1920287609100342, "learning_rate": 6.500779067133524e-05, "loss": 0.5967, "step": 12542 }, { "epoch": 0.8418509445991745, "grad_norm": 1.018536925315857, "learning_rate": 6.499742314603659e-05, "loss": 0.5903, "step": 12544 }, { "epoch": 0.8419851682829436, "grad_norm": 1.0969058275222778, "learning_rate": 6.498705491213917e-05, "loss": 0.6266, "step": 12546 }, { "epoch": 0.8421193919667125, "grad_norm": 0.929409921169281, "learning_rate": 6.497668597013289e-05, "loss": 0.6614, "step": 12548 }, { "epoch": 0.8422536156504815, "grad_norm": 1.1739845275878906, "learning_rate": 6.496631632050763e-05, "loss": 0.6104, "step": 12550 }, { "epoch": 0.8423878393342505, "grad_norm": 1.0775952339172363, "learning_rate": 6.495594596375338e-05, "loss": 0.5348, "step": 12552 }, { "epoch": 0.8425220630180196, "grad_norm": 1.0988727807998657, "learning_rate": 6.494557490036009e-05, "loss": 0.586, "step": 12554 }, { "epoch": 0.8426562867017885, "grad_norm": 1.0093157291412354, "learning_rate": 6.493520313081775e-05, "loss": 0.6403, "step": 12556 }, { "epoch": 0.8427905103855575, "grad_norm": 1.5681852102279663, "learning_rate": 6.492483065561645e-05, "loss": 0.5489, "step": 12558 }, { "epoch": 0.8429247340693266, "grad_norm": 0.8495414853096008, "learning_rate": 6.491445747524627e-05, "loss": 0.5446, "step": 12560 }, { "epoch": 0.8430589577530956, "grad_norm": 0.9610116481781006, "learning_rate": 6.490408359019726e-05, "loss": 0.6112, "step": 12562 }, { "epoch": 0.8431931814368645, "grad_norm": 0.9789390563964844, "learning_rate": 6.489370900095963e-05, "loss": 0.5811, "step": 12564 }, { "epoch": 0.8433274051206335, "grad_norm": 1.0481964349746704, "learning_rate": 6.488333370802354e-05, "loss": 0.6228, "step": 12566 }, { "epoch": 0.8434616288044026, "grad_norm": 1.030542254447937, "learning_rate": 6.487295771187919e-05, "loss": 0.5577, "step": 12568 }, { "epoch": 0.8435958524881715, "grad_norm": 0.9535661935806274, "learning_rate": 6.486258101301685e-05, "loss": 0.6148, "step": 12570 }, { "epoch": 0.8437300761719405, "grad_norm": 0.8901979923248291, "learning_rate": 6.485220361192677e-05, "loss": 0.5021, "step": 12572 }, { "epoch": 0.8438642998557095, "grad_norm": 1.042995572090149, "learning_rate": 6.484182550909927e-05, "loss": 0.5677, "step": 12574 }, { "epoch": 0.8439985235394786, "grad_norm": 1.0647839307785034, "learning_rate": 6.48314467050247e-05, "loss": 0.5217, "step": 12576 }, { "epoch": 0.8441327472232475, "grad_norm": 1.0358980894088745, "learning_rate": 6.482106720019344e-05, "loss": 0.5932, "step": 12578 }, { "epoch": 0.8442669709070165, "grad_norm": 0.9523314833641052, "learning_rate": 6.481068699509591e-05, "loss": 0.6175, "step": 12580 }, { "epoch": 0.8444011945907856, "grad_norm": 1.0470632314682007, "learning_rate": 6.480030609022253e-05, "loss": 0.5218, "step": 12582 }, { "epoch": 0.8445354182745546, "grad_norm": 1.5232722759246826, "learning_rate": 6.478992448606381e-05, "loss": 0.5523, "step": 12584 }, { "epoch": 0.8446696419583235, "grad_norm": 1.0707812309265137, "learning_rate": 6.477954218311021e-05, "loss": 0.5717, "step": 12586 }, { "epoch": 0.8448038656420925, "grad_norm": 1.0655757188796997, "learning_rate": 6.476915918185234e-05, "loss": 0.592, "step": 12588 }, { "epoch": 0.8449380893258616, "grad_norm": 1.2582905292510986, "learning_rate": 6.475877548278073e-05, "loss": 0.5673, "step": 12590 }, { "epoch": 0.8450723130096306, "grad_norm": 1.0590770244598389, "learning_rate": 6.474839108638598e-05, "loss": 0.6062, "step": 12592 }, { "epoch": 0.8452065366933995, "grad_norm": 0.9540066719055176, "learning_rate": 6.473800599315878e-05, "loss": 0.5378, "step": 12594 }, { "epoch": 0.8453407603771685, "grad_norm": 0.9626739025115967, "learning_rate": 6.472762020358976e-05, "loss": 0.5251, "step": 12596 }, { "epoch": 0.8454749840609376, "grad_norm": 1.1993706226348877, "learning_rate": 6.471723371816965e-05, "loss": 0.6221, "step": 12598 }, { "epoch": 0.8456092077447066, "grad_norm": 1.2055251598358154, "learning_rate": 6.470684653738919e-05, "loss": 0.5546, "step": 12600 }, { "epoch": 0.8457434314284755, "grad_norm": 1.076897144317627, "learning_rate": 6.469645866173916e-05, "loss": 0.5965, "step": 12602 }, { "epoch": 0.8458776551122446, "grad_norm": 0.9254967570304871, "learning_rate": 6.468607009171035e-05, "loss": 0.582, "step": 12604 }, { "epoch": 0.8460118787960136, "grad_norm": 1.0577062368392944, "learning_rate": 6.46756808277936e-05, "loss": 0.6295, "step": 12606 }, { "epoch": 0.8461461024797825, "grad_norm": 1.069785237312317, "learning_rate": 6.46652908704798e-05, "loss": 0.6006, "step": 12608 }, { "epoch": 0.8462803261635515, "grad_norm": 1.0423345565795898, "learning_rate": 6.465490022025984e-05, "loss": 0.5394, "step": 12610 }, { "epoch": 0.8464145498473206, "grad_norm": 0.9489768743515015, "learning_rate": 6.464450887762465e-05, "loss": 0.5713, "step": 12612 }, { "epoch": 0.8465487735310896, "grad_norm": 1.0175904035568237, "learning_rate": 6.463411684306522e-05, "loss": 0.5555, "step": 12614 }, { "epoch": 0.8466829972148585, "grad_norm": 0.989685595035553, "learning_rate": 6.462372411707255e-05, "loss": 0.5475, "step": 12616 }, { "epoch": 0.8468172208986275, "grad_norm": 0.989132285118103, "learning_rate": 6.461333070013767e-05, "loss": 0.5425, "step": 12618 }, { "epoch": 0.8469514445823966, "grad_norm": 0.9655634760856628, "learning_rate": 6.460293659275165e-05, "loss": 0.6008, "step": 12620 }, { "epoch": 0.8470856682661656, "grad_norm": 1.1773788928985596, "learning_rate": 6.45925417954056e-05, "loss": 0.5406, "step": 12622 }, { "epoch": 0.8472198919499345, "grad_norm": 1.1722278594970703, "learning_rate": 6.458214630859064e-05, "loss": 0.5549, "step": 12624 }, { "epoch": 0.8473541156337036, "grad_norm": 0.9735685586929321, "learning_rate": 6.457175013279792e-05, "loss": 0.5718, "step": 12626 }, { "epoch": 0.8474883393174726, "grad_norm": 1.0683695077896118, "learning_rate": 6.45613532685187e-05, "loss": 0.657, "step": 12628 }, { "epoch": 0.8476225630012416, "grad_norm": 0.9994304776191711, "learning_rate": 6.455095571624414e-05, "loss": 0.6243, "step": 12630 }, { "epoch": 0.8477567866850105, "grad_norm": 1.0363658666610718, "learning_rate": 6.454055747646555e-05, "loss": 0.609, "step": 12632 }, { "epoch": 0.8478910103687796, "grad_norm": 1.0402913093566895, "learning_rate": 6.45301585496742e-05, "loss": 0.5772, "step": 12634 }, { "epoch": 0.8480252340525486, "grad_norm": 0.9271230101585388, "learning_rate": 6.451975893636144e-05, "loss": 0.6032, "step": 12636 }, { "epoch": 0.8481594577363176, "grad_norm": 1.1815871000289917, "learning_rate": 6.450935863701863e-05, "loss": 0.5725, "step": 12638 }, { "epoch": 0.8482936814200865, "grad_norm": 1.3115997314453125, "learning_rate": 6.449895765213713e-05, "loss": 0.6197, "step": 12640 }, { "epoch": 0.8484279051038556, "grad_norm": 0.9992296695709229, "learning_rate": 6.448855598220842e-05, "loss": 0.5907, "step": 12642 }, { "epoch": 0.8485621287876246, "grad_norm": 0.9422286748886108, "learning_rate": 6.447815362772392e-05, "loss": 0.578, "step": 12644 }, { "epoch": 0.8486963524713935, "grad_norm": 1.1738027334213257, "learning_rate": 6.446775058917514e-05, "loss": 0.6038, "step": 12646 }, { "epoch": 0.8488305761551626, "grad_norm": 1.52314293384552, "learning_rate": 6.445734686705358e-05, "loss": 0.6368, "step": 12648 }, { "epoch": 0.8489647998389316, "grad_norm": 1.1008228063583374, "learning_rate": 6.444694246185082e-05, "loss": 0.6337, "step": 12650 }, { "epoch": 0.8490990235227006, "grad_norm": 0.9162123799324036, "learning_rate": 6.443653737405842e-05, "loss": 0.5293, "step": 12652 }, { "epoch": 0.8492332472064695, "grad_norm": 1.2840626239776611, "learning_rate": 6.442613160416803e-05, "loss": 0.5571, "step": 12654 }, { "epoch": 0.8493674708902386, "grad_norm": 1.146270513534546, "learning_rate": 6.44157251526713e-05, "loss": 0.5627, "step": 12656 }, { "epoch": 0.8495016945740076, "grad_norm": 0.9955841898918152, "learning_rate": 6.440531802005988e-05, "loss": 0.5282, "step": 12658 }, { "epoch": 0.8496359182577766, "grad_norm": 0.9514544606208801, "learning_rate": 6.439491020682553e-05, "loss": 0.5606, "step": 12660 }, { "epoch": 0.8497701419415455, "grad_norm": 1.0206433534622192, "learning_rate": 6.438450171345996e-05, "loss": 0.5182, "step": 12662 }, { "epoch": 0.8499043656253146, "grad_norm": 0.9411662817001343, "learning_rate": 6.437409254045499e-05, "loss": 0.503, "step": 12664 }, { "epoch": 0.8500385893090836, "grad_norm": 0.9811124205589294, "learning_rate": 6.436368268830241e-05, "loss": 0.6004, "step": 12666 }, { "epoch": 0.8501728129928526, "grad_norm": 0.9797987937927246, "learning_rate": 6.435327215749405e-05, "loss": 0.5661, "step": 12668 }, { "epoch": 0.8503070366766216, "grad_norm": 1.0006541013717651, "learning_rate": 6.434286094852184e-05, "loss": 0.5741, "step": 12670 }, { "epoch": 0.8504412603603906, "grad_norm": 1.0398980379104614, "learning_rate": 6.433244906187763e-05, "loss": 0.5918, "step": 12672 }, { "epoch": 0.8505754840441596, "grad_norm": 1.1455544233322144, "learning_rate": 6.432203649805337e-05, "loss": 0.6423, "step": 12674 }, { "epoch": 0.8507097077279286, "grad_norm": 1.1146358251571655, "learning_rate": 6.431162325754108e-05, "loss": 0.596, "step": 12676 }, { "epoch": 0.8508439314116976, "grad_norm": 1.1827868223190308, "learning_rate": 6.430120934083271e-05, "loss": 0.6371, "step": 12678 }, { "epoch": 0.8509781550954666, "grad_norm": 1.0701404809951782, "learning_rate": 6.429079474842035e-05, "loss": 0.5365, "step": 12680 }, { "epoch": 0.8511123787792356, "grad_norm": 1.3694677352905273, "learning_rate": 6.428037948079602e-05, "loss": 0.5899, "step": 12682 }, { "epoch": 0.8512466024630045, "grad_norm": 1.1600438356399536, "learning_rate": 6.426996353845184e-05, "loss": 0.5719, "step": 12684 }, { "epoch": 0.8513808261467736, "grad_norm": 1.0400867462158203, "learning_rate": 6.425954692187995e-05, "loss": 0.5678, "step": 12686 }, { "epoch": 0.8515150498305426, "grad_norm": 0.9638028144836426, "learning_rate": 6.424912963157249e-05, "loss": 0.5384, "step": 12688 }, { "epoch": 0.8516492735143116, "grad_norm": 1.2834314107894897, "learning_rate": 6.423871166802169e-05, "loss": 0.5771, "step": 12690 }, { "epoch": 0.8517834971980806, "grad_norm": 1.7242958545684814, "learning_rate": 6.422829303171976e-05, "loss": 0.5238, "step": 12692 }, { "epoch": 0.8519177208818496, "grad_norm": 0.9400202035903931, "learning_rate": 6.421787372315897e-05, "loss": 0.615, "step": 12694 }, { "epoch": 0.8520519445656186, "grad_norm": 1.0088517665863037, "learning_rate": 6.420745374283159e-05, "loss": 0.6427, "step": 12696 }, { "epoch": 0.8521861682493876, "grad_norm": 0.9345404505729675, "learning_rate": 6.419703309122997e-05, "loss": 0.5377, "step": 12698 }, { "epoch": 0.8523203919331566, "grad_norm": 1.0751745700836182, "learning_rate": 6.418661176884644e-05, "loss": 0.5821, "step": 12700 }, { "epoch": 0.8524546156169256, "grad_norm": 1.0044862031936646, "learning_rate": 6.417618977617342e-05, "loss": 0.5448, "step": 12702 }, { "epoch": 0.8525888393006946, "grad_norm": 0.9652642011642456, "learning_rate": 6.41657671137033e-05, "loss": 0.6212, "step": 12704 }, { "epoch": 0.8527230629844637, "grad_norm": 0.835157036781311, "learning_rate": 6.415534378192853e-05, "loss": 0.648, "step": 12706 }, { "epoch": 0.8528572866682326, "grad_norm": 1.0281325578689575, "learning_rate": 6.414491978134161e-05, "loss": 0.5735, "step": 12708 }, { "epoch": 0.8529915103520016, "grad_norm": 0.9508633017539978, "learning_rate": 6.413449511243505e-05, "loss": 0.5511, "step": 12710 }, { "epoch": 0.8531257340357706, "grad_norm": 2.6051974296569824, "learning_rate": 6.41240697757014e-05, "loss": 0.6665, "step": 12712 }, { "epoch": 0.8532599577195397, "grad_norm": 0.9451747536659241, "learning_rate": 6.41136437716332e-05, "loss": 0.6176, "step": 12714 }, { "epoch": 0.8533941814033086, "grad_norm": 0.9181392192840576, "learning_rate": 6.41032171007231e-05, "loss": 0.5604, "step": 12716 }, { "epoch": 0.8535284050870776, "grad_norm": 1.065870761871338, "learning_rate": 6.409278976346373e-05, "loss": 0.5658, "step": 12718 }, { "epoch": 0.8536626287708466, "grad_norm": 1.0549818277359009, "learning_rate": 6.408236176034774e-05, "loss": 0.6213, "step": 12720 }, { "epoch": 0.8537968524546156, "grad_norm": 0.9842135906219482, "learning_rate": 6.407193309186787e-05, "loss": 0.522, "step": 12722 }, { "epoch": 0.8539310761383846, "grad_norm": 1.1176663637161255, "learning_rate": 6.406150375851682e-05, "loss": 0.6356, "step": 12724 }, { "epoch": 0.8540652998221536, "grad_norm": 0.992957592010498, "learning_rate": 6.405107376078737e-05, "loss": 0.6202, "step": 12726 }, { "epoch": 0.8541995235059227, "grad_norm": 1.0676531791687012, "learning_rate": 6.404064309917231e-05, "loss": 0.5738, "step": 12728 }, { "epoch": 0.8543337471896916, "grad_norm": 1.1479809284210205, "learning_rate": 6.403021177416447e-05, "loss": 0.6201, "step": 12730 }, { "epoch": 0.8544679708734606, "grad_norm": 0.9011659026145935, "learning_rate": 6.401977978625672e-05, "loss": 0.5655, "step": 12732 }, { "epoch": 0.8546021945572296, "grad_norm": 0.9919249415397644, "learning_rate": 6.400934713594194e-05, "loss": 0.6284, "step": 12734 }, { "epoch": 0.8547364182409987, "grad_norm": 0.9688147306442261, "learning_rate": 6.399891382371308e-05, "loss": 0.543, "step": 12736 }, { "epoch": 0.8548706419247676, "grad_norm": 1.1079168319702148, "learning_rate": 6.398847985006304e-05, "loss": 0.5324, "step": 12738 }, { "epoch": 0.8550048656085366, "grad_norm": 1.1334136724472046, "learning_rate": 6.397804521548486e-05, "loss": 0.6161, "step": 12740 }, { "epoch": 0.8551390892923056, "grad_norm": 1.054489016532898, "learning_rate": 6.39676099204715e-05, "loss": 0.5932, "step": 12742 }, { "epoch": 0.8552733129760747, "grad_norm": 1.4015049934387207, "learning_rate": 6.395717396551606e-05, "loss": 0.6064, "step": 12744 }, { "epoch": 0.8554075366598436, "grad_norm": 1.6162772178649902, "learning_rate": 6.394673735111158e-05, "loss": 0.5492, "step": 12746 }, { "epoch": 0.8555417603436126, "grad_norm": 1.143741488456726, "learning_rate": 6.39363000777512e-05, "loss": 0.6125, "step": 12748 }, { "epoch": 0.8556759840273817, "grad_norm": 2.3074216842651367, "learning_rate": 6.392586214592804e-05, "loss": 0.6516, "step": 12750 }, { "epoch": 0.8558102077111507, "grad_norm": 1.0103665590286255, "learning_rate": 6.391542355613526e-05, "loss": 0.591, "step": 12752 }, { "epoch": 0.8559444313949196, "grad_norm": 1.2626314163208008, "learning_rate": 6.390498430886611e-05, "loss": 0.5789, "step": 12754 }, { "epoch": 0.8560786550786886, "grad_norm": 1.0897712707519531, "learning_rate": 6.389454440461378e-05, "loss": 0.5849, "step": 12756 }, { "epoch": 0.8562128787624577, "grad_norm": 1.0769323110580444, "learning_rate": 6.388410384387154e-05, "loss": 0.6044, "step": 12758 }, { "epoch": 0.8563471024462266, "grad_norm": 1.125114917755127, "learning_rate": 6.387366262713272e-05, "loss": 0.6041, "step": 12760 }, { "epoch": 0.8564813261299956, "grad_norm": 0.8260295987129211, "learning_rate": 6.38632207548906e-05, "loss": 0.5575, "step": 12762 }, { "epoch": 0.8566155498137646, "grad_norm": 1.0780701637268066, "learning_rate": 6.385277822763857e-05, "loss": 0.6076, "step": 12764 }, { "epoch": 0.8567497734975337, "grad_norm": 0.9742283225059509, "learning_rate": 6.384233504587001e-05, "loss": 0.6826, "step": 12766 }, { "epoch": 0.8568839971813026, "grad_norm": 0.9774967432022095, "learning_rate": 6.383189121007833e-05, "loss": 0.5658, "step": 12768 }, { "epoch": 0.8570182208650716, "grad_norm": 1.245073914527893, "learning_rate": 6.382144672075701e-05, "loss": 0.5882, "step": 12770 }, { "epoch": 0.8571524445488407, "grad_norm": 1.0366885662078857, "learning_rate": 6.381100157839948e-05, "loss": 0.5367, "step": 12772 }, { "epoch": 0.8572866682326097, "grad_norm": 0.9845640659332275, "learning_rate": 6.380055578349931e-05, "loss": 0.6538, "step": 12774 }, { "epoch": 0.8574208919163786, "grad_norm": 0.9645702242851257, "learning_rate": 6.379010933655001e-05, "loss": 0.5434, "step": 12776 }, { "epoch": 0.8575551156001476, "grad_norm": 1.019515872001648, "learning_rate": 6.377966223804515e-05, "loss": 0.6037, "step": 12778 }, { "epoch": 0.8576893392839167, "grad_norm": 0.8961287140846252, "learning_rate": 6.376921448847836e-05, "loss": 0.5209, "step": 12780 }, { "epoch": 0.8578235629676857, "grad_norm": 0.9874623417854309, "learning_rate": 6.375876608834324e-05, "loss": 0.5967, "step": 12782 }, { "epoch": 0.8579577866514546, "grad_norm": 0.9893937110900879, "learning_rate": 6.374831703813351e-05, "loss": 0.5701, "step": 12784 }, { "epoch": 0.8580920103352236, "grad_norm": 0.9620664715766907, "learning_rate": 6.37378673383428e-05, "loss": 0.5999, "step": 12786 }, { "epoch": 0.8582262340189927, "grad_norm": 0.9989723563194275, "learning_rate": 6.372741698946491e-05, "loss": 0.5446, "step": 12788 }, { "epoch": 0.8583604577027617, "grad_norm": 1.1004778146743774, "learning_rate": 6.371696599199353e-05, "loss": 0.6531, "step": 12790 }, { "epoch": 0.8584946813865306, "grad_norm": 1.4766899347305298, "learning_rate": 6.370651434642248e-05, "loss": 0.6686, "step": 12792 }, { "epoch": 0.8586289050702997, "grad_norm": 1.1126197576522827, "learning_rate": 6.36960620532456e-05, "loss": 0.5572, "step": 12794 }, { "epoch": 0.8587631287540687, "grad_norm": 1.0039613246917725, "learning_rate": 6.368560911295671e-05, "loss": 0.6369, "step": 12796 }, { "epoch": 0.8588973524378376, "grad_norm": 0.9731637239456177, "learning_rate": 6.367515552604971e-05, "loss": 0.5513, "step": 12798 }, { "epoch": 0.8590315761216066, "grad_norm": 0.9223629832267761, "learning_rate": 6.36647012930185e-05, "loss": 0.5354, "step": 12800 }, { "epoch": 0.8591657998053757, "grad_norm": 1.0003281831741333, "learning_rate": 6.365424641435702e-05, "loss": 0.6371, "step": 12802 }, { "epoch": 0.8593000234891447, "grad_norm": 1.3436627388000488, "learning_rate": 6.364379089055923e-05, "loss": 0.5978, "step": 12804 }, { "epoch": 0.8594342471729136, "grad_norm": 0.9740346074104309, "learning_rate": 6.363333472211917e-05, "loss": 0.5426, "step": 12806 }, { "epoch": 0.8595684708566826, "grad_norm": 0.9096996784210205, "learning_rate": 6.362287790953086e-05, "loss": 0.5132, "step": 12808 }, { "epoch": 0.8597026945404517, "grad_norm": 1.1795580387115479, "learning_rate": 6.361242045328834e-05, "loss": 0.5476, "step": 12810 }, { "epoch": 0.8598369182242207, "grad_norm": 1.140982747077942, "learning_rate": 6.360196235388574e-05, "loss": 0.5635, "step": 12812 }, { "epoch": 0.8599711419079896, "grad_norm": 1.1558973789215088, "learning_rate": 6.359150361181715e-05, "loss": 0.5672, "step": 12814 }, { "epoch": 0.8601053655917587, "grad_norm": 1.0639357566833496, "learning_rate": 6.358104422757673e-05, "loss": 0.6059, "step": 12816 }, { "epoch": 0.8602395892755277, "grad_norm": 1.082284688949585, "learning_rate": 6.357058420165872e-05, "loss": 0.5668, "step": 12818 }, { "epoch": 0.8603738129592967, "grad_norm": 1.2438132762908936, "learning_rate": 6.356012353455725e-05, "loss": 0.5804, "step": 12820 }, { "epoch": 0.8605080366430656, "grad_norm": 0.989525318145752, "learning_rate": 6.354966222676661e-05, "loss": 0.5388, "step": 12822 }, { "epoch": 0.8606422603268347, "grad_norm": 1.0800150632858276, "learning_rate": 6.353920027878108e-05, "loss": 0.5278, "step": 12824 }, { "epoch": 0.8607764840106037, "grad_norm": 1.059510588645935, "learning_rate": 6.352873769109498e-05, "loss": 0.5665, "step": 12826 }, { "epoch": 0.8609107076943727, "grad_norm": 1.0791071653366089, "learning_rate": 6.351827446420261e-05, "loss": 0.5637, "step": 12828 }, { "epoch": 0.8610449313781416, "grad_norm": 1.167945384979248, "learning_rate": 6.350781059859835e-05, "loss": 0.6062, "step": 12830 }, { "epoch": 0.8611791550619107, "grad_norm": 0.9925161600112915, "learning_rate": 6.349734609477661e-05, "loss": 0.5524, "step": 12832 }, { "epoch": 0.8613133787456797, "grad_norm": 1.006649374961853, "learning_rate": 6.348688095323181e-05, "loss": 0.5375, "step": 12834 }, { "epoch": 0.8614476024294486, "grad_norm": 2.9027419090270996, "learning_rate": 6.34764151744584e-05, "loss": 0.5353, "step": 12836 }, { "epoch": 0.8615818261132177, "grad_norm": 0.9968249797821045, "learning_rate": 6.346594875895088e-05, "loss": 0.5111, "step": 12838 }, { "epoch": 0.8617160497969867, "grad_norm": 1.0156421661376953, "learning_rate": 6.345548170720376e-05, "loss": 0.5524, "step": 12840 }, { "epoch": 0.8618502734807557, "grad_norm": 0.8986196517944336, "learning_rate": 6.344501401971157e-05, "loss": 0.5402, "step": 12842 }, { "epoch": 0.8619844971645246, "grad_norm": 1.3014073371887207, "learning_rate": 6.343454569696891e-05, "loss": 0.6122, "step": 12844 }, { "epoch": 0.8621187208482937, "grad_norm": 1.0112860202789307, "learning_rate": 6.34240767394704e-05, "loss": 0.5491, "step": 12846 }, { "epoch": 0.8622529445320627, "grad_norm": 1.1823545694351196, "learning_rate": 6.341360714771066e-05, "loss": 0.5835, "step": 12848 }, { "epoch": 0.8623871682158317, "grad_norm": 0.9981070756912231, "learning_rate": 6.340313692218435e-05, "loss": 0.646, "step": 12850 }, { "epoch": 0.8625213918996006, "grad_norm": 1.1531411409378052, "learning_rate": 6.339266606338619e-05, "loss": 0.6488, "step": 12852 }, { "epoch": 0.8626556155833697, "grad_norm": 0.8666185736656189, "learning_rate": 6.338219457181088e-05, "loss": 0.5265, "step": 12854 }, { "epoch": 0.8627898392671387, "grad_norm": 0.9972176551818848, "learning_rate": 6.33717224479532e-05, "loss": 0.6623, "step": 12856 }, { "epoch": 0.8629240629509077, "grad_norm": 1.0785518884658813, "learning_rate": 6.336124969230792e-05, "loss": 0.5663, "step": 12858 }, { "epoch": 0.8630582866346767, "grad_norm": 2.482130527496338, "learning_rate": 6.335077630536988e-05, "loss": 0.5221, "step": 12860 }, { "epoch": 0.8631925103184457, "grad_norm": 1.1714829206466675, "learning_rate": 6.33403022876339e-05, "loss": 0.5932, "step": 12862 }, { "epoch": 0.8633267340022147, "grad_norm": 1.0527064800262451, "learning_rate": 6.332982763959487e-05, "loss": 0.5974, "step": 12864 }, { "epoch": 0.8634609576859837, "grad_norm": 0.9523221850395203, "learning_rate": 6.33193523617477e-05, "loss": 0.5482, "step": 12866 }, { "epoch": 0.8635951813697527, "grad_norm": 1.0951387882232666, "learning_rate": 6.330887645458732e-05, "loss": 0.5733, "step": 12868 }, { "epoch": 0.8637294050535217, "grad_norm": 0.9741978049278259, "learning_rate": 6.32983999186087e-05, "loss": 0.5611, "step": 12870 }, { "epoch": 0.8638636287372907, "grad_norm": 1.015071988105774, "learning_rate": 6.328792275430682e-05, "loss": 0.5379, "step": 12872 }, { "epoch": 0.8639978524210596, "grad_norm": 1.0906835794448853, "learning_rate": 6.327744496217675e-05, "loss": 0.5316, "step": 12874 }, { "epoch": 0.8641320761048287, "grad_norm": 1.0001736879348755, "learning_rate": 6.32669665427135e-05, "loss": 0.5752, "step": 12876 }, { "epoch": 0.8642662997885977, "grad_norm": 0.9937818646430969, "learning_rate": 6.325648749641217e-05, "loss": 0.5676, "step": 12878 }, { "epoch": 0.8644005234723667, "grad_norm": 0.9688825607299805, "learning_rate": 6.324600782376788e-05, "loss": 0.5647, "step": 12880 }, { "epoch": 0.8645347471561357, "grad_norm": 1.5446380376815796, "learning_rate": 6.323552752527575e-05, "loss": 0.5929, "step": 12882 }, { "epoch": 0.8646689708399047, "grad_norm": 1.0048898458480835, "learning_rate": 6.322504660143099e-05, "loss": 0.5474, "step": 12884 }, { "epoch": 0.8648031945236737, "grad_norm": 1.2681150436401367, "learning_rate": 6.321456505272876e-05, "loss": 0.5835, "step": 12886 }, { "epoch": 0.8649374182074427, "grad_norm": 0.9933623671531677, "learning_rate": 6.320408287966433e-05, "loss": 0.6576, "step": 12888 }, { "epoch": 0.8650716418912117, "grad_norm": 1.0742228031158447, "learning_rate": 6.319360008273294e-05, "loss": 0.5655, "step": 12890 }, { "epoch": 0.8652058655749807, "grad_norm": 0.95859295129776, "learning_rate": 6.31831166624299e-05, "loss": 0.5351, "step": 12892 }, { "epoch": 0.8653400892587497, "grad_norm": 1.0715645551681519, "learning_rate": 6.317263261925052e-05, "loss": 0.6147, "step": 12894 }, { "epoch": 0.8654743129425188, "grad_norm": 1.0217869281768799, "learning_rate": 6.316214795369016e-05, "loss": 0.5772, "step": 12896 }, { "epoch": 0.8656085366262877, "grad_norm": 0.9852344989776611, "learning_rate": 6.315166266624418e-05, "loss": 0.5352, "step": 12898 }, { "epoch": 0.8657427603100567, "grad_norm": 1.04218590259552, "learning_rate": 6.314117675740801e-05, "loss": 0.6268, "step": 12900 }, { "epoch": 0.8658769839938257, "grad_norm": 1.0136308670043945, "learning_rate": 6.313069022767707e-05, "loss": 0.6501, "step": 12902 }, { "epoch": 0.8660112076775948, "grad_norm": 0.9950689673423767, "learning_rate": 6.312020307754684e-05, "loss": 0.632, "step": 12904 }, { "epoch": 0.8661454313613637, "grad_norm": 1.0392811298370361, "learning_rate": 6.31097153075128e-05, "loss": 0.6059, "step": 12906 }, { "epoch": 0.8662796550451327, "grad_norm": 1.0014914274215698, "learning_rate": 6.30992269180705e-05, "loss": 0.6234, "step": 12908 }, { "epoch": 0.8664138787289017, "grad_norm": 1.0886601209640503, "learning_rate": 6.30887379097155e-05, "loss": 0.5978, "step": 12910 }, { "epoch": 0.8665481024126707, "grad_norm": 1.0177531242370605, "learning_rate": 6.307824828294338e-05, "loss": 0.6496, "step": 12912 }, { "epoch": 0.8666823260964397, "grad_norm": 0.9890040159225464, "learning_rate": 6.306775803824971e-05, "loss": 0.5529, "step": 12914 }, { "epoch": 0.8668165497802087, "grad_norm": 1.1515992879867554, "learning_rate": 6.30572671761302e-05, "loss": 0.5655, "step": 12916 }, { "epoch": 0.8669507734639778, "grad_norm": 0.9655346274375916, "learning_rate": 6.304677569708047e-05, "loss": 0.551, "step": 12918 }, { "epoch": 0.8670849971477467, "grad_norm": 1.0232477188110352, "learning_rate": 6.303628360159625e-05, "loss": 0.5514, "step": 12920 }, { "epoch": 0.8672192208315157, "grad_norm": 0.9819419384002686, "learning_rate": 6.302579089017327e-05, "loss": 0.5802, "step": 12922 }, { "epoch": 0.8673534445152847, "grad_norm": 1.0228992700576782, "learning_rate": 6.301529756330728e-05, "loss": 0.5945, "step": 12924 }, { "epoch": 0.8674876681990538, "grad_norm": 1.1202229261398315, "learning_rate": 6.300480362149409e-05, "loss": 0.5603, "step": 12926 }, { "epoch": 0.8676218918828227, "grad_norm": 0.8481611609458923, "learning_rate": 6.299430906522948e-05, "loss": 0.5345, "step": 12928 }, { "epoch": 0.8677561155665917, "grad_norm": 0.9928791522979736, "learning_rate": 6.298381389500933e-05, "loss": 0.5645, "step": 12930 }, { "epoch": 0.8678903392503607, "grad_norm": 1.0613855123519897, "learning_rate": 6.297331811132951e-05, "loss": 0.6182, "step": 12932 }, { "epoch": 0.8680245629341298, "grad_norm": 0.9913033843040466, "learning_rate": 6.29628217146859e-05, "loss": 0.6105, "step": 12934 }, { "epoch": 0.8681587866178987, "grad_norm": 0.9503318071365356, "learning_rate": 6.295232470557447e-05, "loss": 0.6953, "step": 12936 }, { "epoch": 0.8682930103016677, "grad_norm": 1.128720760345459, "learning_rate": 6.294182708449117e-05, "loss": 0.5917, "step": 12938 }, { "epoch": 0.8684272339854368, "grad_norm": 1.2992370128631592, "learning_rate": 6.293132885193198e-05, "loss": 0.6229, "step": 12940 }, { "epoch": 0.8685614576692058, "grad_norm": 0.9851817488670349, "learning_rate": 6.292083000839292e-05, "loss": 0.5255, "step": 12942 }, { "epoch": 0.8686956813529747, "grad_norm": 1.0424292087554932, "learning_rate": 6.291033055437008e-05, "loss": 0.5911, "step": 12944 }, { "epoch": 0.8688299050367437, "grad_norm": 1.0020283460617065, "learning_rate": 6.28998304903595e-05, "loss": 0.5693, "step": 12946 }, { "epoch": 0.8689641287205128, "grad_norm": 1.1239510774612427, "learning_rate": 6.288932981685728e-05, "loss": 0.6025, "step": 12948 }, { "epoch": 0.8690983524042817, "grad_norm": 1.115296721458435, "learning_rate": 6.287882853435959e-05, "loss": 0.5697, "step": 12950 }, { "epoch": 0.8692325760880507, "grad_norm": 0.9717348217964172, "learning_rate": 6.286832664336255e-05, "loss": 0.5958, "step": 12952 }, { "epoch": 0.8693667997718197, "grad_norm": 1.0190660953521729, "learning_rate": 6.28578241443624e-05, "loss": 0.5641, "step": 12954 }, { "epoch": 0.8695010234555888, "grad_norm": 1.0290762186050415, "learning_rate": 6.284732103785535e-05, "loss": 0.5396, "step": 12956 }, { "epoch": 0.8696352471393577, "grad_norm": 0.993732750415802, "learning_rate": 6.283681732433762e-05, "loss": 0.5894, "step": 12958 }, { "epoch": 0.8697694708231267, "grad_norm": 0.9604769349098206, "learning_rate": 6.282631300430554e-05, "loss": 0.6063, "step": 12960 }, { "epoch": 0.8699036945068958, "grad_norm": 1.0046974420547485, "learning_rate": 6.281580807825538e-05, "loss": 0.5806, "step": 12962 }, { "epoch": 0.8700379181906648, "grad_norm": 1.1138249635696411, "learning_rate": 6.280530254668349e-05, "loss": 0.5356, "step": 12964 }, { "epoch": 0.8701721418744337, "grad_norm": 1.022376298904419, "learning_rate": 6.279479641008623e-05, "loss": 0.6455, "step": 12966 }, { "epoch": 0.8703063655582027, "grad_norm": 1.090772032737732, "learning_rate": 6.278428966896e-05, "loss": 0.6026, "step": 12968 }, { "epoch": 0.8704405892419718, "grad_norm": 0.9981614947319031, "learning_rate": 6.277378232380123e-05, "loss": 0.5845, "step": 12970 }, { "epoch": 0.8705748129257408, "grad_norm": 1.0998585224151611, "learning_rate": 6.276327437510636e-05, "loss": 0.6177, "step": 12972 }, { "epoch": 0.8707090366095097, "grad_norm": 0.9015064239501953, "learning_rate": 6.275276582337188e-05, "loss": 0.5417, "step": 12974 }, { "epoch": 0.8708432602932787, "grad_norm": 1.05171799659729, "learning_rate": 6.274225666909428e-05, "loss": 0.5902, "step": 12976 }, { "epoch": 0.8709774839770478, "grad_norm": 2.3501741886138916, "learning_rate": 6.273174691277012e-05, "loss": 0.5347, "step": 12978 }, { "epoch": 0.8711117076608168, "grad_norm": 1.110366940498352, "learning_rate": 6.272123655489595e-05, "loss": 0.546, "step": 12980 }, { "epoch": 0.8712459313445857, "grad_norm": 1.0878968238830566, "learning_rate": 6.271072559596836e-05, "loss": 0.673, "step": 12982 }, { "epoch": 0.8713801550283548, "grad_norm": 1.2534525394439697, "learning_rate": 6.270021403648399e-05, "loss": 0.5889, "step": 12984 }, { "epoch": 0.8715143787121238, "grad_norm": 1.1351861953735352, "learning_rate": 6.268970187693947e-05, "loss": 0.5402, "step": 12986 }, { "epoch": 0.8716486023958927, "grad_norm": 0.9459824562072754, "learning_rate": 6.267918911783152e-05, "loss": 0.5777, "step": 12988 }, { "epoch": 0.8717828260796617, "grad_norm": 0.9665629863739014, "learning_rate": 6.266867575965679e-05, "loss": 0.5718, "step": 12990 }, { "epoch": 0.8719170497634308, "grad_norm": 0.8945533633232117, "learning_rate": 6.265816180291205e-05, "loss": 0.5887, "step": 12992 }, { "epoch": 0.8720512734471998, "grad_norm": 1.2175227403640747, "learning_rate": 6.264764724809404e-05, "loss": 0.5936, "step": 12994 }, { "epoch": 0.8721854971309687, "grad_norm": 0.9384693503379822, "learning_rate": 6.263713209569958e-05, "loss": 0.5728, "step": 12996 }, { "epoch": 0.8723197208147377, "grad_norm": 1.171692132949829, "learning_rate": 6.262661634622549e-05, "loss": 0.5447, "step": 12998 }, { "epoch": 0.8724539444985068, "grad_norm": 0.9675195813179016, "learning_rate": 6.26161000001686e-05, "loss": 0.5513, "step": 13000 }, { "epoch": 0.8725881681822758, "grad_norm": 0.999107301235199, "learning_rate": 6.260558305802578e-05, "loss": 0.5405, "step": 13002 }, { "epoch": 0.8727223918660447, "grad_norm": 1.1518807411193848, "learning_rate": 6.259506552029396e-05, "loss": 0.6876, "step": 13004 }, { "epoch": 0.8728566155498138, "grad_norm": 1.0599391460418701, "learning_rate": 6.258454738747006e-05, "loss": 0.5701, "step": 13006 }, { "epoch": 0.8729908392335828, "grad_norm": 1.0742309093475342, "learning_rate": 6.257402866005105e-05, "loss": 0.5679, "step": 13008 }, { "epoch": 0.8731250629173518, "grad_norm": 1.1204947233200073, "learning_rate": 6.25635093385339e-05, "loss": 0.5685, "step": 13010 }, { "epoch": 0.8732592866011207, "grad_norm": 0.9725003838539124, "learning_rate": 6.255298942341564e-05, "loss": 0.5354, "step": 13012 }, { "epoch": 0.8733935102848898, "grad_norm": 1.033810019493103, "learning_rate": 6.254246891519332e-05, "loss": 0.6126, "step": 13014 }, { "epoch": 0.8735277339686588, "grad_norm": 1.3697429895401, "learning_rate": 6.2531947814364e-05, "loss": 0.5436, "step": 13016 }, { "epoch": 0.8736619576524278, "grad_norm": 1.0568641424179077, "learning_rate": 6.25214261214248e-05, "loss": 0.5932, "step": 13018 }, { "epoch": 0.8737961813361967, "grad_norm": 1.1706132888793945, "learning_rate": 6.251090383687283e-05, "loss": 0.5681, "step": 13020 }, { "epoch": 0.8739304050199658, "grad_norm": 0.9343190789222717, "learning_rate": 6.250038096120526e-05, "loss": 0.5876, "step": 13022 }, { "epoch": 0.8740646287037348, "grad_norm": 1.0320550203323364, "learning_rate": 6.248985749491926e-05, "loss": 0.6106, "step": 13024 }, { "epoch": 0.8741988523875037, "grad_norm": 1.1263083219528198, "learning_rate": 6.247933343851207e-05, "loss": 0.6176, "step": 13026 }, { "epoch": 0.8743330760712728, "grad_norm": 1.0735199451446533, "learning_rate": 6.246880879248089e-05, "loss": 0.5905, "step": 13028 }, { "epoch": 0.8744672997550418, "grad_norm": 1.0393165349960327, "learning_rate": 6.245828355732303e-05, "loss": 0.5268, "step": 13030 }, { "epoch": 0.8746015234388108, "grad_norm": 0.9082031846046448, "learning_rate": 6.244775773353577e-05, "loss": 0.5749, "step": 13032 }, { "epoch": 0.8747357471225797, "grad_norm": 1.1140505075454712, "learning_rate": 6.243723132161643e-05, "loss": 0.5703, "step": 13034 }, { "epoch": 0.8748699708063488, "grad_norm": 0.9214529991149902, "learning_rate": 6.242670432206238e-05, "loss": 0.5326, "step": 13036 }, { "epoch": 0.8750041944901178, "grad_norm": 0.938310444355011, "learning_rate": 6.241617673537097e-05, "loss": 0.6143, "step": 13038 }, { "epoch": 0.8751384181738868, "grad_norm": 1.0400673151016235, "learning_rate": 6.240564856203964e-05, "loss": 0.5455, "step": 13040 }, { "epoch": 0.8752726418576557, "grad_norm": 0.957240641117096, "learning_rate": 6.239511980256579e-05, "loss": 0.6131, "step": 13042 }, { "epoch": 0.8754068655414248, "grad_norm": 1.0981582403182983, "learning_rate": 6.238459045744691e-05, "loss": 0.5857, "step": 13044 }, { "epoch": 0.8755410892251938, "grad_norm": 0.9009153842926025, "learning_rate": 6.23740605271805e-05, "loss": 0.5194, "step": 13046 }, { "epoch": 0.8756753129089628, "grad_norm": 1.1389052867889404, "learning_rate": 6.236353001226404e-05, "loss": 0.5983, "step": 13048 }, { "epoch": 0.8758095365927318, "grad_norm": 1.2221170663833618, "learning_rate": 6.235299891319512e-05, "loss": 0.5634, "step": 13050 }, { "epoch": 0.8759437602765008, "grad_norm": 1.0262128114700317, "learning_rate": 6.234246723047128e-05, "loss": 0.5843, "step": 13052 }, { "epoch": 0.8760779839602698, "grad_norm": 1.1158145666122437, "learning_rate": 6.233193496459015e-05, "loss": 0.6331, "step": 13054 }, { "epoch": 0.8762122076440388, "grad_norm": 0.9433540105819702, "learning_rate": 6.232140211604932e-05, "loss": 0.5434, "step": 13056 }, { "epoch": 0.8763464313278078, "grad_norm": 0.9933052659034729, "learning_rate": 6.231086868534647e-05, "loss": 0.5048, "step": 13058 }, { "epoch": 0.8764806550115768, "grad_norm": 0.9645562767982483, "learning_rate": 6.23003346729793e-05, "loss": 0.5822, "step": 13060 }, { "epoch": 0.8766148786953458, "grad_norm": 0.9403725862503052, "learning_rate": 6.22898000794455e-05, "loss": 0.5308, "step": 13062 }, { "epoch": 0.8767491023791147, "grad_norm": 1.009995937347412, "learning_rate": 6.22792649052428e-05, "loss": 0.5839, "step": 13064 }, { "epoch": 0.8768833260628838, "grad_norm": 0.9146338105201721, "learning_rate": 6.226872915086899e-05, "loss": 0.5643, "step": 13066 }, { "epoch": 0.8770175497466528, "grad_norm": 2.256757974624634, "learning_rate": 6.225819281682186e-05, "loss": 0.5551, "step": 13068 }, { "epoch": 0.8771517734304218, "grad_norm": 1.0844676494598389, "learning_rate": 6.22476559035992e-05, "loss": 0.5995, "step": 13070 }, { "epoch": 0.8772859971141908, "grad_norm": 1.0413196086883545, "learning_rate": 6.22371184116989e-05, "loss": 0.6019, "step": 13072 }, { "epoch": 0.8774202207979598, "grad_norm": 1.3758057355880737, "learning_rate": 6.22265803416188e-05, "loss": 0.5993, "step": 13074 }, { "epoch": 0.8775544444817288, "grad_norm": 0.9704021215438843, "learning_rate": 6.221604169385685e-05, "loss": 0.5646, "step": 13076 }, { "epoch": 0.8776886681654978, "grad_norm": 1.0348174571990967, "learning_rate": 6.220550246891095e-05, "loss": 0.656, "step": 13078 }, { "epoch": 0.8778228918492668, "grad_norm": 1.1081150770187378, "learning_rate": 6.219496266727904e-05, "loss": 0.6423, "step": 13080 }, { "epoch": 0.8779571155330358, "grad_norm": 1.0098882913589478, "learning_rate": 6.218442228945914e-05, "loss": 0.5286, "step": 13082 }, { "epoch": 0.8780913392168048, "grad_norm": 1.0357965230941772, "learning_rate": 6.217388133594925e-05, "loss": 0.6186, "step": 13084 }, { "epoch": 0.8782255629005739, "grad_norm": 0.8720493316650391, "learning_rate": 6.216333980724738e-05, "loss": 0.5894, "step": 13086 }, { "epoch": 0.8783597865843428, "grad_norm": 0.8772887587547302, "learning_rate": 6.215279770385165e-05, "loss": 0.5529, "step": 13088 }, { "epoch": 0.8784940102681118, "grad_norm": 0.9870423078536987, "learning_rate": 6.21422550262601e-05, "loss": 0.6042, "step": 13090 }, { "epoch": 0.8786282339518808, "grad_norm": 0.9107111692428589, "learning_rate": 6.213171177497089e-05, "loss": 0.5348, "step": 13092 }, { "epoch": 0.8787624576356499, "grad_norm": 0.9593957662582397, "learning_rate": 6.212116795048216e-05, "loss": 0.5888, "step": 13094 }, { "epoch": 0.8788966813194188, "grad_norm": 0.9927076697349548, "learning_rate": 6.211062355329206e-05, "loss": 0.5603, "step": 13096 }, { "epoch": 0.8790309050031878, "grad_norm": 1.0677847862243652, "learning_rate": 6.210007858389885e-05, "loss": 0.5757, "step": 13098 }, { "epoch": 0.8791651286869568, "grad_norm": 1.0729224681854248, "learning_rate": 6.208953304280069e-05, "loss": 0.6295, "step": 13100 }, { "epoch": 0.8792993523707258, "grad_norm": 1.0005987882614136, "learning_rate": 6.207898693049588e-05, "loss": 0.5564, "step": 13102 }, { "epoch": 0.8794335760544948, "grad_norm": 0.9121020436286926, "learning_rate": 6.206844024748268e-05, "loss": 0.5318, "step": 13104 }, { "epoch": 0.8795677997382638, "grad_norm": 0.9779648780822754, "learning_rate": 6.205789299425941e-05, "loss": 0.5582, "step": 13106 }, { "epoch": 0.8797020234220329, "grad_norm": 1.0180506706237793, "learning_rate": 6.20473451713244e-05, "loss": 0.5855, "step": 13108 }, { "epoch": 0.8798362471058018, "grad_norm": 0.9515188932418823, "learning_rate": 6.203679677917603e-05, "loss": 0.5627, "step": 13110 }, { "epoch": 0.8799704707895708, "grad_norm": 1.2502775192260742, "learning_rate": 6.202624781831268e-05, "loss": 0.6602, "step": 13112 }, { "epoch": 0.8801046944733398, "grad_norm": 1.1253612041473389, "learning_rate": 6.201569828923277e-05, "loss": 0.6438, "step": 13114 }, { "epoch": 0.8802389181571089, "grad_norm": 1.0836929082870483, "learning_rate": 6.200514819243476e-05, "loss": 0.5762, "step": 13116 }, { "epoch": 0.8803731418408778, "grad_norm": 1.1149053573608398, "learning_rate": 6.199459752841709e-05, "loss": 0.5762, "step": 13118 }, { "epoch": 0.8805073655246468, "grad_norm": 1.0289145708084106, "learning_rate": 6.198404629767825e-05, "loss": 0.5773, "step": 13120 }, { "epoch": 0.8806415892084158, "grad_norm": 1.2018251419067383, "learning_rate": 6.197349450071683e-05, "loss": 0.5535, "step": 13122 }, { "epoch": 0.8807758128921849, "grad_norm": 0.9826356172561646, "learning_rate": 6.196294213803131e-05, "loss": 0.5523, "step": 13124 }, { "epoch": 0.8809100365759538, "grad_norm": 0.9500182867050171, "learning_rate": 6.195238921012032e-05, "loss": 0.5726, "step": 13126 }, { "epoch": 0.8810442602597228, "grad_norm": 1.1312055587768555, "learning_rate": 6.194183571748242e-05, "loss": 0.5705, "step": 13128 }, { "epoch": 0.8811784839434919, "grad_norm": 0.9225857257843018, "learning_rate": 6.193128166061629e-05, "loss": 0.5924, "step": 13130 }, { "epoch": 0.8813127076272609, "grad_norm": 1.067992925643921, "learning_rate": 6.192072704002054e-05, "loss": 0.4947, "step": 13132 }, { "epoch": 0.8814469313110298, "grad_norm": 0.8770390748977661, "learning_rate": 6.191017185619389e-05, "loss": 0.5097, "step": 13134 }, { "epoch": 0.8815811549947988, "grad_norm": 1.0908836126327515, "learning_rate": 6.189961610963504e-05, "loss": 0.5721, "step": 13136 }, { "epoch": 0.8817153786785679, "grad_norm": 1.0002118349075317, "learning_rate": 6.188905980084273e-05, "loss": 0.545, "step": 13138 }, { "epoch": 0.8818496023623368, "grad_norm": 1.091551661491394, "learning_rate": 6.187850293031571e-05, "loss": 0.6389, "step": 13140 }, { "epoch": 0.8819838260461058, "grad_norm": 0.9426711797714233, "learning_rate": 6.18679454985528e-05, "loss": 0.5249, "step": 13142 }, { "epoch": 0.8821180497298748, "grad_norm": 1.0135663747787476, "learning_rate": 6.185738750605281e-05, "loss": 0.5869, "step": 13144 }, { "epoch": 0.8822522734136439, "grad_norm": 0.9642865061759949, "learning_rate": 6.184682895331456e-05, "loss": 0.5078, "step": 13146 }, { "epoch": 0.8823864970974128, "grad_norm": 1.0576778650283813, "learning_rate": 6.183626984083694e-05, "loss": 0.5404, "step": 13148 }, { "epoch": 0.8825207207811818, "grad_norm": 1.1490014791488647, "learning_rate": 6.182571016911886e-05, "loss": 0.6374, "step": 13150 }, { "epoch": 0.8826549444649509, "grad_norm": 1.1898303031921387, "learning_rate": 6.181514993865922e-05, "loss": 0.5432, "step": 13152 }, { "epoch": 0.8827891681487199, "grad_norm": 1.1843942403793335, "learning_rate": 6.180458914995696e-05, "loss": 0.6375, "step": 13154 }, { "epoch": 0.8829233918324888, "grad_norm": 1.141417145729065, "learning_rate": 6.17940278035111e-05, "loss": 0.5919, "step": 13156 }, { "epoch": 0.8830576155162578, "grad_norm": 0.9610231518745422, "learning_rate": 6.178346589982061e-05, "loss": 0.5513, "step": 13158 }, { "epoch": 0.8831918392000269, "grad_norm": 1.13080632686615, "learning_rate": 6.177290343938452e-05, "loss": 0.6417, "step": 13160 }, { "epoch": 0.8833260628837959, "grad_norm": 1.237656593322754, "learning_rate": 6.17623404227019e-05, "loss": 0.5621, "step": 13162 }, { "epoch": 0.8834602865675648, "grad_norm": 1.0151150226593018, "learning_rate": 6.175177685027183e-05, "loss": 0.4725, "step": 13164 }, { "epoch": 0.8835945102513338, "grad_norm": 1.2543600797653198, "learning_rate": 6.17412127225934e-05, "loss": 0.568, "step": 13166 }, { "epoch": 0.8837287339351029, "grad_norm": 1.0492841005325317, "learning_rate": 6.173064804016577e-05, "loss": 0.5572, "step": 13168 }, { "epoch": 0.8838629576188719, "grad_norm": 0.9958816170692444, "learning_rate": 6.172008280348808e-05, "loss": 0.6348, "step": 13170 }, { "epoch": 0.8839971813026408, "grad_norm": 0.9536067843437195, "learning_rate": 6.170951701305951e-05, "loss": 0.5068, "step": 13172 }, { "epoch": 0.8841314049864099, "grad_norm": 0.9734018445014954, "learning_rate": 6.169895066937932e-05, "loss": 0.5774, "step": 13174 }, { "epoch": 0.8842656286701789, "grad_norm": 0.978851854801178, "learning_rate": 6.168838377294669e-05, "loss": 0.5614, "step": 13176 }, { "epoch": 0.8843998523539478, "grad_norm": 0.9927994012832642, "learning_rate": 6.167781632426093e-05, "loss": 0.7294, "step": 13178 }, { "epoch": 0.8845340760377168, "grad_norm": 1.0294978618621826, "learning_rate": 6.16672483238213e-05, "loss": 0.6088, "step": 13180 }, { "epoch": 0.8846682997214859, "grad_norm": 1.0444254875183105, "learning_rate": 6.165667977212712e-05, "loss": 0.5596, "step": 13182 }, { "epoch": 0.8848025234052549, "grad_norm": 0.9945619702339172, "learning_rate": 6.164611066967775e-05, "loss": 0.5794, "step": 13184 }, { "epoch": 0.8849367470890238, "grad_norm": 1.567557692527771, "learning_rate": 6.163554101697256e-05, "loss": 0.6003, "step": 13186 }, { "epoch": 0.8850709707727928, "grad_norm": 1.062455177307129, "learning_rate": 6.162497081451093e-05, "loss": 0.5674, "step": 13188 }, { "epoch": 0.8852051944565619, "grad_norm": 1.147298812866211, "learning_rate": 6.161440006279228e-05, "loss": 0.5886, "step": 13190 }, { "epoch": 0.8853394181403309, "grad_norm": 1.4330294132232666, "learning_rate": 6.16038287623161e-05, "loss": 0.5848, "step": 13192 }, { "epoch": 0.8854736418240998, "grad_norm": 0.9991915225982666, "learning_rate": 6.15932569135818e-05, "loss": 0.594, "step": 13194 }, { "epoch": 0.8856078655078689, "grad_norm": 1.1148768663406372, "learning_rate": 6.15826845170889e-05, "loss": 0.6451, "step": 13196 }, { "epoch": 0.8857420891916379, "grad_norm": 1.01297128200531, "learning_rate": 6.157211157333695e-05, "loss": 0.5847, "step": 13198 }, { "epoch": 0.8858763128754069, "grad_norm": 1.046197533607483, "learning_rate": 6.156153808282548e-05, "loss": 0.5754, "step": 13200 }, { "epoch": 0.8860105365591758, "grad_norm": 2.0050272941589355, "learning_rate": 6.155096404605408e-05, "loss": 0.6177, "step": 13202 }, { "epoch": 0.8861447602429449, "grad_norm": 1.025641918182373, "learning_rate": 6.154038946352232e-05, "loss": 0.5976, "step": 13204 }, { "epoch": 0.8862789839267139, "grad_norm": 1.0258253812789917, "learning_rate": 6.152981433572987e-05, "loss": 0.5656, "step": 13206 }, { "epoch": 0.8864132076104829, "grad_norm": 1.1024008989334106, "learning_rate": 6.151923866317636e-05, "loss": 0.6362, "step": 13208 }, { "epoch": 0.8865474312942518, "grad_norm": 0.9120305180549622, "learning_rate": 6.150866244636146e-05, "loss": 0.6099, "step": 13210 }, { "epoch": 0.8866816549780209, "grad_norm": 0.9940522313117981, "learning_rate": 6.149808568578492e-05, "loss": 0.6147, "step": 13212 }, { "epoch": 0.8868158786617899, "grad_norm": 0.9856640696525574, "learning_rate": 6.148750838194642e-05, "loss": 0.559, "step": 13214 }, { "epoch": 0.8869501023455588, "grad_norm": 1.1462854146957397, "learning_rate": 6.147693053534575e-05, "loss": 0.5442, "step": 13216 }, { "epoch": 0.8870843260293279, "grad_norm": 1.1151323318481445, "learning_rate": 6.146635214648267e-05, "loss": 0.6204, "step": 13218 }, { "epoch": 0.8872185497130969, "grad_norm": 1.0104786157608032, "learning_rate": 6.145577321585701e-05, "loss": 0.6074, "step": 13220 }, { "epoch": 0.8873527733968659, "grad_norm": 0.9442871809005737, "learning_rate": 6.144519374396859e-05, "loss": 0.5557, "step": 13222 }, { "epoch": 0.8874869970806348, "grad_norm": 1.0627042055130005, "learning_rate": 6.143461373131725e-05, "loss": 0.5659, "step": 13224 }, { "epoch": 0.8876212207644039, "grad_norm": 0.9679484963417053, "learning_rate": 6.142403317840292e-05, "loss": 0.5486, "step": 13226 }, { "epoch": 0.8877554444481729, "grad_norm": 1.0377732515335083, "learning_rate": 6.14134520857255e-05, "loss": 0.6623, "step": 13228 }, { "epoch": 0.8878896681319419, "grad_norm": 0.8365923762321472, "learning_rate": 6.140287045378488e-05, "loss": 0.5505, "step": 13230 }, { "epoch": 0.8880238918157108, "grad_norm": 1.0632461309432983, "learning_rate": 6.139228828308107e-05, "loss": 0.585, "step": 13232 }, { "epoch": 0.8881581154994799, "grad_norm": 0.8229578733444214, "learning_rate": 6.138170557411403e-05, "loss": 0.6767, "step": 13234 }, { "epoch": 0.8882923391832489, "grad_norm": 0.9627971649169922, "learning_rate": 6.13711223273838e-05, "loss": 0.5549, "step": 13236 }, { "epoch": 0.8884265628670179, "grad_norm": 1.4396005868911743, "learning_rate": 6.136053854339039e-05, "loss": 0.5491, "step": 13238 }, { "epoch": 0.8885607865507869, "grad_norm": 1.045028805732727, "learning_rate": 6.134995422263388e-05, "loss": 0.6166, "step": 13240 }, { "epoch": 0.8886950102345559, "grad_norm": 1.0063236951828003, "learning_rate": 6.133936936561432e-05, "loss": 0.5512, "step": 13242 }, { "epoch": 0.8888292339183249, "grad_norm": 1.044127345085144, "learning_rate": 6.132878397283189e-05, "loss": 0.5556, "step": 13244 }, { "epoch": 0.8889634576020939, "grad_norm": 1.0469932556152344, "learning_rate": 6.131819804478669e-05, "loss": 0.5538, "step": 13246 }, { "epoch": 0.8890976812858629, "grad_norm": 0.9266387224197388, "learning_rate": 6.130761158197888e-05, "loss": 0.5829, "step": 13248 }, { "epoch": 0.8892319049696319, "grad_norm": 1.0345057249069214, "learning_rate": 6.129702458490867e-05, "loss": 0.5726, "step": 13250 }, { "epoch": 0.8893661286534009, "grad_norm": 1.0493334531784058, "learning_rate": 6.128643705407625e-05, "loss": 0.5445, "step": 13252 }, { "epoch": 0.8895003523371698, "grad_norm": 1.1147289276123047, "learning_rate": 6.127584898998188e-05, "loss": 0.643, "step": 13254 }, { "epoch": 0.8896345760209389, "grad_norm": 1.0473716259002686, "learning_rate": 6.126526039312581e-05, "loss": 0.6176, "step": 13256 }, { "epoch": 0.8897687997047079, "grad_norm": 1.0527476072311401, "learning_rate": 6.125467126400835e-05, "loss": 0.6052, "step": 13258 }, { "epoch": 0.8899030233884769, "grad_norm": 1.0677533149719238, "learning_rate": 6.12440816031298e-05, "loss": 0.6273, "step": 13260 }, { "epoch": 0.8900372470722459, "grad_norm": 1.0257129669189453, "learning_rate": 6.12334914109905e-05, "loss": 0.606, "step": 13262 }, { "epoch": 0.8901714707560149, "grad_norm": 1.0978323221206665, "learning_rate": 6.122290068809083e-05, "loss": 0.6126, "step": 13264 }, { "epoch": 0.8903056944397839, "grad_norm": 1.1365458965301514, "learning_rate": 6.121230943493117e-05, "loss": 0.5748, "step": 13266 }, { "epoch": 0.8904399181235529, "grad_norm": 0.9947246313095093, "learning_rate": 6.120171765201194e-05, "loss": 0.5155, "step": 13268 }, { "epoch": 0.8905741418073219, "grad_norm": 1.1812353134155273, "learning_rate": 6.119112533983355e-05, "loss": 0.6175, "step": 13270 }, { "epoch": 0.8907083654910909, "grad_norm": 1.1084691286087036, "learning_rate": 6.118053249889652e-05, "loss": 0.5496, "step": 13272 }, { "epoch": 0.8908425891748599, "grad_norm": 1.022020936012268, "learning_rate": 6.116993912970132e-05, "loss": 0.5587, "step": 13274 }, { "epoch": 0.890976812858629, "grad_norm": 0.9779108166694641, "learning_rate": 6.115934523274845e-05, "loss": 0.5614, "step": 13276 }, { "epoch": 0.8911110365423979, "grad_norm": 1.0704686641693115, "learning_rate": 6.114875080853846e-05, "loss": 0.5601, "step": 13278 }, { "epoch": 0.8912452602261669, "grad_norm": 1.691609263420105, "learning_rate": 6.113815585757192e-05, "loss": 0.5689, "step": 13280 }, { "epoch": 0.8913794839099359, "grad_norm": 1.0745458602905273, "learning_rate": 6.112756038034942e-05, "loss": 0.585, "step": 13282 }, { "epoch": 0.891513707593705, "grad_norm": 1.0736366510391235, "learning_rate": 6.111696437737157e-05, "loss": 0.573, "step": 13284 }, { "epoch": 0.8916479312774739, "grad_norm": 0.9195927977561951, "learning_rate": 6.110636784913901e-05, "loss": 0.5705, "step": 13286 }, { "epoch": 0.8917821549612429, "grad_norm": 0.9990829825401306, "learning_rate": 6.109577079615243e-05, "loss": 0.6391, "step": 13288 }, { "epoch": 0.8919163786450119, "grad_norm": 0.964238703250885, "learning_rate": 6.108517321891247e-05, "loss": 0.6249, "step": 13290 }, { "epoch": 0.8920506023287809, "grad_norm": 1.1120233535766602, "learning_rate": 6.10745751179199e-05, "loss": 0.5978, "step": 13292 }, { "epoch": 0.8921848260125499, "grad_norm": 0.9856030941009521, "learning_rate": 6.106397649367541e-05, "loss": 0.5742, "step": 13294 }, { "epoch": 0.8923190496963189, "grad_norm": 1.4189636707305908, "learning_rate": 6.10533773466798e-05, "loss": 0.554, "step": 13296 }, { "epoch": 0.892453273380088, "grad_norm": 1.0698367357254028, "learning_rate": 6.104277767743385e-05, "loss": 0.5626, "step": 13298 }, { "epoch": 0.8925874970638569, "grad_norm": 0.9755097031593323, "learning_rate": 6.1032177486438355e-05, "loss": 0.5531, "step": 13300 }, { "epoch": 0.8927217207476259, "grad_norm": 1.0601940155029297, "learning_rate": 6.102157677419418e-05, "loss": 0.5307, "step": 13302 }, { "epoch": 0.8928559444313949, "grad_norm": 1.0407485961914062, "learning_rate": 6.101097554120216e-05, "loss": 0.544, "step": 13304 }, { "epoch": 0.892990168115164, "grad_norm": 1.1194372177124023, "learning_rate": 6.100037378796321e-05, "loss": 0.6405, "step": 13306 }, { "epoch": 0.8931243917989329, "grad_norm": 1.0852293968200684, "learning_rate": 6.0989771514978235e-05, "loss": 0.6079, "step": 13308 }, { "epoch": 0.8932586154827019, "grad_norm": 0.9562682509422302, "learning_rate": 6.097916872274815e-05, "loss": 0.5276, "step": 13310 }, { "epoch": 0.8933928391664709, "grad_norm": 1.0830674171447754, "learning_rate": 6.096856541177395e-05, "loss": 0.5368, "step": 13312 }, { "epoch": 0.89352706285024, "grad_norm": 1.0606974363327026, "learning_rate": 6.095796158255659e-05, "loss": 0.5347, "step": 13314 }, { "epoch": 0.8936612865340089, "grad_norm": 1.005293846130371, "learning_rate": 6.09473572355971e-05, "loss": 0.5241, "step": 13316 }, { "epoch": 0.8937955102177779, "grad_norm": 1.0626174211502075, "learning_rate": 6.09367523713965e-05, "loss": 0.6296, "step": 13318 }, { "epoch": 0.893929733901547, "grad_norm": 1.0723947286605835, "learning_rate": 6.092614699045587e-05, "loss": 0.5691, "step": 13320 }, { "epoch": 0.894063957585316, "grad_norm": 1.091525673866272, "learning_rate": 6.091554109327626e-05, "loss": 0.6199, "step": 13322 }, { "epoch": 0.8941981812690849, "grad_norm": 1.0702077150344849, "learning_rate": 6.090493468035882e-05, "loss": 0.5419, "step": 13324 }, { "epoch": 0.8943324049528539, "grad_norm": 1.0261483192443848, "learning_rate": 6.089432775220465e-05, "loss": 0.6493, "step": 13326 }, { "epoch": 0.894466628636623, "grad_norm": 0.816967785358429, "learning_rate": 6.088372030931491e-05, "loss": 0.5044, "step": 13328 }, { "epoch": 0.8946008523203919, "grad_norm": 1.0482172966003418, "learning_rate": 6.0873112352190795e-05, "loss": 0.6522, "step": 13330 }, { "epoch": 0.8947350760041609, "grad_norm": 1.0886648893356323, "learning_rate": 6.0862503881333496e-05, "loss": 0.5491, "step": 13332 }, { "epoch": 0.8948692996879299, "grad_norm": 1.7727330923080444, "learning_rate": 6.085189489724426e-05, "loss": 0.5805, "step": 13334 }, { "epoch": 0.895003523371699, "grad_norm": 0.9612252712249756, "learning_rate": 6.084128540042432e-05, "loss": 0.5659, "step": 13336 }, { "epoch": 0.8951377470554679, "grad_norm": 0.930299699306488, "learning_rate": 6.083067539137497e-05, "loss": 0.5867, "step": 13338 }, { "epoch": 0.8952719707392369, "grad_norm": 0.8667044639587402, "learning_rate": 6.0820064870597504e-05, "loss": 0.5225, "step": 13340 }, { "epoch": 0.895406194423006, "grad_norm": 1.191062092781067, "learning_rate": 6.0809453838593246e-05, "loss": 0.6223, "step": 13342 }, { "epoch": 0.895540418106775, "grad_norm": 0.9350284934043884, "learning_rate": 6.079884229586355e-05, "loss": 0.5615, "step": 13344 }, { "epoch": 0.8956746417905439, "grad_norm": 1.1160637140274048, "learning_rate": 6.0788230242909795e-05, "loss": 0.5936, "step": 13346 }, { "epoch": 0.8958088654743129, "grad_norm": 0.9354270100593567, "learning_rate": 6.077761768023337e-05, "loss": 0.5426, "step": 13348 }, { "epoch": 0.895943089158082, "grad_norm": 0.9717685580253601, "learning_rate": 6.076700460833571e-05, "loss": 0.5489, "step": 13350 }, { "epoch": 0.896077312841851, "grad_norm": 1.0427420139312744, "learning_rate": 6.075639102771824e-05, "loss": 0.4835, "step": 13352 }, { "epoch": 0.8962115365256199, "grad_norm": 1.0258979797363281, "learning_rate": 6.074577693888246e-05, "loss": 0.5937, "step": 13354 }, { "epoch": 0.8963457602093889, "grad_norm": 0.9997778534889221, "learning_rate": 6.073516234232985e-05, "loss": 0.594, "step": 13356 }, { "epoch": 0.896479983893158, "grad_norm": 0.9338006973266602, "learning_rate": 6.072454723856192e-05, "loss": 0.5378, "step": 13358 }, { "epoch": 0.896614207576927, "grad_norm": 1.1041646003723145, "learning_rate": 6.0713931628080236e-05, "loss": 0.5606, "step": 13360 }, { "epoch": 0.8967484312606959, "grad_norm": 1.0136053562164307, "learning_rate": 6.0703315511386336e-05, "loss": 0.5599, "step": 13362 }, { "epoch": 0.896882654944465, "grad_norm": 0.8840245604515076, "learning_rate": 6.069269888898184e-05, "loss": 0.5809, "step": 13364 }, { "epoch": 0.897016878628234, "grad_norm": 0.9898674488067627, "learning_rate": 6.0682081761368325e-05, "loss": 0.5791, "step": 13366 }, { "epoch": 0.8971511023120029, "grad_norm": 1.0291520357131958, "learning_rate": 6.0671464129047474e-05, "loss": 0.5538, "step": 13368 }, { "epoch": 0.8972853259957719, "grad_norm": 0.9146878123283386, "learning_rate": 6.066084599252092e-05, "loss": 0.5134, "step": 13370 }, { "epoch": 0.897419549679541, "grad_norm": 1.0117591619491577, "learning_rate": 6.0650227352290345e-05, "loss": 0.5649, "step": 13372 }, { "epoch": 0.89755377336331, "grad_norm": 1.171174168586731, "learning_rate": 6.063960820885749e-05, "loss": 0.5771, "step": 13374 }, { "epoch": 0.8976879970470789, "grad_norm": 1.0561974048614502, "learning_rate": 6.0628988562724054e-05, "loss": 0.5866, "step": 13376 }, { "epoch": 0.8978222207308479, "grad_norm": 1.0028769969940186, "learning_rate": 6.061836841439182e-05, "loss": 0.5668, "step": 13378 }, { "epoch": 0.897956444414617, "grad_norm": 1.009484887123108, "learning_rate": 6.060774776436255e-05, "loss": 0.5868, "step": 13380 }, { "epoch": 0.898090668098386, "grad_norm": 1.0336222648620605, "learning_rate": 6.059712661313807e-05, "loss": 0.5579, "step": 13382 }, { "epoch": 0.8982248917821549, "grad_norm": 0.9472618103027344, "learning_rate": 6.058650496122018e-05, "loss": 0.5292, "step": 13384 }, { "epoch": 0.898359115465924, "grad_norm": 1.9172183275222778, "learning_rate": 6.057588280911075e-05, "loss": 0.5298, "step": 13386 }, { "epoch": 0.898493339149693, "grad_norm": 1.035021424293518, "learning_rate": 6.056526015731166e-05, "loss": 0.5675, "step": 13388 }, { "epoch": 0.898627562833462, "grad_norm": 0.9642541408538818, "learning_rate": 6.05546370063248e-05, "loss": 0.4864, "step": 13390 }, { "epoch": 0.8987617865172309, "grad_norm": 2.293220043182373, "learning_rate": 6.054401335665211e-05, "loss": 0.563, "step": 13392 }, { "epoch": 0.898896010201, "grad_norm": 1.1741487979888916, "learning_rate": 6.05333892087955e-05, "loss": 0.5971, "step": 13394 }, { "epoch": 0.899030233884769, "grad_norm": 1.116698980331421, "learning_rate": 6.0522764563256985e-05, "loss": 0.6076, "step": 13396 }, { "epoch": 0.899164457568538, "grad_norm": 1.0133905410766602, "learning_rate": 6.0512139420538515e-05, "loss": 0.6022, "step": 13398 }, { "epoch": 0.8992986812523069, "grad_norm": 0.9247488975524902, "learning_rate": 6.050151378114214e-05, "loss": 0.6255, "step": 13400 }, { "epoch": 0.899432904936076, "grad_norm": 1.062780737876892, "learning_rate": 6.049088764556989e-05, "loss": 0.6129, "step": 13402 }, { "epoch": 0.899567128619845, "grad_norm": 1.0486704111099243, "learning_rate": 6.0480261014323826e-05, "loss": 0.5423, "step": 13404 }, { "epoch": 0.8997013523036139, "grad_norm": 1.0059887170791626, "learning_rate": 6.046963388790604e-05, "loss": 0.5905, "step": 13406 }, { "epoch": 0.899835575987383, "grad_norm": 3.0994739532470703, "learning_rate": 6.045900626681864e-05, "loss": 0.5409, "step": 13408 }, { "epoch": 0.899969799671152, "grad_norm": 1.0882151126861572, "learning_rate": 6.044837815156377e-05, "loss": 0.5289, "step": 13410 }, { "epoch": 0.900104023354921, "grad_norm": 1.0200914144515991, "learning_rate": 6.043774954264355e-05, "loss": 0.5814, "step": 13412 }, { "epoch": 0.9002382470386899, "grad_norm": 1.0724719762802124, "learning_rate": 6.0427120440560204e-05, "loss": 0.5838, "step": 13414 }, { "epoch": 0.900372470722459, "grad_norm": 0.8868136405944824, "learning_rate": 6.041649084581593e-05, "loss": 0.5186, "step": 13416 }, { "epoch": 0.900506694406228, "grad_norm": 1.1544818878173828, "learning_rate": 6.040586075891293e-05, "loss": 0.6216, "step": 13418 }, { "epoch": 0.900640918089997, "grad_norm": 0.9154133200645447, "learning_rate": 6.0395230180353504e-05, "loss": 0.6012, "step": 13420 }, { "epoch": 0.9007751417737659, "grad_norm": 1.0481500625610352, "learning_rate": 6.038459911063986e-05, "loss": 0.5635, "step": 13422 }, { "epoch": 0.900909365457535, "grad_norm": 0.9151366353034973, "learning_rate": 6.0373967550274336e-05, "loss": 0.5563, "step": 13424 }, { "epoch": 0.901043589141304, "grad_norm": 1.0725773572921753, "learning_rate": 6.0363335499759265e-05, "loss": 0.5719, "step": 13426 }, { "epoch": 0.901177812825073, "grad_norm": 1.158782958984375, "learning_rate": 6.035270295959695e-05, "loss": 0.5878, "step": 13428 }, { "epoch": 0.901312036508842, "grad_norm": 1.0959148406982422, "learning_rate": 6.034206993028979e-05, "loss": 0.5669, "step": 13430 }, { "epoch": 0.901446260192611, "grad_norm": 0.9470241069793701, "learning_rate": 6.0331436412340147e-05, "loss": 0.6154, "step": 13432 }, { "epoch": 0.90158048387638, "grad_norm": 1.0807803869247437, "learning_rate": 6.032080240625045e-05, "loss": 0.5643, "step": 13434 }, { "epoch": 0.901714707560149, "grad_norm": 0.996346652507782, "learning_rate": 6.031016791252315e-05, "loss": 0.6318, "step": 13436 }, { "epoch": 0.901848931243918, "grad_norm": 0.9364991188049316, "learning_rate": 6.0299532931660675e-05, "loss": 0.5382, "step": 13438 }, { "epoch": 0.901983154927687, "grad_norm": 1.0150824785232544, "learning_rate": 6.028889746416553e-05, "loss": 0.5568, "step": 13440 }, { "epoch": 0.902117378611456, "grad_norm": 1.0036660432815552, "learning_rate": 6.0278261510540214e-05, "loss": 0.5679, "step": 13442 }, { "epoch": 0.9022516022952249, "grad_norm": 0.9905140995979309, "learning_rate": 6.026762507128725e-05, "loss": 0.5577, "step": 13444 }, { "epoch": 0.902385825978994, "grad_norm": 1.0355485677719116, "learning_rate": 6.025698814690919e-05, "loss": 0.5882, "step": 13446 }, { "epoch": 0.902520049662763, "grad_norm": 1.0384221076965332, "learning_rate": 6.0246350737908605e-05, "loss": 0.6298, "step": 13448 }, { "epoch": 0.902654273346532, "grad_norm": 0.9836850762367249, "learning_rate": 6.02357128447881e-05, "loss": 0.5671, "step": 13450 }, { "epoch": 0.902788497030301, "grad_norm": 0.9347818493843079, "learning_rate": 6.022507446805029e-05, "loss": 0.5293, "step": 13452 }, { "epoch": 0.90292272071407, "grad_norm": 0.9558823704719543, "learning_rate": 6.0214435608197825e-05, "loss": 0.5222, "step": 13454 }, { "epoch": 0.903056944397839, "grad_norm": 7.560023307800293, "learning_rate": 6.020379626573336e-05, "loss": 0.5481, "step": 13456 }, { "epoch": 0.903191168081608, "grad_norm": 1.25991690158844, "learning_rate": 6.019315644115959e-05, "loss": 0.5572, "step": 13458 }, { "epoch": 0.903325391765377, "grad_norm": 0.9077351093292236, "learning_rate": 6.018251613497922e-05, "loss": 0.595, "step": 13460 }, { "epoch": 0.903459615449146, "grad_norm": 1.0868170261383057, "learning_rate": 6.0171875347694974e-05, "loss": 0.5016, "step": 13462 }, { "epoch": 0.903593839132915, "grad_norm": 0.8331081867218018, "learning_rate": 6.0161234079809635e-05, "loss": 0.53, "step": 13464 }, { "epoch": 0.903728062816684, "grad_norm": 1.0237597227096558, "learning_rate": 6.015059233182596e-05, "loss": 0.6166, "step": 13466 }, { "epoch": 0.903862286500453, "grad_norm": 0.9407981038093567, "learning_rate": 6.013995010424676e-05, "loss": 0.5412, "step": 13468 }, { "epoch": 0.903996510184222, "grad_norm": 0.8044424057006836, "learning_rate": 6.012930739757485e-05, "loss": 0.5489, "step": 13470 }, { "epoch": 0.904130733867991, "grad_norm": 1.1213793754577637, "learning_rate": 6.011866421231309e-05, "loss": 0.5933, "step": 13472 }, { "epoch": 0.9042649575517601, "grad_norm": 1.1691128015518188, "learning_rate": 6.010802054896435e-05, "loss": 0.6067, "step": 13474 }, { "epoch": 0.904399181235529, "grad_norm": 1.0560898780822754, "learning_rate": 6.0097376408031504e-05, "loss": 0.5358, "step": 13476 }, { "epoch": 0.904533404919298, "grad_norm": 1.1528799533843994, "learning_rate": 6.008673179001748e-05, "loss": 0.6544, "step": 13478 }, { "epoch": 0.904667628603067, "grad_norm": 0.9637367129325867, "learning_rate": 6.007608669542522e-05, "loss": 0.5733, "step": 13480 }, { "epoch": 0.904801852286836, "grad_norm": 1.0823876857757568, "learning_rate": 6.006544112475767e-05, "loss": 0.5851, "step": 13482 }, { "epoch": 0.904936075970605, "grad_norm": 1.0340442657470703, "learning_rate": 6.0054795078517826e-05, "loss": 0.6026, "step": 13484 }, { "epoch": 0.905070299654374, "grad_norm": 1.4215126037597656, "learning_rate": 6.004414855720869e-05, "loss": 0.5864, "step": 13486 }, { "epoch": 0.905204523338143, "grad_norm": 1.04523503780365, "learning_rate": 6.003350156133327e-05, "loss": 0.5672, "step": 13488 }, { "epoch": 0.905338747021912, "grad_norm": 0.9753003716468811, "learning_rate": 6.002285409139464e-05, "loss": 0.6201, "step": 13490 }, { "epoch": 0.905472970705681, "grad_norm": 1.0781129598617554, "learning_rate": 6.001220614789587e-05, "loss": 0.6049, "step": 13492 }, { "epoch": 0.90560719438945, "grad_norm": 0.9975566267967224, "learning_rate": 6.000155773134005e-05, "loss": 0.5173, "step": 13494 }, { "epoch": 0.9057414180732191, "grad_norm": 1.2302011251449585, "learning_rate": 5.999090884223029e-05, "loss": 0.6304, "step": 13496 }, { "epoch": 0.905875641756988, "grad_norm": 1.0563715696334839, "learning_rate": 5.998025948106973e-05, "loss": 0.5148, "step": 13498 }, { "epoch": 0.906009865440757, "grad_norm": 1.1134657859802246, "learning_rate": 5.9969609648361526e-05, "loss": 0.588, "step": 13500 }, { "epoch": 0.906144089124526, "grad_norm": 1.0876972675323486, "learning_rate": 5.995895934460889e-05, "loss": 0.6915, "step": 13502 }, { "epoch": 0.9062783128082951, "grad_norm": 0.9595404267311096, "learning_rate": 5.994830857031499e-05, "loss": 0.4885, "step": 13504 }, { "epoch": 0.906412536492064, "grad_norm": 0.8634196519851685, "learning_rate": 5.9937657325983086e-05, "loss": 0.4999, "step": 13506 }, { "epoch": 0.906546760175833, "grad_norm": 1.1488295793533325, "learning_rate": 5.992700561211641e-05, "loss": 0.6347, "step": 13508 }, { "epoch": 0.906680983859602, "grad_norm": 1.0268058776855469, "learning_rate": 5.991635342921823e-05, "loss": 0.5978, "step": 13510 }, { "epoch": 0.9068152075433711, "grad_norm": 1.0089184045791626, "learning_rate": 5.9905700777791864e-05, "loss": 0.5651, "step": 13512 }, { "epoch": 0.90694943122714, "grad_norm": 1.0301094055175781, "learning_rate": 5.989504765834061e-05, "loss": 0.5734, "step": 13514 }, { "epoch": 0.907083654910909, "grad_norm": 0.872610867023468, "learning_rate": 5.9884394071367814e-05, "loss": 0.5492, "step": 13516 }, { "epoch": 0.9072178785946781, "grad_norm": 0.9029110074043274, "learning_rate": 5.9873740017376825e-05, "loss": 0.5531, "step": 13518 }, { "epoch": 0.907352102278447, "grad_norm": 1.2001795768737793, "learning_rate": 5.986308549687105e-05, "loss": 0.5644, "step": 13520 }, { "epoch": 0.907486325962216, "grad_norm": 1.0047690868377686, "learning_rate": 5.9852430510353876e-05, "loss": 0.5968, "step": 13522 }, { "epoch": 0.907620549645985, "grad_norm": 0.9836839437484741, "learning_rate": 5.984177505832872e-05, "loss": 0.6166, "step": 13524 }, { "epoch": 0.9077547733297541, "grad_norm": 1.0714616775512695, "learning_rate": 5.9831119141299064e-05, "loss": 0.608, "step": 13526 }, { "epoch": 0.907888997013523, "grad_norm": 1.010029911994934, "learning_rate": 5.982046275976836e-05, "loss": 0.5505, "step": 13528 }, { "epoch": 0.908023220697292, "grad_norm": 1.1323999166488647, "learning_rate": 5.9809805914240104e-05, "loss": 0.5769, "step": 13530 }, { "epoch": 0.908157444381061, "grad_norm": 1.1161619424819946, "learning_rate": 5.979914860521779e-05, "loss": 0.6016, "step": 13532 }, { "epoch": 0.9082916680648301, "grad_norm": 0.9982798099517822, "learning_rate": 5.978849083320499e-05, "loss": 0.5088, "step": 13534 }, { "epoch": 0.908425891748599, "grad_norm": 0.9984214305877686, "learning_rate": 5.977783259870524e-05, "loss": 0.561, "step": 13536 }, { "epoch": 0.908560115432368, "grad_norm": 1.203289270401001, "learning_rate": 5.9767173902222116e-05, "loss": 0.6009, "step": 13538 }, { "epoch": 0.9086943391161371, "grad_norm": 1.0529379844665527, "learning_rate": 5.975651474425925e-05, "loss": 0.5708, "step": 13540 }, { "epoch": 0.9088285627999061, "grad_norm": 0.941807746887207, "learning_rate": 5.9745855125320236e-05, "loss": 0.5475, "step": 13542 }, { "epoch": 0.908962786483675, "grad_norm": 0.9625012278556824, "learning_rate": 5.973519504590874e-05, "loss": 0.5515, "step": 13544 }, { "epoch": 0.909097010167444, "grad_norm": 1.1061104536056519, "learning_rate": 5.972453450652842e-05, "loss": 0.6741, "step": 13546 }, { "epoch": 0.9092312338512131, "grad_norm": 1.1080636978149414, "learning_rate": 5.971387350768297e-05, "loss": 0.5872, "step": 13548 }, { "epoch": 0.9093654575349821, "grad_norm": 1.0175831317901611, "learning_rate": 5.970321204987609e-05, "loss": 0.5422, "step": 13550 }, { "epoch": 0.909499681218751, "grad_norm": 1.123096227645874, "learning_rate": 5.969255013361153e-05, "loss": 0.6282, "step": 13552 }, { "epoch": 0.90963390490252, "grad_norm": 0.9660599827766418, "learning_rate": 5.968188775939303e-05, "loss": 0.5618, "step": 13554 }, { "epoch": 0.9097681285862891, "grad_norm": 1.6226638555526733, "learning_rate": 5.9671224927724366e-05, "loss": 0.5281, "step": 13556 }, { "epoch": 0.909902352270058, "grad_norm": 1.0370644330978394, "learning_rate": 5.966056163910936e-05, "loss": 0.6161, "step": 13558 }, { "epoch": 0.910036575953827, "grad_norm": 0.9805968403816223, "learning_rate": 5.9649897894051785e-05, "loss": 0.5821, "step": 13560 }, { "epoch": 0.9101707996375961, "grad_norm": 1.1472930908203125, "learning_rate": 5.963923369305554e-05, "loss": 0.5705, "step": 13562 }, { "epoch": 0.9103050233213651, "grad_norm": 1.0938692092895508, "learning_rate": 5.9628569036624446e-05, "loss": 0.5824, "step": 13564 }, { "epoch": 0.910439247005134, "grad_norm": 1.1195660829544067, "learning_rate": 5.9617903925262395e-05, "loss": 0.5666, "step": 13566 }, { "epoch": 0.910573470688903, "grad_norm": 1.0558995008468628, "learning_rate": 5.960723835947331e-05, "loss": 0.6008, "step": 13568 }, { "epoch": 0.9107076943726721, "grad_norm": 1.1444767713546753, "learning_rate": 5.959657233976108e-05, "loss": 0.5366, "step": 13570 }, { "epoch": 0.9108419180564411, "grad_norm": 0.8425387144088745, "learning_rate": 5.9585905866629687e-05, "loss": 0.5151, "step": 13572 }, { "epoch": 0.91097614174021, "grad_norm": 1.035324215888977, "learning_rate": 5.95752389405831e-05, "loss": 0.5506, "step": 13574 }, { "epoch": 0.911110365423979, "grad_norm": 1.067969560623169, "learning_rate": 5.9564571562125294e-05, "loss": 0.5854, "step": 13576 }, { "epoch": 0.9112445891077481, "grad_norm": 1.0317015647888184, "learning_rate": 5.9553903731760295e-05, "loss": 0.5845, "step": 13578 }, { "epoch": 0.9113788127915171, "grad_norm": 1.012310266494751, "learning_rate": 5.954323544999213e-05, "loss": 0.5982, "step": 13580 }, { "epoch": 0.911513036475286, "grad_norm": 0.8723168969154358, "learning_rate": 5.953256671732487e-05, "loss": 0.5747, "step": 13582 }, { "epoch": 0.9116472601590551, "grad_norm": 0.97370445728302, "learning_rate": 5.952189753426255e-05, "loss": 0.5772, "step": 13584 }, { "epoch": 0.9117814838428241, "grad_norm": 1.5741316080093384, "learning_rate": 5.951122790130931e-05, "loss": 0.5825, "step": 13586 }, { "epoch": 0.9119157075265931, "grad_norm": 0.9457995891571045, "learning_rate": 5.950055781896926e-05, "loss": 0.5331, "step": 13588 }, { "epoch": 0.912049931210362, "grad_norm": 0.8744380474090576, "learning_rate": 5.948988728774652e-05, "loss": 0.5643, "step": 13590 }, { "epoch": 0.9121841548941311, "grad_norm": 0.9800755381584167, "learning_rate": 5.947921630814528e-05, "loss": 0.5018, "step": 13592 }, { "epoch": 0.9123183785779001, "grad_norm": 1.0435247421264648, "learning_rate": 5.94685448806697e-05, "loss": 0.6186, "step": 13594 }, { "epoch": 0.912452602261669, "grad_norm": 1.0551824569702148, "learning_rate": 5.945787300582401e-05, "loss": 0.6018, "step": 13596 }, { "epoch": 0.912586825945438, "grad_norm": 0.974545955657959, "learning_rate": 5.944720068411239e-05, "loss": 0.5904, "step": 13598 }, { "epoch": 0.9127210496292071, "grad_norm": 1.0696234703063965, "learning_rate": 5.943652791603913e-05, "loss": 0.5361, "step": 13600 }, { "epoch": 0.9128552733129761, "grad_norm": 1.0218651294708252, "learning_rate": 5.9425854702108494e-05, "loss": 0.5753, "step": 13602 }, { "epoch": 0.912989496996745, "grad_norm": 0.9226881861686707, "learning_rate": 5.941518104282474e-05, "loss": 0.5257, "step": 13604 }, { "epoch": 0.9131237206805141, "grad_norm": 0.9766383171081543, "learning_rate": 5.940450693869222e-05, "loss": 0.5771, "step": 13606 }, { "epoch": 0.9132579443642831, "grad_norm": 1.0304303169250488, "learning_rate": 5.939383239021521e-05, "loss": 0.5821, "step": 13608 }, { "epoch": 0.9133921680480521, "grad_norm": 1.0369923114776611, "learning_rate": 5.938315739789812e-05, "loss": 0.5398, "step": 13610 }, { "epoch": 0.913526391731821, "grad_norm": 0.9135134816169739, "learning_rate": 5.9372481962245274e-05, "loss": 0.5373, "step": 13612 }, { "epoch": 0.9136606154155901, "grad_norm": 1.0946376323699951, "learning_rate": 5.9361806083761084e-05, "loss": 0.5835, "step": 13614 }, { "epoch": 0.9137948390993591, "grad_norm": 0.9748871922492981, "learning_rate": 5.935112976294997e-05, "loss": 0.6036, "step": 13616 }, { "epoch": 0.9139290627831281, "grad_norm": 1.154877781867981, "learning_rate": 5.934045300031637e-05, "loss": 0.5261, "step": 13618 }, { "epoch": 0.914063286466897, "grad_norm": 1.0215812921524048, "learning_rate": 5.932977579636474e-05, "loss": 0.5244, "step": 13620 }, { "epoch": 0.9141975101506661, "grad_norm": 0.9126201868057251, "learning_rate": 5.9319098151599525e-05, "loss": 0.5494, "step": 13622 }, { "epoch": 0.9143317338344351, "grad_norm": 1.1465352773666382, "learning_rate": 5.9308420066525274e-05, "loss": 0.5866, "step": 13624 }, { "epoch": 0.9144659575182041, "grad_norm": 1.2223682403564453, "learning_rate": 5.9297741541646465e-05, "loss": 0.5492, "step": 13626 }, { "epoch": 0.9146001812019731, "grad_norm": 0.9171653985977173, "learning_rate": 5.9287062577467646e-05, "loss": 0.5318, "step": 13628 }, { "epoch": 0.9147344048857421, "grad_norm": 0.9185870289802551, "learning_rate": 5.92763831744934e-05, "loss": 0.6155, "step": 13630 }, { "epoch": 0.9148686285695111, "grad_norm": 1.004479169845581, "learning_rate": 5.926570333322828e-05, "loss": 0.5913, "step": 13632 }, { "epoch": 0.91500285225328, "grad_norm": 1.0511237382888794, "learning_rate": 5.9255023054176895e-05, "loss": 0.5884, "step": 13634 }, { "epoch": 0.9151370759370491, "grad_norm": 1.1387178897857666, "learning_rate": 5.924434233784388e-05, "loss": 0.6764, "step": 13636 }, { "epoch": 0.9152712996208181, "grad_norm": 0.9619917273521423, "learning_rate": 5.9233661184733856e-05, "loss": 0.5727, "step": 13638 }, { "epoch": 0.9154055233045871, "grad_norm": 0.9369518160820007, "learning_rate": 5.92229795953515e-05, "loss": 0.5183, "step": 13640 }, { "epoch": 0.915539746988356, "grad_norm": 0.9075946807861328, "learning_rate": 5.92122975702015e-05, "loss": 0.469, "step": 13642 }, { "epoch": 0.9156739706721251, "grad_norm": 0.9302048087120056, "learning_rate": 5.9201615109788555e-05, "loss": 0.5152, "step": 13644 }, { "epoch": 0.9158081943558941, "grad_norm": 1.0235142707824707, "learning_rate": 5.9190932214617376e-05, "loss": 0.5364, "step": 13646 }, { "epoch": 0.9159424180396631, "grad_norm": 1.1028952598571777, "learning_rate": 5.9180248885192735e-05, "loss": 0.6569, "step": 13648 }, { "epoch": 0.9160766417234321, "grad_norm": 1.1462273597717285, "learning_rate": 5.9169565122019386e-05, "loss": 0.5516, "step": 13650 }, { "epoch": 0.9162108654072011, "grad_norm": 1.035347819328308, "learning_rate": 5.915888092560212e-05, "loss": 0.5598, "step": 13652 }, { "epoch": 0.9163450890909701, "grad_norm": 0.9631198048591614, "learning_rate": 5.914819629644574e-05, "loss": 0.5827, "step": 13654 }, { "epoch": 0.9164793127747392, "grad_norm": 0.9795407652854919, "learning_rate": 5.913751123505506e-05, "loss": 0.606, "step": 13656 }, { "epoch": 0.9166135364585081, "grad_norm": 1.05812406539917, "learning_rate": 5.912682574193497e-05, "loss": 0.6198, "step": 13658 }, { "epoch": 0.9167477601422771, "grad_norm": 0.9368353486061096, "learning_rate": 5.911613981759029e-05, "loss": 0.5508, "step": 13660 }, { "epoch": 0.9168819838260461, "grad_norm": 1.0112839937210083, "learning_rate": 5.9105453462525915e-05, "loss": 0.5814, "step": 13662 }, { "epoch": 0.9170162075098152, "grad_norm": 1.1456836462020874, "learning_rate": 5.90947666772468e-05, "loss": 0.5903, "step": 13664 }, { "epoch": 0.9171504311935841, "grad_norm": 0.9879085421562195, "learning_rate": 5.9084079462257824e-05, "loss": 0.5769, "step": 13666 }, { "epoch": 0.9172846548773531, "grad_norm": 1.080277919769287, "learning_rate": 5.907339181806397e-05, "loss": 0.5787, "step": 13668 }, { "epoch": 0.9174188785611221, "grad_norm": 0.9881849884986877, "learning_rate": 5.906270374517019e-05, "loss": 0.5421, "step": 13670 }, { "epoch": 0.9175531022448911, "grad_norm": 1.0958977937698364, "learning_rate": 5.905201524408148e-05, "loss": 0.5934, "step": 13672 }, { "epoch": 0.9176873259286601, "grad_norm": 0.9930013418197632, "learning_rate": 5.9041326315302835e-05, "loss": 0.5825, "step": 13674 }, { "epoch": 0.9178215496124291, "grad_norm": 1.0344637632369995, "learning_rate": 5.903063695933931e-05, "loss": 0.6249, "step": 13676 }, { "epoch": 0.9179557732961982, "grad_norm": 1.037278413772583, "learning_rate": 5.9019947176695954e-05, "loss": 0.6541, "step": 13678 }, { "epoch": 0.9180899969799671, "grad_norm": 1.1754828691482544, "learning_rate": 5.900925696787783e-05, "loss": 0.5919, "step": 13680 }, { "epoch": 0.9182242206637361, "grad_norm": 0.9693192839622498, "learning_rate": 5.899856633339003e-05, "loss": 0.5956, "step": 13682 }, { "epoch": 0.9183584443475051, "grad_norm": 1.4770317077636719, "learning_rate": 5.898787527373766e-05, "loss": 0.5479, "step": 13684 }, { "epoch": 0.9184926680312742, "grad_norm": 1.070414662361145, "learning_rate": 5.897718378942586e-05, "loss": 0.5375, "step": 13686 }, { "epoch": 0.9186268917150431, "grad_norm": 1.149235725402832, "learning_rate": 5.8966491880959775e-05, "loss": 0.5922, "step": 13688 }, { "epoch": 0.9187611153988121, "grad_norm": 0.9528571367263794, "learning_rate": 5.895579954884458e-05, "loss": 0.5335, "step": 13690 }, { "epoch": 0.9188953390825811, "grad_norm": 1.0059806108474731, "learning_rate": 5.894510679358547e-05, "loss": 0.5089, "step": 13692 }, { "epoch": 0.9190295627663502, "grad_norm": 0.9775877594947815, "learning_rate": 5.8934413615687655e-05, "loss": 0.547, "step": 13694 }, { "epoch": 0.9191637864501191, "grad_norm": 1.1173903942108154, "learning_rate": 5.892372001565637e-05, "loss": 0.5916, "step": 13696 }, { "epoch": 0.9192980101338881, "grad_norm": 1.368699550628662, "learning_rate": 5.891302599399685e-05, "loss": 0.5281, "step": 13698 }, { "epoch": 0.9194322338176572, "grad_norm": 0.9711874127388, "learning_rate": 5.8902331551214387e-05, "loss": 0.5552, "step": 13700 }, { "epoch": 0.9195664575014262, "grad_norm": 1.1150704622268677, "learning_rate": 5.8891636687814276e-05, "loss": 0.5741, "step": 13702 }, { "epoch": 0.9197006811851951, "grad_norm": 1.1072198152542114, "learning_rate": 5.8880941404301795e-05, "loss": 0.5677, "step": 13704 }, { "epoch": 0.9198349048689641, "grad_norm": 1.1020877361297607, "learning_rate": 5.887024570118231e-05, "loss": 0.5875, "step": 13706 }, { "epoch": 0.9199691285527332, "grad_norm": 0.941753089427948, "learning_rate": 5.885954957896115e-05, "loss": 0.533, "step": 13708 }, { "epoch": 0.9201033522365021, "grad_norm": 0.9490309357643127, "learning_rate": 5.884885303814369e-05, "loss": 0.5541, "step": 13710 }, { "epoch": 0.9202375759202711, "grad_norm": 0.9945219159126282, "learning_rate": 5.8838156079235326e-05, "loss": 0.5433, "step": 13712 }, { "epoch": 0.9203717996040401, "grad_norm": 1.004615068435669, "learning_rate": 5.8827458702741465e-05, "loss": 0.5921, "step": 13714 }, { "epoch": 0.9205060232878092, "grad_norm": 0.942687451839447, "learning_rate": 5.881676090916756e-05, "loss": 0.5633, "step": 13716 }, { "epoch": 0.9206402469715781, "grad_norm": 1.033555507659912, "learning_rate": 5.880606269901902e-05, "loss": 0.5705, "step": 13718 }, { "epoch": 0.9207744706553471, "grad_norm": 1.065773367881775, "learning_rate": 5.879536407280134e-05, "loss": 0.5556, "step": 13720 }, { "epoch": 0.9209086943391162, "grad_norm": 1.1345205307006836, "learning_rate": 5.8784665031020004e-05, "loss": 0.6009, "step": 13722 }, { "epoch": 0.9210429180228852, "grad_norm": 1.0204540491104126, "learning_rate": 5.87739655741805e-05, "loss": 0.5334, "step": 13724 }, { "epoch": 0.9211771417066541, "grad_norm": 0.9084629416465759, "learning_rate": 5.8763265702788404e-05, "loss": 0.4941, "step": 13726 }, { "epoch": 0.9213113653904231, "grad_norm": 1.0169886350631714, "learning_rate": 5.8752565417349215e-05, "loss": 0.5576, "step": 13728 }, { "epoch": 0.9214455890741922, "grad_norm": 0.9902750849723816, "learning_rate": 5.874186471836854e-05, "loss": 0.6228, "step": 13730 }, { "epoch": 0.9215798127579612, "grad_norm": 1.0019904375076294, "learning_rate": 5.8731163606351933e-05, "loss": 0.6261, "step": 13732 }, { "epoch": 0.9217140364417301, "grad_norm": 0.9995107650756836, "learning_rate": 5.872046208180503e-05, "loss": 0.5553, "step": 13734 }, { "epoch": 0.9218482601254991, "grad_norm": 1.1056007146835327, "learning_rate": 5.8709760145233416e-05, "loss": 0.5882, "step": 13736 }, { "epoch": 0.9219824838092682, "grad_norm": 0.9298726916313171, "learning_rate": 5.869905779714278e-05, "loss": 0.5271, "step": 13738 }, { "epoch": 0.9221167074930371, "grad_norm": 1.219878911972046, "learning_rate": 5.868835503803877e-05, "loss": 0.6204, "step": 13740 }, { "epoch": 0.9222509311768061, "grad_norm": 1.0206249952316284, "learning_rate": 5.867765186842706e-05, "loss": 0.5647, "step": 13742 }, { "epoch": 0.9223851548605752, "grad_norm": 0.9591622948646545, "learning_rate": 5.866694828881337e-05, "loss": 0.5728, "step": 13744 }, { "epoch": 0.9225193785443442, "grad_norm": 1.1038897037506104, "learning_rate": 5.8656244299703414e-05, "loss": 0.591, "step": 13746 }, { "epoch": 0.9226536022281131, "grad_norm": 1.1192244291305542, "learning_rate": 5.864553990160294e-05, "loss": 0.6176, "step": 13748 }, { "epoch": 0.9227878259118821, "grad_norm": 1.0274847745895386, "learning_rate": 5.86348350950177e-05, "loss": 0.6095, "step": 13750 }, { "epoch": 0.9229220495956512, "grad_norm": 0.9790551662445068, "learning_rate": 5.8624129880453485e-05, "loss": 0.5717, "step": 13752 }, { "epoch": 0.9230562732794202, "grad_norm": 1.0101349353790283, "learning_rate": 5.8613424258416094e-05, "loss": 0.6067, "step": 13754 }, { "epoch": 0.9231904969631891, "grad_norm": 1.0220857858657837, "learning_rate": 5.8602718229411335e-05, "loss": 0.5313, "step": 13756 }, { "epoch": 0.9233247206469581, "grad_norm": 0.9489063620567322, "learning_rate": 5.859201179394508e-05, "loss": 0.5373, "step": 13758 }, { "epoch": 0.9234589443307272, "grad_norm": 0.9276555180549622, "learning_rate": 5.858130495252314e-05, "loss": 0.5868, "step": 13760 }, { "epoch": 0.9235931680144962, "grad_norm": 1.358279824256897, "learning_rate": 5.857059770565142e-05, "loss": 0.5832, "step": 13762 }, { "epoch": 0.9237273916982651, "grad_norm": 1.062290072441101, "learning_rate": 5.855989005383581e-05, "loss": 0.5398, "step": 13764 }, { "epoch": 0.9238616153820342, "grad_norm": 1.0557093620300293, "learning_rate": 5.8549181997582225e-05, "loss": 0.6448, "step": 13766 }, { "epoch": 0.9239958390658032, "grad_norm": 1.2883045673370361, "learning_rate": 5.853847353739663e-05, "loss": 0.5191, "step": 13768 }, { "epoch": 0.9241300627495722, "grad_norm": 1.1146330833435059, "learning_rate": 5.852776467378492e-05, "loss": 0.6236, "step": 13770 }, { "epoch": 0.9242642864333411, "grad_norm": 1.016501784324646, "learning_rate": 5.8517055407253115e-05, "loss": 0.5872, "step": 13772 }, { "epoch": 0.9243985101171102, "grad_norm": 0.9463066458702087, "learning_rate": 5.850634573830718e-05, "loss": 0.5085, "step": 13774 }, { "epoch": 0.9245327338008792, "grad_norm": 0.9416144490242004, "learning_rate": 5.849563566745313e-05, "loss": 0.567, "step": 13776 }, { "epoch": 0.9246669574846481, "grad_norm": 1.0586539506912231, "learning_rate": 5.8484925195197016e-05, "loss": 0.5298, "step": 13778 }, { "epoch": 0.9248011811684171, "grad_norm": 0.986104428768158, "learning_rate": 5.847421432204486e-05, "loss": 0.5542, "step": 13780 }, { "epoch": 0.9249354048521862, "grad_norm": 1.1029913425445557, "learning_rate": 5.846350304850274e-05, "loss": 0.5755, "step": 13782 }, { "epoch": 0.9250696285359552, "grad_norm": 0.9016444087028503, "learning_rate": 5.845279137507675e-05, "loss": 0.5321, "step": 13784 }, { "epoch": 0.9252038522197241, "grad_norm": 0.9199962615966797, "learning_rate": 5.8442079302273e-05, "loss": 0.5333, "step": 13786 }, { "epoch": 0.9253380759034932, "grad_norm": 0.9962599277496338, "learning_rate": 5.843136683059758e-05, "loss": 0.5798, "step": 13788 }, { "epoch": 0.9254722995872622, "grad_norm": 1.060033917427063, "learning_rate": 5.842065396055667e-05, "loss": 0.573, "step": 13790 }, { "epoch": 0.9256065232710312, "grad_norm": 1.0801200866699219, "learning_rate": 5.8409940692656416e-05, "loss": 0.5885, "step": 13792 }, { "epoch": 0.9257407469548001, "grad_norm": 1.0969904661178589, "learning_rate": 5.8399227027403e-05, "loss": 0.5462, "step": 13794 }, { "epoch": 0.9258749706385692, "grad_norm": 1.4458987712860107, "learning_rate": 5.838851296530263e-05, "loss": 0.5204, "step": 13796 }, { "epoch": 0.9260091943223382, "grad_norm": 0.9598033428192139, "learning_rate": 5.837779850686152e-05, "loss": 0.5761, "step": 13798 }, { "epoch": 0.9261434180061072, "grad_norm": 1.0324113368988037, "learning_rate": 5.836708365258589e-05, "loss": 0.6112, "step": 13800 }, { "epoch": 0.9262776416898761, "grad_norm": 0.886147141456604, "learning_rate": 5.835636840298202e-05, "loss": 0.5548, "step": 13802 }, { "epoch": 0.9264118653736452, "grad_norm": 1.0079913139343262, "learning_rate": 5.834565275855617e-05, "loss": 0.55, "step": 13804 }, { "epoch": 0.9265460890574142, "grad_norm": 1.0281902551651, "learning_rate": 5.833493671981465e-05, "loss": 0.5206, "step": 13806 }, { "epoch": 0.9266803127411832, "grad_norm": 0.8789565563201904, "learning_rate": 5.832422028726375e-05, "loss": 0.4976, "step": 13808 }, { "epoch": 0.9268145364249522, "grad_norm": 1.0310882329940796, "learning_rate": 5.8313503461409826e-05, "loss": 0.5597, "step": 13810 }, { "epoch": 0.9269487601087212, "grad_norm": 1.0418280363082886, "learning_rate": 5.830278624275919e-05, "loss": 0.5402, "step": 13812 }, { "epoch": 0.9270829837924902, "grad_norm": 1.0995814800262451, "learning_rate": 5.829206863181823e-05, "loss": 0.5538, "step": 13814 }, { "epoch": 0.9272172074762591, "grad_norm": 0.9341369867324829, "learning_rate": 5.8281350629093346e-05, "loss": 0.5543, "step": 13816 }, { "epoch": 0.9273514311600282, "grad_norm": 1.0317434072494507, "learning_rate": 5.8270632235090916e-05, "loss": 0.5834, "step": 13818 }, { "epoch": 0.9274856548437972, "grad_norm": 1.0883063077926636, "learning_rate": 5.825991345031739e-05, "loss": 0.5688, "step": 13820 }, { "epoch": 0.9276198785275662, "grad_norm": 1.025431752204895, "learning_rate": 5.82491942752792e-05, "loss": 0.6176, "step": 13822 }, { "epoch": 0.9277541022113351, "grad_norm": 0.9473832845687866, "learning_rate": 5.823847471048279e-05, "loss": 0.5329, "step": 13824 }, { "epoch": 0.9278883258951042, "grad_norm": 0.9456143379211426, "learning_rate": 5.822775475643465e-05, "loss": 0.7219, "step": 13826 }, { "epoch": 0.9280225495788732, "grad_norm": 0.8828773498535156, "learning_rate": 5.821703441364128e-05, "loss": 0.6389, "step": 13828 }, { "epoch": 0.9281567732626422, "grad_norm": 0.9927616715431213, "learning_rate": 5.820631368260919e-05, "loss": 0.5574, "step": 13830 }, { "epoch": 0.9282909969464112, "grad_norm": 1.0232497453689575, "learning_rate": 5.819559256384492e-05, "loss": 0.5249, "step": 13832 }, { "epoch": 0.9284252206301802, "grad_norm": 0.9136453866958618, "learning_rate": 5.818487105785502e-05, "loss": 0.5522, "step": 13834 }, { "epoch": 0.9285594443139492, "grad_norm": 1.032614827156067, "learning_rate": 5.8174149165146044e-05, "loss": 0.5188, "step": 13836 }, { "epoch": 0.9286936679977182, "grad_norm": 1.0510674715042114, "learning_rate": 5.816342688622462e-05, "loss": 0.604, "step": 13838 }, { "epoch": 0.9288278916814872, "grad_norm": 1.0441523790359497, "learning_rate": 5.815270422159731e-05, "loss": 0.556, "step": 13840 }, { "epoch": 0.9289621153652562, "grad_norm": 1.0888450145721436, "learning_rate": 5.8141981171770755e-05, "loss": 0.6212, "step": 13842 }, { "epoch": 0.9290963390490252, "grad_norm": 0.9734417796134949, "learning_rate": 5.81312577372516e-05, "loss": 0.6006, "step": 13844 }, { "epoch": 0.9292305627327943, "grad_norm": 1.0175516605377197, "learning_rate": 5.8120533918546506e-05, "loss": 0.5665, "step": 13846 }, { "epoch": 0.9293647864165632, "grad_norm": 1.0254477262496948, "learning_rate": 5.8109809716162164e-05, "loss": 0.5918, "step": 13848 }, { "epoch": 0.9294990101003322, "grad_norm": 0.9375723600387573, "learning_rate": 5.809908513060524e-05, "loss": 0.5815, "step": 13850 }, { "epoch": 0.9296332337841012, "grad_norm": 0.9494642019271851, "learning_rate": 5.8088360162382486e-05, "loss": 0.5414, "step": 13852 }, { "epoch": 0.9297674574678702, "grad_norm": 1.0334968566894531, "learning_rate": 5.8077634812000614e-05, "loss": 0.566, "step": 13854 }, { "epoch": 0.9299016811516392, "grad_norm": 1.381908893585205, "learning_rate": 5.806690907996638e-05, "loss": 0.5649, "step": 13856 }, { "epoch": 0.9300359048354082, "grad_norm": 1.1403449773788452, "learning_rate": 5.8056182966786566e-05, "loss": 0.6084, "step": 13858 }, { "epoch": 0.9301701285191772, "grad_norm": 1.0421216487884521, "learning_rate": 5.804545647296793e-05, "loss": 0.5579, "step": 13860 }, { "epoch": 0.9303043522029462, "grad_norm": 1.0358749628067017, "learning_rate": 5.803472959901731e-05, "loss": 0.591, "step": 13862 }, { "epoch": 0.9304385758867152, "grad_norm": 1.1746695041656494, "learning_rate": 5.802400234544152e-05, "loss": 0.6525, "step": 13864 }, { "epoch": 0.9305727995704842, "grad_norm": 1.0326679944992065, "learning_rate": 5.801327471274738e-05, "loss": 0.5865, "step": 13866 }, { "epoch": 0.9307070232542533, "grad_norm": 1.0468127727508545, "learning_rate": 5.8002546701441785e-05, "loss": 0.6163, "step": 13868 }, { "epoch": 0.9308412469380222, "grad_norm": 0.9539506435394287, "learning_rate": 5.7991818312031575e-05, "loss": 0.5444, "step": 13870 }, { "epoch": 0.9309754706217912, "grad_norm": 0.9731017351150513, "learning_rate": 5.798108954502368e-05, "loss": 0.5387, "step": 13872 }, { "epoch": 0.9311096943055602, "grad_norm": 0.9137818813323975, "learning_rate": 5.7970360400924994e-05, "loss": 0.5162, "step": 13874 }, { "epoch": 0.9312439179893293, "grad_norm": 1.7687433958053589, "learning_rate": 5.795963088024247e-05, "loss": 0.584, "step": 13876 }, { "epoch": 0.9313781416730982, "grad_norm": 1.2483657598495483, "learning_rate": 5.794890098348301e-05, "loss": 0.597, "step": 13878 }, { "epoch": 0.9315123653568672, "grad_norm": 1.0281192064285278, "learning_rate": 5.7938170711153614e-05, "loss": 0.5919, "step": 13880 }, { "epoch": 0.9316465890406362, "grad_norm": 1.559960126876831, "learning_rate": 5.792744006376127e-05, "loss": 0.5408, "step": 13882 }, { "epoch": 0.9317808127244053, "grad_norm": 1.0883233547210693, "learning_rate": 5.791670904181297e-05, "loss": 0.546, "step": 13884 }, { "epoch": 0.9319150364081742, "grad_norm": 1.0710606575012207, "learning_rate": 5.7905977645815745e-05, "loss": 0.595, "step": 13886 }, { "epoch": 0.9320492600919432, "grad_norm": 1.2496391534805298, "learning_rate": 5.789524587627661e-05, "loss": 0.5146, "step": 13888 }, { "epoch": 0.9321834837757123, "grad_norm": 0.9544695615768433, "learning_rate": 5.788451373370263e-05, "loss": 0.583, "step": 13890 }, { "epoch": 0.9323177074594812, "grad_norm": 1.0018619298934937, "learning_rate": 5.78737812186009e-05, "loss": 0.615, "step": 13892 }, { "epoch": 0.9324519311432502, "grad_norm": 1.0168393850326538, "learning_rate": 5.7863048331478466e-05, "loss": 0.5867, "step": 13894 }, { "epoch": 0.9325861548270192, "grad_norm": 0.9545914530754089, "learning_rate": 5.785231507284248e-05, "loss": 0.6116, "step": 13896 }, { "epoch": 0.9327203785107883, "grad_norm": 1.0262305736541748, "learning_rate": 5.7841581443200035e-05, "loss": 0.5901, "step": 13898 }, { "epoch": 0.9328546021945572, "grad_norm": 0.9550249576568604, "learning_rate": 5.783084744305829e-05, "loss": 0.5452, "step": 13900 }, { "epoch": 0.9329888258783262, "grad_norm": 0.8978448510169983, "learning_rate": 5.7820113072924395e-05, "loss": 0.5158, "step": 13902 }, { "epoch": 0.9331230495620952, "grad_norm": 1.11091148853302, "learning_rate": 5.780937833330554e-05, "loss": 0.5729, "step": 13904 }, { "epoch": 0.9332572732458643, "grad_norm": 0.9957830905914307, "learning_rate": 5.779864322470894e-05, "loss": 0.6102, "step": 13906 }, { "epoch": 0.9333914969296332, "grad_norm": 1.2966653108596802, "learning_rate": 5.778790774764176e-05, "loss": 0.5261, "step": 13908 }, { "epoch": 0.9335257206134022, "grad_norm": 1.0487574338912964, "learning_rate": 5.777717190261125e-05, "loss": 0.5847, "step": 13910 }, { "epoch": 0.9336599442971713, "grad_norm": 1.1108134984970093, "learning_rate": 5.7766435690124667e-05, "loss": 0.5592, "step": 13912 }, { "epoch": 0.9337941679809403, "grad_norm": 0.9877544045448303, "learning_rate": 5.775569911068925e-05, "loss": 0.4801, "step": 13914 }, { "epoch": 0.9339283916647092, "grad_norm": 1.107453465461731, "learning_rate": 5.774496216481233e-05, "loss": 0.5566, "step": 13916 }, { "epoch": 0.9340626153484782, "grad_norm": 1.3392894268035889, "learning_rate": 5.773422485300116e-05, "loss": 0.5968, "step": 13918 }, { "epoch": 0.9341968390322473, "grad_norm": 1.0226224660873413, "learning_rate": 5.772348717576309e-05, "loss": 0.6067, "step": 13920 }, { "epoch": 0.9343310627160163, "grad_norm": 1.0401580333709717, "learning_rate": 5.771274913360543e-05, "loss": 0.5746, "step": 13922 }, { "epoch": 0.9344652863997852, "grad_norm": 1.0642728805541992, "learning_rate": 5.7702010727035536e-05, "loss": 0.5197, "step": 13924 }, { "epoch": 0.9345995100835542, "grad_norm": 1.141188383102417, "learning_rate": 5.769127195656079e-05, "loss": 0.593, "step": 13926 }, { "epoch": 0.9347337337673233, "grad_norm": 1.0031726360321045, "learning_rate": 5.768053282268855e-05, "loss": 0.5746, "step": 13928 }, { "epoch": 0.9348679574510922, "grad_norm": 0.9465569853782654, "learning_rate": 5.766979332592626e-05, "loss": 0.5283, "step": 13930 }, { "epoch": 0.9350021811348612, "grad_norm": 1.2472999095916748, "learning_rate": 5.76590534667813e-05, "loss": 0.522, "step": 13932 }, { "epoch": 0.9351364048186303, "grad_norm": 1.7484946250915527, "learning_rate": 5.764831324576113e-05, "loss": 0.5881, "step": 13934 }, { "epoch": 0.9352706285023993, "grad_norm": 1.056246042251587, "learning_rate": 5.7637572663373194e-05, "loss": 0.5656, "step": 13936 }, { "epoch": 0.9354048521861682, "grad_norm": 0.9071460366249084, "learning_rate": 5.762683172012498e-05, "loss": 0.4649, "step": 13938 }, { "epoch": 0.9355390758699372, "grad_norm": 1.0678855180740356, "learning_rate": 5.761609041652396e-05, "loss": 0.5451, "step": 13940 }, { "epoch": 0.9356732995537063, "grad_norm": 0.9900966286659241, "learning_rate": 5.7605348753077634e-05, "loss": 0.5007, "step": 13942 }, { "epoch": 0.9358075232374753, "grad_norm": 1.0609707832336426, "learning_rate": 5.7594606730293554e-05, "loss": 0.5787, "step": 13944 }, { "epoch": 0.9359417469212442, "grad_norm": 1.087057113647461, "learning_rate": 5.7583864348679226e-05, "loss": 0.6216, "step": 13946 }, { "epoch": 0.9360759706050132, "grad_norm": 1.0868579149246216, "learning_rate": 5.7573121608742234e-05, "loss": 0.6006, "step": 13948 }, { "epoch": 0.9362101942887823, "grad_norm": 1.0762981176376343, "learning_rate": 5.7562378510990125e-05, "loss": 0.58, "step": 13950 }, { "epoch": 0.9363444179725513, "grad_norm": 1.1625175476074219, "learning_rate": 5.755163505593051e-05, "loss": 0.5484, "step": 13952 }, { "epoch": 0.9364786416563202, "grad_norm": 1.0515735149383545, "learning_rate": 5.754089124407097e-05, "loss": 0.5732, "step": 13954 }, { "epoch": 0.9366128653400893, "grad_norm": 0.9707502126693726, "learning_rate": 5.753014707591916e-05, "loss": 0.5438, "step": 13956 }, { "epoch": 0.9367470890238583, "grad_norm": 1.2039716243743896, "learning_rate": 5.751940255198272e-05, "loss": 0.6074, "step": 13958 }, { "epoch": 0.9368813127076273, "grad_norm": 1.8390384912490845, "learning_rate": 5.750865767276927e-05, "loss": 0.5314, "step": 13960 }, { "epoch": 0.9370155363913962, "grad_norm": 1.0250873565673828, "learning_rate": 5.7497912438786536e-05, "loss": 0.6513, "step": 13962 }, { "epoch": 0.9371497600751653, "grad_norm": 1.1512795686721802, "learning_rate": 5.7487166850542165e-05, "loss": 0.564, "step": 13964 }, { "epoch": 0.9372839837589343, "grad_norm": 1.3163058757781982, "learning_rate": 5.74764209085439e-05, "loss": 0.5528, "step": 13966 }, { "epoch": 0.9374182074427032, "grad_norm": 1.0196266174316406, "learning_rate": 5.746567461329943e-05, "loss": 0.5459, "step": 13968 }, { "epoch": 0.9375524311264722, "grad_norm": 1.0521317720413208, "learning_rate": 5.7454927965316516e-05, "loss": 0.5021, "step": 13970 }, { "epoch": 0.9376866548102413, "grad_norm": 0.9628283381462097, "learning_rate": 5.7444180965102936e-05, "loss": 0.5569, "step": 13972 }, { "epoch": 0.9378208784940103, "grad_norm": 1.032279372215271, "learning_rate": 5.743343361316644e-05, "loss": 0.6062, "step": 13974 }, { "epoch": 0.9379551021777792, "grad_norm": 1.5314966440200806, "learning_rate": 5.742268591001481e-05, "loss": 0.6062, "step": 13976 }, { "epoch": 0.9380893258615483, "grad_norm": 1.048116683959961, "learning_rate": 5.741193785615587e-05, "loss": 0.5767, "step": 13978 }, { "epoch": 0.9382235495453173, "grad_norm": 1.027986764907837, "learning_rate": 5.740118945209744e-05, "loss": 0.5113, "step": 13980 }, { "epoch": 0.9383577732290863, "grad_norm": 0.9534016847610474, "learning_rate": 5.739044069834737e-05, "loss": 0.5451, "step": 13982 }, { "epoch": 0.9384919969128552, "grad_norm": 0.9986965656280518, "learning_rate": 5.73796915954135e-05, "loss": 0.5946, "step": 13984 }, { "epoch": 0.9386262205966243, "grad_norm": 0.9526494145393372, "learning_rate": 5.7368942143803725e-05, "loss": 0.616, "step": 13986 }, { "epoch": 0.9387604442803933, "grad_norm": 0.9984648823738098, "learning_rate": 5.735819234402591e-05, "loss": 0.5458, "step": 13988 }, { "epoch": 0.9388946679641623, "grad_norm": 1.1599220037460327, "learning_rate": 5.7347442196587986e-05, "loss": 0.5715, "step": 13990 }, { "epoch": 0.9390288916479312, "grad_norm": 1.0007301568984985, "learning_rate": 5.7336691701997866e-05, "loss": 0.5822, "step": 13992 }, { "epoch": 0.9391631153317003, "grad_norm": 1.06472647190094, "learning_rate": 5.732594086076348e-05, "loss": 0.6006, "step": 13994 }, { "epoch": 0.9392973390154693, "grad_norm": 0.8864287734031677, "learning_rate": 5.731518967339281e-05, "loss": 0.6629, "step": 13996 }, { "epoch": 0.9394315626992383, "grad_norm": 1.0640459060668945, "learning_rate": 5.730443814039379e-05, "loss": 0.5738, "step": 13998 }, { "epoch": 0.9395657863830073, "grad_norm": 1.1999764442443848, "learning_rate": 5.729368626227446e-05, "loss": 0.5897, "step": 14000 }, { "epoch": 0.9397000100667763, "grad_norm": 0.9244864583015442, "learning_rate": 5.728293403954278e-05, "loss": 0.5637, "step": 14002 }, { "epoch": 0.9398342337505453, "grad_norm": 1.094678282737732, "learning_rate": 5.727218147270678e-05, "loss": 0.6543, "step": 14004 }, { "epoch": 0.9399684574343142, "grad_norm": 1.1416062116622925, "learning_rate": 5.726142856227452e-05, "loss": 0.5896, "step": 14006 }, { "epoch": 0.9401026811180833, "grad_norm": 1.0241265296936035, "learning_rate": 5.725067530875403e-05, "loss": 0.578, "step": 14008 }, { "epoch": 0.9402369048018523, "grad_norm": 1.0072925090789795, "learning_rate": 5.72399217126534e-05, "loss": 0.539, "step": 14010 }, { "epoch": 0.9403711284856213, "grad_norm": 0.9036150574684143, "learning_rate": 5.722916777448069e-05, "loss": 0.5875, "step": 14012 }, { "epoch": 0.9405053521693902, "grad_norm": 0.9075189828872681, "learning_rate": 5.721841349474404e-05, "loss": 0.5424, "step": 14014 }, { "epoch": 0.9406395758531593, "grad_norm": 0.9381653070449829, "learning_rate": 5.720765887395153e-05, "loss": 0.5986, "step": 14016 }, { "epoch": 0.9407737995369283, "grad_norm": 1.1812331676483154, "learning_rate": 5.719690391261131e-05, "loss": 0.586, "step": 14018 }, { "epoch": 0.9409080232206973, "grad_norm": 1.0900657176971436, "learning_rate": 5.718614861123155e-05, "loss": 0.5922, "step": 14020 }, { "epoch": 0.9410422469044663, "grad_norm": 1.0432283878326416, "learning_rate": 5.717539297032039e-05, "loss": 0.6569, "step": 14022 }, { "epoch": 0.9411764705882353, "grad_norm": 1.0058923959732056, "learning_rate": 5.716463699038602e-05, "loss": 0.5696, "step": 14024 }, { "epoch": 0.9413106942720043, "grad_norm": 1.1516404151916504, "learning_rate": 5.7153880671936635e-05, "loss": 0.6802, "step": 14026 }, { "epoch": 0.9414449179557733, "grad_norm": 1.3103623390197754, "learning_rate": 5.7143124015480466e-05, "loss": 0.6526, "step": 14028 }, { "epoch": 0.9415791416395423, "grad_norm": 0.9740951657295227, "learning_rate": 5.713236702152572e-05, "loss": 0.6088, "step": 14030 }, { "epoch": 0.9417133653233113, "grad_norm": 0.9947975873947144, "learning_rate": 5.7121609690580666e-05, "loss": 0.5089, "step": 14032 }, { "epoch": 0.9418475890070803, "grad_norm": 1.1044875383377075, "learning_rate": 5.711085202315356e-05, "loss": 0.5827, "step": 14034 }, { "epoch": 0.9419818126908494, "grad_norm": 1.043334722518921, "learning_rate": 5.710009401975268e-05, "loss": 0.5639, "step": 14036 }, { "epoch": 0.9421160363746183, "grad_norm": 0.9688626527786255, "learning_rate": 5.708933568088632e-05, "loss": 0.5751, "step": 14038 }, { "epoch": 0.9422502600583873, "grad_norm": 1.048002004623413, "learning_rate": 5.707857700706278e-05, "loss": 0.5697, "step": 14040 }, { "epoch": 0.9423844837421563, "grad_norm": 1.013647198677063, "learning_rate": 5.706781799879041e-05, "loss": 0.5997, "step": 14042 }, { "epoch": 0.9425187074259253, "grad_norm": 0.9789992570877075, "learning_rate": 5.705705865657753e-05, "loss": 0.5131, "step": 14044 }, { "epoch": 0.9426529311096943, "grad_norm": 0.9202343821525574, "learning_rate": 5.704629898093251e-05, "loss": 0.4936, "step": 14046 }, { "epoch": 0.9427871547934633, "grad_norm": 1.0707128047943115, "learning_rate": 5.703553897236372e-05, "loss": 0.6015, "step": 14048 }, { "epoch": 0.9429213784772323, "grad_norm": 1.1972709894180298, "learning_rate": 5.702477863137954e-05, "loss": 0.5278, "step": 14050 }, { "epoch": 0.9430556021610013, "grad_norm": 0.9796330332756042, "learning_rate": 5.7014017958488375e-05, "loss": 0.5181, "step": 14052 }, { "epoch": 0.9431898258447703, "grad_norm": 1.0746116638183594, "learning_rate": 5.700325695419868e-05, "loss": 0.5927, "step": 14054 }, { "epoch": 0.9433240495285393, "grad_norm": 0.8973516821861267, "learning_rate": 5.699249561901884e-05, "loss": 0.5777, "step": 14056 }, { "epoch": 0.9434582732123084, "grad_norm": 1.0385633707046509, "learning_rate": 5.698173395345735e-05, "loss": 0.6789, "step": 14058 }, { "epoch": 0.9435924968960773, "grad_norm": 1.1208579540252686, "learning_rate": 5.6970971958022644e-05, "loss": 0.5817, "step": 14060 }, { "epoch": 0.9437267205798463, "grad_norm": 1.0253500938415527, "learning_rate": 5.696020963322324e-05, "loss": 0.5461, "step": 14062 }, { "epoch": 0.9438609442636153, "grad_norm": 0.8999269604682922, "learning_rate": 5.69494469795676e-05, "loss": 0.5908, "step": 14064 }, { "epoch": 0.9439951679473844, "grad_norm": 1.0557160377502441, "learning_rate": 5.693868399756426e-05, "loss": 0.6127, "step": 14066 }, { "epoch": 0.9441293916311533, "grad_norm": 0.8857409954071045, "learning_rate": 5.692792068772176e-05, "loss": 0.5033, "step": 14068 }, { "epoch": 0.9442636153149223, "grad_norm": 0.8972952365875244, "learning_rate": 5.691715705054861e-05, "loss": 0.549, "step": 14070 }, { "epoch": 0.9443978389986913, "grad_norm": 1.5294030904769897, "learning_rate": 5.69063930865534e-05, "loss": 0.6064, "step": 14072 }, { "epoch": 0.9445320626824604, "grad_norm": 0.9840405583381653, "learning_rate": 5.689562879624469e-05, "loss": 0.5188, "step": 14074 }, { "epoch": 0.9446662863662293, "grad_norm": 1.0034092664718628, "learning_rate": 5.68848641801311e-05, "loss": 0.5347, "step": 14076 }, { "epoch": 0.9448005100499983, "grad_norm": 1.0162858963012695, "learning_rate": 5.68740992387212e-05, "loss": 0.5567, "step": 14078 }, { "epoch": 0.9449347337337674, "grad_norm": 1.1284905672073364, "learning_rate": 5.686333397252363e-05, "loss": 0.5736, "step": 14080 }, { "epoch": 0.9450689574175363, "grad_norm": 0.9935622811317444, "learning_rate": 5.6852568382047026e-05, "loss": 0.5967, "step": 14082 }, { "epoch": 0.9452031811013053, "grad_norm": 0.9152777791023254, "learning_rate": 5.684180246780004e-05, "loss": 0.5198, "step": 14084 }, { "epoch": 0.9453374047850743, "grad_norm": 1.104828953742981, "learning_rate": 5.683103623029135e-05, "loss": 0.5803, "step": 14086 }, { "epoch": 0.9454716284688434, "grad_norm": 0.9986449480056763, "learning_rate": 5.6820269670029615e-05, "loss": 0.5851, "step": 14088 }, { "epoch": 0.9456058521526123, "grad_norm": 0.8983079791069031, "learning_rate": 5.680950278752356e-05, "loss": 0.5776, "step": 14090 }, { "epoch": 0.9457400758363813, "grad_norm": 1.0416613817214966, "learning_rate": 5.679873558328188e-05, "loss": 0.5336, "step": 14092 }, { "epoch": 0.9458742995201503, "grad_norm": 0.8977410793304443, "learning_rate": 5.6787968057813324e-05, "loss": 0.6275, "step": 14094 }, { "epoch": 0.9460085232039194, "grad_norm": 1.108778953552246, "learning_rate": 5.6777200211626624e-05, "loss": 0.6378, "step": 14096 }, { "epoch": 0.9461427468876883, "grad_norm": 1.1451404094696045, "learning_rate": 5.6766432045230536e-05, "loss": 0.631, "step": 14098 }, { "epoch": 0.9462769705714573, "grad_norm": 1.8046618700027466, "learning_rate": 5.675566355913384e-05, "loss": 0.6229, "step": 14100 }, { "epoch": 0.9464111942552264, "grad_norm": 0.9053519368171692, "learning_rate": 5.6744894753845326e-05, "loss": 0.4854, "step": 14102 }, { "epoch": 0.9465454179389954, "grad_norm": 1.076835036277771, "learning_rate": 5.67341256298738e-05, "loss": 0.5935, "step": 14104 }, { "epoch": 0.9466796416227643, "grad_norm": 1.1154593229293823, "learning_rate": 5.672335618772808e-05, "loss": 0.6073, "step": 14106 }, { "epoch": 0.9468138653065333, "grad_norm": 0.9786189794540405, "learning_rate": 5.671258642791699e-05, "loss": 0.5601, "step": 14108 }, { "epoch": 0.9469480889903024, "grad_norm": 0.944292426109314, "learning_rate": 5.670181635094941e-05, "loss": 0.5244, "step": 14110 }, { "epoch": 0.9470823126740714, "grad_norm": 0.9197020530700684, "learning_rate": 5.669104595733419e-05, "loss": 0.5687, "step": 14112 }, { "epoch": 0.9472165363578403, "grad_norm": 1.0978754758834839, "learning_rate": 5.668027524758019e-05, "loss": 0.6745, "step": 14114 }, { "epoch": 0.9473507600416093, "grad_norm": 1.1043089628219604, "learning_rate": 5.6669504222196327e-05, "loss": 0.5164, "step": 14116 }, { "epoch": 0.9474849837253784, "grad_norm": 0.9744171500205994, "learning_rate": 5.665873288169149e-05, "loss": 0.5805, "step": 14118 }, { "epoch": 0.9476192074091473, "grad_norm": 0.8978694081306458, "learning_rate": 5.664796122657463e-05, "loss": 0.6074, "step": 14120 }, { "epoch": 0.9477534310929163, "grad_norm": 1.1278657913208008, "learning_rate": 5.6637189257354675e-05, "loss": 0.6419, "step": 14122 }, { "epoch": 0.9478876547766854, "grad_norm": 0.9175683259963989, "learning_rate": 5.6626416974540585e-05, "loss": 0.4927, "step": 14124 }, { "epoch": 0.9480218784604544, "grad_norm": 1.0380014181137085, "learning_rate": 5.6615644378641306e-05, "loss": 0.547, "step": 14126 }, { "epoch": 0.9481561021442233, "grad_norm": 1.1204023361206055, "learning_rate": 5.660487147016584e-05, "loss": 0.5597, "step": 14128 }, { "epoch": 0.9482903258279923, "grad_norm": 1.1506484746932983, "learning_rate": 5.659409824962321e-05, "loss": 0.523, "step": 14130 }, { "epoch": 0.9484245495117614, "grad_norm": 1.0465366840362549, "learning_rate": 5.658332471752239e-05, "loss": 0.5114, "step": 14132 }, { "epoch": 0.9485587731955304, "grad_norm": 0.9276459813117981, "learning_rate": 5.6572550874372424e-05, "loss": 0.6087, "step": 14134 }, { "epoch": 0.9486929968792993, "grad_norm": 1.038851261138916, "learning_rate": 5.656177672068235e-05, "loss": 0.6461, "step": 14136 }, { "epoch": 0.9488272205630683, "grad_norm": 0.9065415263175964, "learning_rate": 5.655100225696123e-05, "loss": 0.5719, "step": 14138 }, { "epoch": 0.9489614442468374, "grad_norm": 1.2550922632217407, "learning_rate": 5.654022748371813e-05, "loss": 0.5503, "step": 14140 }, { "epoch": 0.9490956679306064, "grad_norm": 1.0652170181274414, "learning_rate": 5.6529452401462145e-05, "loss": 0.5942, "step": 14142 }, { "epoch": 0.9492298916143753, "grad_norm": 1.002002239227295, "learning_rate": 5.651867701070238e-05, "loss": 0.6353, "step": 14144 }, { "epoch": 0.9493641152981444, "grad_norm": 0.9674058556556702, "learning_rate": 5.650790131194794e-05, "loss": 0.5837, "step": 14146 }, { "epoch": 0.9494983389819134, "grad_norm": 1.1481109857559204, "learning_rate": 5.649712530570797e-05, "loss": 0.5894, "step": 14148 }, { "epoch": 0.9496325626656824, "grad_norm": 1.0205858945846558, "learning_rate": 5.648634899249159e-05, "loss": 0.5757, "step": 14150 }, { "epoch": 0.9497667863494513, "grad_norm": 0.9188188314437866, "learning_rate": 5.647557237280798e-05, "loss": 0.5892, "step": 14152 }, { "epoch": 0.9499010100332204, "grad_norm": 1.0300636291503906, "learning_rate": 5.6464795447166306e-05, "loss": 0.5722, "step": 14154 }, { "epoch": 0.9500352337169894, "grad_norm": 1.0198545455932617, "learning_rate": 5.645401821607577e-05, "loss": 0.5481, "step": 14156 }, { "epoch": 0.9501694574007583, "grad_norm": 1.0702418088912964, "learning_rate": 5.644324068004556e-05, "loss": 0.6041, "step": 14158 }, { "epoch": 0.9503036810845273, "grad_norm": 0.919014573097229, "learning_rate": 5.6432462839584886e-05, "loss": 0.5284, "step": 14160 }, { "epoch": 0.9504379047682964, "grad_norm": 1.0594698190689087, "learning_rate": 5.6421684695203015e-05, "loss": 0.551, "step": 14162 }, { "epoch": 0.9505721284520654, "grad_norm": 1.2445323467254639, "learning_rate": 5.6410906247409144e-05, "loss": 0.5357, "step": 14164 }, { "epoch": 0.9507063521358343, "grad_norm": 0.9506213665008545, "learning_rate": 5.6400127496712585e-05, "loss": 0.5968, "step": 14166 }, { "epoch": 0.9508405758196034, "grad_norm": 1.0716971158981323, "learning_rate": 5.6389348443622566e-05, "loss": 0.5523, "step": 14168 }, { "epoch": 0.9509747995033724, "grad_norm": 1.083978295326233, "learning_rate": 5.6378569088648406e-05, "loss": 0.5733, "step": 14170 }, { "epoch": 0.9511090231871414, "grad_norm": 1.1101188659667969, "learning_rate": 5.63677894322994e-05, "loss": 0.576, "step": 14172 }, { "epoch": 0.9512432468709103, "grad_norm": 0.9841271638870239, "learning_rate": 5.635700947508486e-05, "loss": 0.5641, "step": 14174 }, { "epoch": 0.9513774705546794, "grad_norm": 0.9433690309524536, "learning_rate": 5.6346229217514136e-05, "loss": 0.5221, "step": 14176 }, { "epoch": 0.9515116942384484, "grad_norm": 1.0756908655166626, "learning_rate": 5.6335448660096544e-05, "loss": 0.5917, "step": 14178 }, { "epoch": 0.9516459179222174, "grad_norm": 1.024658203125, "learning_rate": 5.632466780334148e-05, "loss": 0.5805, "step": 14180 }, { "epoch": 0.9517801416059863, "grad_norm": 1.002120018005371, "learning_rate": 5.6313886647758294e-05, "loss": 0.5101, "step": 14182 }, { "epoch": 0.9519143652897554, "grad_norm": 1.178074598312378, "learning_rate": 5.6303105193856365e-05, "loss": 0.5554, "step": 14184 }, { "epoch": 0.9520485889735244, "grad_norm": 1.1281471252441406, "learning_rate": 5.6292323442145126e-05, "loss": 0.5906, "step": 14186 }, { "epoch": 0.9521828126572934, "grad_norm": 1.529188871383667, "learning_rate": 5.6281541393133976e-05, "loss": 0.5639, "step": 14188 }, { "epoch": 0.9523170363410624, "grad_norm": 1.301809549331665, "learning_rate": 5.6270759047332334e-05, "loss": 0.6673, "step": 14190 }, { "epoch": 0.9524512600248314, "grad_norm": 1.0367953777313232, "learning_rate": 5.625997640524967e-05, "loss": 0.5203, "step": 14192 }, { "epoch": 0.9525854837086004, "grad_norm": 0.9069284200668335, "learning_rate": 5.624919346739542e-05, "loss": 0.6059, "step": 14194 }, { "epoch": 0.9527197073923693, "grad_norm": 0.9075580835342407, "learning_rate": 5.623841023427908e-05, "loss": 0.544, "step": 14196 }, { "epoch": 0.9528539310761384, "grad_norm": 1.0331097841262817, "learning_rate": 5.622762670641011e-05, "loss": 0.4943, "step": 14198 }, { "epoch": 0.9529881547599074, "grad_norm": 0.9860792756080627, "learning_rate": 5.6216842884298035e-05, "loss": 0.5721, "step": 14200 }, { "epoch": 0.9531223784436764, "grad_norm": 0.9954896569252014, "learning_rate": 5.6206058768452355e-05, "loss": 0.5632, "step": 14202 }, { "epoch": 0.9532566021274453, "grad_norm": 1.2916309833526611, "learning_rate": 5.61952743593826e-05, "loss": 0.6002, "step": 14204 }, { "epoch": 0.9533908258112144, "grad_norm": 1.2309260368347168, "learning_rate": 5.618448965759832e-05, "loss": 0.6084, "step": 14206 }, { "epoch": 0.9535250494949834, "grad_norm": 1.2513889074325562, "learning_rate": 5.617370466360905e-05, "loss": 0.5453, "step": 14208 }, { "epoch": 0.9536592731787524, "grad_norm": 1.052930474281311, "learning_rate": 5.616291937792439e-05, "loss": 0.5588, "step": 14210 }, { "epoch": 0.9537934968625214, "grad_norm": 0.9992319941520691, "learning_rate": 5.615213380105391e-05, "loss": 0.5131, "step": 14212 }, { "epoch": 0.9539277205462904, "grad_norm": 0.9948887825012207, "learning_rate": 5.6141347933507204e-05, "loss": 0.5847, "step": 14214 }, { "epoch": 0.9540619442300594, "grad_norm": 0.9883646368980408, "learning_rate": 5.613056177579388e-05, "loss": 0.5812, "step": 14216 }, { "epoch": 0.9541961679138284, "grad_norm": 0.9431707262992859, "learning_rate": 5.611977532842355e-05, "loss": 0.5341, "step": 14218 }, { "epoch": 0.9543303915975974, "grad_norm": 0.9274953007698059, "learning_rate": 5.61089885919059e-05, "loss": 0.5018, "step": 14220 }, { "epoch": 0.9544646152813664, "grad_norm": 0.9960140585899353, "learning_rate": 5.609820156675053e-05, "loss": 0.6009, "step": 14222 }, { "epoch": 0.9545988389651354, "grad_norm": 1.1744269132614136, "learning_rate": 5.608741425346714e-05, "loss": 0.5912, "step": 14224 }, { "epoch": 0.9547330626489045, "grad_norm": 1.336578369140625, "learning_rate": 5.607662665256539e-05, "loss": 0.5872, "step": 14226 }, { "epoch": 0.9548672863326734, "grad_norm": 1.1415321826934814, "learning_rate": 5.606583876455499e-05, "loss": 0.5724, "step": 14228 }, { "epoch": 0.9550015100164424, "grad_norm": 1.2327148914337158, "learning_rate": 5.605505058994562e-05, "loss": 0.5643, "step": 14230 }, { "epoch": 0.9551357337002114, "grad_norm": 1.079979658126831, "learning_rate": 5.604426212924703e-05, "loss": 0.5841, "step": 14232 }, { "epoch": 0.9552699573839804, "grad_norm": 1.0726792812347412, "learning_rate": 5.6033473382968936e-05, "loss": 0.6108, "step": 14234 }, { "epoch": 0.9554041810677494, "grad_norm": 1.0051898956298828, "learning_rate": 5.6022684351621094e-05, "loss": 0.5602, "step": 14236 }, { "epoch": 0.9555384047515184, "grad_norm": 0.9825963973999023, "learning_rate": 5.601189503571326e-05, "loss": 0.5093, "step": 14238 }, { "epoch": 0.9556726284352874, "grad_norm": 1.103543758392334, "learning_rate": 5.6001105435755194e-05, "loss": 0.5362, "step": 14240 }, { "epoch": 0.9558068521190564, "grad_norm": 0.9990962147712708, "learning_rate": 5.599031555225671e-05, "loss": 0.5821, "step": 14242 }, { "epoch": 0.9559410758028254, "grad_norm": 1.0620709657669067, "learning_rate": 5.597952538572758e-05, "loss": 0.5948, "step": 14244 }, { "epoch": 0.9560752994865944, "grad_norm": 0.8622420430183411, "learning_rate": 5.596873493667765e-05, "loss": 0.5126, "step": 14246 }, { "epoch": 0.9562095231703635, "grad_norm": 0.9423746466636658, "learning_rate": 5.595794420561673e-05, "loss": 0.4744, "step": 14248 }, { "epoch": 0.9563437468541324, "grad_norm": 0.9781889319419861, "learning_rate": 5.5947153193054655e-05, "loss": 0.5743, "step": 14250 }, { "epoch": 0.9564779705379014, "grad_norm": 1.4569517374038696, "learning_rate": 5.59363618995013e-05, "loss": 0.5426, "step": 14252 }, { "epoch": 0.9566121942216704, "grad_norm": 1.0532081127166748, "learning_rate": 5.5925570325466504e-05, "loss": 0.5646, "step": 14254 }, { "epoch": 0.9567464179054395, "grad_norm": 1.0339252948760986, "learning_rate": 5.591477847146016e-05, "loss": 0.6094, "step": 14256 }, { "epoch": 0.9568806415892084, "grad_norm": 1.1620696783065796, "learning_rate": 5.5903986337992174e-05, "loss": 0.5773, "step": 14258 }, { "epoch": 0.9570148652729774, "grad_norm": 1.038197636604309, "learning_rate": 5.589319392557244e-05, "loss": 0.5293, "step": 14260 }, { "epoch": 0.9571490889567464, "grad_norm": 1.1082844734191895, "learning_rate": 5.588240123471088e-05, "loss": 0.5785, "step": 14262 }, { "epoch": 0.9572833126405155, "grad_norm": 1.0165228843688965, "learning_rate": 5.587160826591743e-05, "loss": 0.568, "step": 14264 }, { "epoch": 0.9574175363242844, "grad_norm": 0.986341655254364, "learning_rate": 5.586081501970203e-05, "loss": 0.5844, "step": 14266 }, { "epoch": 0.9575517600080534, "grad_norm": 1.1666079759597778, "learning_rate": 5.585002149657466e-05, "loss": 0.5321, "step": 14268 }, { "epoch": 0.9576859836918225, "grad_norm": 1.0912145376205444, "learning_rate": 5.583922769704526e-05, "loss": 0.5657, "step": 14270 }, { "epoch": 0.9578202073755914, "grad_norm": 1.0317710638046265, "learning_rate": 5.5828433621623845e-05, "loss": 0.5876, "step": 14272 }, { "epoch": 0.9579544310593604, "grad_norm": 1.0638327598571777, "learning_rate": 5.5817639270820386e-05, "loss": 0.583, "step": 14274 }, { "epoch": 0.9580886547431294, "grad_norm": 0.9866347908973694, "learning_rate": 5.580684464514494e-05, "loss": 0.5578, "step": 14276 }, { "epoch": 0.9582228784268985, "grad_norm": 0.9564821124076843, "learning_rate": 5.579604974510748e-05, "loss": 0.52, "step": 14278 }, { "epoch": 0.9583571021106674, "grad_norm": 0.9945246577262878, "learning_rate": 5.578525457121807e-05, "loss": 0.6157, "step": 14280 }, { "epoch": 0.9584913257944364, "grad_norm": 1.0693161487579346, "learning_rate": 5.5774459123986766e-05, "loss": 0.5211, "step": 14282 }, { "epoch": 0.9586255494782054, "grad_norm": 0.8919063806533813, "learning_rate": 5.5763663403923614e-05, "loss": 0.5595, "step": 14284 }, { "epoch": 0.9587597731619745, "grad_norm": 1.1274627447128296, "learning_rate": 5.575286741153871e-05, "loss": 0.578, "step": 14286 }, { "epoch": 0.9588939968457434, "grad_norm": 0.9995314478874207, "learning_rate": 5.574207114734212e-05, "loss": 0.5078, "step": 14288 }, { "epoch": 0.9590282205295124, "grad_norm": 0.99176424741745, "learning_rate": 5.573127461184398e-05, "loss": 0.6396, "step": 14290 }, { "epoch": 0.9591624442132815, "grad_norm": 1.0251235961914062, "learning_rate": 5.5720477805554374e-05, "loss": 0.5035, "step": 14292 }, { "epoch": 0.9592966678970505, "grad_norm": 1.1442151069641113, "learning_rate": 5.570968072898343e-05, "loss": 0.5987, "step": 14294 }, { "epoch": 0.9594308915808194, "grad_norm": 0.9356967806816101, "learning_rate": 5.569888338264131e-05, "loss": 0.6108, "step": 14296 }, { "epoch": 0.9595651152645884, "grad_norm": 0.9598607420921326, "learning_rate": 5.568808576703816e-05, "loss": 0.5645, "step": 14298 }, { "epoch": 0.9596993389483575, "grad_norm": 1.0034985542297363, "learning_rate": 5.567728788268414e-05, "loss": 0.5511, "step": 14300 }, { "epoch": 0.9598335626321265, "grad_norm": 1.1254874467849731, "learning_rate": 5.566648973008942e-05, "loss": 0.5516, "step": 14302 }, { "epoch": 0.9599677863158954, "grad_norm": 1.0316601991653442, "learning_rate": 5.565569130976422e-05, "loss": 0.5374, "step": 14304 }, { "epoch": 0.9601020099996644, "grad_norm": 0.9164896011352539, "learning_rate": 5.564489262221871e-05, "loss": 0.5546, "step": 14306 }, { "epoch": 0.9602362336834335, "grad_norm": 0.9961039423942566, "learning_rate": 5.563409366796314e-05, "loss": 0.5399, "step": 14308 }, { "epoch": 0.9603704573672024, "grad_norm": 1.1713188886642456, "learning_rate": 5.562329444750771e-05, "loss": 0.5298, "step": 14310 }, { "epoch": 0.9605046810509714, "grad_norm": 1.0865498781204224, "learning_rate": 5.561249496136268e-05, "loss": 0.6398, "step": 14312 }, { "epoch": 0.9606389047347405, "grad_norm": 0.9350690245628357, "learning_rate": 5.5601695210038306e-05, "loss": 0.5508, "step": 14314 }, { "epoch": 0.9607731284185095, "grad_norm": 1.061317801475525, "learning_rate": 5.559089519404484e-05, "loss": 0.5835, "step": 14316 }, { "epoch": 0.9609073521022784, "grad_norm": 1.0611228942871094, "learning_rate": 5.5580094913892575e-05, "loss": 0.5825, "step": 14318 }, { "epoch": 0.9610415757860474, "grad_norm": 1.0069859027862549, "learning_rate": 5.556929437009181e-05, "loss": 0.5683, "step": 14320 }, { "epoch": 0.9611757994698165, "grad_norm": 1.1307488679885864, "learning_rate": 5.555849356315281e-05, "loss": 0.5182, "step": 14322 }, { "epoch": 0.9613100231535855, "grad_norm": 1.115998387336731, "learning_rate": 5.554769249358595e-05, "loss": 0.5551, "step": 14324 }, { "epoch": 0.9614442468373544, "grad_norm": 1.0153923034667969, "learning_rate": 5.553689116190151e-05, "loss": 0.5525, "step": 14326 }, { "epoch": 0.9615784705211234, "grad_norm": 1.0945372581481934, "learning_rate": 5.552608956860985e-05, "loss": 0.5292, "step": 14328 }, { "epoch": 0.9617126942048925, "grad_norm": 0.8611571192741394, "learning_rate": 5.551528771422133e-05, "loss": 0.5004, "step": 14330 }, { "epoch": 0.9618469178886615, "grad_norm": 1.0803242921829224, "learning_rate": 5.55044855992463e-05, "loss": 0.5747, "step": 14332 }, { "epoch": 0.9619811415724304, "grad_norm": 0.9125853776931763, "learning_rate": 5.549368322419517e-05, "loss": 0.5362, "step": 14334 }, { "epoch": 0.9621153652561995, "grad_norm": 1.0588202476501465, "learning_rate": 5.548288058957829e-05, "loss": 0.5546, "step": 14336 }, { "epoch": 0.9622495889399685, "grad_norm": 0.9635347723960876, "learning_rate": 5.54720776959061e-05, "loss": 0.506, "step": 14338 }, { "epoch": 0.9623838126237375, "grad_norm": 1.0806838274002075, "learning_rate": 5.546127454368898e-05, "loss": 0.5233, "step": 14340 }, { "epoch": 0.9625180363075064, "grad_norm": 1.066923975944519, "learning_rate": 5.54504711334374e-05, "loss": 0.5789, "step": 14342 }, { "epoch": 0.9626522599912755, "grad_norm": 1.0439131259918213, "learning_rate": 5.543966746566176e-05, "loss": 0.5167, "step": 14344 }, { "epoch": 0.9627864836750445, "grad_norm": 0.9134027361869812, "learning_rate": 5.542886354087252e-05, "loss": 0.5634, "step": 14346 }, { "epoch": 0.9629207073588134, "grad_norm": 0.922843337059021, "learning_rate": 5.5418059359580175e-05, "loss": 0.5137, "step": 14348 }, { "epoch": 0.9630549310425824, "grad_norm": 0.9720866680145264, "learning_rate": 5.5407254922295174e-05, "loss": 0.6233, "step": 14350 }, { "epoch": 0.9631891547263515, "grad_norm": 0.9502121210098267, "learning_rate": 5.5396450229528006e-05, "loss": 0.5707, "step": 14352 }, { "epoch": 0.9633233784101205, "grad_norm": 0.9958463311195374, "learning_rate": 5.5385645281789176e-05, "loss": 0.5451, "step": 14354 }, { "epoch": 0.9634576020938894, "grad_norm": 0.9970313310623169, "learning_rate": 5.537484007958921e-05, "loss": 0.5506, "step": 14356 }, { "epoch": 0.9635918257776585, "grad_norm": 0.9940379858016968, "learning_rate": 5.53640346234386e-05, "loss": 0.561, "step": 14358 }, { "epoch": 0.9637260494614275, "grad_norm": 0.8887693285942078, "learning_rate": 5.535322891384791e-05, "loss": 0.5119, "step": 14360 }, { "epoch": 0.9638602731451965, "grad_norm": 1.103327989578247, "learning_rate": 5.534242295132769e-05, "loss": 0.5505, "step": 14362 }, { "epoch": 0.9639944968289654, "grad_norm": 0.986727237701416, "learning_rate": 5.533161673638847e-05, "loss": 0.5266, "step": 14364 }, { "epoch": 0.9641287205127345, "grad_norm": 1.071113109588623, "learning_rate": 5.532081026954087e-05, "loss": 0.5083, "step": 14366 }, { "epoch": 0.9642629441965035, "grad_norm": 1.021314263343811, "learning_rate": 5.531000355129543e-05, "loss": 0.5189, "step": 14368 }, { "epoch": 0.9643971678802725, "grad_norm": 1.0355725288391113, "learning_rate": 5.529919658216276e-05, "loss": 0.6481, "step": 14370 }, { "epoch": 0.9645313915640414, "grad_norm": 1.4715527296066284, "learning_rate": 5.5288389362653484e-05, "loss": 0.5981, "step": 14372 }, { "epoch": 0.9646656152478105, "grad_norm": 1.6728663444519043, "learning_rate": 5.52775818932782e-05, "loss": 0.5561, "step": 14374 }, { "epoch": 0.9647998389315795, "grad_norm": 0.9639118313789368, "learning_rate": 5.5266774174547564e-05, "loss": 0.6246, "step": 14376 }, { "epoch": 0.9649340626153485, "grad_norm": 1.2170542478561401, "learning_rate": 5.525596620697219e-05, "loss": 0.529, "step": 14378 }, { "epoch": 0.9650682862991175, "grad_norm": 1.3546918630599976, "learning_rate": 5.5245157991062755e-05, "loss": 0.4982, "step": 14380 }, { "epoch": 0.9652025099828865, "grad_norm": 0.9821697473526001, "learning_rate": 5.523434952732991e-05, "loss": 0.5468, "step": 14382 }, { "epoch": 0.9653367336666555, "grad_norm": 1.0652328729629517, "learning_rate": 5.522354081628435e-05, "loss": 0.5706, "step": 14384 }, { "epoch": 0.9654709573504244, "grad_norm": 1.3068857192993164, "learning_rate": 5.5212731858436774e-05, "loss": 0.5942, "step": 14386 }, { "epoch": 0.9656051810341935, "grad_norm": 0.922817587852478, "learning_rate": 5.520192265429784e-05, "loss": 0.5204, "step": 14388 }, { "epoch": 0.9657394047179625, "grad_norm": 1.0467361211776733, "learning_rate": 5.519111320437832e-05, "loss": 0.6672, "step": 14390 }, { "epoch": 0.9658736284017315, "grad_norm": 1.0627771615982056, "learning_rate": 5.518030350918888e-05, "loss": 0.5945, "step": 14392 }, { "epoch": 0.9660078520855004, "grad_norm": 1.2258409261703491, "learning_rate": 5.5169493569240295e-05, "loss": 0.6113, "step": 14394 }, { "epoch": 0.9661420757692695, "grad_norm": 0.9771142601966858, "learning_rate": 5.5158683385043307e-05, "loss": 0.629, "step": 14396 }, { "epoch": 0.9662762994530385, "grad_norm": 1.0257612466812134, "learning_rate": 5.514787295710867e-05, "loss": 0.5419, "step": 14398 }, { "epoch": 0.9664105231368075, "grad_norm": 1.0048645734786987, "learning_rate": 5.513706228594717e-05, "loss": 0.51, "step": 14400 }, { "epoch": 0.9665447468205765, "grad_norm": 0.9502382278442383, "learning_rate": 5.512625137206957e-05, "loss": 0.5781, "step": 14402 }, { "epoch": 0.9666789705043455, "grad_norm": 0.9471548199653625, "learning_rate": 5.5115440215986666e-05, "loss": 0.5382, "step": 14404 }, { "epoch": 0.9668131941881145, "grad_norm": 0.8059168457984924, "learning_rate": 5.510462881820928e-05, "loss": 0.5111, "step": 14406 }, { "epoch": 0.9669474178718835, "grad_norm": 1.1791728734970093, "learning_rate": 5.509381717924822e-05, "loss": 0.5111, "step": 14408 }, { "epoch": 0.9670816415556525, "grad_norm": 1.0287694931030273, "learning_rate": 5.508300529961431e-05, "loss": 0.5906, "step": 14410 }, { "epoch": 0.9672158652394215, "grad_norm": 1.289045810699463, "learning_rate": 5.50721931798184e-05, "loss": 0.5613, "step": 14412 }, { "epoch": 0.9673500889231905, "grad_norm": 1.0925270318984985, "learning_rate": 5.506138082037133e-05, "loss": 0.6117, "step": 14414 }, { "epoch": 0.9674843126069596, "grad_norm": 0.950714111328125, "learning_rate": 5.505056822178397e-05, "loss": 0.5338, "step": 14416 }, { "epoch": 0.9676185362907285, "grad_norm": 0.9248968362808228, "learning_rate": 5.5039755384567207e-05, "loss": 0.6021, "step": 14418 }, { "epoch": 0.9677527599744975, "grad_norm": 0.9010964632034302, "learning_rate": 5.502894230923189e-05, "loss": 0.5709, "step": 14420 }, { "epoch": 0.9678869836582665, "grad_norm": 1.3293532133102417, "learning_rate": 5.501812899628895e-05, "loss": 0.5097, "step": 14422 }, { "epoch": 0.9680212073420355, "grad_norm": 1.0516278743743896, "learning_rate": 5.5007315446249285e-05, "loss": 0.6134, "step": 14424 }, { "epoch": 0.9681554310258045, "grad_norm": 0.9383218288421631, "learning_rate": 5.49965016596238e-05, "loss": 0.5647, "step": 14426 }, { "epoch": 0.9682896547095735, "grad_norm": 0.9845718145370483, "learning_rate": 5.498568763692345e-05, "loss": 0.6191, "step": 14428 }, { "epoch": 0.9684238783933425, "grad_norm": 0.9353554844856262, "learning_rate": 5.497487337865916e-05, "loss": 0.5185, "step": 14430 }, { "epoch": 0.9685581020771115, "grad_norm": 1.1654001474380493, "learning_rate": 5.4964058885341886e-05, "loss": 0.5456, "step": 14432 }, { "epoch": 0.9686923257608805, "grad_norm": 1.0446298122406006, "learning_rate": 5.495324415748259e-05, "loss": 0.5631, "step": 14434 }, { "epoch": 0.9688265494446495, "grad_norm": 1.0155998468399048, "learning_rate": 5.494242919559224e-05, "loss": 0.5683, "step": 14436 }, { "epoch": 0.9689607731284186, "grad_norm": 1.6033741235733032, "learning_rate": 5.493161400018184e-05, "loss": 0.5528, "step": 14438 }, { "epoch": 0.9690949968121875, "grad_norm": 1.1973978281021118, "learning_rate": 5.492079857176236e-05, "loss": 0.5742, "step": 14440 }, { "epoch": 0.9692292204959565, "grad_norm": 1.5843385457992554, "learning_rate": 5.490998291084485e-05, "loss": 0.6305, "step": 14442 }, { "epoch": 0.9693634441797255, "grad_norm": 0.9376939535140991, "learning_rate": 5.489916701794028e-05, "loss": 0.5295, "step": 14444 }, { "epoch": 0.9694976678634946, "grad_norm": 1.0814796686172485, "learning_rate": 5.488835089355971e-05, "loss": 0.5446, "step": 14446 }, { "epoch": 0.9696318915472635, "grad_norm": 1.1428366899490356, "learning_rate": 5.487753453821418e-05, "loss": 0.6206, "step": 14448 }, { "epoch": 0.9697661152310325, "grad_norm": 1.3953572511672974, "learning_rate": 5.4866717952414716e-05, "loss": 0.5119, "step": 14450 }, { "epoch": 0.9699003389148015, "grad_norm": 0.9352095723152161, "learning_rate": 5.485590113667242e-05, "loss": 0.5478, "step": 14452 }, { "epoch": 0.9700345625985706, "grad_norm": 1.2032496929168701, "learning_rate": 5.484508409149833e-05, "loss": 0.5643, "step": 14454 }, { "epoch": 0.9701687862823395, "grad_norm": 1.0390560626983643, "learning_rate": 5.483426681740356e-05, "loss": 0.5129, "step": 14456 }, { "epoch": 0.9703030099661085, "grad_norm": 1.0590581893920898, "learning_rate": 5.482344931489918e-05, "loss": 0.4824, "step": 14458 }, { "epoch": 0.9704372336498776, "grad_norm": 1.2861895561218262, "learning_rate": 5.48126315844963e-05, "loss": 0.525, "step": 14460 }, { "epoch": 0.9705714573336465, "grad_norm": 0.983880877494812, "learning_rate": 5.480181362670605e-05, "loss": 0.5981, "step": 14462 }, { "epoch": 0.9707056810174155, "grad_norm": 0.8727746605873108, "learning_rate": 5.4790995442039537e-05, "loss": 0.5825, "step": 14464 }, { "epoch": 0.9708399047011845, "grad_norm": 0.8624457716941833, "learning_rate": 5.4780177031007916e-05, "loss": 0.5023, "step": 14466 }, { "epoch": 0.9709741283849536, "grad_norm": 1.4579488039016724, "learning_rate": 5.4769358394122326e-05, "loss": 0.5932, "step": 14468 }, { "epoch": 0.9711083520687225, "grad_norm": 1.1287567615509033, "learning_rate": 5.475853953189393e-05, "loss": 0.5976, "step": 14470 }, { "epoch": 0.9712425757524915, "grad_norm": 1.1077579259872437, "learning_rate": 5.474772044483391e-05, "loss": 0.5965, "step": 14472 }, { "epoch": 0.9713767994362605, "grad_norm": 0.9818617701530457, "learning_rate": 5.473690113345342e-05, "loss": 0.5347, "step": 14474 }, { "epoch": 0.9715110231200296, "grad_norm": 1.0572015047073364, "learning_rate": 5.472608159826368e-05, "loss": 0.5764, "step": 14476 }, { "epoch": 0.9716452468037985, "grad_norm": 1.1121615171432495, "learning_rate": 5.471526183977587e-05, "loss": 0.586, "step": 14478 }, { "epoch": 0.9717794704875675, "grad_norm": 0.9781137108802795, "learning_rate": 5.470444185850121e-05, "loss": 0.5684, "step": 14480 }, { "epoch": 0.9719136941713366, "grad_norm": 0.9827112555503845, "learning_rate": 5.4693621654950925e-05, "loss": 0.6294, "step": 14482 }, { "epoch": 0.9720479178551056, "grad_norm": 0.9477865099906921, "learning_rate": 5.4682801229636236e-05, "loss": 0.6169, "step": 14484 }, { "epoch": 0.9721821415388745, "grad_norm": 1.0273088216781616, "learning_rate": 5.467198058306842e-05, "loss": 0.482, "step": 14486 }, { "epoch": 0.9723163652226435, "grad_norm": 1.1097997426986694, "learning_rate": 5.466115971575869e-05, "loss": 0.5889, "step": 14488 }, { "epoch": 0.9724505889064126, "grad_norm": 1.070196509361267, "learning_rate": 5.465033862821835e-05, "loss": 0.6092, "step": 14490 }, { "epoch": 0.9725848125901816, "grad_norm": 1.0266796350479126, "learning_rate": 5.463951732095862e-05, "loss": 0.5122, "step": 14492 }, { "epoch": 0.9727190362739505, "grad_norm": 0.8820559978485107, "learning_rate": 5.462869579449085e-05, "loss": 0.5148, "step": 14494 }, { "epoch": 0.9728532599577195, "grad_norm": 0.9386069178581238, "learning_rate": 5.461787404932629e-05, "loss": 0.6387, "step": 14496 }, { "epoch": 0.9729874836414886, "grad_norm": 1.0852800607681274, "learning_rate": 5.460705208597626e-05, "loss": 0.5331, "step": 14498 }, { "epoch": 0.9731217073252575, "grad_norm": 1.0661139488220215, "learning_rate": 5.459622990495209e-05, "loss": 0.6374, "step": 14500 }, { "epoch": 0.9732559310090265, "grad_norm": 1.011678695678711, "learning_rate": 5.458540750676509e-05, "loss": 0.5799, "step": 14502 }, { "epoch": 0.9733901546927956, "grad_norm": 0.9369387626647949, "learning_rate": 5.457458489192661e-05, "loss": 0.551, "step": 14504 }, { "epoch": 0.9735243783765646, "grad_norm": 1.2200850248336792, "learning_rate": 5.4563762060947975e-05, "loss": 0.5718, "step": 14506 }, { "epoch": 0.9736586020603335, "grad_norm": 0.9318606853485107, "learning_rate": 5.455293901434056e-05, "loss": 0.6158, "step": 14508 }, { "epoch": 0.9737928257441025, "grad_norm": 1.0478731393814087, "learning_rate": 5.454211575261574e-05, "loss": 0.5957, "step": 14510 }, { "epoch": 0.9739270494278716, "grad_norm": 1.01186203956604, "learning_rate": 5.453129227628487e-05, "loss": 0.5676, "step": 14512 }, { "epoch": 0.9740612731116406, "grad_norm": 0.862026572227478, "learning_rate": 5.4520468585859364e-05, "loss": 0.5009, "step": 14514 }, { "epoch": 0.9741954967954095, "grad_norm": 0.9080526828765869, "learning_rate": 5.450964468185059e-05, "loss": 0.5186, "step": 14516 }, { "epoch": 0.9743297204791785, "grad_norm": 0.9445207715034485, "learning_rate": 5.4498820564769994e-05, "loss": 0.5707, "step": 14518 }, { "epoch": 0.9744639441629476, "grad_norm": 1.0916472673416138, "learning_rate": 5.4487996235128946e-05, "loss": 0.6095, "step": 14520 }, { "epoch": 0.9745981678467166, "grad_norm": 1.0981478691101074, "learning_rate": 5.447717169343892e-05, "loss": 0.5759, "step": 14522 }, { "epoch": 0.9747323915304855, "grad_norm": 0.9555610418319702, "learning_rate": 5.4466346940211345e-05, "loss": 0.4984, "step": 14524 }, { "epoch": 0.9748666152142546, "grad_norm": 0.9845288991928101, "learning_rate": 5.4455521975957635e-05, "loss": 0.5628, "step": 14526 }, { "epoch": 0.9750008388980236, "grad_norm": 0.9806883335113525, "learning_rate": 5.444469680118929e-05, "loss": 0.6304, "step": 14528 }, { "epoch": 0.9751350625817926, "grad_norm": 1.305885910987854, "learning_rate": 5.443387141641775e-05, "loss": 0.494, "step": 14530 }, { "epoch": 0.9752692862655615, "grad_norm": 1.145750641822815, "learning_rate": 5.4423045822154506e-05, "loss": 0.55, "step": 14532 }, { "epoch": 0.9754035099493306, "grad_norm": 1.3389184474945068, "learning_rate": 5.4412220018911056e-05, "loss": 0.6095, "step": 14534 }, { "epoch": 0.9755377336330996, "grad_norm": 1.0017625093460083, "learning_rate": 5.4401394007198866e-05, "loss": 0.5358, "step": 14536 }, { "epoch": 0.9756719573168685, "grad_norm": 1.0230153799057007, "learning_rate": 5.4390567787529476e-05, "loss": 0.563, "step": 14538 }, { "epoch": 0.9758061810006375, "grad_norm": 0.8892274498939514, "learning_rate": 5.437974136041439e-05, "loss": 0.5384, "step": 14540 }, { "epoch": 0.9759404046844066, "grad_norm": 1.2023080587387085, "learning_rate": 5.436891472636515e-05, "loss": 0.6027, "step": 14542 }, { "epoch": 0.9760746283681756, "grad_norm": 1.0622117519378662, "learning_rate": 5.435808788589326e-05, "loss": 0.5884, "step": 14544 }, { "epoch": 0.9762088520519445, "grad_norm": 0.9289963841438293, "learning_rate": 5.4347260839510296e-05, "loss": 0.5231, "step": 14546 }, { "epoch": 0.9763430757357136, "grad_norm": 0.9408954381942749, "learning_rate": 5.433643358772781e-05, "loss": 0.5371, "step": 14548 }, { "epoch": 0.9764772994194826, "grad_norm": 1.0064287185668945, "learning_rate": 5.432560613105736e-05, "loss": 0.6041, "step": 14550 }, { "epoch": 0.9766115231032516, "grad_norm": 1.038871169090271, "learning_rate": 5.431477847001053e-05, "loss": 0.5369, "step": 14552 }, { "epoch": 0.9767457467870205, "grad_norm": 1.119463324546814, "learning_rate": 5.4303950605098894e-05, "loss": 0.5582, "step": 14554 }, { "epoch": 0.9768799704707896, "grad_norm": 0.9053308963775635, "learning_rate": 5.429312253683406e-05, "loss": 0.5107, "step": 14556 }, { "epoch": 0.9770141941545586, "grad_norm": 1.0307782888412476, "learning_rate": 5.428229426572763e-05, "loss": 0.5846, "step": 14558 }, { "epoch": 0.9771484178383276, "grad_norm": 1.3395590782165527, "learning_rate": 5.427146579229121e-05, "loss": 0.5869, "step": 14560 }, { "epoch": 0.9772826415220965, "grad_norm": 1.043879747390747, "learning_rate": 5.426063711703644e-05, "loss": 0.5227, "step": 14562 }, { "epoch": 0.9774168652058656, "grad_norm": 1.0789127349853516, "learning_rate": 5.424980824047493e-05, "loss": 0.4961, "step": 14564 }, { "epoch": 0.9775510888896346, "grad_norm": 0.843816876411438, "learning_rate": 5.4238979163118355e-05, "loss": 0.5498, "step": 14566 }, { "epoch": 0.9776853125734036, "grad_norm": 1.064896821975708, "learning_rate": 5.422814988547834e-05, "loss": 0.5129, "step": 14568 }, { "epoch": 0.9778195362571726, "grad_norm": 1.0035399198532104, "learning_rate": 5.421732040806656e-05, "loss": 0.5442, "step": 14570 }, { "epoch": 0.9779537599409416, "grad_norm": 0.9078490734100342, "learning_rate": 5.420649073139469e-05, "loss": 0.5404, "step": 14572 }, { "epoch": 0.9780879836247106, "grad_norm": 1.3465280532836914, "learning_rate": 5.419566085597439e-05, "loss": 0.5681, "step": 14574 }, { "epoch": 0.9782222073084795, "grad_norm": 1.1271129846572876, "learning_rate": 5.418483078231737e-05, "loss": 0.5544, "step": 14576 }, { "epoch": 0.9783564309922486, "grad_norm": 1.011531114578247, "learning_rate": 5.417400051093533e-05, "loss": 0.5734, "step": 14578 }, { "epoch": 0.9784906546760176, "grad_norm": 1.0569415092468262, "learning_rate": 5.416317004233997e-05, "loss": 0.5964, "step": 14580 }, { "epoch": 0.9786248783597866, "grad_norm": 1.0962992906570435, "learning_rate": 5.4152339377043015e-05, "loss": 0.6113, "step": 14582 }, { "epoch": 0.9787591020435555, "grad_norm": 1.5367470979690552, "learning_rate": 5.41415085155562e-05, "loss": 0.57, "step": 14584 }, { "epoch": 0.9788933257273246, "grad_norm": 0.9857184290885925, "learning_rate": 5.413067745839123e-05, "loss": 0.5527, "step": 14586 }, { "epoch": 0.9790275494110936, "grad_norm": 1.040051817893982, "learning_rate": 5.4119846206059876e-05, "loss": 0.4739, "step": 14588 }, { "epoch": 0.9791617730948626, "grad_norm": 1.4327423572540283, "learning_rate": 5.410901475907391e-05, "loss": 0.6003, "step": 14590 }, { "epoch": 0.9792959967786316, "grad_norm": 1.0983351469039917, "learning_rate": 5.409818311794506e-05, "loss": 0.5514, "step": 14592 }, { "epoch": 0.9794302204624006, "grad_norm": 0.9713186025619507, "learning_rate": 5.4087351283185116e-05, "loss": 0.5475, "step": 14594 }, { "epoch": 0.9795644441461696, "grad_norm": 1.1389487981796265, "learning_rate": 5.407651925530587e-05, "loss": 0.5809, "step": 14596 }, { "epoch": 0.9796986678299386, "grad_norm": 1.1033782958984375, "learning_rate": 5.406568703481909e-05, "loss": 0.5278, "step": 14598 }, { "epoch": 0.9798328915137076, "grad_norm": 1.0787391662597656, "learning_rate": 5.4054854622236614e-05, "loss": 0.6168, "step": 14600 }, { "epoch": 0.9799671151974766, "grad_norm": 1.0263839960098267, "learning_rate": 5.4044022018070214e-05, "loss": 0.6141, "step": 14602 }, { "epoch": 0.9801013388812456, "grad_norm": 1.0619577169418335, "learning_rate": 5.4033189222831735e-05, "loss": 0.5731, "step": 14604 }, { "epoch": 0.9802355625650147, "grad_norm": 0.9042320847511292, "learning_rate": 5.402235623703299e-05, "loss": 0.5621, "step": 14606 }, { "epoch": 0.9803697862487836, "grad_norm": 1.096408724784851, "learning_rate": 5.401152306118582e-05, "loss": 0.5749, "step": 14608 }, { "epoch": 0.9805040099325526, "grad_norm": 0.9234085083007812, "learning_rate": 5.400068969580209e-05, "loss": 0.5746, "step": 14610 }, { "epoch": 0.9806382336163216, "grad_norm": 1.074016809463501, "learning_rate": 5.398985614139361e-05, "loss": 0.5535, "step": 14612 }, { "epoch": 0.9807724573000905, "grad_norm": 1.464599609375, "learning_rate": 5.3979022398472304e-05, "loss": 0.5655, "step": 14614 }, { "epoch": 0.9809066809838596, "grad_norm": 1.0339241027832031, "learning_rate": 5.396818846754999e-05, "loss": 0.5225, "step": 14616 }, { "epoch": 0.9810409046676286, "grad_norm": 1.2816940546035767, "learning_rate": 5.39573543491386e-05, "loss": 0.5979, "step": 14618 }, { "epoch": 0.9811751283513976, "grad_norm": 1.0052075386047363, "learning_rate": 5.3946520043749974e-05, "loss": 0.5497, "step": 14620 }, { "epoch": 0.9813093520351666, "grad_norm": 1.0185097455978394, "learning_rate": 5.3935685551896045e-05, "loss": 0.5869, "step": 14622 }, { "epoch": 0.9814435757189356, "grad_norm": 1.0448133945465088, "learning_rate": 5.392485087408872e-05, "loss": 0.5521, "step": 14624 }, { "epoch": 0.9815777994027046, "grad_norm": 1.0752489566802979, "learning_rate": 5.39140160108399e-05, "loss": 0.5742, "step": 14626 }, { "epoch": 0.9817120230864737, "grad_norm": 1.1034209728240967, "learning_rate": 5.390318096266152e-05, "loss": 0.5859, "step": 14628 }, { "epoch": 0.9818462467702426, "grad_norm": 0.8905879855155945, "learning_rate": 5.3892345730065506e-05, "loss": 0.5794, "step": 14630 }, { "epoch": 0.9819804704540116, "grad_norm": 0.9774551391601562, "learning_rate": 5.3881510313563835e-05, "loss": 0.5619, "step": 14632 }, { "epoch": 0.9821146941377806, "grad_norm": 0.9077772498130798, "learning_rate": 5.387067471366841e-05, "loss": 0.5547, "step": 14634 }, { "epoch": 0.9822489178215497, "grad_norm": 0.9890025854110718, "learning_rate": 5.385983893089123e-05, "loss": 0.5386, "step": 14636 }, { "epoch": 0.9823831415053186, "grad_norm": 1.1164836883544922, "learning_rate": 5.384900296574425e-05, "loss": 0.6223, "step": 14638 }, { "epoch": 0.9825173651890876, "grad_norm": 0.9294412732124329, "learning_rate": 5.383816681873944e-05, "loss": 0.5361, "step": 14640 }, { "epoch": 0.9826515888728566, "grad_norm": 1.1060147285461426, "learning_rate": 5.3827330490388804e-05, "loss": 0.5258, "step": 14642 }, { "epoch": 0.9827858125566257, "grad_norm": 1.018481969833374, "learning_rate": 5.381649398120433e-05, "loss": 0.533, "step": 14644 }, { "epoch": 0.9829200362403946, "grad_norm": 1.081281304359436, "learning_rate": 5.380565729169802e-05, "loss": 0.5637, "step": 14646 }, { "epoch": 0.9830542599241636, "grad_norm": 1.0529273748397827, "learning_rate": 5.379482042238187e-05, "loss": 0.5198, "step": 14648 }, { "epoch": 0.9831884836079327, "grad_norm": 1.2811602354049683, "learning_rate": 5.378398337376793e-05, "loss": 0.6008, "step": 14650 }, { "epoch": 0.9833227072917016, "grad_norm": 0.9522740840911865, "learning_rate": 5.377314614636822e-05, "loss": 0.5655, "step": 14652 }, { "epoch": 0.9834569309754706, "grad_norm": 0.9688047170639038, "learning_rate": 5.3762308740694765e-05, "loss": 0.5304, "step": 14654 }, { "epoch": 0.9835911546592396, "grad_norm": 0.9964549541473389, "learning_rate": 5.375147115725964e-05, "loss": 0.5505, "step": 14656 }, { "epoch": 0.9837253783430087, "grad_norm": 1.0041929483413696, "learning_rate": 5.374063339657486e-05, "loss": 0.5451, "step": 14658 }, { "epoch": 0.9838596020267776, "grad_norm": 0.9628466367721558, "learning_rate": 5.372979545915252e-05, "loss": 0.4683, "step": 14660 }, { "epoch": 0.9839938257105466, "grad_norm": 0.9024425148963928, "learning_rate": 5.371895734550467e-05, "loss": 0.5437, "step": 14662 }, { "epoch": 0.9841280493943156, "grad_norm": 0.9593817591667175, "learning_rate": 5.3708119056143404e-05, "loss": 0.6366, "step": 14664 }, { "epoch": 0.9842622730780847, "grad_norm": 1.0373190641403198, "learning_rate": 5.3697280591580804e-05, "loss": 0.5662, "step": 14666 }, { "epoch": 0.9843964967618536, "grad_norm": 1.1384583711624146, "learning_rate": 5.368644195232896e-05, "loss": 0.5668, "step": 14668 }, { "epoch": 0.9845307204456226, "grad_norm": 0.8748607039451599, "learning_rate": 5.367560313889998e-05, "loss": 0.5586, "step": 14670 }, { "epoch": 0.9846649441293917, "grad_norm": 0.9103529453277588, "learning_rate": 5.366476415180599e-05, "loss": 0.5621, "step": 14672 }, { "epoch": 0.9847991678131607, "grad_norm": 1.006399393081665, "learning_rate": 5.365392499155909e-05, "loss": 0.548, "step": 14674 }, { "epoch": 0.9849333914969296, "grad_norm": 1.0762403011322021, "learning_rate": 5.3643085658671435e-05, "loss": 0.5084, "step": 14676 }, { "epoch": 0.9850676151806986, "grad_norm": 1.1388792991638184, "learning_rate": 5.363224615365513e-05, "loss": 0.5364, "step": 14678 }, { "epoch": 0.9852018388644677, "grad_norm": 0.9625356793403625, "learning_rate": 5.362140647702235e-05, "loss": 0.5392, "step": 14680 }, { "epoch": 0.9853360625482367, "grad_norm": 1.1637482643127441, "learning_rate": 5.361056662928522e-05, "loss": 0.5696, "step": 14682 }, { "epoch": 0.9854702862320056, "grad_norm": 0.9725779294967651, "learning_rate": 5.3599726610955926e-05, "loss": 0.5755, "step": 14684 }, { "epoch": 0.9856045099157746, "grad_norm": 1.0699341297149658, "learning_rate": 5.358888642254663e-05, "loss": 0.5043, "step": 14686 }, { "epoch": 0.9857387335995437, "grad_norm": 1.067064881324768, "learning_rate": 5.35780460645695e-05, "loss": 0.4789, "step": 14688 }, { "epoch": 0.9858729572833126, "grad_norm": 0.9795387387275696, "learning_rate": 5.3567205537536735e-05, "loss": 0.6033, "step": 14690 }, { "epoch": 0.9860071809670816, "grad_norm": 1.0278023481369019, "learning_rate": 5.3556364841960524e-05, "loss": 0.5999, "step": 14692 }, { "epoch": 0.9861414046508507, "grad_norm": 1.0064140558242798, "learning_rate": 5.3545523978353073e-05, "loss": 0.5623, "step": 14694 }, { "epoch": 0.9862756283346197, "grad_norm": 1.255136251449585, "learning_rate": 5.3534682947226576e-05, "loss": 0.5479, "step": 14696 }, { "epoch": 0.9864098520183886, "grad_norm": 0.9306426644325256, "learning_rate": 5.352384174909325e-05, "loss": 0.5063, "step": 14698 }, { "epoch": 0.9865440757021576, "grad_norm": 1.0653096437454224, "learning_rate": 5.351300038446535e-05, "loss": 0.5808, "step": 14700 }, { "epoch": 0.9866782993859267, "grad_norm": 1.0382238626480103, "learning_rate": 5.350215885385509e-05, "loss": 0.6025, "step": 14702 }, { "epoch": 0.9868125230696957, "grad_norm": 1.0285108089447021, "learning_rate": 5.34913171577747e-05, "loss": 0.565, "step": 14704 }, { "epoch": 0.9869467467534646, "grad_norm": 1.061592936515808, "learning_rate": 5.348047529673644e-05, "loss": 0.5551, "step": 14706 }, { "epoch": 0.9870809704372336, "grad_norm": 1.0288335084915161, "learning_rate": 5.346963327125258e-05, "loss": 0.6314, "step": 14708 }, { "epoch": 0.9872151941210027, "grad_norm": 0.976237416267395, "learning_rate": 5.3458791081835356e-05, "loss": 0.5668, "step": 14710 }, { "epoch": 0.9873494178047717, "grad_norm": 1.0939053297042847, "learning_rate": 5.3447948728997054e-05, "loss": 0.6221, "step": 14712 }, { "epoch": 0.9874836414885406, "grad_norm": 0.9784379005432129, "learning_rate": 5.3437106213249964e-05, "loss": 0.6294, "step": 14714 }, { "epoch": 0.9876178651723097, "grad_norm": 1.2871649265289307, "learning_rate": 5.3426263535106355e-05, "loss": 0.6035, "step": 14716 }, { "epoch": 0.9877520888560787, "grad_norm": 1.0066332817077637, "learning_rate": 5.3415420695078555e-05, "loss": 0.5382, "step": 14718 }, { "epoch": 0.9878863125398477, "grad_norm": 0.8645907640457153, "learning_rate": 5.3404577693678815e-05, "loss": 0.5568, "step": 14720 }, { "epoch": 0.9880205362236166, "grad_norm": 0.8626660108566284, "learning_rate": 5.3393734531419496e-05, "loss": 0.5174, "step": 14722 }, { "epoch": 0.9881547599073857, "grad_norm": 1.0427008867263794, "learning_rate": 5.338289120881287e-05, "loss": 0.5293, "step": 14724 }, { "epoch": 0.9882889835911547, "grad_norm": 1.114475131034851, "learning_rate": 5.33720477263713e-05, "loss": 0.6314, "step": 14726 }, { "epoch": 0.9884232072749236, "grad_norm": 0.9445342421531677, "learning_rate": 5.336120408460711e-05, "loss": 0.5164, "step": 14728 }, { "epoch": 0.9885574309586926, "grad_norm": 0.9713104963302612, "learning_rate": 5.335036028403264e-05, "loss": 0.4932, "step": 14730 }, { "epoch": 0.9886916546424617, "grad_norm": 0.9602975249290466, "learning_rate": 5.333951632516023e-05, "loss": 0.5653, "step": 14732 }, { "epoch": 0.9888258783262307, "grad_norm": 1.1221202611923218, "learning_rate": 5.332867220850223e-05, "loss": 0.5752, "step": 14734 }, { "epoch": 0.9889601020099996, "grad_norm": 1.0036790370941162, "learning_rate": 5.331782793457102e-05, "loss": 0.5924, "step": 14736 }, { "epoch": 0.9890943256937687, "grad_norm": 1.0514169931411743, "learning_rate": 5.330698350387897e-05, "loss": 0.4869, "step": 14738 }, { "epoch": 0.9892285493775377, "grad_norm": 0.9604849815368652, "learning_rate": 5.329613891693843e-05, "loss": 0.5585, "step": 14740 }, { "epoch": 0.9893627730613067, "grad_norm": 0.91641765832901, "learning_rate": 5.3285294174261836e-05, "loss": 0.5653, "step": 14742 }, { "epoch": 0.9894969967450756, "grad_norm": 0.9412071108818054, "learning_rate": 5.3274449276361535e-05, "loss": 0.5276, "step": 14744 }, { "epoch": 0.9896312204288447, "grad_norm": 1.0002440214157104, "learning_rate": 5.3263604223749954e-05, "loss": 0.5431, "step": 14746 }, { "epoch": 0.9897654441126137, "grad_norm": 1.19967520236969, "learning_rate": 5.3252759016939494e-05, "loss": 0.5119, "step": 14748 }, { "epoch": 0.9898996677963827, "grad_norm": 1.0427453517913818, "learning_rate": 5.3241913656442557e-05, "loss": 0.5676, "step": 14750 }, { "epoch": 0.9900338914801516, "grad_norm": 0.9265989661216736, "learning_rate": 5.3231068142771576e-05, "loss": 0.525, "step": 14752 }, { "epoch": 0.9901681151639207, "grad_norm": 1.0798743963241577, "learning_rate": 5.322022247643897e-05, "loss": 0.5615, "step": 14754 }, { "epoch": 0.9903023388476897, "grad_norm": 1.1865473985671997, "learning_rate": 5.320937665795721e-05, "loss": 0.5714, "step": 14756 }, { "epoch": 0.9904365625314587, "grad_norm": 0.9494701027870178, "learning_rate": 5.31985306878387e-05, "loss": 0.5346, "step": 14758 }, { "epoch": 0.9905707862152276, "grad_norm": 1.1785624027252197, "learning_rate": 5.31876845665959e-05, "loss": 0.5902, "step": 14760 }, { "epoch": 0.9907050098989967, "grad_norm": 1.032488465309143, "learning_rate": 5.317683829474127e-05, "loss": 0.5525, "step": 14762 }, { "epoch": 0.9908392335827657, "grad_norm": 1.1259870529174805, "learning_rate": 5.3165991872787276e-05, "loss": 0.5442, "step": 14764 }, { "epoch": 0.9909734572665346, "grad_norm": 0.9362767934799194, "learning_rate": 5.315514530124641e-05, "loss": 0.5118, "step": 14766 }, { "epoch": 0.9911076809503037, "grad_norm": 0.9072252511978149, "learning_rate": 5.3144298580631124e-05, "loss": 0.5172, "step": 14768 }, { "epoch": 0.9912419046340727, "grad_norm": 1.0078966617584229, "learning_rate": 5.313345171145392e-05, "loss": 0.5842, "step": 14770 }, { "epoch": 0.9913761283178417, "grad_norm": 1.0020571947097778, "learning_rate": 5.3122604694227265e-05, "loss": 0.5355, "step": 14772 }, { "epoch": 0.9915103520016106, "grad_norm": 1.0428295135498047, "learning_rate": 5.311175752946369e-05, "loss": 0.5712, "step": 14774 }, { "epoch": 0.9916445756853797, "grad_norm": 0.929182767868042, "learning_rate": 5.310091021767569e-05, "loss": 0.5704, "step": 14776 }, { "epoch": 0.9917787993691487, "grad_norm": 1.0764765739440918, "learning_rate": 5.309006275937578e-05, "loss": 0.5831, "step": 14778 }, { "epoch": 0.9919130230529177, "grad_norm": 0.9372965693473816, "learning_rate": 5.307921515507649e-05, "loss": 0.5835, "step": 14780 }, { "epoch": 0.9920472467366866, "grad_norm": 0.9608304500579834, "learning_rate": 5.306836740529032e-05, "loss": 0.5846, "step": 14782 }, { "epoch": 0.9921814704204557, "grad_norm": 1.067068099975586, "learning_rate": 5.305751951052984e-05, "loss": 0.5471, "step": 14784 }, { "epoch": 0.9923156941042247, "grad_norm": 0.9308376908302307, "learning_rate": 5.3046671471307566e-05, "loss": 0.5986, "step": 14786 }, { "epoch": 0.9924499177879937, "grad_norm": 1.5615354776382446, "learning_rate": 5.303582328813605e-05, "loss": 0.5178, "step": 14788 }, { "epoch": 0.9925841414717627, "grad_norm": 1.149412751197815, "learning_rate": 5.302497496152787e-05, "loss": 0.5592, "step": 14790 }, { "epoch": 0.9927183651555317, "grad_norm": 1.004371166229248, "learning_rate": 5.3014126491995566e-05, "loss": 0.5396, "step": 14792 }, { "epoch": 0.9928525888393007, "grad_norm": 0.9424512386322021, "learning_rate": 5.300327788005171e-05, "loss": 0.5271, "step": 14794 }, { "epoch": 0.9929868125230698, "grad_norm": 0.9469589591026306, "learning_rate": 5.2992429126208875e-05, "loss": 0.5848, "step": 14796 }, { "epoch": 0.9931210362068387, "grad_norm": 1.0328859090805054, "learning_rate": 5.2981580230979664e-05, "loss": 0.5462, "step": 14798 }, { "epoch": 0.9932552598906077, "grad_norm": 1.0615562200546265, "learning_rate": 5.297073119487664e-05, "loss": 0.5663, "step": 14800 }, { "epoch": 0.9933894835743767, "grad_norm": 0.840920627117157, "learning_rate": 5.295988201841241e-05, "loss": 0.5381, "step": 14802 }, { "epoch": 0.9935237072581456, "grad_norm": 0.9971423149108887, "learning_rate": 5.294903270209958e-05, "loss": 0.6138, "step": 14804 }, { "epoch": 0.9936579309419147, "grad_norm": 0.92223060131073, "learning_rate": 5.2938183246450735e-05, "loss": 0.5215, "step": 14806 }, { "epoch": 0.9937921546256837, "grad_norm": 1.0310229063034058, "learning_rate": 5.2927333651978526e-05, "loss": 0.513, "step": 14808 }, { "epoch": 0.9939263783094527, "grad_norm": 0.995556116104126, "learning_rate": 5.291648391919555e-05, "loss": 0.5637, "step": 14810 }, { "epoch": 0.9940606019932217, "grad_norm": 1.0456836223602295, "learning_rate": 5.290563404861445e-05, "loss": 0.5778, "step": 14812 }, { "epoch": 0.9941948256769907, "grad_norm": 1.004275918006897, "learning_rate": 5.289478404074786e-05, "loss": 0.5476, "step": 14814 }, { "epoch": 0.9943290493607597, "grad_norm": 0.9248266220092773, "learning_rate": 5.28839338961084e-05, "loss": 0.4775, "step": 14816 }, { "epoch": 0.9944632730445288, "grad_norm": 0.8902478218078613, "learning_rate": 5.287308361520875e-05, "loss": 0.5171, "step": 14818 }, { "epoch": 0.9945974967282977, "grad_norm": 0.9084435105323792, "learning_rate": 5.2862233198561525e-05, "loss": 0.5864, "step": 14820 }, { "epoch": 0.9947317204120667, "grad_norm": 0.9576614499092102, "learning_rate": 5.285138264667944e-05, "loss": 0.5086, "step": 14822 }, { "epoch": 0.9948659440958357, "grad_norm": 0.906507134437561, "learning_rate": 5.284053196007511e-05, "loss": 0.5744, "step": 14824 }, { "epoch": 0.9950001677796048, "grad_norm": 0.9690229296684265, "learning_rate": 5.282968113926123e-05, "loss": 0.5425, "step": 14826 }, { "epoch": 0.9951343914633737, "grad_norm": 1.0923490524291992, "learning_rate": 5.28188301847505e-05, "loss": 0.5831, "step": 14828 }, { "epoch": 0.9952686151471427, "grad_norm": 0.9221930503845215, "learning_rate": 5.280797909705557e-05, "loss": 0.5376, "step": 14830 }, { "epoch": 0.9954028388309117, "grad_norm": 1.0290145874023438, "learning_rate": 5.279712787668917e-05, "loss": 0.5997, "step": 14832 }, { "epoch": 0.9955370625146808, "grad_norm": 1.0450056791305542, "learning_rate": 5.278627652416396e-05, "loss": 0.5372, "step": 14834 }, { "epoch": 0.9956712861984497, "grad_norm": 1.0815494060516357, "learning_rate": 5.277542503999265e-05, "loss": 0.6227, "step": 14836 }, { "epoch": 0.9958055098822187, "grad_norm": 1.0937553644180298, "learning_rate": 5.2764573424688e-05, "loss": 0.5485, "step": 14838 }, { "epoch": 0.9959397335659878, "grad_norm": 1.1829911470413208, "learning_rate": 5.275372167876267e-05, "loss": 0.5357, "step": 14840 }, { "epoch": 0.9960739572497567, "grad_norm": 0.966904878616333, "learning_rate": 5.274286980272941e-05, "loss": 0.5506, "step": 14842 }, { "epoch": 0.9962081809335257, "grad_norm": 1.0377508401870728, "learning_rate": 5.273201779710094e-05, "loss": 0.5892, "step": 14844 }, { "epoch": 0.9963424046172947, "grad_norm": 1.8484457731246948, "learning_rate": 5.2721165662390014e-05, "loss": 0.5935, "step": 14846 }, { "epoch": 0.9964766283010638, "grad_norm": 0.9310474991798401, "learning_rate": 5.2710313399109346e-05, "loss": 0.5466, "step": 14848 }, { "epoch": 0.9966108519848327, "grad_norm": 1.152030348777771, "learning_rate": 5.26994610077717e-05, "loss": 0.5936, "step": 14850 }, { "epoch": 0.9967450756686017, "grad_norm": 1.1568481922149658, "learning_rate": 5.2688608488889844e-05, "loss": 0.5835, "step": 14852 }, { "epoch": 0.9968792993523707, "grad_norm": 2.007110834121704, "learning_rate": 5.2677755842976516e-05, "loss": 0.5801, "step": 14854 }, { "epoch": 0.9970135230361398, "grad_norm": 1.1071622371673584, "learning_rate": 5.266690307054449e-05, "loss": 0.4839, "step": 14856 }, { "epoch": 0.9971477467199087, "grad_norm": 1.1917561292648315, "learning_rate": 5.265605017210654e-05, "loss": 0.6797, "step": 14858 }, { "epoch": 0.9972819704036777, "grad_norm": 0.9590668082237244, "learning_rate": 5.264519714817544e-05, "loss": 0.5584, "step": 14860 }, { "epoch": 0.9974161940874468, "grad_norm": 1.018331527709961, "learning_rate": 5.263434399926398e-05, "loss": 0.5471, "step": 14862 }, { "epoch": 0.9975504177712158, "grad_norm": 1.0494593381881714, "learning_rate": 5.2623490725884936e-05, "loss": 0.629, "step": 14864 }, { "epoch": 0.9976846414549847, "grad_norm": 1.1049399375915527, "learning_rate": 5.261263732855113e-05, "loss": 0.5519, "step": 14866 }, { "epoch": 0.9978188651387537, "grad_norm": 1.0918635129928589, "learning_rate": 5.260178380777534e-05, "loss": 0.5383, "step": 14868 }, { "epoch": 0.9979530888225228, "grad_norm": 1.036096453666687, "learning_rate": 5.2590930164070384e-05, "loss": 0.5483, "step": 14870 }, { "epoch": 0.9980873125062918, "grad_norm": 1.629939317703247, "learning_rate": 5.258007639794907e-05, "loss": 0.5799, "step": 14872 }, { "epoch": 0.9982215361900607, "grad_norm": 0.9630006551742554, "learning_rate": 5.256922250992421e-05, "loss": 0.5127, "step": 14874 }, { "epoch": 0.9983557598738297, "grad_norm": 1.0506491661071777, "learning_rate": 5.255836850050866e-05, "loss": 0.5702, "step": 14876 }, { "epoch": 0.9984899835575988, "grad_norm": 1.0084110498428345, "learning_rate": 5.2547514370215214e-05, "loss": 0.547, "step": 14878 }, { "epoch": 0.9986242072413677, "grad_norm": 0.9572237133979797, "learning_rate": 5.253666011955673e-05, "loss": 0.5034, "step": 14880 }, { "epoch": 0.9987584309251367, "grad_norm": 0.914639949798584, "learning_rate": 5.252580574904603e-05, "loss": 0.5491, "step": 14882 }, { "epoch": 0.9988926546089058, "grad_norm": 1.0369341373443604, "learning_rate": 5.2514951259195985e-05, "loss": 0.5681, "step": 14884 }, { "epoch": 0.9990268782926748, "grad_norm": 1.4023351669311523, "learning_rate": 5.250409665051944e-05, "loss": 0.6777, "step": 14886 }, { "epoch": 0.9991611019764437, "grad_norm": 1.0324677228927612, "learning_rate": 5.2493241923529245e-05, "loss": 0.607, "step": 14888 }, { "epoch": 0.9992953256602127, "grad_norm": 0.9444645643234253, "learning_rate": 5.248238707873828e-05, "loss": 0.5395, "step": 14890 }, { "epoch": 0.9994295493439818, "grad_norm": 1.0191946029663086, "learning_rate": 5.24715321166594e-05, "loss": 0.5444, "step": 14892 }, { "epoch": 0.9995637730277508, "grad_norm": 1.0185309648513794, "learning_rate": 5.246067703780551e-05, "loss": 0.5728, "step": 14894 }, { "epoch": 0.9996979967115197, "grad_norm": 1.0221612453460693, "learning_rate": 5.244982184268944e-05, "loss": 0.5277, "step": 14896 }, { "epoch": 0.9998322203952887, "grad_norm": 0.8939405083656311, "learning_rate": 5.243896653182412e-05, "loss": 0.5152, "step": 14898 }, { "epoch": 0.9999664440790578, "grad_norm": 0.9363802075386047, "learning_rate": 5.242811110572242e-05, "loss": 0.5334, "step": 14900 }, { "epoch": 1.0000671118418845, "grad_norm": 0.926179051399231, "learning_rate": 5.241725556489724e-05, "loss": 0.3938, "step": 14902 }, { "epoch": 1.0002013355256536, "grad_norm": 1.053067684173584, "learning_rate": 5.2406399909861505e-05, "loss": 0.5814, "step": 14904 }, { "epoch": 1.0003355592094225, "grad_norm": 1.0768216848373413, "learning_rate": 5.239554414112809e-05, "loss": 0.539, "step": 14906 }, { "epoch": 1.0004697828931914, "grad_norm": 0.8976664543151855, "learning_rate": 5.238468825920994e-05, "loss": 0.5091, "step": 14908 }, { "epoch": 1.0006040065769606, "grad_norm": 1.118083119392395, "learning_rate": 5.237383226461995e-05, "loss": 0.5734, "step": 14910 }, { "epoch": 1.0007382302607295, "grad_norm": 0.9272267818450928, "learning_rate": 5.236297615787106e-05, "loss": 0.5075, "step": 14912 }, { "epoch": 1.0008724539444984, "grad_norm": 0.9571478366851807, "learning_rate": 5.235211993947617e-05, "loss": 0.5533, "step": 14914 }, { "epoch": 1.0010066776282676, "grad_norm": 1.372754693031311, "learning_rate": 5.2341263609948254e-05, "loss": 0.5162, "step": 14916 }, { "epoch": 1.0011409013120365, "grad_norm": 0.9124377965927124, "learning_rate": 5.2330407169800245e-05, "loss": 0.4917, "step": 14918 }, { "epoch": 1.0012751249958054, "grad_norm": 1.0203857421875, "learning_rate": 5.2319550619545074e-05, "loss": 0.5309, "step": 14920 }, { "epoch": 1.0014093486795745, "grad_norm": 1.0425175428390503, "learning_rate": 5.23086939596957e-05, "loss": 0.539, "step": 14922 }, { "epoch": 1.0015435723633435, "grad_norm": 0.9315015077590942, "learning_rate": 5.229783719076506e-05, "loss": 0.5044, "step": 14924 }, { "epoch": 1.0016777960471126, "grad_norm": 1.1324846744537354, "learning_rate": 5.2286980313266145e-05, "loss": 0.5346, "step": 14926 }, { "epoch": 1.0018120197308815, "grad_norm": 1.1176323890686035, "learning_rate": 5.227612332771191e-05, "loss": 0.5452, "step": 14928 }, { "epoch": 1.0019462434146504, "grad_norm": 0.9105541706085205, "learning_rate": 5.2265266234615315e-05, "loss": 0.5263, "step": 14930 }, { "epoch": 1.0020804670984196, "grad_norm": 0.9921098351478577, "learning_rate": 5.225440903448936e-05, "loss": 0.4861, "step": 14932 }, { "epoch": 1.0022146907821885, "grad_norm": 0.9832185506820679, "learning_rate": 5.2243551727847e-05, "loss": 0.5168, "step": 14934 }, { "epoch": 1.0023489144659574, "grad_norm": 0.9794383645057678, "learning_rate": 5.223269431520124e-05, "loss": 0.5213, "step": 14936 }, { "epoch": 1.0024831381497266, "grad_norm": 1.0107759237289429, "learning_rate": 5.222183679706506e-05, "loss": 0.5449, "step": 14938 }, { "epoch": 1.0026173618334955, "grad_norm": 0.9340489506721497, "learning_rate": 5.221097917395146e-05, "loss": 0.5418, "step": 14940 }, { "epoch": 1.0027515855172646, "grad_norm": 1.0198230743408203, "learning_rate": 5.220012144637344e-05, "loss": 0.4777, "step": 14942 }, { "epoch": 1.0028858092010335, "grad_norm": 1.0641562938690186, "learning_rate": 5.2189263614843995e-05, "loss": 0.592, "step": 14944 }, { "epoch": 1.0030200328848025, "grad_norm": 1.0694611072540283, "learning_rate": 5.2178405679876174e-05, "loss": 0.4984, "step": 14946 }, { "epoch": 1.0031542565685716, "grad_norm": 1.0537960529327393, "learning_rate": 5.216754764198295e-05, "loss": 0.541, "step": 14948 }, { "epoch": 1.0032884802523405, "grad_norm": 1.0442579984664917, "learning_rate": 5.215668950167737e-05, "loss": 0.5575, "step": 14950 }, { "epoch": 1.0034227039361094, "grad_norm": 0.9863391518592834, "learning_rate": 5.214583125947245e-05, "loss": 0.5007, "step": 14952 }, { "epoch": 1.0035569276198786, "grad_norm": 0.8672400712966919, "learning_rate": 5.213497291588122e-05, "loss": 0.5325, "step": 14954 }, { "epoch": 1.0036911513036475, "grad_norm": 1.105807900428772, "learning_rate": 5.212411447141673e-05, "loss": 0.5258, "step": 14956 }, { "epoch": 1.0038253749874164, "grad_norm": 1.0071481466293335, "learning_rate": 5.2113255926592e-05, "loss": 0.5268, "step": 14958 }, { "epoch": 1.0039595986711856, "grad_norm": 1.0454081296920776, "learning_rate": 5.210239728192008e-05, "loss": 0.5541, "step": 14960 }, { "epoch": 1.0040938223549545, "grad_norm": 0.9871719479560852, "learning_rate": 5.209153853791402e-05, "loss": 0.5713, "step": 14962 }, { "epoch": 1.0042280460387236, "grad_norm": 1.0889171361923218, "learning_rate": 5.208067969508688e-05, "loss": 0.5697, "step": 14964 }, { "epoch": 1.0043622697224925, "grad_norm": 1.007730484008789, "learning_rate": 5.2069820753951726e-05, "loss": 0.5457, "step": 14966 }, { "epoch": 1.0044964934062615, "grad_norm": 1.1622997522354126, "learning_rate": 5.205896171502159e-05, "loss": 0.6189, "step": 14968 }, { "epoch": 1.0046307170900306, "grad_norm": 1.2704561948776245, "learning_rate": 5.204810257880959e-05, "loss": 0.477, "step": 14970 }, { "epoch": 1.0047649407737995, "grad_norm": 1.12416672706604, "learning_rate": 5.203724334582875e-05, "loss": 0.5957, "step": 14972 }, { "epoch": 1.0048991644575684, "grad_norm": 1.3205918073654175, "learning_rate": 5.202638401659218e-05, "loss": 0.5413, "step": 14974 }, { "epoch": 1.0050333881413376, "grad_norm": 1.1126925945281982, "learning_rate": 5.201552459161293e-05, "loss": 0.5216, "step": 14976 }, { "epoch": 1.0051676118251065, "grad_norm": 1.7548534870147705, "learning_rate": 5.200466507140411e-05, "loss": 0.5176, "step": 14978 }, { "epoch": 1.0053018355088756, "grad_norm": 1.009682536125183, "learning_rate": 5.1993805456478826e-05, "loss": 0.5166, "step": 14980 }, { "epoch": 1.0054360591926446, "grad_norm": 1.3446884155273438, "learning_rate": 5.198294574735014e-05, "loss": 0.5367, "step": 14982 }, { "epoch": 1.0055702828764135, "grad_norm": 1.5239311456680298, "learning_rate": 5.197208594453118e-05, "loss": 0.5185, "step": 14984 }, { "epoch": 1.0057045065601826, "grad_norm": 1.0466574430465698, "learning_rate": 5.1961226048535026e-05, "loss": 0.5689, "step": 14986 }, { "epoch": 1.0058387302439515, "grad_norm": 0.9727661609649658, "learning_rate": 5.195036605987481e-05, "loss": 0.5662, "step": 14988 }, { "epoch": 1.0059729539277205, "grad_norm": 1.0696989297866821, "learning_rate": 5.193950597906362e-05, "loss": 0.6016, "step": 14990 }, { "epoch": 1.0061071776114896, "grad_norm": 0.9576994776725769, "learning_rate": 5.192864580661459e-05, "loss": 0.4973, "step": 14992 }, { "epoch": 1.0062414012952585, "grad_norm": 1.0753369331359863, "learning_rate": 5.1917785543040844e-05, "loss": 0.5364, "step": 14994 }, { "epoch": 1.0063756249790274, "grad_norm": 1.042697548866272, "learning_rate": 5.1906925188855494e-05, "loss": 0.5297, "step": 14996 }, { "epoch": 1.0065098486627966, "grad_norm": 0.9751782417297363, "learning_rate": 5.189606474457168e-05, "loss": 0.546, "step": 14998 }, { "epoch": 1.0066440723465655, "grad_norm": 0.9660375714302063, "learning_rate": 5.188520421070253e-05, "loss": 0.5445, "step": 15000 }, { "epoch": 1.0067782960303346, "grad_norm": 1.259419560432434, "learning_rate": 5.187434358776121e-05, "loss": 0.5468, "step": 15002 }, { "epoch": 1.0069125197141036, "grad_norm": 1.0952138900756836, "learning_rate": 5.186348287626083e-05, "loss": 0.5825, "step": 15004 }, { "epoch": 1.0070467433978725, "grad_norm": 1.0939810276031494, "learning_rate": 5.185262207671454e-05, "loss": 0.5217, "step": 15006 }, { "epoch": 1.0071809670816416, "grad_norm": 1.4713177680969238, "learning_rate": 5.184176118963551e-05, "loss": 0.5227, "step": 15008 }, { "epoch": 1.0073151907654105, "grad_norm": 0.9648658633232117, "learning_rate": 5.183090021553687e-05, "loss": 0.4656, "step": 15010 }, { "epoch": 1.0074494144491795, "grad_norm": 1.2975503206253052, "learning_rate": 5.182003915493181e-05, "loss": 0.4875, "step": 15012 }, { "epoch": 1.0075836381329486, "grad_norm": 1.0438868999481201, "learning_rate": 5.180917800833347e-05, "loss": 0.5622, "step": 15014 }, { "epoch": 1.0077178618167175, "grad_norm": 0.9114359021186829, "learning_rate": 5.179831677625503e-05, "loss": 0.5518, "step": 15016 }, { "epoch": 1.0078520855004867, "grad_norm": 0.9837536215782166, "learning_rate": 5.1787455459209665e-05, "loss": 0.5371, "step": 15018 }, { "epoch": 1.0079863091842556, "grad_norm": 1.137822151184082, "learning_rate": 5.1776594057710535e-05, "loss": 0.5564, "step": 15020 }, { "epoch": 1.0081205328680245, "grad_norm": 1.026039958000183, "learning_rate": 5.176573257227084e-05, "loss": 0.493, "step": 15022 }, { "epoch": 1.0082547565517936, "grad_norm": 0.9626202583312988, "learning_rate": 5.175487100340374e-05, "loss": 0.4941, "step": 15024 }, { "epoch": 1.0083889802355626, "grad_norm": 0.8714377284049988, "learning_rate": 5.1744009351622434e-05, "loss": 0.4707, "step": 15026 }, { "epoch": 1.0085232039193315, "grad_norm": 0.9464874267578125, "learning_rate": 5.173314761744014e-05, "loss": 0.4727, "step": 15028 }, { "epoch": 1.0086574276031006, "grad_norm": 0.9182068705558777, "learning_rate": 5.172228580137001e-05, "loss": 0.5544, "step": 15030 }, { "epoch": 1.0087916512868695, "grad_norm": 0.984942615032196, "learning_rate": 5.171142390392526e-05, "loss": 0.5588, "step": 15032 }, { "epoch": 1.0089258749706387, "grad_norm": 0.9684733748435974, "learning_rate": 5.1700561925619096e-05, "loss": 0.5343, "step": 15034 }, { "epoch": 1.0090600986544076, "grad_norm": 0.8601505160331726, "learning_rate": 5.1689699866964735e-05, "loss": 0.5005, "step": 15036 }, { "epoch": 1.0091943223381765, "grad_norm": 1.0310051441192627, "learning_rate": 5.167883772847537e-05, "loss": 0.4845, "step": 15038 }, { "epoch": 1.0093285460219457, "grad_norm": 1.2362431287765503, "learning_rate": 5.166797551066422e-05, "loss": 0.5716, "step": 15040 }, { "epoch": 1.0094627697057146, "grad_norm": 0.9385756850242615, "learning_rate": 5.165711321404453e-05, "loss": 0.5015, "step": 15042 }, { "epoch": 1.0095969933894835, "grad_norm": 0.9549235701560974, "learning_rate": 5.1646250839129476e-05, "loss": 0.5317, "step": 15044 }, { "epoch": 1.0097312170732526, "grad_norm": 1.0057629346847534, "learning_rate": 5.163538838643232e-05, "loss": 0.511, "step": 15046 }, { "epoch": 1.0098654407570216, "grad_norm": 1.0021158456802368, "learning_rate": 5.162452585646627e-05, "loss": 0.5398, "step": 15048 }, { "epoch": 1.0099996644407905, "grad_norm": 1.0679807662963867, "learning_rate": 5.161366324974457e-05, "loss": 0.5315, "step": 15050 }, { "epoch": 1.0101338881245596, "grad_norm": 0.9807422161102295, "learning_rate": 5.160280056678045e-05, "loss": 0.5698, "step": 15052 }, { "epoch": 1.0102681118083285, "grad_norm": 0.9693793058395386, "learning_rate": 5.159193780808716e-05, "loss": 0.4875, "step": 15054 }, { "epoch": 1.0104023354920977, "grad_norm": 0.9912816286087036, "learning_rate": 5.158107497417795e-05, "loss": 0.5324, "step": 15056 }, { "epoch": 1.0105365591758666, "grad_norm": 0.9343096017837524, "learning_rate": 5.157021206556604e-05, "loss": 0.5687, "step": 15058 }, { "epoch": 1.0106707828596355, "grad_norm": 0.8785936832427979, "learning_rate": 5.155934908276471e-05, "loss": 0.4924, "step": 15060 }, { "epoch": 1.0108050065434047, "grad_norm": 1.003708004951477, "learning_rate": 5.154848602628719e-05, "loss": 0.5324, "step": 15062 }, { "epoch": 1.0109392302271736, "grad_norm": 0.9341418147087097, "learning_rate": 5.153762289664676e-05, "loss": 0.5417, "step": 15064 }, { "epoch": 1.0110734539109425, "grad_norm": 0.9692851901054382, "learning_rate": 5.1526759694356665e-05, "loss": 0.5643, "step": 15066 }, { "epoch": 1.0112076775947116, "grad_norm": 0.9379444122314453, "learning_rate": 5.151589641993018e-05, "loss": 0.55, "step": 15068 }, { "epoch": 1.0113419012784806, "grad_norm": 1.1084202527999878, "learning_rate": 5.150503307388057e-05, "loss": 0.5434, "step": 15070 }, { "epoch": 1.0114761249622495, "grad_norm": 0.9925508499145508, "learning_rate": 5.1494169656721104e-05, "loss": 0.5256, "step": 15072 }, { "epoch": 1.0116103486460186, "grad_norm": 1.0956413745880127, "learning_rate": 5.148330616896506e-05, "loss": 0.5348, "step": 15074 }, { "epoch": 1.0117445723297875, "grad_norm": 1.0557767152786255, "learning_rate": 5.147244261112573e-05, "loss": 0.5522, "step": 15076 }, { "epoch": 1.0118787960135567, "grad_norm": 1.0028269290924072, "learning_rate": 5.1461578983716365e-05, "loss": 0.4744, "step": 15078 }, { "epoch": 1.0120130196973256, "grad_norm": 1.0762214660644531, "learning_rate": 5.145071528725027e-05, "loss": 0.5087, "step": 15080 }, { "epoch": 1.0121472433810945, "grad_norm": 0.9658424258232117, "learning_rate": 5.143985152224073e-05, "loss": 0.5149, "step": 15082 }, { "epoch": 1.0122814670648637, "grad_norm": 1.0201389789581299, "learning_rate": 5.1428987689201046e-05, "loss": 0.5152, "step": 15084 }, { "epoch": 1.0124156907486326, "grad_norm": 1.068712830543518, "learning_rate": 5.1418123788644504e-05, "loss": 0.5835, "step": 15086 }, { "epoch": 1.0125499144324015, "grad_norm": 1.046278715133667, "learning_rate": 5.1407259821084384e-05, "loss": 0.5451, "step": 15088 }, { "epoch": 1.0126841381161706, "grad_norm": 1.3016585111618042, "learning_rate": 5.139639578703403e-05, "loss": 0.5872, "step": 15090 }, { "epoch": 1.0128183617999396, "grad_norm": 1.1036945581436157, "learning_rate": 5.1385531687006695e-05, "loss": 0.5226, "step": 15092 }, { "epoch": 1.0129525854837087, "grad_norm": 1.1816788911819458, "learning_rate": 5.137466752151573e-05, "loss": 0.5442, "step": 15094 }, { "epoch": 1.0130868091674776, "grad_norm": 1.0297253131866455, "learning_rate": 5.136380329107442e-05, "loss": 0.5106, "step": 15096 }, { "epoch": 1.0132210328512465, "grad_norm": 1.035953164100647, "learning_rate": 5.1352938996196084e-05, "loss": 0.5429, "step": 15098 }, { "epoch": 1.0133552565350157, "grad_norm": 0.8848439455032349, "learning_rate": 5.134207463739404e-05, "loss": 0.5334, "step": 15100 }, { "epoch": 1.0134894802187846, "grad_norm": 1.233180046081543, "learning_rate": 5.1331210215181613e-05, "loss": 0.5403, "step": 15102 }, { "epoch": 1.0136237039025535, "grad_norm": 0.961600124835968, "learning_rate": 5.132034573007213e-05, "loss": 0.4906, "step": 15104 }, { "epoch": 1.0137579275863227, "grad_norm": 0.9835965037345886, "learning_rate": 5.130948118257889e-05, "loss": 0.478, "step": 15106 }, { "epoch": 1.0138921512700916, "grad_norm": 1.0651036500930786, "learning_rate": 5.129861657321525e-05, "loss": 0.553, "step": 15108 }, { "epoch": 1.0140263749538605, "grad_norm": 1.226927399635315, "learning_rate": 5.128775190249453e-05, "loss": 0.5391, "step": 15110 }, { "epoch": 1.0141605986376296, "grad_norm": 1.0278208255767822, "learning_rate": 5.1276887170930064e-05, "loss": 0.5731, "step": 15112 }, { "epoch": 1.0142948223213986, "grad_norm": 1.1334031820297241, "learning_rate": 5.1266022379035196e-05, "loss": 0.5337, "step": 15114 }, { "epoch": 1.0144290460051677, "grad_norm": 1.5725743770599365, "learning_rate": 5.125515752732324e-05, "loss": 0.522, "step": 15116 }, { "epoch": 1.0145632696889366, "grad_norm": 1.0932778120040894, "learning_rate": 5.1244292616307574e-05, "loss": 0.5713, "step": 15118 }, { "epoch": 1.0146974933727055, "grad_norm": 0.8721126317977905, "learning_rate": 5.123342764650153e-05, "loss": 0.4632, "step": 15120 }, { "epoch": 1.0148317170564747, "grad_norm": 0.9035164713859558, "learning_rate": 5.122256261841846e-05, "loss": 0.4953, "step": 15122 }, { "epoch": 1.0149659407402436, "grad_norm": 1.0134105682373047, "learning_rate": 5.12116975325717e-05, "loss": 0.5629, "step": 15124 }, { "epoch": 1.0151001644240125, "grad_norm": 0.9002230763435364, "learning_rate": 5.120083238947462e-05, "loss": 0.497, "step": 15126 }, { "epoch": 1.0152343881077817, "grad_norm": 1.014272928237915, "learning_rate": 5.1189967189640566e-05, "loss": 0.4951, "step": 15128 }, { "epoch": 1.0153686117915506, "grad_norm": 1.0982106924057007, "learning_rate": 5.11791019335829e-05, "loss": 0.5601, "step": 15130 }, { "epoch": 1.0155028354753197, "grad_norm": 1.020804762840271, "learning_rate": 5.116823662181499e-05, "loss": 0.5075, "step": 15132 }, { "epoch": 1.0156370591590886, "grad_norm": 0.7800991535186768, "learning_rate": 5.11573712548502e-05, "loss": 0.4879, "step": 15134 }, { "epoch": 1.0157712828428576, "grad_norm": 1.0576050281524658, "learning_rate": 5.11465058332019e-05, "loss": 0.5168, "step": 15136 }, { "epoch": 1.0159055065266267, "grad_norm": 0.9641649723052979, "learning_rate": 5.113564035738343e-05, "loss": 0.5333, "step": 15138 }, { "epoch": 1.0160397302103956, "grad_norm": 1.0422450304031372, "learning_rate": 5.1124774827908215e-05, "loss": 0.4981, "step": 15140 }, { "epoch": 1.0161739538941645, "grad_norm": 0.9134299755096436, "learning_rate": 5.111390924528959e-05, "loss": 0.5266, "step": 15142 }, { "epoch": 1.0163081775779337, "grad_norm": 0.8555437326431274, "learning_rate": 5.1103043610040934e-05, "loss": 0.5251, "step": 15144 }, { "epoch": 1.0164424012617026, "grad_norm": 0.8712595701217651, "learning_rate": 5.1092177922675655e-05, "loss": 0.5532, "step": 15146 }, { "epoch": 1.0165766249454715, "grad_norm": 0.9501728415489197, "learning_rate": 5.1081312183707095e-05, "loss": 0.5305, "step": 15148 }, { "epoch": 1.0167108486292407, "grad_norm": 1.0660793781280518, "learning_rate": 5.107044639364867e-05, "loss": 0.5201, "step": 15150 }, { "epoch": 1.0168450723130096, "grad_norm": 1.0098044872283936, "learning_rate": 5.1059580553013764e-05, "loss": 0.516, "step": 15152 }, { "epoch": 1.0169792959967787, "grad_norm": 1.2454719543457031, "learning_rate": 5.104871466231575e-05, "loss": 0.511, "step": 15154 }, { "epoch": 1.0171135196805476, "grad_norm": 0.9541716575622559, "learning_rate": 5.103784872206805e-05, "loss": 0.5125, "step": 15156 }, { "epoch": 1.0172477433643166, "grad_norm": 1.0367459058761597, "learning_rate": 5.1026982732784014e-05, "loss": 0.5279, "step": 15158 }, { "epoch": 1.0173819670480857, "grad_norm": 1.1213152408599854, "learning_rate": 5.101611669497708e-05, "loss": 0.6091, "step": 15160 }, { "epoch": 1.0175161907318546, "grad_norm": 0.9246627688407898, "learning_rate": 5.1005250609160625e-05, "loss": 0.4449, "step": 15162 }, { "epoch": 1.0176504144156235, "grad_norm": 0.9593459963798523, "learning_rate": 5.099438447584804e-05, "loss": 0.5165, "step": 15164 }, { "epoch": 1.0177846380993927, "grad_norm": 1.5162698030471802, "learning_rate": 5.098351829555277e-05, "loss": 0.5046, "step": 15166 }, { "epoch": 1.0179188617831616, "grad_norm": 0.9656062126159668, "learning_rate": 5.097265206878817e-05, "loss": 0.5115, "step": 15168 }, { "epoch": 1.0180530854669307, "grad_norm": 1.1321715116500854, "learning_rate": 5.0961785796067696e-05, "loss": 0.5514, "step": 15170 }, { "epoch": 1.0181873091506997, "grad_norm": 0.9948189854621887, "learning_rate": 5.095091947790472e-05, "loss": 0.552, "step": 15172 }, { "epoch": 1.0183215328344686, "grad_norm": 1.1939868927001953, "learning_rate": 5.094005311481268e-05, "loss": 0.6173, "step": 15174 }, { "epoch": 1.0184557565182377, "grad_norm": 1.0231444835662842, "learning_rate": 5.0929186707304964e-05, "loss": 0.5605, "step": 15176 }, { "epoch": 1.0185899802020066, "grad_norm": 0.9155842065811157, "learning_rate": 5.0918320255895014e-05, "loss": 0.5419, "step": 15178 }, { "epoch": 1.0187242038857756, "grad_norm": 1.0757856369018555, "learning_rate": 5.0907453761096235e-05, "loss": 0.6102, "step": 15180 }, { "epoch": 1.0188584275695447, "grad_norm": 0.9137722253799438, "learning_rate": 5.0896587223422044e-05, "loss": 0.538, "step": 15182 }, { "epoch": 1.0189926512533136, "grad_norm": 0.923180341720581, "learning_rate": 5.088572064338588e-05, "loss": 0.5489, "step": 15184 }, { "epoch": 1.0191268749370828, "grad_norm": 1.0486664772033691, "learning_rate": 5.087485402150115e-05, "loss": 0.5017, "step": 15186 }, { "epoch": 1.0192610986208517, "grad_norm": 0.9409403204917908, "learning_rate": 5.086398735828129e-05, "loss": 0.5356, "step": 15188 }, { "epoch": 1.0193953223046206, "grad_norm": 1.0220710039138794, "learning_rate": 5.085312065423973e-05, "loss": 0.5132, "step": 15190 }, { "epoch": 1.0195295459883897, "grad_norm": 0.9534757733345032, "learning_rate": 5.084225390988989e-05, "loss": 0.479, "step": 15192 }, { "epoch": 1.0196637696721587, "grad_norm": 1.086157202720642, "learning_rate": 5.083138712574522e-05, "loss": 0.4957, "step": 15194 }, { "epoch": 1.0197979933559276, "grad_norm": 1.0429636240005493, "learning_rate": 5.082052030231913e-05, "loss": 0.5382, "step": 15196 }, { "epoch": 1.0199322170396967, "grad_norm": 1.0217652320861816, "learning_rate": 5.080965344012508e-05, "loss": 0.5078, "step": 15198 }, { "epoch": 1.0200664407234656, "grad_norm": 1.0648549795150757, "learning_rate": 5.079878653967649e-05, "loss": 0.5477, "step": 15200 }, { "epoch": 1.0202006644072346, "grad_norm": 1.0137075185775757, "learning_rate": 5.078791960148681e-05, "loss": 0.5905, "step": 15202 }, { "epoch": 1.0203348880910037, "grad_norm": 0.9725910425186157, "learning_rate": 5.0777052626069476e-05, "loss": 0.4865, "step": 15204 }, { "epoch": 1.0204691117747726, "grad_norm": 0.9284847974777222, "learning_rate": 5.076618561393793e-05, "loss": 0.5589, "step": 15206 }, { "epoch": 1.0206033354585418, "grad_norm": 0.9524595737457275, "learning_rate": 5.075531856560563e-05, "loss": 0.584, "step": 15208 }, { "epoch": 1.0207375591423107, "grad_norm": 1.1114261150360107, "learning_rate": 5.074445148158601e-05, "loss": 0.5155, "step": 15210 }, { "epoch": 1.0208717828260796, "grad_norm": 0.8127722144126892, "learning_rate": 5.073358436239252e-05, "loss": 0.5009, "step": 15212 }, { "epoch": 1.0210060065098487, "grad_norm": 1.8480273485183716, "learning_rate": 5.072271720853862e-05, "loss": 0.5675, "step": 15214 }, { "epoch": 1.0211402301936177, "grad_norm": 1.0490953922271729, "learning_rate": 5.071185002053773e-05, "loss": 0.5311, "step": 15216 }, { "epoch": 1.0212744538773866, "grad_norm": 0.9619256258010864, "learning_rate": 5.0700982798903355e-05, "loss": 0.5182, "step": 15218 }, { "epoch": 1.0214086775611557, "grad_norm": 1.031275987625122, "learning_rate": 5.06901155441489e-05, "loss": 0.5755, "step": 15220 }, { "epoch": 1.0215429012449246, "grad_norm": 0.938673734664917, "learning_rate": 5.067924825678785e-05, "loss": 0.4703, "step": 15222 }, { "epoch": 1.0216771249286936, "grad_norm": 1.615932822227478, "learning_rate": 5.066838093733365e-05, "loss": 0.5154, "step": 15224 }, { "epoch": 1.0218113486124627, "grad_norm": 0.8763120770454407, "learning_rate": 5.065751358629976e-05, "loss": 0.4773, "step": 15226 }, { "epoch": 1.0219455722962316, "grad_norm": 1.0280568599700928, "learning_rate": 5.0646646204199663e-05, "loss": 0.4761, "step": 15228 }, { "epoch": 1.0220797959800008, "grad_norm": 0.9118609428405762, "learning_rate": 5.0635778791546794e-05, "loss": 0.4487, "step": 15230 }, { "epoch": 1.0222140196637697, "grad_norm": 1.0934200286865234, "learning_rate": 5.062491134885463e-05, "loss": 0.5314, "step": 15232 }, { "epoch": 1.0223482433475386, "grad_norm": 0.9376586079597473, "learning_rate": 5.061404387663663e-05, "loss": 0.4847, "step": 15234 }, { "epoch": 1.0224824670313077, "grad_norm": 1.2409019470214844, "learning_rate": 5.060317637540628e-05, "loss": 0.5537, "step": 15236 }, { "epoch": 1.0226166907150767, "grad_norm": 1.0247794389724731, "learning_rate": 5.059230884567702e-05, "loss": 0.51, "step": 15238 }, { "epoch": 1.0227509143988456, "grad_norm": 0.9596056938171387, "learning_rate": 5.058144128796232e-05, "loss": 0.5794, "step": 15240 }, { "epoch": 1.0228851380826147, "grad_norm": 1.0091639757156372, "learning_rate": 5.057057370277568e-05, "loss": 0.5205, "step": 15242 }, { "epoch": 1.0230193617663836, "grad_norm": 1.0142924785614014, "learning_rate": 5.0559706090630546e-05, "loss": 0.5319, "step": 15244 }, { "epoch": 1.0231535854501528, "grad_norm": 1.8155488967895508, "learning_rate": 5.0548838452040407e-05, "loss": 0.5499, "step": 15246 }, { "epoch": 1.0232878091339217, "grad_norm": 1.1085665225982666, "learning_rate": 5.053797078751872e-05, "loss": 0.5811, "step": 15248 }, { "epoch": 1.0234220328176906, "grad_norm": 1.546950101852417, "learning_rate": 5.052710309757899e-05, "loss": 0.5264, "step": 15250 }, { "epoch": 1.0235562565014598, "grad_norm": 1.0630773305892944, "learning_rate": 5.0516235382734654e-05, "loss": 0.5004, "step": 15252 }, { "epoch": 1.0236904801852287, "grad_norm": 1.086484432220459, "learning_rate": 5.050536764349921e-05, "loss": 0.5448, "step": 15254 }, { "epoch": 1.0238247038689976, "grad_norm": 0.962790846824646, "learning_rate": 5.049449988038616e-05, "loss": 0.4521, "step": 15256 }, { "epoch": 1.0239589275527667, "grad_norm": 1.0335437059402466, "learning_rate": 5.048363209390895e-05, "loss": 0.5506, "step": 15258 }, { "epoch": 1.0240931512365357, "grad_norm": 0.9776301980018616, "learning_rate": 5.047276428458108e-05, "loss": 0.5172, "step": 15260 }, { "epoch": 1.0242273749203046, "grad_norm": 1.1775176525115967, "learning_rate": 5.046189645291601e-05, "loss": 0.6449, "step": 15262 }, { "epoch": 1.0243615986040737, "grad_norm": 1.103438138961792, "learning_rate": 5.0451028599427265e-05, "loss": 0.5563, "step": 15264 }, { "epoch": 1.0244958222878426, "grad_norm": 3.5545434951782227, "learning_rate": 5.044016072462829e-05, "loss": 0.565, "step": 15266 }, { "epoch": 1.0246300459716118, "grad_norm": 0.9224105477333069, "learning_rate": 5.0429292829032596e-05, "loss": 0.5518, "step": 15268 }, { "epoch": 1.0247642696553807, "grad_norm": 0.9298921823501587, "learning_rate": 5.041842491315365e-05, "loss": 0.5029, "step": 15270 }, { "epoch": 1.0248984933391496, "grad_norm": 1.0152496099472046, "learning_rate": 5.040755697750496e-05, "loss": 0.5202, "step": 15272 }, { "epoch": 1.0250327170229188, "grad_norm": 1.5045030117034912, "learning_rate": 5.03966890226e-05, "loss": 0.5188, "step": 15274 }, { "epoch": 1.0251669407066877, "grad_norm": 1.1592066287994385, "learning_rate": 5.0385821048952255e-05, "loss": 0.5209, "step": 15276 }, { "epoch": 1.0253011643904566, "grad_norm": 0.8823243379592896, "learning_rate": 5.037495305707524e-05, "loss": 0.5019, "step": 15278 }, { "epoch": 1.0254353880742257, "grad_norm": 0.9502339363098145, "learning_rate": 5.0364085047482424e-05, "loss": 0.5374, "step": 15280 }, { "epoch": 1.0255696117579947, "grad_norm": 1.0712730884552002, "learning_rate": 5.035321702068731e-05, "loss": 0.5501, "step": 15282 }, { "epoch": 1.0257038354417638, "grad_norm": 0.9488974213600159, "learning_rate": 5.034234897720338e-05, "loss": 0.5536, "step": 15284 }, { "epoch": 1.0258380591255327, "grad_norm": 1.146383285522461, "learning_rate": 5.0331480917544126e-05, "loss": 0.5045, "step": 15286 }, { "epoch": 1.0259722828093016, "grad_norm": 1.014808177947998, "learning_rate": 5.032061284222307e-05, "loss": 0.5006, "step": 15288 }, { "epoch": 1.0261065064930708, "grad_norm": 1.062474012374878, "learning_rate": 5.0309744751753664e-05, "loss": 0.537, "step": 15290 }, { "epoch": 1.0262407301768397, "grad_norm": 1.1122550964355469, "learning_rate": 5.029887664664944e-05, "loss": 0.5578, "step": 15292 }, { "epoch": 1.0263749538606086, "grad_norm": 0.9269875288009644, "learning_rate": 5.028800852742388e-05, "loss": 0.4753, "step": 15294 }, { "epoch": 1.0265091775443778, "grad_norm": 1.066814661026001, "learning_rate": 5.0277140394590484e-05, "loss": 0.5387, "step": 15296 }, { "epoch": 1.0266434012281467, "grad_norm": 1.0602961778640747, "learning_rate": 5.026627224866276e-05, "loss": 0.5202, "step": 15298 }, { "epoch": 1.0267776249119156, "grad_norm": 1.077146053314209, "learning_rate": 5.025540409015418e-05, "loss": 0.5219, "step": 15300 }, { "epoch": 1.0269118485956847, "grad_norm": 1.0115227699279785, "learning_rate": 5.024453591957826e-05, "loss": 0.6292, "step": 15302 }, { "epoch": 1.0270460722794537, "grad_norm": 1.0494619607925415, "learning_rate": 5.023366773744851e-05, "loss": 0.539, "step": 15304 }, { "epoch": 1.0271802959632228, "grad_norm": 1.0594550371170044, "learning_rate": 5.022279954427841e-05, "loss": 0.5376, "step": 15306 }, { "epoch": 1.0273145196469917, "grad_norm": 0.9713077545166016, "learning_rate": 5.0211931340581476e-05, "loss": 0.5806, "step": 15308 }, { "epoch": 1.0274487433307606, "grad_norm": 0.9390246272087097, "learning_rate": 5.020106312687119e-05, "loss": 0.5121, "step": 15310 }, { "epoch": 1.0275829670145298, "grad_norm": 1.2264317274093628, "learning_rate": 5.0190194903661094e-05, "loss": 0.5358, "step": 15312 }, { "epoch": 1.0277171906982987, "grad_norm": 2.0339279174804688, "learning_rate": 5.0179326671464634e-05, "loss": 0.5839, "step": 15314 }, { "epoch": 1.0278514143820676, "grad_norm": 0.9309272766113281, "learning_rate": 5.016845843079535e-05, "loss": 0.5548, "step": 15316 }, { "epoch": 1.0279856380658368, "grad_norm": 0.9557270407676697, "learning_rate": 5.015759018216675e-05, "loss": 0.5128, "step": 15318 }, { "epoch": 1.0281198617496057, "grad_norm": 1.0359398126602173, "learning_rate": 5.01467219260923e-05, "loss": 0.5008, "step": 15320 }, { "epoch": 1.0282540854333748, "grad_norm": 0.9772992134094238, "learning_rate": 5.0135853663085554e-05, "loss": 0.4431, "step": 15322 }, { "epoch": 1.0283883091171437, "grad_norm": 1.0367960929870605, "learning_rate": 5.012498539365996e-05, "loss": 0.498, "step": 15324 }, { "epoch": 1.0285225328009127, "grad_norm": 0.9311704635620117, "learning_rate": 5.0114117118329075e-05, "loss": 0.5364, "step": 15326 }, { "epoch": 1.0286567564846818, "grad_norm": 1.004698634147644, "learning_rate": 5.0103248837606375e-05, "loss": 0.523, "step": 15328 }, { "epoch": 1.0287909801684507, "grad_norm": 0.9973604083061218, "learning_rate": 5.0092380552005355e-05, "loss": 0.5805, "step": 15330 }, { "epoch": 1.0289252038522196, "grad_norm": 1.0765000581741333, "learning_rate": 5.008151226203957e-05, "loss": 0.584, "step": 15332 }, { "epoch": 1.0290594275359888, "grad_norm": 0.9316189289093018, "learning_rate": 5.007064396822245e-05, "loss": 0.5076, "step": 15334 }, { "epoch": 1.0291936512197577, "grad_norm": 0.9343737959861755, "learning_rate": 5.0059775671067575e-05, "loss": 0.5298, "step": 15336 }, { "epoch": 1.0293278749035266, "grad_norm": 0.8994764685630798, "learning_rate": 5.0048907371088394e-05, "loss": 0.5493, "step": 15338 }, { "epoch": 1.0294620985872958, "grad_norm": 0.9307758212089539, "learning_rate": 5.0038039068798446e-05, "loss": 0.5207, "step": 15340 }, { "epoch": 1.0295963222710647, "grad_norm": 1.009665846824646, "learning_rate": 5.002717076471123e-05, "loss": 0.5141, "step": 15342 }, { "epoch": 1.0297305459548338, "grad_norm": 1.037687063217163, "learning_rate": 5.001630245934024e-05, "loss": 0.4918, "step": 15344 }, { "epoch": 1.0298647696386027, "grad_norm": 0.9489524364471436, "learning_rate": 5.000543415319901e-05, "loss": 0.5337, "step": 15346 }, { "epoch": 1.0299989933223717, "grad_norm": 0.9218322038650513, "learning_rate": 4.999456584680101e-05, "loss": 0.441, "step": 15348 }, { "epoch": 1.0301332170061408, "grad_norm": 1.0558072328567505, "learning_rate": 4.998369754065976e-05, "loss": 0.5311, "step": 15350 }, { "epoch": 1.0302674406899097, "grad_norm": 0.8926640152931213, "learning_rate": 4.997282923528878e-05, "loss": 0.487, "step": 15352 }, { "epoch": 1.0304016643736786, "grad_norm": 0.9766799807548523, "learning_rate": 4.996196093120156e-05, "loss": 0.4965, "step": 15354 }, { "epoch": 1.0305358880574478, "grad_norm": 0.9748830199241638, "learning_rate": 4.995109262891162e-05, "loss": 0.498, "step": 15356 }, { "epoch": 1.0306701117412167, "grad_norm": 1.2828388214111328, "learning_rate": 4.994022432893243e-05, "loss": 0.5085, "step": 15358 }, { "epoch": 1.0308043354249858, "grad_norm": 1.0764567852020264, "learning_rate": 4.9929356031777555e-05, "loss": 0.4781, "step": 15360 }, { "epoch": 1.0309385591087548, "grad_norm": 1.0225658416748047, "learning_rate": 4.991848773796044e-05, "loss": 0.5247, "step": 15362 }, { "epoch": 1.0310727827925237, "grad_norm": 1.0597457885742188, "learning_rate": 4.990761944799465e-05, "loss": 0.5318, "step": 15364 }, { "epoch": 1.0312070064762928, "grad_norm": 0.9888409972190857, "learning_rate": 4.989675116239363e-05, "loss": 0.5106, "step": 15366 }, { "epoch": 1.0313412301600617, "grad_norm": 1.002458095550537, "learning_rate": 4.988588288167093e-05, "loss": 0.5645, "step": 15368 }, { "epoch": 1.0314754538438307, "grad_norm": 1.1779125928878784, "learning_rate": 4.9875014606340035e-05, "loss": 0.6012, "step": 15370 }, { "epoch": 1.0316096775275998, "grad_norm": 1.0960805416107178, "learning_rate": 4.9864146336914465e-05, "loss": 0.5152, "step": 15372 }, { "epoch": 1.0317439012113687, "grad_norm": 1.0984114408493042, "learning_rate": 4.98532780739077e-05, "loss": 0.5857, "step": 15374 }, { "epoch": 1.0318781248951376, "grad_norm": 1.0565415620803833, "learning_rate": 4.984240981783327e-05, "loss": 0.5191, "step": 15376 }, { "epoch": 1.0320123485789068, "grad_norm": 0.964955747127533, "learning_rate": 4.9831541569204654e-05, "loss": 0.5259, "step": 15378 }, { "epoch": 1.0321465722626757, "grad_norm": 0.8490574359893799, "learning_rate": 4.982067332853538e-05, "loss": 0.5113, "step": 15380 }, { "epoch": 1.0322807959464448, "grad_norm": 1.0339528322219849, "learning_rate": 4.980980509633894e-05, "loss": 0.5321, "step": 15382 }, { "epoch": 1.0324150196302138, "grad_norm": 1.0278650522232056, "learning_rate": 4.979893687312881e-05, "loss": 0.5367, "step": 15384 }, { "epoch": 1.0325492433139827, "grad_norm": 0.9023627638816833, "learning_rate": 4.978806865941855e-05, "loss": 0.5279, "step": 15386 }, { "epoch": 1.0326834669977518, "grad_norm": 1.0649601221084595, "learning_rate": 4.97772004557216e-05, "loss": 0.5181, "step": 15388 }, { "epoch": 1.0328176906815207, "grad_norm": 0.95375657081604, "learning_rate": 4.9766332262551515e-05, "loss": 0.5993, "step": 15390 }, { "epoch": 1.0329519143652897, "grad_norm": 1.07741379737854, "learning_rate": 4.975546408042175e-05, "loss": 0.5013, "step": 15392 }, { "epoch": 1.0330861380490588, "grad_norm": 1.0350271463394165, "learning_rate": 4.9744595909845845e-05, "loss": 0.5081, "step": 15394 }, { "epoch": 1.0332203617328277, "grad_norm": 0.9855200052261353, "learning_rate": 4.973372775133725e-05, "loss": 0.6056, "step": 15396 }, { "epoch": 1.0333545854165969, "grad_norm": 1.0609577894210815, "learning_rate": 4.972285960540954e-05, "loss": 0.49, "step": 15398 }, { "epoch": 1.0334888091003658, "grad_norm": 0.8649885058403015, "learning_rate": 4.9711991472576124e-05, "loss": 0.4648, "step": 15400 }, { "epoch": 1.0336230327841347, "grad_norm": 1.088198184967041, "learning_rate": 4.9701123353350575e-05, "loss": 0.5857, "step": 15402 }, { "epoch": 1.0337572564679038, "grad_norm": 0.9336490631103516, "learning_rate": 4.969025524824634e-05, "loss": 0.5333, "step": 15404 }, { "epoch": 1.0338914801516728, "grad_norm": 0.9601561427116394, "learning_rate": 4.967938715777696e-05, "loss": 0.5048, "step": 15406 }, { "epoch": 1.0340257038354417, "grad_norm": 1.0431386232376099, "learning_rate": 4.9668519082455885e-05, "loss": 0.4623, "step": 15408 }, { "epoch": 1.0341599275192108, "grad_norm": 1.0056498050689697, "learning_rate": 4.965765102279664e-05, "loss": 0.5427, "step": 15410 }, { "epoch": 1.0342941512029797, "grad_norm": 1.0134047269821167, "learning_rate": 4.964678297931271e-05, "loss": 0.5011, "step": 15412 }, { "epoch": 1.0344283748867489, "grad_norm": 1.0183725357055664, "learning_rate": 4.9635914952517594e-05, "loss": 0.5092, "step": 15414 }, { "epoch": 1.0345625985705178, "grad_norm": 1.4094773530960083, "learning_rate": 4.9625046942924774e-05, "loss": 0.6091, "step": 15416 }, { "epoch": 1.0346968222542867, "grad_norm": 0.8718349933624268, "learning_rate": 4.9614178951047756e-05, "loss": 0.5464, "step": 15418 }, { "epoch": 1.0348310459380559, "grad_norm": 0.9513201117515564, "learning_rate": 4.9603310977400006e-05, "loss": 0.4983, "step": 15420 }, { "epoch": 1.0349652696218248, "grad_norm": 1.1781314611434937, "learning_rate": 4.959244302249506e-05, "loss": 0.5312, "step": 15422 }, { "epoch": 1.0350994933055937, "grad_norm": 1.0829561948776245, "learning_rate": 4.958157508684635e-05, "loss": 0.4857, "step": 15424 }, { "epoch": 1.0352337169893628, "grad_norm": 1.045479416847229, "learning_rate": 4.957070717096742e-05, "loss": 0.4792, "step": 15426 }, { "epoch": 1.0353679406731318, "grad_norm": 0.9355472326278687, "learning_rate": 4.955983927537171e-05, "loss": 0.514, "step": 15428 }, { "epoch": 1.0355021643569007, "grad_norm": 1.0759092569351196, "learning_rate": 4.9548971400572747e-05, "loss": 0.5504, "step": 15430 }, { "epoch": 1.0356363880406698, "grad_norm": 0.847141683101654, "learning_rate": 4.9538103547083985e-05, "loss": 0.5492, "step": 15432 }, { "epoch": 1.0357706117244387, "grad_norm": 0.9926348328590393, "learning_rate": 4.952723571541893e-05, "loss": 0.5228, "step": 15434 }, { "epoch": 1.0359048354082079, "grad_norm": 1.4465930461883545, "learning_rate": 4.9516367906091055e-05, "loss": 0.5384, "step": 15436 }, { "epoch": 1.0360390590919768, "grad_norm": 0.8559713959693909, "learning_rate": 4.950550011961385e-05, "loss": 0.5147, "step": 15438 }, { "epoch": 1.0361732827757457, "grad_norm": 1.0673739910125732, "learning_rate": 4.9494632356500785e-05, "loss": 0.5402, "step": 15440 }, { "epoch": 1.0363075064595149, "grad_norm": 0.8725389838218689, "learning_rate": 4.948376461726536e-05, "loss": 0.5465, "step": 15442 }, { "epoch": 1.0364417301432838, "grad_norm": 1.2083971500396729, "learning_rate": 4.947289690242102e-05, "loss": 0.565, "step": 15444 }, { "epoch": 1.0365759538270527, "grad_norm": 1.0119904279708862, "learning_rate": 4.9462029212481286e-05, "loss": 0.5557, "step": 15446 }, { "epoch": 1.0367101775108218, "grad_norm": 1.090747594833374, "learning_rate": 4.945116154795959e-05, "loss": 0.5061, "step": 15448 }, { "epoch": 1.0368444011945908, "grad_norm": 0.9607612490653992, "learning_rate": 4.9440293909369465e-05, "loss": 0.477, "step": 15450 }, { "epoch": 1.0369786248783597, "grad_norm": 0.9650788307189941, "learning_rate": 4.942942629722434e-05, "loss": 0.4996, "step": 15452 }, { "epoch": 1.0371128485621288, "grad_norm": 0.9474275708198547, "learning_rate": 4.941855871203769e-05, "loss": 0.5445, "step": 15454 }, { "epoch": 1.0372470722458977, "grad_norm": 1.009928584098816, "learning_rate": 4.940769115432301e-05, "loss": 0.5052, "step": 15456 }, { "epoch": 1.0373812959296669, "grad_norm": 1.0166889429092407, "learning_rate": 4.939682362459374e-05, "loss": 0.5236, "step": 15458 }, { "epoch": 1.0375155196134358, "grad_norm": 1.0336209535598755, "learning_rate": 4.938595612336339e-05, "loss": 0.5955, "step": 15460 }, { "epoch": 1.0376497432972047, "grad_norm": 0.9604952931404114, "learning_rate": 4.937508865114537e-05, "loss": 0.4671, "step": 15462 }, { "epoch": 1.0377839669809739, "grad_norm": 0.8933176398277283, "learning_rate": 4.936422120845322e-05, "loss": 0.5717, "step": 15464 }, { "epoch": 1.0379181906647428, "grad_norm": 1.7485871315002441, "learning_rate": 4.935335379580034e-05, "loss": 0.4836, "step": 15466 }, { "epoch": 1.0380524143485117, "grad_norm": 0.983931303024292, "learning_rate": 4.934248641370025e-05, "loss": 0.5091, "step": 15468 }, { "epoch": 1.0381866380322808, "grad_norm": 0.856342077255249, "learning_rate": 4.933161906266636e-05, "loss": 0.4919, "step": 15470 }, { "epoch": 1.0383208617160498, "grad_norm": 0.995201051235199, "learning_rate": 4.9320751743212176e-05, "loss": 0.5338, "step": 15472 }, { "epoch": 1.038455085399819, "grad_norm": 0.9791795015335083, "learning_rate": 4.930988445585112e-05, "loss": 0.5643, "step": 15474 }, { "epoch": 1.0385893090835878, "grad_norm": 0.9542690515518188, "learning_rate": 4.929901720109668e-05, "loss": 0.5248, "step": 15476 }, { "epoch": 1.0387235327673567, "grad_norm": 0.95274817943573, "learning_rate": 4.928814997946228e-05, "loss": 0.5349, "step": 15478 }, { "epoch": 1.0388577564511259, "grad_norm": 1.0679212808609009, "learning_rate": 4.927728279146141e-05, "loss": 0.5314, "step": 15480 }, { "epoch": 1.0389919801348948, "grad_norm": 1.0120261907577515, "learning_rate": 4.9266415637607494e-05, "loss": 0.5584, "step": 15482 }, { "epoch": 1.0391262038186637, "grad_norm": 0.9586861729621887, "learning_rate": 4.925554851841401e-05, "loss": 0.5243, "step": 15484 }, { "epoch": 1.0392604275024329, "grad_norm": 0.850306510925293, "learning_rate": 4.924468143439437e-05, "loss": 0.4676, "step": 15486 }, { "epoch": 1.0393946511862018, "grad_norm": 0.9537341594696045, "learning_rate": 4.923381438606208e-05, "loss": 0.4883, "step": 15488 }, { "epoch": 1.039528874869971, "grad_norm": 0.8852509260177612, "learning_rate": 4.922294737393053e-05, "loss": 0.532, "step": 15490 }, { "epoch": 1.0396630985537398, "grad_norm": 0.929013192653656, "learning_rate": 4.92120803985132e-05, "loss": 0.5263, "step": 15492 }, { "epoch": 1.0397973222375088, "grad_norm": 1.0598328113555908, "learning_rate": 4.920121346032351e-05, "loss": 0.5332, "step": 15494 }, { "epoch": 1.039931545921278, "grad_norm": 1.0350052118301392, "learning_rate": 4.919034655987493e-05, "loss": 0.494, "step": 15496 }, { "epoch": 1.0400657696050468, "grad_norm": 1.0005286931991577, "learning_rate": 4.917947969768087e-05, "loss": 0.5654, "step": 15498 }, { "epoch": 1.0401999932888157, "grad_norm": 1.1021864414215088, "learning_rate": 4.916861287425479e-05, "loss": 0.4904, "step": 15500 }, { "epoch": 1.0403342169725849, "grad_norm": 4.215618133544922, "learning_rate": 4.915774609011011e-05, "loss": 0.5055, "step": 15502 }, { "epoch": 1.0404684406563538, "grad_norm": 1.0950462818145752, "learning_rate": 4.914687934576028e-05, "loss": 0.6111, "step": 15504 }, { "epoch": 1.0406026643401227, "grad_norm": 2.1147496700286865, "learning_rate": 4.913601264171871e-05, "loss": 0.5393, "step": 15506 }, { "epoch": 1.0407368880238919, "grad_norm": 1.1324349641799927, "learning_rate": 4.912514597849886e-05, "loss": 0.5148, "step": 15508 }, { "epoch": 1.0408711117076608, "grad_norm": 1.0787025690078735, "learning_rate": 4.911427935661411e-05, "loss": 0.4627, "step": 15510 }, { "epoch": 1.04100533539143, "grad_norm": 1.0233891010284424, "learning_rate": 4.910341277657796e-05, "loss": 0.5605, "step": 15512 }, { "epoch": 1.0411395590751988, "grad_norm": 1.246164321899414, "learning_rate": 4.909254623890376e-05, "loss": 0.5491, "step": 15514 }, { "epoch": 1.0412737827589678, "grad_norm": 0.9711535573005676, "learning_rate": 4.9081679744105e-05, "loss": 0.5522, "step": 15516 }, { "epoch": 1.041408006442737, "grad_norm": 0.989448606967926, "learning_rate": 4.907081329269503e-05, "loss": 0.4817, "step": 15518 }, { "epoch": 1.0415422301265058, "grad_norm": 1.0142803192138672, "learning_rate": 4.905994688518733e-05, "loss": 0.5904, "step": 15520 }, { "epoch": 1.0416764538102747, "grad_norm": 0.8768779039382935, "learning_rate": 4.9049080522095296e-05, "loss": 0.5051, "step": 15522 }, { "epoch": 1.0418106774940439, "grad_norm": 0.9926018118858337, "learning_rate": 4.903821420393231e-05, "loss": 0.4733, "step": 15524 }, { "epoch": 1.0419449011778128, "grad_norm": 0.9845841526985168, "learning_rate": 4.902734793121183e-05, "loss": 0.5254, "step": 15526 }, { "epoch": 1.0420791248615817, "grad_norm": 0.9354475736618042, "learning_rate": 4.9016481704447245e-05, "loss": 0.4868, "step": 15528 }, { "epoch": 1.0422133485453509, "grad_norm": 1.0436205863952637, "learning_rate": 4.900561552415197e-05, "loss": 0.5086, "step": 15530 }, { "epoch": 1.0423475722291198, "grad_norm": 1.009522557258606, "learning_rate": 4.899474939083939e-05, "loss": 0.528, "step": 15532 }, { "epoch": 1.042481795912889, "grad_norm": 1.0397756099700928, "learning_rate": 4.8983883305022945e-05, "loss": 0.6065, "step": 15534 }, { "epoch": 1.0426160195966578, "grad_norm": 1.0646541118621826, "learning_rate": 4.8973017267216e-05, "loss": 0.5222, "step": 15536 }, { "epoch": 1.0427502432804268, "grad_norm": 1.0897085666656494, "learning_rate": 4.896215127793198e-05, "loss": 0.4996, "step": 15538 }, { "epoch": 1.042884466964196, "grad_norm": 0.9040811657905579, "learning_rate": 4.895128533768426e-05, "loss": 0.5078, "step": 15540 }, { "epoch": 1.0430186906479648, "grad_norm": 1.010197639465332, "learning_rate": 4.894041944698626e-05, "loss": 0.5786, "step": 15542 }, { "epoch": 1.0431529143317337, "grad_norm": 1.1387351751327515, "learning_rate": 4.892955360635134e-05, "loss": 0.4843, "step": 15544 }, { "epoch": 1.0432871380155029, "grad_norm": 1.2616219520568848, "learning_rate": 4.891868781629292e-05, "loss": 0.5211, "step": 15546 }, { "epoch": 1.0434213616992718, "grad_norm": 0.9757934808731079, "learning_rate": 4.890782207732436e-05, "loss": 0.5223, "step": 15548 }, { "epoch": 1.043555585383041, "grad_norm": 1.0289428234100342, "learning_rate": 4.8896956389959084e-05, "loss": 0.5326, "step": 15550 }, { "epoch": 1.0436898090668099, "grad_norm": 1.0326473712921143, "learning_rate": 4.888609075471042e-05, "loss": 0.5655, "step": 15552 }, { "epoch": 1.0438240327505788, "grad_norm": 0.9574761390686035, "learning_rate": 4.88752251720918e-05, "loss": 0.4885, "step": 15554 }, { "epoch": 1.043958256434348, "grad_norm": 1.0071873664855957, "learning_rate": 4.886435964261656e-05, "loss": 0.5466, "step": 15556 }, { "epoch": 1.0440924801181168, "grad_norm": 0.9385173916816711, "learning_rate": 4.885349416679812e-05, "loss": 0.5591, "step": 15558 }, { "epoch": 1.0442267038018858, "grad_norm": 1.1447536945343018, "learning_rate": 4.884262874514981e-05, "loss": 0.5402, "step": 15560 }, { "epoch": 1.044360927485655, "grad_norm": 1.043013095855713, "learning_rate": 4.883176337818502e-05, "loss": 0.4843, "step": 15562 }, { "epoch": 1.0444951511694238, "grad_norm": 0.9053693413734436, "learning_rate": 4.88208980664171e-05, "loss": 0.5234, "step": 15564 }, { "epoch": 1.044629374853193, "grad_norm": 1.0035746097564697, "learning_rate": 4.8810032810359446e-05, "loss": 0.5079, "step": 15566 }, { "epoch": 1.0447635985369619, "grad_norm": 1.0047979354858398, "learning_rate": 4.879916761052539e-05, "loss": 0.5525, "step": 15568 }, { "epoch": 1.0448978222207308, "grad_norm": 1.0252243280410767, "learning_rate": 4.8788302467428316e-05, "loss": 0.5254, "step": 15570 }, { "epoch": 1.0450320459045, "grad_norm": 1.0482876300811768, "learning_rate": 4.877743738158155e-05, "loss": 0.5176, "step": 15572 }, { "epoch": 1.0451662695882689, "grad_norm": 1.153091311454773, "learning_rate": 4.876657235349848e-05, "loss": 0.564, "step": 15574 }, { "epoch": 1.0453004932720378, "grad_norm": 0.9788694977760315, "learning_rate": 4.875570738369242e-05, "loss": 0.482, "step": 15576 }, { "epoch": 1.045434716955807, "grad_norm": 1.1284476518630981, "learning_rate": 4.8744842472676764e-05, "loss": 0.5772, "step": 15578 }, { "epoch": 1.0455689406395758, "grad_norm": 0.9172835350036621, "learning_rate": 4.873397762096481e-05, "loss": 0.504, "step": 15580 }, { "epoch": 1.0457031643233448, "grad_norm": 0.9032467007637024, "learning_rate": 4.872311282906994e-05, "loss": 0.5281, "step": 15582 }, { "epoch": 1.045837388007114, "grad_norm": 1.1164354085922241, "learning_rate": 4.871224809750547e-05, "loss": 0.4898, "step": 15584 }, { "epoch": 1.0459716116908828, "grad_norm": 1.059873104095459, "learning_rate": 4.8701383426784755e-05, "loss": 0.4963, "step": 15586 }, { "epoch": 1.046105835374652, "grad_norm": 1.0384563207626343, "learning_rate": 4.869051881742112e-05, "loss": 0.5859, "step": 15588 }, { "epoch": 1.0462400590584209, "grad_norm": 1.0289987325668335, "learning_rate": 4.867965426992788e-05, "loss": 0.5095, "step": 15590 }, { "epoch": 1.0463742827421898, "grad_norm": 0.8439013957977295, "learning_rate": 4.86687897848184e-05, "loss": 0.5785, "step": 15592 }, { "epoch": 1.046508506425959, "grad_norm": 0.995211124420166, "learning_rate": 4.8657925362605964e-05, "loss": 0.5559, "step": 15594 }, { "epoch": 1.0466427301097279, "grad_norm": 0.8534039258956909, "learning_rate": 4.8647061003803934e-05, "loss": 0.5191, "step": 15596 }, { "epoch": 1.0467769537934968, "grad_norm": 0.8605908751487732, "learning_rate": 4.86361967089256e-05, "loss": 0.5426, "step": 15598 }, { "epoch": 1.046911177477266, "grad_norm": 0.9673829078674316, "learning_rate": 4.8625332478484295e-05, "loss": 0.4796, "step": 15600 }, { "epoch": 1.0470454011610348, "grad_norm": 1.0090407133102417, "learning_rate": 4.861446831299332e-05, "loss": 0.5467, "step": 15602 }, { "epoch": 1.0471796248448038, "grad_norm": 0.9353126287460327, "learning_rate": 4.8603604212966005e-05, "loss": 0.5637, "step": 15604 }, { "epoch": 1.047313848528573, "grad_norm": 1.1131153106689453, "learning_rate": 4.859274017891563e-05, "loss": 0.5538, "step": 15606 }, { "epoch": 1.0474480722123418, "grad_norm": 0.9784945845603943, "learning_rate": 4.858187621135553e-05, "loss": 0.5159, "step": 15608 }, { "epoch": 1.047582295896111, "grad_norm": 2.116542100906372, "learning_rate": 4.8571012310798966e-05, "loss": 0.4381, "step": 15610 }, { "epoch": 1.0477165195798799, "grad_norm": 0.9533230066299438, "learning_rate": 4.8560148477759285e-05, "loss": 0.5108, "step": 15612 }, { "epoch": 1.0478507432636488, "grad_norm": 0.8793781399726868, "learning_rate": 4.854928471274974e-05, "loss": 0.4968, "step": 15614 }, { "epoch": 1.047984966947418, "grad_norm": 1.0545098781585693, "learning_rate": 4.8538421016283646e-05, "loss": 0.5109, "step": 15616 }, { "epoch": 1.0481191906311869, "grad_norm": 0.9535578489303589, "learning_rate": 4.8527557388874285e-05, "loss": 0.5156, "step": 15618 }, { "epoch": 1.0482534143149558, "grad_norm": 0.9787843823432922, "learning_rate": 4.851669383103495e-05, "loss": 0.5216, "step": 15620 }, { "epoch": 1.048387637998725, "grad_norm": 1.1390225887298584, "learning_rate": 4.85058303432789e-05, "loss": 0.5536, "step": 15622 }, { "epoch": 1.0485218616824938, "grad_norm": 0.929733395576477, "learning_rate": 4.849496692611945e-05, "loss": 0.5519, "step": 15624 }, { "epoch": 1.048656085366263, "grad_norm": 0.8988843560218811, "learning_rate": 4.8484103580069826e-05, "loss": 0.5773, "step": 15626 }, { "epoch": 1.048790309050032, "grad_norm": 0.9410921335220337, "learning_rate": 4.8473240305643346e-05, "loss": 0.5147, "step": 15628 }, { "epoch": 1.0489245327338008, "grad_norm": 1.0105018615722656, "learning_rate": 4.846237710335325e-05, "loss": 0.559, "step": 15630 }, { "epoch": 1.04905875641757, "grad_norm": 1.3036748170852661, "learning_rate": 4.845151397371283e-05, "loss": 0.5141, "step": 15632 }, { "epoch": 1.0491929801013389, "grad_norm": 0.9526950716972351, "learning_rate": 4.8440650917235304e-05, "loss": 0.497, "step": 15634 }, { "epoch": 1.0493272037851078, "grad_norm": 1.0017216205596924, "learning_rate": 4.8429787934433974e-05, "loss": 0.5319, "step": 15636 }, { "epoch": 1.049461427468877, "grad_norm": 0.922514796257019, "learning_rate": 4.841892502582206e-05, "loss": 0.5426, "step": 15638 }, { "epoch": 1.0495956511526459, "grad_norm": 0.8724059462547302, "learning_rate": 4.840806219191285e-05, "loss": 0.4598, "step": 15640 }, { "epoch": 1.0497298748364148, "grad_norm": 0.9697558879852295, "learning_rate": 4.839719943321954e-05, "loss": 0.4628, "step": 15642 }, { "epoch": 1.049864098520184, "grad_norm": 0.9984668493270874, "learning_rate": 4.8386336750255435e-05, "loss": 0.4958, "step": 15644 }, { "epoch": 1.0499983222039528, "grad_norm": 1.0505653619766235, "learning_rate": 4.837547414353373e-05, "loss": 0.5643, "step": 15646 }, { "epoch": 1.050132545887722, "grad_norm": 1.0480382442474365, "learning_rate": 4.8364611613567686e-05, "loss": 0.5502, "step": 15648 }, { "epoch": 1.050266769571491, "grad_norm": 0.9821338057518005, "learning_rate": 4.835374916087053e-05, "loss": 0.5248, "step": 15650 }, { "epoch": 1.0504009932552598, "grad_norm": 1.1076489686965942, "learning_rate": 4.8342886785955484e-05, "loss": 0.5948, "step": 15652 }, { "epoch": 1.050535216939029, "grad_norm": 0.9364720582962036, "learning_rate": 4.8332024489335775e-05, "loss": 0.5359, "step": 15654 }, { "epoch": 1.0506694406227979, "grad_norm": 0.8961271643638611, "learning_rate": 4.8321162271524634e-05, "loss": 0.5273, "step": 15656 }, { "epoch": 1.0508036643065668, "grad_norm": 0.9937961101531982, "learning_rate": 4.831030013303528e-05, "loss": 0.5199, "step": 15658 }, { "epoch": 1.050937887990336, "grad_norm": 1.0300461053848267, "learning_rate": 4.829943807438091e-05, "loss": 0.498, "step": 15660 }, { "epoch": 1.0510721116741049, "grad_norm": 0.9919794797897339, "learning_rate": 4.828857609607476e-05, "loss": 0.5201, "step": 15662 }, { "epoch": 1.051206335357874, "grad_norm": 0.9712582230567932, "learning_rate": 4.827771419863001e-05, "loss": 0.5073, "step": 15664 }, { "epoch": 1.051340559041643, "grad_norm": 0.9613664746284485, "learning_rate": 4.826685238255989e-05, "loss": 0.4974, "step": 15666 }, { "epoch": 1.0514747827254118, "grad_norm": 1.1144733428955078, "learning_rate": 4.825599064837758e-05, "loss": 0.573, "step": 15668 }, { "epoch": 1.051609006409181, "grad_norm": 1.1156517267227173, "learning_rate": 4.824512899659628e-05, "loss": 0.5103, "step": 15670 }, { "epoch": 1.05174323009295, "grad_norm": 0.9424161911010742, "learning_rate": 4.823426742772917e-05, "loss": 0.5447, "step": 15672 }, { "epoch": 1.0518774537767188, "grad_norm": 1.1218960285186768, "learning_rate": 4.822340594228948e-05, "loss": 0.4991, "step": 15674 }, { "epoch": 1.052011677460488, "grad_norm": 1.262618899345398, "learning_rate": 4.821254454079035e-05, "loss": 0.5725, "step": 15676 }, { "epoch": 1.0521459011442569, "grad_norm": 1.4871364831924438, "learning_rate": 4.820168322374498e-05, "loss": 0.5486, "step": 15678 }, { "epoch": 1.0522801248280258, "grad_norm": 0.847819447517395, "learning_rate": 4.819082199166654e-05, "loss": 0.4634, "step": 15680 }, { "epoch": 1.052414348511795, "grad_norm": 1.0007290840148926, "learning_rate": 4.8179960845068204e-05, "loss": 0.5471, "step": 15682 }, { "epoch": 1.0525485721955639, "grad_norm": 0.8427039384841919, "learning_rate": 4.816909978446314e-05, "loss": 0.4646, "step": 15684 }, { "epoch": 1.052682795879333, "grad_norm": 1.072413444519043, "learning_rate": 4.815823881036451e-05, "loss": 0.4995, "step": 15686 }, { "epoch": 1.052817019563102, "grad_norm": 0.9217433929443359, "learning_rate": 4.814737792328547e-05, "loss": 0.5128, "step": 15688 }, { "epoch": 1.0529512432468708, "grad_norm": 0.9816290140151978, "learning_rate": 4.8136517123739197e-05, "loss": 0.5168, "step": 15690 }, { "epoch": 1.05308546693064, "grad_norm": 1.0718098878860474, "learning_rate": 4.8125656412238805e-05, "loss": 0.5763, "step": 15692 }, { "epoch": 1.053219690614409, "grad_norm": 0.9611361026763916, "learning_rate": 4.8114795789297476e-05, "loss": 0.4981, "step": 15694 }, { "epoch": 1.0533539142981778, "grad_norm": 1.0672252178192139, "learning_rate": 4.810393525542833e-05, "loss": 0.5391, "step": 15696 }, { "epoch": 1.053488137981947, "grad_norm": 1.0438896417617798, "learning_rate": 4.8093074811144525e-05, "loss": 0.5363, "step": 15698 }, { "epoch": 1.0536223616657159, "grad_norm": 1.0529118776321411, "learning_rate": 4.808221445695916e-05, "loss": 0.5853, "step": 15700 }, { "epoch": 1.053756585349485, "grad_norm": 0.7470693588256836, "learning_rate": 4.8071354193385424e-05, "loss": 0.4655, "step": 15702 }, { "epoch": 1.053890809033254, "grad_norm": 1.0209118127822876, "learning_rate": 4.806049402093638e-05, "loss": 0.4974, "step": 15704 }, { "epoch": 1.0540250327170229, "grad_norm": 0.9176881909370422, "learning_rate": 4.80496339401252e-05, "loss": 0.5345, "step": 15706 }, { "epoch": 1.054159256400792, "grad_norm": 0.9125534296035767, "learning_rate": 4.803877395146497e-05, "loss": 0.4683, "step": 15708 }, { "epoch": 1.054293480084561, "grad_norm": 1.0628985166549683, "learning_rate": 4.8027914055468826e-05, "loss": 0.565, "step": 15710 }, { "epoch": 1.0544277037683298, "grad_norm": 1.0448428392410278, "learning_rate": 4.8017054252649856e-05, "loss": 0.5217, "step": 15712 }, { "epoch": 1.054561927452099, "grad_norm": 1.2996362447738647, "learning_rate": 4.800619454352118e-05, "loss": 0.5223, "step": 15714 }, { "epoch": 1.054696151135868, "grad_norm": 0.9758515954017639, "learning_rate": 4.799533492859588e-05, "loss": 0.5306, "step": 15716 }, { "epoch": 1.0548303748196368, "grad_norm": 0.9228617548942566, "learning_rate": 4.798447540838708e-05, "loss": 0.5784, "step": 15718 }, { "epoch": 1.054964598503406, "grad_norm": 0.9225476384162903, "learning_rate": 4.797361598340784e-05, "loss": 0.5606, "step": 15720 }, { "epoch": 1.0550988221871749, "grad_norm": 0.8702499866485596, "learning_rate": 4.7962756654171264e-05, "loss": 0.4912, "step": 15722 }, { "epoch": 1.055233045870944, "grad_norm": 0.9524227976799011, "learning_rate": 4.795189742119043e-05, "loss": 0.4948, "step": 15724 }, { "epoch": 1.055367269554713, "grad_norm": 0.9143248796463013, "learning_rate": 4.794103828497842e-05, "loss": 0.5853, "step": 15726 }, { "epoch": 1.0555014932384819, "grad_norm": 1.1705297231674194, "learning_rate": 4.79301792460483e-05, "loss": 0.4788, "step": 15728 }, { "epoch": 1.055635716922251, "grad_norm": 0.9776636362075806, "learning_rate": 4.791932030491313e-05, "loss": 0.5441, "step": 15730 }, { "epoch": 1.05576994060602, "grad_norm": 1.2618513107299805, "learning_rate": 4.7908461462086e-05, "loss": 0.5419, "step": 15732 }, { "epoch": 1.0559041642897888, "grad_norm": 0.7628543972969055, "learning_rate": 4.789760271807993e-05, "loss": 0.5262, "step": 15734 }, { "epoch": 1.056038387973558, "grad_norm": 1.0558035373687744, "learning_rate": 4.788674407340803e-05, "loss": 0.5234, "step": 15736 }, { "epoch": 1.056172611657327, "grad_norm": 1.4487251043319702, "learning_rate": 4.787588552858328e-05, "loss": 0.5369, "step": 15738 }, { "epoch": 1.056306835341096, "grad_norm": 0.9106961488723755, "learning_rate": 4.7865027084118805e-05, "loss": 0.4481, "step": 15740 }, { "epoch": 1.056441059024865, "grad_norm": 1.0827953815460205, "learning_rate": 4.785416874052756e-05, "loss": 0.4801, "step": 15742 }, { "epoch": 1.0565752827086339, "grad_norm": 1.0268560647964478, "learning_rate": 4.7843310498322646e-05, "loss": 0.5095, "step": 15744 }, { "epoch": 1.056709506392403, "grad_norm": 1.1408530473709106, "learning_rate": 4.783245235801706e-05, "loss": 0.5395, "step": 15746 }, { "epoch": 1.056843730076172, "grad_norm": 1.4307409524917603, "learning_rate": 4.7821594320123844e-05, "loss": 0.4855, "step": 15748 }, { "epoch": 1.0569779537599409, "grad_norm": 1.0708088874816895, "learning_rate": 4.781073638515601e-05, "loss": 0.521, "step": 15750 }, { "epoch": 1.05711217744371, "grad_norm": 1.113745927810669, "learning_rate": 4.779987855362658e-05, "loss": 0.5703, "step": 15752 }, { "epoch": 1.057246401127479, "grad_norm": 1.1250548362731934, "learning_rate": 4.778902082604856e-05, "loss": 0.5019, "step": 15754 }, { "epoch": 1.0573806248112478, "grad_norm": 0.9557268619537354, "learning_rate": 4.777816320293496e-05, "loss": 0.4917, "step": 15756 }, { "epoch": 1.057514848495017, "grad_norm": 0.9044797420501709, "learning_rate": 4.7767305684798775e-05, "loss": 0.5486, "step": 15758 }, { "epoch": 1.057649072178786, "grad_norm": 0.9962960481643677, "learning_rate": 4.7756448272153014e-05, "loss": 0.533, "step": 15760 }, { "epoch": 1.057783295862555, "grad_norm": 0.907252848148346, "learning_rate": 4.774559096551064e-05, "loss": 0.5375, "step": 15762 }, { "epoch": 1.057917519546324, "grad_norm": 1.0109272003173828, "learning_rate": 4.773473376538469e-05, "loss": 0.5223, "step": 15764 }, { "epoch": 1.0580517432300929, "grad_norm": 1.1772733926773071, "learning_rate": 4.772387667228808e-05, "loss": 0.521, "step": 15766 }, { "epoch": 1.058185966913862, "grad_norm": 1.018127202987671, "learning_rate": 4.771301968673386e-05, "loss": 0.4875, "step": 15768 }, { "epoch": 1.058320190597631, "grad_norm": 1.1795014142990112, "learning_rate": 4.770216280923493e-05, "loss": 0.5322, "step": 15770 }, { "epoch": 1.0584544142813999, "grad_norm": 0.9482703804969788, "learning_rate": 4.7691306040304306e-05, "loss": 0.4502, "step": 15772 }, { "epoch": 1.058588637965169, "grad_norm": 0.9544832110404968, "learning_rate": 4.768044938045493e-05, "loss": 0.5145, "step": 15774 }, { "epoch": 1.058722861648938, "grad_norm": 1.0192888975143433, "learning_rate": 4.766959283019976e-05, "loss": 0.522, "step": 15776 }, { "epoch": 1.058857085332707, "grad_norm": 1.009538173675537, "learning_rate": 4.765873639005174e-05, "loss": 0.5029, "step": 15778 }, { "epoch": 1.058991309016476, "grad_norm": 1.0669879913330078, "learning_rate": 4.764788006052383e-05, "loss": 0.5343, "step": 15780 }, { "epoch": 1.059125532700245, "grad_norm": 0.9777652025222778, "learning_rate": 4.763702384212896e-05, "loss": 0.5338, "step": 15782 }, { "epoch": 1.059259756384014, "grad_norm": 1.0392273664474487, "learning_rate": 4.762616773538007e-05, "loss": 0.5249, "step": 15784 }, { "epoch": 1.059393980067783, "grad_norm": 0.9199720025062561, "learning_rate": 4.7615311740790074e-05, "loss": 0.4959, "step": 15786 }, { "epoch": 1.0595282037515519, "grad_norm": 0.9195120334625244, "learning_rate": 4.760445585887192e-05, "loss": 0.5123, "step": 15788 }, { "epoch": 1.059662427435321, "grad_norm": 1.0891872644424438, "learning_rate": 4.759360009013849e-05, "loss": 0.4989, "step": 15790 }, { "epoch": 1.05979665111909, "grad_norm": 0.9747539162635803, "learning_rate": 4.7582744435102774e-05, "loss": 0.4794, "step": 15792 }, { "epoch": 1.0599308748028589, "grad_norm": 1.2288579940795898, "learning_rate": 4.7571888894277604e-05, "loss": 0.5171, "step": 15794 }, { "epoch": 1.060065098486628, "grad_norm": 1.1269773244857788, "learning_rate": 4.756103346817589e-05, "loss": 0.5108, "step": 15796 }, { "epoch": 1.060199322170397, "grad_norm": 1.2069462537765503, "learning_rate": 4.755017815731059e-05, "loss": 0.5832, "step": 15798 }, { "epoch": 1.060333545854166, "grad_norm": 1.3380992412567139, "learning_rate": 4.753932296219451e-05, "loss": 0.5867, "step": 15800 }, { "epoch": 1.060467769537935, "grad_norm": 1.0082257986068726, "learning_rate": 4.7528467883340615e-05, "loss": 0.5294, "step": 15802 }, { "epoch": 1.060601993221704, "grad_norm": 1.0993695259094238, "learning_rate": 4.751761292126173e-05, "loss": 0.5601, "step": 15804 }, { "epoch": 1.060736216905473, "grad_norm": 1.235148310661316, "learning_rate": 4.750675807647077e-05, "loss": 0.5, "step": 15806 }, { "epoch": 1.060870440589242, "grad_norm": 0.9587987661361694, "learning_rate": 4.749590334948057e-05, "loss": 0.521, "step": 15808 }, { "epoch": 1.0610046642730109, "grad_norm": 1.074412226676941, "learning_rate": 4.7485048740804026e-05, "loss": 0.5009, "step": 15810 }, { "epoch": 1.06113888795678, "grad_norm": 1.0490909814834595, "learning_rate": 4.747419425095398e-05, "loss": 0.5007, "step": 15812 }, { "epoch": 1.061273111640549, "grad_norm": 1.0220736265182495, "learning_rate": 4.74633398804433e-05, "loss": 0.4692, "step": 15814 }, { "epoch": 1.061407335324318, "grad_norm": 0.9102790355682373, "learning_rate": 4.7452485629784804e-05, "loss": 0.4877, "step": 15816 }, { "epoch": 1.061541559008087, "grad_norm": 1.1566509008407593, "learning_rate": 4.744163149949136e-05, "loss": 0.5361, "step": 15818 }, { "epoch": 1.061675782691856, "grad_norm": 1.052136778831482, "learning_rate": 4.74307774900758e-05, "loss": 0.5177, "step": 15820 }, { "epoch": 1.061810006375625, "grad_norm": 1.163494348526001, "learning_rate": 4.7419923602050954e-05, "loss": 0.5434, "step": 15822 }, { "epoch": 1.061944230059394, "grad_norm": 1.0164376497268677, "learning_rate": 4.740906983592962e-05, "loss": 0.5123, "step": 15824 }, { "epoch": 1.062078453743163, "grad_norm": 1.0817575454711914, "learning_rate": 4.739821619222468e-05, "loss": 0.506, "step": 15826 }, { "epoch": 1.062212677426932, "grad_norm": 1.079735517501831, "learning_rate": 4.738736267144887e-05, "loss": 0.5029, "step": 15828 }, { "epoch": 1.062346901110701, "grad_norm": 1.0035247802734375, "learning_rate": 4.7376509274115075e-05, "loss": 0.5282, "step": 15830 }, { "epoch": 1.0624811247944699, "grad_norm": 0.8570907115936279, "learning_rate": 4.736565600073602e-05, "loss": 0.4403, "step": 15832 }, { "epoch": 1.062615348478239, "grad_norm": 1.206586241722107, "learning_rate": 4.735480285182457e-05, "loss": 0.5351, "step": 15834 }, { "epoch": 1.062749572162008, "grad_norm": 0.960978627204895, "learning_rate": 4.7343949827893465e-05, "loss": 0.5426, "step": 15836 }, { "epoch": 1.062883795845777, "grad_norm": 1.1588352918624878, "learning_rate": 4.733309692945552e-05, "loss": 0.5676, "step": 15838 }, { "epoch": 1.063018019529546, "grad_norm": 1.031522274017334, "learning_rate": 4.732224415702349e-05, "loss": 0.5163, "step": 15840 }, { "epoch": 1.063152243213315, "grad_norm": 0.9539149403572083, "learning_rate": 4.731139151111017e-05, "loss": 0.5197, "step": 15842 }, { "epoch": 1.063286466897084, "grad_norm": 0.9566182494163513, "learning_rate": 4.7300538992228295e-05, "loss": 0.5026, "step": 15844 }, { "epoch": 1.063420690580853, "grad_norm": 0.9178019762039185, "learning_rate": 4.7289686600890666e-05, "loss": 0.5097, "step": 15846 }, { "epoch": 1.063554914264622, "grad_norm": 1.049057126045227, "learning_rate": 4.727883433761e-05, "loss": 0.546, "step": 15848 }, { "epoch": 1.063689137948391, "grad_norm": 1.0182324647903442, "learning_rate": 4.726798220289907e-05, "loss": 0.5669, "step": 15850 }, { "epoch": 1.06382336163216, "grad_norm": 0.9539602994918823, "learning_rate": 4.7257130197270586e-05, "loss": 0.5234, "step": 15852 }, { "epoch": 1.063957585315929, "grad_norm": 0.981003999710083, "learning_rate": 4.724627832123735e-05, "loss": 0.5796, "step": 15854 }, { "epoch": 1.064091808999698, "grad_norm": 0.9285516142845154, "learning_rate": 4.7235426575312004e-05, "loss": 0.5216, "step": 15856 }, { "epoch": 1.064226032683467, "grad_norm": 1.002652883529663, "learning_rate": 4.722457496000735e-05, "loss": 0.4857, "step": 15858 }, { "epoch": 1.064360256367236, "grad_norm": 1.072216510772705, "learning_rate": 4.721372347583604e-05, "loss": 0.508, "step": 15860 }, { "epoch": 1.064494480051005, "grad_norm": 0.9234919548034668, "learning_rate": 4.7202872123310845e-05, "loss": 0.51, "step": 15862 }, { "epoch": 1.064628703734774, "grad_norm": 0.9576901197433472, "learning_rate": 4.7192020902944446e-05, "loss": 0.5355, "step": 15864 }, { "epoch": 1.064762927418543, "grad_norm": 0.9486761689186096, "learning_rate": 4.71811698152495e-05, "loss": 0.4915, "step": 15866 }, { "epoch": 1.064897151102312, "grad_norm": 1.001057744026184, "learning_rate": 4.7170318860738774e-05, "loss": 0.5064, "step": 15868 }, { "epoch": 1.0650313747860811, "grad_norm": 1.1781103610992432, "learning_rate": 4.71594680399249e-05, "loss": 0.551, "step": 15870 }, { "epoch": 1.06516559846985, "grad_norm": 0.9904471635818481, "learning_rate": 4.714861735332058e-05, "loss": 0.481, "step": 15872 }, { "epoch": 1.065299822153619, "grad_norm": 0.966191291809082, "learning_rate": 4.713776680143848e-05, "loss": 0.5074, "step": 15874 }, { "epoch": 1.065434045837388, "grad_norm": 1.0541632175445557, "learning_rate": 4.7126916384791274e-05, "loss": 0.4729, "step": 15876 }, { "epoch": 1.065568269521157, "grad_norm": 1.1319295167922974, "learning_rate": 4.7116066103891615e-05, "loss": 0.5112, "step": 15878 }, { "epoch": 1.065702493204926, "grad_norm": 1.2292240858078003, "learning_rate": 4.710521595925217e-05, "loss": 0.534, "step": 15880 }, { "epoch": 1.065836716888695, "grad_norm": 1.2403215169906616, "learning_rate": 4.709436595138557e-05, "loss": 0.5141, "step": 15882 }, { "epoch": 1.065970940572464, "grad_norm": 1.0871567726135254, "learning_rate": 4.708351608080447e-05, "loss": 0.5905, "step": 15884 }, { "epoch": 1.066105164256233, "grad_norm": 0.9792987108230591, "learning_rate": 4.7072666348021485e-05, "loss": 0.5363, "step": 15886 }, { "epoch": 1.066239387940002, "grad_norm": 1.1306519508361816, "learning_rate": 4.706181675354929e-05, "loss": 0.5359, "step": 15888 }, { "epoch": 1.066373611623771, "grad_norm": 0.9030818343162537, "learning_rate": 4.7050967297900434e-05, "loss": 0.4874, "step": 15890 }, { "epoch": 1.0665078353075401, "grad_norm": 1.0452027320861816, "learning_rate": 4.704011798158762e-05, "loss": 0.5392, "step": 15892 }, { "epoch": 1.066642058991309, "grad_norm": 1.0378111600875854, "learning_rate": 4.702926880512337e-05, "loss": 0.5701, "step": 15894 }, { "epoch": 1.066776282675078, "grad_norm": 0.915477454662323, "learning_rate": 4.701841976902035e-05, "loss": 0.4822, "step": 15896 }, { "epoch": 1.066910506358847, "grad_norm": 0.9939043521881104, "learning_rate": 4.700757087379112e-05, "loss": 0.4608, "step": 15898 }, { "epoch": 1.067044730042616, "grad_norm": 0.9939762949943542, "learning_rate": 4.69967221199483e-05, "loss": 0.6015, "step": 15900 }, { "epoch": 1.067178953726385, "grad_norm": 0.9916800856590271, "learning_rate": 4.698587350800444e-05, "loss": 0.5067, "step": 15902 }, { "epoch": 1.067313177410154, "grad_norm": 0.9965695738792419, "learning_rate": 4.697502503847214e-05, "loss": 0.4918, "step": 15904 }, { "epoch": 1.067447401093923, "grad_norm": 0.9177051782608032, "learning_rate": 4.696417671186395e-05, "loss": 0.5107, "step": 15906 }, { "epoch": 1.067581624777692, "grad_norm": 1.0016969442367554, "learning_rate": 4.6953328528692446e-05, "loss": 0.5405, "step": 15908 }, { "epoch": 1.067715848461461, "grad_norm": 1.0871280431747437, "learning_rate": 4.694248048947017e-05, "loss": 0.4547, "step": 15910 }, { "epoch": 1.06785007214523, "grad_norm": 0.9711535573005676, "learning_rate": 4.693163259470969e-05, "loss": 0.5125, "step": 15912 }, { "epoch": 1.0679842958289991, "grad_norm": 1.0115195512771606, "learning_rate": 4.692078484492351e-05, "loss": 0.5022, "step": 15914 }, { "epoch": 1.068118519512768, "grad_norm": 1.0912082195281982, "learning_rate": 4.6909937240624236e-05, "loss": 0.5468, "step": 15916 }, { "epoch": 1.068252743196537, "grad_norm": 1.298017978668213, "learning_rate": 4.689908978232431e-05, "loss": 0.5389, "step": 15918 }, { "epoch": 1.068386966880306, "grad_norm": 1.0747569799423218, "learning_rate": 4.688824247053633e-05, "loss": 0.4934, "step": 15920 }, { "epoch": 1.068521190564075, "grad_norm": 0.9096031188964844, "learning_rate": 4.6877395305772733e-05, "loss": 0.4718, "step": 15922 }, { "epoch": 1.068655414247844, "grad_norm": 1.0086485147476196, "learning_rate": 4.68665482885461e-05, "loss": 0.5542, "step": 15924 }, { "epoch": 1.068789637931613, "grad_norm": 1.085276484489441, "learning_rate": 4.685570141936888e-05, "loss": 0.585, "step": 15926 }, { "epoch": 1.068923861615382, "grad_norm": 1.0500186681747437, "learning_rate": 4.684485469875359e-05, "loss": 0.5445, "step": 15928 }, { "epoch": 1.069058085299151, "grad_norm": 1.138278841972351, "learning_rate": 4.6834008127212715e-05, "loss": 0.4745, "step": 15930 }, { "epoch": 1.06919230898292, "grad_norm": 1.0815253257751465, "learning_rate": 4.682316170525874e-05, "loss": 0.5361, "step": 15932 }, { "epoch": 1.069326532666689, "grad_norm": 0.9723836183547974, "learning_rate": 4.681231543340412e-05, "loss": 0.5012, "step": 15934 }, { "epoch": 1.0694607563504581, "grad_norm": 0.908646821975708, "learning_rate": 4.680146931216132e-05, "loss": 0.5123, "step": 15936 }, { "epoch": 1.069594980034227, "grad_norm": 1.2489705085754395, "learning_rate": 4.679062334204281e-05, "loss": 0.5196, "step": 15938 }, { "epoch": 1.069729203717996, "grad_norm": 1.1271929740905762, "learning_rate": 4.677977752356103e-05, "loss": 0.5203, "step": 15940 }, { "epoch": 1.069863427401765, "grad_norm": 1.1925686597824097, "learning_rate": 4.676893185722844e-05, "loss": 0.5434, "step": 15942 }, { "epoch": 1.069997651085534, "grad_norm": 1.0192997455596924, "learning_rate": 4.675808634355746e-05, "loss": 0.5289, "step": 15944 }, { "epoch": 1.0701318747693032, "grad_norm": 1.0960651636123657, "learning_rate": 4.674724098306053e-05, "loss": 0.5115, "step": 15946 }, { "epoch": 1.070266098453072, "grad_norm": 0.9219456315040588, "learning_rate": 4.6736395776250065e-05, "loss": 0.4884, "step": 15948 }, { "epoch": 1.070400322136841, "grad_norm": 1.0570820569992065, "learning_rate": 4.6725550723638484e-05, "loss": 0.5441, "step": 15950 }, { "epoch": 1.0705345458206101, "grad_norm": 0.9883645176887512, "learning_rate": 4.671470582573817e-05, "loss": 0.5217, "step": 15952 }, { "epoch": 1.070668769504379, "grad_norm": 1.0507816076278687, "learning_rate": 4.670386108306159e-05, "loss": 0.5517, "step": 15954 }, { "epoch": 1.070802993188148, "grad_norm": 0.9801546335220337, "learning_rate": 4.669301649612105e-05, "loss": 0.494, "step": 15956 }, { "epoch": 1.0709372168719171, "grad_norm": 1.0489400625228882, "learning_rate": 4.6682172065428994e-05, "loss": 0.5596, "step": 15958 }, { "epoch": 1.071071440555686, "grad_norm": 0.9993674159049988, "learning_rate": 4.667132779149778e-05, "loss": 0.5661, "step": 15960 }, { "epoch": 1.071205664239455, "grad_norm": 1.1793206930160522, "learning_rate": 4.6660483674839794e-05, "loss": 0.504, "step": 15962 }, { "epoch": 1.071339887923224, "grad_norm": 0.9190689325332642, "learning_rate": 4.664963971596737e-05, "loss": 0.5407, "step": 15964 }, { "epoch": 1.071474111606993, "grad_norm": 0.9717457890510559, "learning_rate": 4.66387959153929e-05, "loss": 0.5289, "step": 15966 }, { "epoch": 1.0716083352907622, "grad_norm": 1.0649359226226807, "learning_rate": 4.6627952273628705e-05, "loss": 0.5045, "step": 15968 }, { "epoch": 1.071742558974531, "grad_norm": 1.1034852266311646, "learning_rate": 4.6617108791187137e-05, "loss": 0.5771, "step": 15970 }, { "epoch": 1.0718767826583, "grad_norm": 0.9633535742759705, "learning_rate": 4.6606265468580516e-05, "loss": 0.5035, "step": 15972 }, { "epoch": 1.0720110063420691, "grad_norm": 0.9629459977149963, "learning_rate": 4.6595422306321196e-05, "loss": 0.5197, "step": 15974 }, { "epoch": 1.072145230025838, "grad_norm": 0.9145321249961853, "learning_rate": 4.658457930492145e-05, "loss": 0.5586, "step": 15976 }, { "epoch": 1.072279453709607, "grad_norm": 1.001462697982788, "learning_rate": 4.657373646489365e-05, "loss": 0.5125, "step": 15978 }, { "epoch": 1.0724136773933761, "grad_norm": 1.06826913356781, "learning_rate": 4.656289378675003e-05, "loss": 0.4817, "step": 15980 }, { "epoch": 1.072547901077145, "grad_norm": 0.9796193838119507, "learning_rate": 4.655205127100296e-05, "loss": 0.4744, "step": 15982 }, { "epoch": 1.072682124760914, "grad_norm": 1.1441547870635986, "learning_rate": 4.654120891816464e-05, "loss": 0.559, "step": 15984 }, { "epoch": 1.072816348444683, "grad_norm": 1.1235517263412476, "learning_rate": 4.653036672874743e-05, "loss": 0.5606, "step": 15986 }, { "epoch": 1.072950572128452, "grad_norm": 0.9840055108070374, "learning_rate": 4.651952470326356e-05, "loss": 0.4739, "step": 15988 }, { "epoch": 1.0730847958122212, "grad_norm": 0.9808876514434814, "learning_rate": 4.6508682842225305e-05, "loss": 0.5439, "step": 15990 }, { "epoch": 1.07321901949599, "grad_norm": 1.0697896480560303, "learning_rate": 4.6497841146144916e-05, "loss": 0.5498, "step": 15992 }, { "epoch": 1.073353243179759, "grad_norm": 1.1211960315704346, "learning_rate": 4.6486999615534656e-05, "loss": 0.5357, "step": 15994 }, { "epoch": 1.0734874668635281, "grad_norm": 1.0424667596817017, "learning_rate": 4.6476158250906743e-05, "loss": 0.5504, "step": 15996 }, { "epoch": 1.073621690547297, "grad_norm": 1.003092646598816, "learning_rate": 4.6465317052773436e-05, "loss": 0.5114, "step": 15998 }, { "epoch": 1.073755914231066, "grad_norm": 1.044110894203186, "learning_rate": 4.645447602164695e-05, "loss": 0.511, "step": 16000 }, { "epoch": 1.0738901379148351, "grad_norm": 1.0264644622802734, "learning_rate": 4.644363515803949e-05, "loss": 0.4896, "step": 16002 }, { "epoch": 1.074024361598604, "grad_norm": 1.0796425342559814, "learning_rate": 4.643279446246328e-05, "loss": 0.5787, "step": 16004 }, { "epoch": 1.074158585282373, "grad_norm": 1.070082187652588, "learning_rate": 4.6421953935430514e-05, "loss": 0.4995, "step": 16006 }, { "epoch": 1.074292808966142, "grad_norm": 1.3933830261230469, "learning_rate": 4.641111357745339e-05, "loss": 0.5326, "step": 16008 }, { "epoch": 1.074427032649911, "grad_norm": 0.9243556261062622, "learning_rate": 4.6400273389044086e-05, "loss": 0.5082, "step": 16010 }, { "epoch": 1.0745612563336802, "grad_norm": 1.024209976196289, "learning_rate": 4.6389433370714795e-05, "loss": 0.5286, "step": 16012 }, { "epoch": 1.074695480017449, "grad_norm": 0.9765271544456482, "learning_rate": 4.637859352297766e-05, "loss": 0.5442, "step": 16014 }, { "epoch": 1.074829703701218, "grad_norm": 1.0625931024551392, "learning_rate": 4.636775384634489e-05, "loss": 0.4908, "step": 16016 }, { "epoch": 1.0749639273849871, "grad_norm": 1.0464589595794678, "learning_rate": 4.635691434132858e-05, "loss": 0.5058, "step": 16018 }, { "epoch": 1.075098151068756, "grad_norm": 0.9293957948684692, "learning_rate": 4.634607500844092e-05, "loss": 0.5146, "step": 16020 }, { "epoch": 1.0752323747525252, "grad_norm": 1.012489676475525, "learning_rate": 4.633523584819402e-05, "loss": 0.5348, "step": 16022 }, { "epoch": 1.0753665984362941, "grad_norm": 1.014211654663086, "learning_rate": 4.632439686110003e-05, "loss": 0.4705, "step": 16024 }, { "epoch": 1.075500822120063, "grad_norm": 0.9691089987754822, "learning_rate": 4.631355804767105e-05, "loss": 0.4996, "step": 16026 }, { "epoch": 1.0756350458038322, "grad_norm": 0.902036190032959, "learning_rate": 4.6302719408419214e-05, "loss": 0.4969, "step": 16028 }, { "epoch": 1.075769269487601, "grad_norm": 1.0455721616744995, "learning_rate": 4.629188094385661e-05, "loss": 0.5705, "step": 16030 }, { "epoch": 1.07590349317137, "grad_norm": 0.8780853748321533, "learning_rate": 4.6281042654495346e-05, "loss": 0.5202, "step": 16032 }, { "epoch": 1.0760377168551392, "grad_norm": 0.939189076423645, "learning_rate": 4.627020454084749e-05, "loss": 0.5331, "step": 16034 }, { "epoch": 1.076171940538908, "grad_norm": 1.0931113958358765, "learning_rate": 4.6259366603425153e-05, "loss": 0.4951, "step": 16036 }, { "epoch": 1.076306164222677, "grad_norm": 1.035915732383728, "learning_rate": 4.624852884274037e-05, "loss": 0.5259, "step": 16038 }, { "epoch": 1.0764403879064461, "grad_norm": 2.711751937866211, "learning_rate": 4.623769125930524e-05, "loss": 0.5106, "step": 16040 }, { "epoch": 1.076574611590215, "grad_norm": 1.0253950357437134, "learning_rate": 4.6226853853631775e-05, "loss": 0.5397, "step": 16042 }, { "epoch": 1.0767088352739842, "grad_norm": 1.050638198852539, "learning_rate": 4.621601662623208e-05, "loss": 0.519, "step": 16044 }, { "epoch": 1.0768430589577531, "grad_norm": 0.9499335885047913, "learning_rate": 4.6205179577618125e-05, "loss": 0.5042, "step": 16046 }, { "epoch": 1.076977282641522, "grad_norm": 1.057031273841858, "learning_rate": 4.619434270830199e-05, "loss": 0.4846, "step": 16048 }, { "epoch": 1.0771115063252912, "grad_norm": 0.9463146328926086, "learning_rate": 4.6183506018795674e-05, "loss": 0.4907, "step": 16050 }, { "epoch": 1.07724573000906, "grad_norm": 0.9645826816558838, "learning_rate": 4.61726695096112e-05, "loss": 0.4425, "step": 16052 }, { "epoch": 1.077379953692829, "grad_norm": 1.0076628923416138, "learning_rate": 4.616183318126056e-05, "loss": 0.4796, "step": 16054 }, { "epoch": 1.0775141773765982, "grad_norm": 0.9495932459831238, "learning_rate": 4.615099703425576e-05, "loss": 0.5142, "step": 16056 }, { "epoch": 1.077648401060367, "grad_norm": 1.01508367061615, "learning_rate": 4.614016106910877e-05, "loss": 0.544, "step": 16058 }, { "epoch": 1.077782624744136, "grad_norm": 0.9466344714164734, "learning_rate": 4.6129325286331594e-05, "loss": 0.529, "step": 16060 }, { "epoch": 1.0779168484279051, "grad_norm": 1.0745893716812134, "learning_rate": 4.611848968643618e-05, "loss": 0.5167, "step": 16062 }, { "epoch": 1.078051072111674, "grad_norm": 1.092829942703247, "learning_rate": 4.61076542699345e-05, "loss": 0.5296, "step": 16064 }, { "epoch": 1.0781852957954432, "grad_norm": 1.1403659582138062, "learning_rate": 4.609681903733848e-05, "loss": 0.5268, "step": 16066 }, { "epoch": 1.0783195194792121, "grad_norm": 0.9573639631271362, "learning_rate": 4.608598398916012e-05, "loss": 0.5114, "step": 16068 }, { "epoch": 1.078453743162981, "grad_norm": 1.0285440683364868, "learning_rate": 4.607514912591131e-05, "loss": 0.5266, "step": 16070 }, { "epoch": 1.0785879668467502, "grad_norm": 1.112507939338684, "learning_rate": 4.6064314448103974e-05, "loss": 0.4869, "step": 16072 }, { "epoch": 1.078722190530519, "grad_norm": 0.9259080290794373, "learning_rate": 4.605347995625005e-05, "loss": 0.4942, "step": 16074 }, { "epoch": 1.078856414214288, "grad_norm": 1.1185704469680786, "learning_rate": 4.604264565086141e-05, "loss": 0.4717, "step": 16076 }, { "epoch": 1.0789906378980572, "grad_norm": 1.0561480522155762, "learning_rate": 4.6031811532450025e-05, "loss": 0.6153, "step": 16078 }, { "epoch": 1.079124861581826, "grad_norm": 1.0150277614593506, "learning_rate": 4.602097760152771e-05, "loss": 0.5489, "step": 16080 }, { "epoch": 1.079259085265595, "grad_norm": 0.9486995935440063, "learning_rate": 4.601014385860639e-05, "loss": 0.4631, "step": 16082 }, { "epoch": 1.0793933089493641, "grad_norm": 0.9660173058509827, "learning_rate": 4.599931030419793e-05, "loss": 0.5461, "step": 16084 }, { "epoch": 1.079527532633133, "grad_norm": 1.0332201719284058, "learning_rate": 4.598847693881419e-05, "loss": 0.5661, "step": 16086 }, { "epoch": 1.0796617563169022, "grad_norm": 0.9627061486244202, "learning_rate": 4.5977643762967024e-05, "loss": 0.4992, "step": 16088 }, { "epoch": 1.0797959800006711, "grad_norm": 1.7318940162658691, "learning_rate": 4.596681077716828e-05, "loss": 0.5161, "step": 16090 }, { "epoch": 1.07993020368444, "grad_norm": 1.0354950428009033, "learning_rate": 4.59559779819298e-05, "loss": 0.5276, "step": 16092 }, { "epoch": 1.0800644273682092, "grad_norm": 1.1396992206573486, "learning_rate": 4.594514537776341e-05, "loss": 0.5819, "step": 16094 }, { "epoch": 1.080198651051978, "grad_norm": 0.9875621199607849, "learning_rate": 4.5934312965180916e-05, "loss": 0.5414, "step": 16096 }, { "epoch": 1.0803328747357472, "grad_norm": 0.8946545124053955, "learning_rate": 4.5923480744694156e-05, "loss": 0.5294, "step": 16098 }, { "epoch": 1.0804670984195162, "grad_norm": 0.9404687285423279, "learning_rate": 4.5912648716814896e-05, "loss": 0.5864, "step": 16100 }, { "epoch": 1.080601322103285, "grad_norm": 0.9598170518875122, "learning_rate": 4.590181688205496e-05, "loss": 0.4466, "step": 16102 }, { "epoch": 1.0807355457870542, "grad_norm": 0.9472155570983887, "learning_rate": 4.58909852409261e-05, "loss": 0.5081, "step": 16104 }, { "epoch": 1.0808697694708231, "grad_norm": 1.1002511978149414, "learning_rate": 4.5880153793940136e-05, "loss": 0.5322, "step": 16106 }, { "epoch": 1.081003993154592, "grad_norm": 1.1255414485931396, "learning_rate": 4.5869322541608775e-05, "loss": 0.5168, "step": 16108 }, { "epoch": 1.0811382168383612, "grad_norm": 1.0046885013580322, "learning_rate": 4.5858491484443815e-05, "loss": 0.4487, "step": 16110 }, { "epoch": 1.0812724405221301, "grad_norm": 0.956312894821167, "learning_rate": 4.584766062295698e-05, "loss": 0.5109, "step": 16112 }, { "epoch": 1.081406664205899, "grad_norm": 1.0430505275726318, "learning_rate": 4.583682995766003e-05, "loss": 0.5431, "step": 16114 }, { "epoch": 1.0815408878896682, "grad_norm": 1.0224894285202026, "learning_rate": 4.5825999489064676e-05, "loss": 0.5609, "step": 16116 }, { "epoch": 1.081675111573437, "grad_norm": 1.0758253335952759, "learning_rate": 4.581516921768263e-05, "loss": 0.529, "step": 16118 }, { "epoch": 1.0818093352572062, "grad_norm": 1.0846209526062012, "learning_rate": 4.580433914402562e-05, "loss": 0.5035, "step": 16120 }, { "epoch": 1.0819435589409752, "grad_norm": 1.0508264303207397, "learning_rate": 4.579350926860533e-05, "loss": 0.5633, "step": 16122 }, { "epoch": 1.082077782624744, "grad_norm": 0.9724158048629761, "learning_rate": 4.5782679591933444e-05, "loss": 0.4924, "step": 16124 }, { "epoch": 1.0822120063085132, "grad_norm": 0.9624977111816406, "learning_rate": 4.577185011452168e-05, "loss": 0.5157, "step": 16126 }, { "epoch": 1.0823462299922821, "grad_norm": 1.0579854249954224, "learning_rate": 4.576102083688165e-05, "loss": 0.4689, "step": 16128 }, { "epoch": 1.082480453676051, "grad_norm": 1.0702515840530396, "learning_rate": 4.575019175952508e-05, "loss": 0.5714, "step": 16130 }, { "epoch": 1.0826146773598202, "grad_norm": 0.9868396520614624, "learning_rate": 4.573936288296356e-05, "loss": 0.5183, "step": 16132 }, { "epoch": 1.0827489010435891, "grad_norm": 1.0361624956130981, "learning_rate": 4.57285342077088e-05, "loss": 0.4942, "step": 16134 }, { "epoch": 1.082883124727358, "grad_norm": 1.1705342531204224, "learning_rate": 4.5717705734272375e-05, "loss": 0.5181, "step": 16136 }, { "epoch": 1.0830173484111272, "grad_norm": 0.8735315203666687, "learning_rate": 4.570687746316595e-05, "loss": 0.5328, "step": 16138 }, { "epoch": 1.083151572094896, "grad_norm": 0.9336830973625183, "learning_rate": 4.569604939490113e-05, "loss": 0.5143, "step": 16140 }, { "epoch": 1.0832857957786652, "grad_norm": 1.2821564674377441, "learning_rate": 4.568522152998948e-05, "loss": 0.5512, "step": 16142 }, { "epoch": 1.0834200194624342, "grad_norm": 1.0232316255569458, "learning_rate": 4.5674393868942666e-05, "loss": 0.504, "step": 16144 }, { "epoch": 1.083554243146203, "grad_norm": 0.9223902225494385, "learning_rate": 4.56635664122722e-05, "loss": 0.5187, "step": 16146 }, { "epoch": 1.0836884668299722, "grad_norm": 0.9724869728088379, "learning_rate": 4.5652739160489715e-05, "loss": 0.5811, "step": 16148 }, { "epoch": 1.0838226905137411, "grad_norm": 1.07927668094635, "learning_rate": 4.564191211410675e-05, "loss": 0.5569, "step": 16150 }, { "epoch": 1.08395691419751, "grad_norm": 0.8426041603088379, "learning_rate": 4.563108527363487e-05, "loss": 0.4821, "step": 16152 }, { "epoch": 1.0840911378812792, "grad_norm": 0.8651317954063416, "learning_rate": 4.562025863958562e-05, "loss": 0.5163, "step": 16154 }, { "epoch": 1.0842253615650481, "grad_norm": 1.1290720701217651, "learning_rate": 4.5609432212470536e-05, "loss": 0.6004, "step": 16156 }, { "epoch": 1.084359585248817, "grad_norm": 0.8488070368766785, "learning_rate": 4.5598605992801145e-05, "loss": 0.5177, "step": 16158 }, { "epoch": 1.0844938089325862, "grad_norm": 1.019907832145691, "learning_rate": 4.5587779981088976e-05, "loss": 0.5181, "step": 16160 }, { "epoch": 1.084628032616355, "grad_norm": 0.998149573802948, "learning_rate": 4.557695417784551e-05, "loss": 0.4685, "step": 16162 }, { "epoch": 1.0847622563001242, "grad_norm": 1.00678288936615, "learning_rate": 4.5566128583582276e-05, "loss": 0.4771, "step": 16164 }, { "epoch": 1.0848964799838932, "grad_norm": 0.9872289299964905, "learning_rate": 4.5555303198810724e-05, "loss": 0.4704, "step": 16166 }, { "epoch": 1.085030703667662, "grad_norm": 1.347158432006836, "learning_rate": 4.5544478024042384e-05, "loss": 0.5665, "step": 16168 }, { "epoch": 1.0851649273514312, "grad_norm": 2.2895302772521973, "learning_rate": 4.553365305978867e-05, "loss": 0.5762, "step": 16170 }, { "epoch": 1.0852991510352001, "grad_norm": 0.9067912697792053, "learning_rate": 4.5522828306561085e-05, "loss": 0.4757, "step": 16172 }, { "epoch": 1.0854333747189693, "grad_norm": 1.1870911121368408, "learning_rate": 4.551200376487105e-05, "loss": 0.5064, "step": 16174 }, { "epoch": 1.0855675984027382, "grad_norm": 0.9698526859283447, "learning_rate": 4.5501179435230024e-05, "loss": 0.5167, "step": 16176 }, { "epoch": 1.0857018220865071, "grad_norm": 1.0006768703460693, "learning_rate": 4.549035531814941e-05, "loss": 0.4965, "step": 16178 }, { "epoch": 1.0858360457702763, "grad_norm": 1.1461081504821777, "learning_rate": 4.5479531414140654e-05, "loss": 0.5515, "step": 16180 }, { "epoch": 1.0859702694540452, "grad_norm": 0.9489409327507019, "learning_rate": 4.5468707723715135e-05, "loss": 0.5092, "step": 16182 }, { "epoch": 1.086104493137814, "grad_norm": 1.11704421043396, "learning_rate": 4.545788424738427e-05, "loss": 0.5481, "step": 16184 }, { "epoch": 1.0862387168215832, "grad_norm": 0.8226884007453918, "learning_rate": 4.544706098565944e-05, "loss": 0.5124, "step": 16186 }, { "epoch": 1.0863729405053522, "grad_norm": 1.157618522644043, "learning_rate": 4.543623793905204e-05, "loss": 0.543, "step": 16188 }, { "epoch": 1.086507164189121, "grad_norm": 1.1042851209640503, "learning_rate": 4.54254151080734e-05, "loss": 0.5354, "step": 16190 }, { "epoch": 1.0866413878728902, "grad_norm": 0.9790997505187988, "learning_rate": 4.541459249323492e-05, "loss": 0.5428, "step": 16192 }, { "epoch": 1.0867756115566591, "grad_norm": 0.9773795008659363, "learning_rate": 4.54037700950479e-05, "loss": 0.4837, "step": 16194 }, { "epoch": 1.0869098352404283, "grad_norm": 1.0129271745681763, "learning_rate": 4.5392947914023745e-05, "loss": 0.5219, "step": 16196 }, { "epoch": 1.0870440589241972, "grad_norm": 0.9165695309638977, "learning_rate": 4.538212595067371e-05, "loss": 0.4536, "step": 16198 }, { "epoch": 1.0871782826079661, "grad_norm": 0.8758540153503418, "learning_rate": 4.5371304205509154e-05, "loss": 0.4542, "step": 16200 }, { "epoch": 1.0873125062917353, "grad_norm": 1.1202819347381592, "learning_rate": 4.5360482679041375e-05, "loss": 0.5094, "step": 16202 }, { "epoch": 1.0874467299755042, "grad_norm": 1.1250008344650269, "learning_rate": 4.534966137178167e-05, "loss": 0.5187, "step": 16204 }, { "epoch": 1.087580953659273, "grad_norm": 1.0680208206176758, "learning_rate": 4.533884028424133e-05, "loss": 0.5682, "step": 16206 }, { "epoch": 1.0877151773430422, "grad_norm": 1.0380007028579712, "learning_rate": 4.5328019416931594e-05, "loss": 0.4994, "step": 16208 }, { "epoch": 1.0878494010268112, "grad_norm": 0.9769343733787537, "learning_rate": 4.531719877036377e-05, "loss": 0.4653, "step": 16210 }, { "epoch": 1.08798362471058, "grad_norm": 0.9960572123527527, "learning_rate": 4.530637834504908e-05, "loss": 0.5058, "step": 16212 }, { "epoch": 1.0881178483943492, "grad_norm": 1.506523609161377, "learning_rate": 4.5295558141498804e-05, "loss": 0.5711, "step": 16214 }, { "epoch": 1.0882520720781181, "grad_norm": 1.0053727626800537, "learning_rate": 4.528473816022414e-05, "loss": 0.4923, "step": 16216 }, { "epoch": 1.0883862957618873, "grad_norm": 0.912634015083313, "learning_rate": 4.5273918401736346e-05, "loss": 0.4597, "step": 16218 }, { "epoch": 1.0885205194456562, "grad_norm": 0.9932961463928223, "learning_rate": 4.5263098866546586e-05, "loss": 0.5446, "step": 16220 }, { "epoch": 1.0886547431294251, "grad_norm": 0.943209171295166, "learning_rate": 4.525227955516612e-05, "loss": 0.5379, "step": 16222 }, { "epoch": 1.0887889668131943, "grad_norm": 1.0366032123565674, "learning_rate": 4.524146046810608e-05, "loss": 0.5927, "step": 16224 }, { "epoch": 1.0889231904969632, "grad_norm": 0.9892827868461609, "learning_rate": 4.523064160587769e-05, "loss": 0.5305, "step": 16226 }, { "epoch": 1.089057414180732, "grad_norm": 1.046505331993103, "learning_rate": 4.5219822968992096e-05, "loss": 0.5157, "step": 16228 }, { "epoch": 1.0891916378645012, "grad_norm": 0.9887049794197083, "learning_rate": 4.520900455796049e-05, "loss": 0.5115, "step": 16230 }, { "epoch": 1.0893258615482702, "grad_norm": 1.1368025541305542, "learning_rate": 4.5198186373293964e-05, "loss": 0.4765, "step": 16232 }, { "epoch": 1.089460085232039, "grad_norm": 0.9224391579627991, "learning_rate": 4.518736841550373e-05, "loss": 0.4793, "step": 16234 }, { "epoch": 1.0895943089158082, "grad_norm": 1.2007701396942139, "learning_rate": 4.5176550685100834e-05, "loss": 0.5259, "step": 16236 }, { "epoch": 1.0897285325995771, "grad_norm": 0.918589174747467, "learning_rate": 4.516573318259646e-05, "loss": 0.5098, "step": 16238 }, { "epoch": 1.0898627562833463, "grad_norm": 0.9159785509109497, "learning_rate": 4.515491590850167e-05, "loss": 0.5297, "step": 16240 }, { "epoch": 1.0899969799671152, "grad_norm": 1.0669053792953491, "learning_rate": 4.5144098863327586e-05, "loss": 0.5326, "step": 16242 }, { "epoch": 1.0901312036508841, "grad_norm": 0.9650025367736816, "learning_rate": 4.513328204758528e-05, "loss": 0.5523, "step": 16244 }, { "epoch": 1.0902654273346533, "grad_norm": 1.2968016862869263, "learning_rate": 4.5122465461785833e-05, "loss": 0.5294, "step": 16246 }, { "epoch": 1.0903996510184222, "grad_norm": 1.3808320760726929, "learning_rate": 4.511164910644029e-05, "loss": 0.5688, "step": 16248 }, { "epoch": 1.0905338747021913, "grad_norm": 0.9764894843101501, "learning_rate": 4.5100832982059724e-05, "loss": 0.5248, "step": 16250 }, { "epoch": 1.0906680983859602, "grad_norm": 1.1142827272415161, "learning_rate": 4.509001708915516e-05, "loss": 0.5147, "step": 16252 }, { "epoch": 1.0908023220697292, "grad_norm": 0.7973035573959351, "learning_rate": 4.507920142823764e-05, "loss": 0.4832, "step": 16254 }, { "epoch": 1.0909365457534983, "grad_norm": 0.8636682033538818, "learning_rate": 4.506838599981816e-05, "loss": 0.5067, "step": 16256 }, { "epoch": 1.0910707694372672, "grad_norm": 0.9508370757102966, "learning_rate": 4.5057570804407774e-05, "loss": 0.4957, "step": 16258 }, { "epoch": 1.0912049931210361, "grad_norm": 1.341615080833435, "learning_rate": 4.5046755842517415e-05, "loss": 0.5189, "step": 16260 }, { "epoch": 1.0913392168048053, "grad_norm": 1.0907329320907593, "learning_rate": 4.503594111465812e-05, "loss": 0.4795, "step": 16262 }, { "epoch": 1.0914734404885742, "grad_norm": 1.1239118576049805, "learning_rate": 4.5025126621340844e-05, "loss": 0.5052, "step": 16264 }, { "epoch": 1.0916076641723431, "grad_norm": 1.1593395471572876, "learning_rate": 4.501431236307655e-05, "loss": 0.5134, "step": 16266 }, { "epoch": 1.0917418878561123, "grad_norm": 1.0171058177947998, "learning_rate": 4.500349834037619e-05, "loss": 0.581, "step": 16268 }, { "epoch": 1.0918761115398812, "grad_norm": 1.9492132663726807, "learning_rate": 4.499268455375073e-05, "loss": 0.5196, "step": 16270 }, { "epoch": 1.0920103352236503, "grad_norm": 0.9708751440048218, "learning_rate": 4.498187100371105e-05, "loss": 0.5253, "step": 16272 }, { "epoch": 1.0921445589074192, "grad_norm": 0.8562864661216736, "learning_rate": 4.497105769076812e-05, "loss": 0.4797, "step": 16274 }, { "epoch": 1.0922787825911882, "grad_norm": 0.9457457661628723, "learning_rate": 4.496024461543282e-05, "loss": 0.5226, "step": 16276 }, { "epoch": 1.0924130062749573, "grad_norm": 1.0705337524414062, "learning_rate": 4.494943177821604e-05, "loss": 0.5482, "step": 16278 }, { "epoch": 1.0925472299587262, "grad_norm": 0.9535156488418579, "learning_rate": 4.493861917962869e-05, "loss": 0.5088, "step": 16280 }, { "epoch": 1.0926814536424951, "grad_norm": 0.9982197284698486, "learning_rate": 4.492780682018162e-05, "loss": 0.5319, "step": 16282 }, { "epoch": 1.0928156773262643, "grad_norm": 0.9514666795730591, "learning_rate": 4.4916994700385714e-05, "loss": 0.4807, "step": 16284 }, { "epoch": 1.0929499010100332, "grad_norm": 0.948200523853302, "learning_rate": 4.4906182820751796e-05, "loss": 0.4654, "step": 16286 }, { "epoch": 1.0930841246938021, "grad_norm": 0.9668101072311401, "learning_rate": 4.489537118179074e-05, "loss": 0.5109, "step": 16288 }, { "epoch": 1.0932183483775713, "grad_norm": 1.168709635734558, "learning_rate": 4.488455978401334e-05, "loss": 0.5406, "step": 16290 }, { "epoch": 1.0933525720613402, "grad_norm": 0.963259220123291, "learning_rate": 4.4873748627930455e-05, "loss": 0.5907, "step": 16292 }, { "epoch": 1.0934867957451093, "grad_norm": 0.9700908660888672, "learning_rate": 4.4862937714052835e-05, "loss": 0.548, "step": 16294 }, { "epoch": 1.0936210194288782, "grad_norm": 0.9730653762817383, "learning_rate": 4.4852127042891354e-05, "loss": 0.5575, "step": 16296 }, { "epoch": 1.0937552431126472, "grad_norm": 0.8968258500099182, "learning_rate": 4.48413166149567e-05, "loss": 0.4486, "step": 16298 }, { "epoch": 1.0938894667964163, "grad_norm": 0.8358473181724548, "learning_rate": 4.483050643075972e-05, "loss": 0.4455, "step": 16300 }, { "epoch": 1.0940236904801852, "grad_norm": 0.9382853507995605, "learning_rate": 4.481969649081112e-05, "loss": 0.5231, "step": 16302 }, { "epoch": 1.0941579141639541, "grad_norm": 0.9947872161865234, "learning_rate": 4.4808886795621705e-05, "loss": 0.5522, "step": 16304 }, { "epoch": 1.0942921378477233, "grad_norm": 1.0441114902496338, "learning_rate": 4.479807734570216e-05, "loss": 0.5215, "step": 16306 }, { "epoch": 1.0944263615314922, "grad_norm": 0.883138120174408, "learning_rate": 4.478726814156325e-05, "loss": 0.5197, "step": 16308 }, { "epoch": 1.0945605852152611, "grad_norm": 0.8907055854797363, "learning_rate": 4.4776459183715654e-05, "loss": 0.4668, "step": 16310 }, { "epoch": 1.0946948088990303, "grad_norm": 1.0268926620483398, "learning_rate": 4.4765650472670094e-05, "loss": 0.4907, "step": 16312 }, { "epoch": 1.0948290325827992, "grad_norm": 0.9086816310882568, "learning_rate": 4.475484200893725e-05, "loss": 0.4902, "step": 16314 }, { "epoch": 1.0949632562665683, "grad_norm": 0.9753834009170532, "learning_rate": 4.474403379302782e-05, "loss": 0.4988, "step": 16316 }, { "epoch": 1.0950974799503372, "grad_norm": 0.960116982460022, "learning_rate": 4.473322582545244e-05, "loss": 0.5771, "step": 16318 }, { "epoch": 1.0952317036341062, "grad_norm": 1.0215414762496948, "learning_rate": 4.472241810672181e-05, "loss": 0.5681, "step": 16320 }, { "epoch": 1.0953659273178753, "grad_norm": 1.0304012298583984, "learning_rate": 4.471161063734652e-05, "loss": 0.5528, "step": 16322 }, { "epoch": 1.0955001510016442, "grad_norm": 1.0913125276565552, "learning_rate": 4.470080341783726e-05, "loss": 0.5906, "step": 16324 }, { "epoch": 1.0956343746854131, "grad_norm": 1.086877465248108, "learning_rate": 4.4689996448704574e-05, "loss": 0.5078, "step": 16326 }, { "epoch": 1.0957685983691823, "grad_norm": 0.9057826995849609, "learning_rate": 4.4679189730459144e-05, "loss": 0.5856, "step": 16328 }, { "epoch": 1.0959028220529512, "grad_norm": 0.9654415845870972, "learning_rate": 4.466838326361152e-05, "loss": 0.5095, "step": 16330 }, { "epoch": 1.0960370457367203, "grad_norm": 0.9577851295471191, "learning_rate": 4.4657577048672324e-05, "loss": 0.5384, "step": 16332 }, { "epoch": 1.0961712694204893, "grad_norm": 0.8881108164787292, "learning_rate": 4.464677108615209e-05, "loss": 0.49, "step": 16334 }, { "epoch": 1.0963054931042582, "grad_norm": 0.9594892263412476, "learning_rate": 4.463596537656141e-05, "loss": 0.5315, "step": 16336 }, { "epoch": 1.0964397167880273, "grad_norm": 1.1049425601959229, "learning_rate": 4.4625159920410804e-05, "loss": 0.5076, "step": 16338 }, { "epoch": 1.0965739404717962, "grad_norm": 1.0198605060577393, "learning_rate": 4.4614354718210835e-05, "loss": 0.5602, "step": 16340 }, { "epoch": 1.0967081641555652, "grad_norm": 1.053015112876892, "learning_rate": 4.4603549770472e-05, "loss": 0.5425, "step": 16342 }, { "epoch": 1.0968423878393343, "grad_norm": 1.0289908647537231, "learning_rate": 4.459274507770484e-05, "loss": 0.5039, "step": 16344 }, { "epoch": 1.0969766115231032, "grad_norm": 0.9137513637542725, "learning_rate": 4.4581940640419836e-05, "loss": 0.4879, "step": 16346 }, { "epoch": 1.0971108352068724, "grad_norm": 0.9542897343635559, "learning_rate": 4.457113645912748e-05, "loss": 0.5342, "step": 16348 }, { "epoch": 1.0972450588906413, "grad_norm": 1.230817198753357, "learning_rate": 4.456033253433826e-05, "loss": 0.534, "step": 16350 }, { "epoch": 1.0973792825744102, "grad_norm": 1.0108160972595215, "learning_rate": 4.4549528866562614e-05, "loss": 0.5424, "step": 16352 }, { "epoch": 1.0975135062581793, "grad_norm": 1.118992567062378, "learning_rate": 4.453872545631104e-05, "loss": 0.5265, "step": 16354 }, { "epoch": 1.0976477299419483, "grad_norm": 1.1093859672546387, "learning_rate": 4.452792230409392e-05, "loss": 0.5922, "step": 16356 }, { "epoch": 1.0977819536257172, "grad_norm": 0.9274280071258545, "learning_rate": 4.451711941042173e-05, "loss": 0.4781, "step": 16358 }, { "epoch": 1.0979161773094863, "grad_norm": 0.9606037735939026, "learning_rate": 4.4506316775804844e-05, "loss": 0.4916, "step": 16360 }, { "epoch": 1.0980504009932552, "grad_norm": 0.9684455394744873, "learning_rate": 4.449551440075371e-05, "loss": 0.4674, "step": 16362 }, { "epoch": 1.0981846246770242, "grad_norm": 0.9659847021102905, "learning_rate": 4.448471228577868e-05, "loss": 0.4583, "step": 16364 }, { "epoch": 1.0983188483607933, "grad_norm": 0.9780763387680054, "learning_rate": 4.447391043139016e-05, "loss": 0.4875, "step": 16366 }, { "epoch": 1.0984530720445622, "grad_norm": 0.9492866396903992, "learning_rate": 4.4463108838098504e-05, "loss": 0.5307, "step": 16368 }, { "epoch": 1.0985872957283314, "grad_norm": 1.2335330247879028, "learning_rate": 4.445230750641407e-05, "loss": 0.5269, "step": 16370 }, { "epoch": 1.0987215194121003, "grad_norm": 1.0171606540679932, "learning_rate": 4.4441506436847194e-05, "loss": 0.6064, "step": 16372 }, { "epoch": 1.0988557430958692, "grad_norm": 0.92994624376297, "learning_rate": 4.443070562990821e-05, "loss": 0.4774, "step": 16374 }, { "epoch": 1.0989899667796383, "grad_norm": 0.9542266130447388, "learning_rate": 4.441990508610743e-05, "loss": 0.531, "step": 16376 }, { "epoch": 1.0991241904634073, "grad_norm": 0.9981623888015747, "learning_rate": 4.440910480595517e-05, "loss": 0.4934, "step": 16378 }, { "epoch": 1.0992584141471762, "grad_norm": 1.0491749048233032, "learning_rate": 4.439830478996169e-05, "loss": 0.4824, "step": 16380 }, { "epoch": 1.0993926378309453, "grad_norm": 1.1588420867919922, "learning_rate": 4.438750503863733e-05, "loss": 0.5313, "step": 16382 }, { "epoch": 1.0995268615147142, "grad_norm": 1.062170147895813, "learning_rate": 4.437670555249228e-05, "loss": 0.5637, "step": 16384 }, { "epoch": 1.0996610851984832, "grad_norm": 1.0424307584762573, "learning_rate": 4.436590633203688e-05, "loss": 0.5015, "step": 16386 }, { "epoch": 1.0997953088822523, "grad_norm": 0.9287711977958679, "learning_rate": 4.4355107377781286e-05, "loss": 0.5714, "step": 16388 }, { "epoch": 1.0999295325660212, "grad_norm": 1.1476563215255737, "learning_rate": 4.434430869023579e-05, "loss": 0.5265, "step": 16390 }, { "epoch": 1.1000637562497904, "grad_norm": 0.9120633006095886, "learning_rate": 4.433351026991058e-05, "loss": 0.5154, "step": 16392 }, { "epoch": 1.1001979799335593, "grad_norm": 1.005014419555664, "learning_rate": 4.4322712117315874e-05, "loss": 0.5392, "step": 16394 }, { "epoch": 1.1003322036173282, "grad_norm": 0.8479301333427429, "learning_rate": 4.4311914232961847e-05, "loss": 0.433, "step": 16396 }, { "epoch": 1.1004664273010973, "grad_norm": 2.0558910369873047, "learning_rate": 4.4301116617358695e-05, "loss": 0.4687, "step": 16398 }, { "epoch": 1.1006006509848663, "grad_norm": 0.9969558119773865, "learning_rate": 4.429031927101657e-05, "loss": 0.5175, "step": 16400 }, { "epoch": 1.1007348746686352, "grad_norm": 0.9618621468544006, "learning_rate": 4.4279522194445644e-05, "loss": 0.4884, "step": 16402 }, { "epoch": 1.1008690983524043, "grad_norm": 1.015475869178772, "learning_rate": 4.4268725388156026e-05, "loss": 0.5621, "step": 16404 }, { "epoch": 1.1010033220361732, "grad_norm": 1.1162946224212646, "learning_rate": 4.425792885265789e-05, "loss": 0.5391, "step": 16406 }, { "epoch": 1.1011375457199424, "grad_norm": 1.1964268684387207, "learning_rate": 4.4247132588461286e-05, "loss": 0.5575, "step": 16408 }, { "epoch": 1.1012717694037113, "grad_norm": 0.8786594867706299, "learning_rate": 4.423633659607639e-05, "loss": 0.5216, "step": 16410 }, { "epoch": 1.1014059930874802, "grad_norm": 1.0007182359695435, "learning_rate": 4.422554087601325e-05, "loss": 0.4838, "step": 16412 }, { "epoch": 1.1015402167712494, "grad_norm": 1.1021301746368408, "learning_rate": 4.421474542878195e-05, "loss": 0.5776, "step": 16414 }, { "epoch": 1.1016744404550183, "grad_norm": 0.9514111876487732, "learning_rate": 4.4203950254892545e-05, "loss": 0.4923, "step": 16416 }, { "epoch": 1.1018086641387872, "grad_norm": 1.0056926012039185, "learning_rate": 4.4193155354855074e-05, "loss": 0.4776, "step": 16418 }, { "epoch": 1.1019428878225563, "grad_norm": 1.1442005634307861, "learning_rate": 4.4182360729179625e-05, "loss": 0.4783, "step": 16420 }, { "epoch": 1.1020771115063253, "grad_norm": 1.4579870700836182, "learning_rate": 4.4171566378376166e-05, "loss": 0.4795, "step": 16422 }, { "epoch": 1.1022113351900944, "grad_norm": 1.0166319608688354, "learning_rate": 4.416077230295475e-05, "loss": 0.4865, "step": 16424 }, { "epoch": 1.1023455588738633, "grad_norm": 1.166882872581482, "learning_rate": 4.414997850342536e-05, "loss": 0.5292, "step": 16426 }, { "epoch": 1.1024797825576322, "grad_norm": 0.9497777819633484, "learning_rate": 4.4139184980297985e-05, "loss": 0.4938, "step": 16428 }, { "epoch": 1.1026140062414014, "grad_norm": 0.9077328443527222, "learning_rate": 4.4128391734082586e-05, "loss": 0.4966, "step": 16430 }, { "epoch": 1.1027482299251703, "grad_norm": 0.9687386155128479, "learning_rate": 4.4117598765289145e-05, "loss": 0.5106, "step": 16432 }, { "epoch": 1.1028824536089392, "grad_norm": 1.446250081062317, "learning_rate": 4.410680607442758e-05, "loss": 0.5049, "step": 16434 }, { "epoch": 1.1030166772927084, "grad_norm": 0.9839146137237549, "learning_rate": 4.4096013662007844e-05, "loss": 0.4981, "step": 16436 }, { "epoch": 1.1031509009764773, "grad_norm": 0.9283409714698792, "learning_rate": 4.408522152853985e-05, "loss": 0.5569, "step": 16438 }, { "epoch": 1.1032851246602462, "grad_norm": 0.952659547328949, "learning_rate": 4.407442967453352e-05, "loss": 0.538, "step": 16440 }, { "epoch": 1.1034193483440153, "grad_norm": 1.0071378946304321, "learning_rate": 4.406363810049871e-05, "loss": 0.5718, "step": 16442 }, { "epoch": 1.1035535720277843, "grad_norm": 1.1124968528747559, "learning_rate": 4.405284680694536e-05, "loss": 0.5117, "step": 16444 }, { "epoch": 1.1036877957115534, "grad_norm": 0.8670393824577332, "learning_rate": 4.4042055794383277e-05, "loss": 0.4914, "step": 16446 }, { "epoch": 1.1038220193953223, "grad_norm": 1.0304452180862427, "learning_rate": 4.4031265063322364e-05, "loss": 0.5183, "step": 16448 }, { "epoch": 1.1039562430790912, "grad_norm": 0.9968014359474182, "learning_rate": 4.402047461427241e-05, "loss": 0.508, "step": 16450 }, { "epoch": 1.1040904667628604, "grad_norm": 1.0950051546096802, "learning_rate": 4.40096844477433e-05, "loss": 0.545, "step": 16452 }, { "epoch": 1.1042246904466293, "grad_norm": 0.7988759279251099, "learning_rate": 4.399889456424481e-05, "loss": 0.4289, "step": 16454 }, { "epoch": 1.1043589141303982, "grad_norm": 1.0534387826919556, "learning_rate": 4.398810496428675e-05, "loss": 0.5391, "step": 16456 }, { "epoch": 1.1044931378141674, "grad_norm": 1.1303116083145142, "learning_rate": 4.397731564837891e-05, "loss": 0.5349, "step": 16458 }, { "epoch": 1.1046273614979363, "grad_norm": 1.1211031675338745, "learning_rate": 4.396652661703107e-05, "loss": 0.5482, "step": 16460 }, { "epoch": 1.1047615851817052, "grad_norm": 1.0421147346496582, "learning_rate": 4.395573787075297e-05, "loss": 0.6276, "step": 16462 }, { "epoch": 1.1048958088654743, "grad_norm": 1.1967524290084839, "learning_rate": 4.394494941005438e-05, "loss": 0.5523, "step": 16464 }, { "epoch": 1.1050300325492433, "grad_norm": 0.9994180202484131, "learning_rate": 4.3934161235445015e-05, "loss": 0.5339, "step": 16466 }, { "epoch": 1.1051642562330124, "grad_norm": 0.9488910436630249, "learning_rate": 4.392337334743461e-05, "loss": 0.4396, "step": 16468 }, { "epoch": 1.1052984799167813, "grad_norm": 0.9103826284408569, "learning_rate": 4.391258574653285e-05, "loss": 0.4677, "step": 16470 }, { "epoch": 1.1054327036005502, "grad_norm": 0.8869611620903015, "learning_rate": 4.390179843324947e-05, "loss": 0.4818, "step": 16472 }, { "epoch": 1.1055669272843194, "grad_norm": 0.9392869472503662, "learning_rate": 4.3891011408094105e-05, "loss": 0.5036, "step": 16474 }, { "epoch": 1.1057011509680883, "grad_norm": 0.919829785823822, "learning_rate": 4.3880224671576455e-05, "loss": 0.4985, "step": 16476 }, { "epoch": 1.1058353746518572, "grad_norm": 1.0271228551864624, "learning_rate": 4.3869438224206126e-05, "loss": 0.587, "step": 16478 }, { "epoch": 1.1059695983356264, "grad_norm": 1.1444569826126099, "learning_rate": 4.3858652066492814e-05, "loss": 0.5378, "step": 16480 }, { "epoch": 1.1061038220193953, "grad_norm": 0.974545955657959, "learning_rate": 4.3847866198946116e-05, "loss": 0.4844, "step": 16482 }, { "epoch": 1.1062380457031644, "grad_norm": 1.0419704914093018, "learning_rate": 4.383708062207562e-05, "loss": 0.5338, "step": 16484 }, { "epoch": 1.1063722693869333, "grad_norm": 3.126567840576172, "learning_rate": 4.382629533639095e-05, "loss": 0.5039, "step": 16486 }, { "epoch": 1.1065064930707023, "grad_norm": 0.9070227146148682, "learning_rate": 4.381551034240169e-05, "loss": 0.4685, "step": 16488 }, { "epoch": 1.1066407167544714, "grad_norm": 1.186069130897522, "learning_rate": 4.3804725640617414e-05, "loss": 0.4914, "step": 16490 }, { "epoch": 1.1067749404382403, "grad_norm": 0.9902550578117371, "learning_rate": 4.379394123154766e-05, "loss": 0.5757, "step": 16492 }, { "epoch": 1.1069091641220092, "grad_norm": 0.9856970906257629, "learning_rate": 4.3783157115701984e-05, "loss": 0.5759, "step": 16494 }, { "epoch": 1.1070433878057784, "grad_norm": 1.0601118803024292, "learning_rate": 4.3772373293589894e-05, "loss": 0.5112, "step": 16496 }, { "epoch": 1.1071776114895473, "grad_norm": 1.0865708589553833, "learning_rate": 4.376158976572094e-05, "loss": 0.5128, "step": 16498 }, { "epoch": 1.1073118351733164, "grad_norm": 1.0044316053390503, "learning_rate": 4.375080653260459e-05, "loss": 0.4788, "step": 16500 }, { "epoch": 1.1074460588570854, "grad_norm": 0.95487380027771, "learning_rate": 4.374002359475036e-05, "loss": 0.4666, "step": 16502 }, { "epoch": 1.1075802825408543, "grad_norm": 0.9853495955467224, "learning_rate": 4.3729240952667684e-05, "loss": 0.4803, "step": 16504 }, { "epoch": 1.1077145062246234, "grad_norm": 1.0185718536376953, "learning_rate": 4.3718458606866056e-05, "loss": 0.4928, "step": 16506 }, { "epoch": 1.1078487299083923, "grad_norm": 1.0792577266693115, "learning_rate": 4.370767655785488e-05, "loss": 0.5683, "step": 16508 }, { "epoch": 1.1079829535921613, "grad_norm": 1.0750395059585571, "learning_rate": 4.3696894806143653e-05, "loss": 0.49, "step": 16510 }, { "epoch": 1.1081171772759304, "grad_norm": 0.9474500417709351, "learning_rate": 4.368611335224172e-05, "loss": 0.5574, "step": 16512 }, { "epoch": 1.1082514009596993, "grad_norm": 1.19668447971344, "learning_rate": 4.367533219665853e-05, "loss": 0.5625, "step": 16514 }, { "epoch": 1.1083856246434682, "grad_norm": 0.8541036248207092, "learning_rate": 4.3664551339903454e-05, "loss": 0.4387, "step": 16516 }, { "epoch": 1.1085198483272374, "grad_norm": 1.0199828147888184, "learning_rate": 4.3653770782485876e-05, "loss": 0.5202, "step": 16518 }, { "epoch": 1.1086540720110063, "grad_norm": 1.1608493328094482, "learning_rate": 4.364299052491514e-05, "loss": 0.4858, "step": 16520 }, { "epoch": 1.1087882956947754, "grad_norm": 0.9933175444602966, "learning_rate": 4.36322105677006e-05, "loss": 0.4959, "step": 16522 }, { "epoch": 1.1089225193785444, "grad_norm": 1.8374931812286377, "learning_rate": 4.36214309113516e-05, "loss": 0.4993, "step": 16524 }, { "epoch": 1.1090567430623133, "grad_norm": 1.0183755159378052, "learning_rate": 4.361065155637744e-05, "loss": 0.5429, "step": 16526 }, { "epoch": 1.1091909667460824, "grad_norm": 1.0194748640060425, "learning_rate": 4.359987250328743e-05, "loss": 0.5263, "step": 16528 }, { "epoch": 1.1093251904298513, "grad_norm": 1.099562644958496, "learning_rate": 4.358909375259086e-05, "loss": 0.5132, "step": 16530 }, { "epoch": 1.1094594141136203, "grad_norm": 0.9151814579963684, "learning_rate": 4.357831530479699e-05, "loss": 0.4958, "step": 16532 }, { "epoch": 1.1095936377973894, "grad_norm": 0.92188560962677, "learning_rate": 4.3567537160415125e-05, "loss": 0.4762, "step": 16534 }, { "epoch": 1.1097278614811583, "grad_norm": 1.0499858856201172, "learning_rate": 4.355675931995444e-05, "loss": 0.5955, "step": 16536 }, { "epoch": 1.1098620851649272, "grad_norm": 1.0483999252319336, "learning_rate": 4.354598178392425e-05, "loss": 0.5233, "step": 16538 }, { "epoch": 1.1099963088486964, "grad_norm": 1.0125541687011719, "learning_rate": 4.353520455283369e-05, "loss": 0.5648, "step": 16540 }, { "epoch": 1.1101305325324653, "grad_norm": 0.927888035774231, "learning_rate": 4.3524427627192024e-05, "loss": 0.4779, "step": 16542 }, { "epoch": 1.1102647562162344, "grad_norm": 0.8777483701705933, "learning_rate": 4.3513651007508414e-05, "loss": 0.4826, "step": 16544 }, { "epoch": 1.1103989799000034, "grad_norm": 0.905463457107544, "learning_rate": 4.3502874694292045e-05, "loss": 0.514, "step": 16546 }, { "epoch": 1.1105332035837723, "grad_norm": 1.000179648399353, "learning_rate": 4.3492098688052064e-05, "loss": 0.494, "step": 16548 }, { "epoch": 1.1106674272675414, "grad_norm": 0.9666675925254822, "learning_rate": 4.3481322989297626e-05, "loss": 0.5082, "step": 16550 }, { "epoch": 1.1108016509513103, "grad_norm": 1.0104807615280151, "learning_rate": 4.347054759853787e-05, "loss": 0.4707, "step": 16552 }, { "epoch": 1.1109358746350793, "grad_norm": 2.4930474758148193, "learning_rate": 4.3459772516281886e-05, "loss": 0.5219, "step": 16554 }, { "epoch": 1.1110700983188484, "grad_norm": 0.97373366355896, "learning_rate": 4.344899774303879e-05, "loss": 0.4301, "step": 16556 }, { "epoch": 1.1112043220026173, "grad_norm": 0.9810713529586792, "learning_rate": 4.343822327931767e-05, "loss": 0.5474, "step": 16558 }, { "epoch": 1.1113385456863865, "grad_norm": 1.073265790939331, "learning_rate": 4.34274491256276e-05, "loss": 0.5042, "step": 16560 }, { "epoch": 1.1114727693701554, "grad_norm": 1.1238216161727905, "learning_rate": 4.3416675282477626e-05, "loss": 0.513, "step": 16562 }, { "epoch": 1.1116069930539243, "grad_norm": 0.9564653038978577, "learning_rate": 4.340590175037681e-05, "loss": 0.5194, "step": 16564 }, { "epoch": 1.1117412167376934, "grad_norm": 1.396801233291626, "learning_rate": 4.339512852983416e-05, "loss": 0.5107, "step": 16566 }, { "epoch": 1.1118754404214624, "grad_norm": 1.0099092721939087, "learning_rate": 4.3384355621358706e-05, "loss": 0.4657, "step": 16568 }, { "epoch": 1.1120096641052313, "grad_norm": 0.8993440270423889, "learning_rate": 4.3373583025459426e-05, "loss": 0.5107, "step": 16570 }, { "epoch": 1.1121438877890004, "grad_norm": 1.0395410060882568, "learning_rate": 4.3362810742645344e-05, "loss": 0.5569, "step": 16572 }, { "epoch": 1.1122781114727693, "grad_norm": 1.009219765663147, "learning_rate": 4.3352038773425376e-05, "loss": 0.4735, "step": 16574 }, { "epoch": 1.1124123351565385, "grad_norm": 0.9954773783683777, "learning_rate": 4.334126711830852e-05, "loss": 0.4676, "step": 16576 }, { "epoch": 1.1125465588403074, "grad_norm": 1.03559148311615, "learning_rate": 4.3330495777803685e-05, "loss": 0.5046, "step": 16578 }, { "epoch": 1.1126807825240763, "grad_norm": 0.972273588180542, "learning_rate": 4.3319724752419825e-05, "loss": 0.4735, "step": 16580 }, { "epoch": 1.1128150062078455, "grad_norm": 0.8986105918884277, "learning_rate": 4.330895404266583e-05, "loss": 0.4736, "step": 16582 }, { "epoch": 1.1129492298916144, "grad_norm": 1.0649678707122803, "learning_rate": 4.32981836490506e-05, "loss": 0.5084, "step": 16584 }, { "epoch": 1.1130834535753833, "grad_norm": 1.1590815782546997, "learning_rate": 4.3287413572083e-05, "loss": 0.4913, "step": 16586 }, { "epoch": 1.1132176772591524, "grad_norm": 1.0562185049057007, "learning_rate": 4.327664381227193e-05, "loss": 0.5497, "step": 16588 }, { "epoch": 1.1133519009429214, "grad_norm": 0.97037672996521, "learning_rate": 4.32658743701262e-05, "loss": 0.5467, "step": 16590 }, { "epoch": 1.1134861246266903, "grad_norm": 1.028220772743225, "learning_rate": 4.325510524615468e-05, "loss": 0.472, "step": 16592 }, { "epoch": 1.1136203483104594, "grad_norm": 1.1340383291244507, "learning_rate": 4.324433644086617e-05, "loss": 0.5538, "step": 16594 }, { "epoch": 1.1137545719942283, "grad_norm": 1.131611704826355, "learning_rate": 4.3233567954769475e-05, "loss": 0.4957, "step": 16596 }, { "epoch": 1.1138887956779975, "grad_norm": 1.2661018371582031, "learning_rate": 4.3222799788373374e-05, "loss": 0.4914, "step": 16598 }, { "epoch": 1.1140230193617664, "grad_norm": 1.0229700803756714, "learning_rate": 4.3212031942186694e-05, "loss": 0.5814, "step": 16600 }, { "epoch": 1.1141572430455353, "grad_norm": 0.9082956314086914, "learning_rate": 4.320126441671811e-05, "loss": 0.5003, "step": 16602 }, { "epoch": 1.1142914667293045, "grad_norm": 1.0306313037872314, "learning_rate": 4.3190497212476444e-05, "loss": 0.4976, "step": 16604 }, { "epoch": 1.1144256904130734, "grad_norm": 0.9325722455978394, "learning_rate": 4.317973032997038e-05, "loss": 0.4844, "step": 16606 }, { "epoch": 1.1145599140968423, "grad_norm": 1.0862069129943848, "learning_rate": 4.316896376970866e-05, "loss": 0.5539, "step": 16608 }, { "epoch": 1.1146941377806114, "grad_norm": 0.9598913192749023, "learning_rate": 4.3158197532199964e-05, "loss": 0.4889, "step": 16610 }, { "epoch": 1.1148283614643804, "grad_norm": 1.1408271789550781, "learning_rate": 4.314743161795298e-05, "loss": 0.5388, "step": 16612 }, { "epoch": 1.1149625851481493, "grad_norm": 0.9544355273246765, "learning_rate": 4.313666602747638e-05, "loss": 0.5197, "step": 16614 }, { "epoch": 1.1150968088319184, "grad_norm": 1.1572375297546387, "learning_rate": 4.3125900761278817e-05, "loss": 0.4982, "step": 16616 }, { "epoch": 1.1152310325156873, "grad_norm": 0.9330697655677795, "learning_rate": 4.3115135819868925e-05, "loss": 0.4749, "step": 16618 }, { "epoch": 1.1153652561994565, "grad_norm": 0.9616682529449463, "learning_rate": 4.3104371203755315e-05, "loss": 0.4508, "step": 16620 }, { "epoch": 1.1154994798832254, "grad_norm": 1.175607681274414, "learning_rate": 4.309360691344662e-05, "loss": 0.5284, "step": 16622 }, { "epoch": 1.1156337035669943, "grad_norm": 1.0348509550094604, "learning_rate": 4.30828429494514e-05, "loss": 0.4945, "step": 16624 }, { "epoch": 1.1157679272507635, "grad_norm": 1.0511419773101807, "learning_rate": 4.3072079312278266e-05, "loss": 0.5344, "step": 16626 }, { "epoch": 1.1159021509345324, "grad_norm": 1.7016751766204834, "learning_rate": 4.306131600243575e-05, "loss": 0.5863, "step": 16628 }, { "epoch": 1.1160363746183013, "grad_norm": 1.1384695768356323, "learning_rate": 4.305055302043242e-05, "loss": 0.5492, "step": 16630 }, { "epoch": 1.1161705983020704, "grad_norm": 1.0545936822891235, "learning_rate": 4.303979036677677e-05, "loss": 0.507, "step": 16632 }, { "epoch": 1.1163048219858394, "grad_norm": 0.8321439623832703, "learning_rate": 4.302902804197737e-05, "loss": 0.4136, "step": 16634 }, { "epoch": 1.1164390456696085, "grad_norm": 1.3442049026489258, "learning_rate": 4.301826604654266e-05, "loss": 0.5373, "step": 16636 }, { "epoch": 1.1165732693533774, "grad_norm": 0.9786088466644287, "learning_rate": 4.3007504380981165e-05, "loss": 0.5555, "step": 16638 }, { "epoch": 1.1167074930371463, "grad_norm": 1.044503927230835, "learning_rate": 4.2996743045801334e-05, "loss": 0.499, "step": 16640 }, { "epoch": 1.1168417167209155, "grad_norm": 1.028637170791626, "learning_rate": 4.298598204151163e-05, "loss": 0.5545, "step": 16642 }, { "epoch": 1.1169759404046844, "grad_norm": 1.0345284938812256, "learning_rate": 4.297522136862047e-05, "loss": 0.5329, "step": 16644 }, { "epoch": 1.1171101640884533, "grad_norm": 0.9361871480941772, "learning_rate": 4.2964461027636305e-05, "loss": 0.5188, "step": 16646 }, { "epoch": 1.1172443877722225, "grad_norm": 1.0022510290145874, "learning_rate": 4.29537010190675e-05, "loss": 0.5554, "step": 16648 }, { "epoch": 1.1173786114559914, "grad_norm": 1.1015814542770386, "learning_rate": 4.2942941343422485e-05, "loss": 0.5083, "step": 16650 }, { "epoch": 1.1175128351397605, "grad_norm": 1.0244308710098267, "learning_rate": 4.2932182001209596e-05, "loss": 0.5631, "step": 16652 }, { "epoch": 1.1176470588235294, "grad_norm": 0.9587509632110596, "learning_rate": 4.2921422992937224e-05, "loss": 0.4849, "step": 16654 }, { "epoch": 1.1177812825072984, "grad_norm": 1.068928837776184, "learning_rate": 4.291066431911369e-05, "loss": 0.4937, "step": 16656 }, { "epoch": 1.1179155061910675, "grad_norm": 0.9799166321754456, "learning_rate": 4.2899905980247336e-05, "loss": 0.4927, "step": 16658 }, { "epoch": 1.1180497298748364, "grad_norm": 0.970655620098114, "learning_rate": 4.2889147976846436e-05, "loss": 0.5538, "step": 16660 }, { "epoch": 1.1181839535586053, "grad_norm": 1.1147160530090332, "learning_rate": 4.287839030941934e-05, "loss": 0.5111, "step": 16662 }, { "epoch": 1.1183181772423745, "grad_norm": 1.0592046976089478, "learning_rate": 4.286763297847428e-05, "loss": 0.6146, "step": 16664 }, { "epoch": 1.1184524009261434, "grad_norm": 1.1099165678024292, "learning_rate": 4.2856875984519545e-05, "loss": 0.5323, "step": 16666 }, { "epoch": 1.1185866246099123, "grad_norm": 0.9070066809654236, "learning_rate": 4.2846119328063363e-05, "loss": 0.5229, "step": 16668 }, { "epoch": 1.1187208482936815, "grad_norm": 1.1833245754241943, "learning_rate": 4.283536300961399e-05, "loss": 0.5909, "step": 16670 }, { "epoch": 1.1188550719774504, "grad_norm": 0.9487655162811279, "learning_rate": 4.282460702967962e-05, "loss": 0.4418, "step": 16672 }, { "epoch": 1.1189892956612195, "grad_norm": 0.9981588125228882, "learning_rate": 4.2813851388768464e-05, "loss": 0.5104, "step": 16674 }, { "epoch": 1.1191235193449884, "grad_norm": 0.9488283395767212, "learning_rate": 4.2803096087388687e-05, "loss": 0.523, "step": 16676 }, { "epoch": 1.1192577430287574, "grad_norm": 0.9185371994972229, "learning_rate": 4.279234112604848e-05, "loss": 0.5018, "step": 16678 }, { "epoch": 1.1193919667125265, "grad_norm": 0.9361635446548462, "learning_rate": 4.2781586505255965e-05, "loss": 0.5053, "step": 16680 }, { "epoch": 1.1195261903962954, "grad_norm": 1.1580199003219604, "learning_rate": 4.277083222551931e-05, "loss": 0.5772, "step": 16682 }, { "epoch": 1.1196604140800643, "grad_norm": 1.0087647438049316, "learning_rate": 4.276007828734661e-05, "loss": 0.5337, "step": 16684 }, { "epoch": 1.1197946377638335, "grad_norm": 0.9543098211288452, "learning_rate": 4.2749324691245977e-05, "loss": 0.5617, "step": 16686 }, { "epoch": 1.1199288614476024, "grad_norm": 0.9907855987548828, "learning_rate": 4.27385714377255e-05, "loss": 0.4678, "step": 16688 }, { "epoch": 1.1200630851313713, "grad_norm": 0.9831494092941284, "learning_rate": 4.2727818527293226e-05, "loss": 0.489, "step": 16690 }, { "epoch": 1.1201973088151405, "grad_norm": 1.1383953094482422, "learning_rate": 4.271706596045725e-05, "loss": 0.5409, "step": 16692 }, { "epoch": 1.1203315324989094, "grad_norm": 1.3033020496368408, "learning_rate": 4.270631373772556e-05, "loss": 0.4914, "step": 16694 }, { "epoch": 1.1204657561826785, "grad_norm": 1.1383192539215088, "learning_rate": 4.269556185960623e-05, "loss": 0.5127, "step": 16696 }, { "epoch": 1.1205999798664474, "grad_norm": 1.1375019550323486, "learning_rate": 4.2684810326607206e-05, "loss": 0.5282, "step": 16698 }, { "epoch": 1.1207342035502164, "grad_norm": 0.9572421312332153, "learning_rate": 4.267405913923654e-05, "loss": 0.4612, "step": 16700 }, { "epoch": 1.1208684272339855, "grad_norm": 1.217170238494873, "learning_rate": 4.2663308298002146e-05, "loss": 0.5451, "step": 16702 }, { "epoch": 1.1210026509177544, "grad_norm": 0.9934802055358887, "learning_rate": 4.265255780341203e-05, "loss": 0.4905, "step": 16704 }, { "epoch": 1.1211368746015233, "grad_norm": 1.1445648670196533, "learning_rate": 4.2641807655974096e-05, "loss": 0.604, "step": 16706 }, { "epoch": 1.1212710982852925, "grad_norm": 1.0030955076217651, "learning_rate": 4.26310578561963e-05, "loss": 0.5055, "step": 16708 }, { "epoch": 1.1214053219690614, "grad_norm": 1.2229998111724854, "learning_rate": 4.262030840458651e-05, "loss": 0.6362, "step": 16710 }, { "epoch": 1.1215395456528305, "grad_norm": 0.9906282424926758, "learning_rate": 4.260955930165265e-05, "loss": 0.4983, "step": 16712 }, { "epoch": 1.1216737693365995, "grad_norm": 0.9919167757034302, "learning_rate": 4.259881054790257e-05, "loss": 0.5159, "step": 16714 }, { "epoch": 1.1218079930203684, "grad_norm": 0.9842211604118347, "learning_rate": 4.258806214384415e-05, "loss": 0.5157, "step": 16716 }, { "epoch": 1.1219422167041375, "grad_norm": 1.0635043382644653, "learning_rate": 4.2577314089985204e-05, "loss": 0.5433, "step": 16718 }, { "epoch": 1.1220764403879064, "grad_norm": 1.0264695882797241, "learning_rate": 4.2566566386833584e-05, "loss": 0.5344, "step": 16720 }, { "epoch": 1.1222106640716754, "grad_norm": 0.9740486145019531, "learning_rate": 4.255581903489707e-05, "loss": 0.5472, "step": 16722 }, { "epoch": 1.1223448877554445, "grad_norm": 0.8539181351661682, "learning_rate": 4.254507203468349e-05, "loss": 0.4583, "step": 16724 }, { "epoch": 1.1224791114392134, "grad_norm": 1.0089354515075684, "learning_rate": 4.253432538670057e-05, "loss": 0.4837, "step": 16726 }, { "epoch": 1.1226133351229826, "grad_norm": 0.9615651965141296, "learning_rate": 4.252357909145611e-05, "loss": 0.5052, "step": 16728 }, { "epoch": 1.1227475588067515, "grad_norm": 1.0212202072143555, "learning_rate": 4.251283314945783e-05, "loss": 0.5558, "step": 16730 }, { "epoch": 1.1228817824905204, "grad_norm": 1.0308191776275635, "learning_rate": 4.2502087561213475e-05, "loss": 0.4758, "step": 16732 }, { "epoch": 1.1230160061742895, "grad_norm": 0.9926826357841492, "learning_rate": 4.249134232723073e-05, "loss": 0.5044, "step": 16734 }, { "epoch": 1.1231502298580585, "grad_norm": 1.4401243925094604, "learning_rate": 4.2480597448017294e-05, "loss": 0.4973, "step": 16736 }, { "epoch": 1.1232844535418274, "grad_norm": 0.9778569340705872, "learning_rate": 4.246985292408084e-05, "loss": 0.5164, "step": 16738 }, { "epoch": 1.1234186772255965, "grad_norm": 0.9825350642204285, "learning_rate": 4.2459108755929034e-05, "loss": 0.5223, "step": 16740 }, { "epoch": 1.1235529009093654, "grad_norm": 1.7382341623306274, "learning_rate": 4.244836494406951e-05, "loss": 0.4909, "step": 16742 }, { "epoch": 1.1236871245931344, "grad_norm": 1.0226861238479614, "learning_rate": 4.2437621489009894e-05, "loss": 0.5395, "step": 16744 }, { "epoch": 1.1238213482769035, "grad_norm": 0.9868961572647095, "learning_rate": 4.242687839125778e-05, "loss": 0.5436, "step": 16746 }, { "epoch": 1.1239555719606724, "grad_norm": 1.6000854969024658, "learning_rate": 4.2416135651320785e-05, "loss": 0.5686, "step": 16748 }, { "epoch": 1.1240897956444416, "grad_norm": 0.9828076958656311, "learning_rate": 4.2405393269706444e-05, "loss": 0.5938, "step": 16750 }, { "epoch": 1.1242240193282105, "grad_norm": 0.9150723814964294, "learning_rate": 4.239465124692237e-05, "loss": 0.443, "step": 16752 }, { "epoch": 1.1243582430119794, "grad_norm": 0.915022075176239, "learning_rate": 4.238390958347604e-05, "loss": 0.4964, "step": 16754 }, { "epoch": 1.1244924666957485, "grad_norm": 0.9428569674491882, "learning_rate": 4.237316827987502e-05, "loss": 0.4481, "step": 16756 }, { "epoch": 1.1246266903795175, "grad_norm": 0.7897739410400391, "learning_rate": 4.236242733662682e-05, "loss": 0.5133, "step": 16758 }, { "epoch": 1.1247609140632864, "grad_norm": 0.892500638961792, "learning_rate": 4.235168675423887e-05, "loss": 0.5001, "step": 16760 }, { "epoch": 1.1248951377470555, "grad_norm": 0.92838454246521, "learning_rate": 4.2340946533218725e-05, "loss": 0.4946, "step": 16762 }, { "epoch": 1.1250293614308244, "grad_norm": 0.9659767150878906, "learning_rate": 4.233020667407375e-05, "loss": 0.5315, "step": 16764 }, { "epoch": 1.1251635851145934, "grad_norm": 1.129931926727295, "learning_rate": 4.2319467177311463e-05, "loss": 0.5051, "step": 16766 }, { "epoch": 1.1252978087983625, "grad_norm": 0.9765834212303162, "learning_rate": 4.2308728043439224e-05, "loss": 0.4413, "step": 16768 }, { "epoch": 1.1254320324821314, "grad_norm": 1.0590866804122925, "learning_rate": 4.2297989272964476e-05, "loss": 0.508, "step": 16770 }, { "epoch": 1.1255662561659006, "grad_norm": 0.9979098439216614, "learning_rate": 4.228725086639458e-05, "loss": 0.5216, "step": 16772 }, { "epoch": 1.1257004798496695, "grad_norm": 0.987686812877655, "learning_rate": 4.227651282423693e-05, "loss": 0.5149, "step": 16774 }, { "epoch": 1.1258347035334384, "grad_norm": 0.9693307876586914, "learning_rate": 4.2265775146998844e-05, "loss": 0.5459, "step": 16776 }, { "epoch": 1.1259689272172075, "grad_norm": 1.0481395721435547, "learning_rate": 4.225503783518769e-05, "loss": 0.5663, "step": 16778 }, { "epoch": 1.1261031509009765, "grad_norm": 0.9288170337677002, "learning_rate": 4.224430088931075e-05, "loss": 0.493, "step": 16780 }, { "epoch": 1.1262373745847456, "grad_norm": 1.0299285650253296, "learning_rate": 4.223356430987536e-05, "loss": 0.5827, "step": 16782 }, { "epoch": 1.1263715982685145, "grad_norm": 1.0432701110839844, "learning_rate": 4.222282809738875e-05, "loss": 0.463, "step": 16784 }, { "epoch": 1.1265058219522834, "grad_norm": 1.214779257774353, "learning_rate": 4.221209225235827e-05, "loss": 0.5146, "step": 16786 }, { "epoch": 1.1266400456360526, "grad_norm": 1.0165801048278809, "learning_rate": 4.220135677529107e-05, "loss": 0.487, "step": 16788 }, { "epoch": 1.1267742693198215, "grad_norm": 0.9639692306518555, "learning_rate": 4.219062166669447e-05, "loss": 0.5206, "step": 16790 }, { "epoch": 1.1269084930035904, "grad_norm": 1.0271775722503662, "learning_rate": 4.21798869270756e-05, "loss": 0.5052, "step": 16792 }, { "epoch": 1.1270427166873596, "grad_norm": 1.0103228092193604, "learning_rate": 4.216915255694172e-05, "loss": 0.5101, "step": 16794 }, { "epoch": 1.1271769403711285, "grad_norm": 1.2222870588302612, "learning_rate": 4.215841855679997e-05, "loss": 0.5024, "step": 16796 }, { "epoch": 1.1273111640548974, "grad_norm": 0.9785430431365967, "learning_rate": 4.214768492715754e-05, "loss": 0.5736, "step": 16798 }, { "epoch": 1.1274453877386665, "grad_norm": 1.064740538597107, "learning_rate": 4.213695166852153e-05, "loss": 0.5272, "step": 16800 }, { "epoch": 1.1275796114224355, "grad_norm": 1.0236676931381226, "learning_rate": 4.212621878139912e-05, "loss": 0.6104, "step": 16802 }, { "epoch": 1.1277138351062046, "grad_norm": 0.9074214696884155, "learning_rate": 4.211548626629737e-05, "loss": 0.5048, "step": 16804 }, { "epoch": 1.1278480587899735, "grad_norm": 0.8793216943740845, "learning_rate": 4.21047541237234e-05, "loss": 0.513, "step": 16806 }, { "epoch": 1.1279822824737424, "grad_norm": 1.0216366052627563, "learning_rate": 4.2094022354184266e-05, "loss": 0.471, "step": 16808 }, { "epoch": 1.1281165061575116, "grad_norm": 1.0820353031158447, "learning_rate": 4.208329095818704e-05, "loss": 0.5181, "step": 16810 }, { "epoch": 1.1282507298412805, "grad_norm": 0.9416481852531433, "learning_rate": 4.207255993623872e-05, "loss": 0.552, "step": 16812 }, { "epoch": 1.1283849535250494, "grad_norm": 0.9594340920448303, "learning_rate": 4.206182928884639e-05, "loss": 0.4678, "step": 16814 }, { "epoch": 1.1285191772088186, "grad_norm": 0.9467558264732361, "learning_rate": 4.205109901651699e-05, "loss": 0.4932, "step": 16816 }, { "epoch": 1.1286534008925875, "grad_norm": 0.9780734181404114, "learning_rate": 4.204036911975755e-05, "loss": 0.4848, "step": 16818 }, { "epoch": 1.1287876245763564, "grad_norm": 1.0107637643814087, "learning_rate": 4.2029639599075004e-05, "loss": 0.5356, "step": 16820 }, { "epoch": 1.1289218482601255, "grad_norm": 1.1758426427841187, "learning_rate": 4.201891045497633e-05, "loss": 0.5661, "step": 16822 }, { "epoch": 1.1290560719438945, "grad_norm": 1.074581503868103, "learning_rate": 4.200818168796844e-05, "loss": 0.498, "step": 16824 }, { "epoch": 1.1291902956276636, "grad_norm": 1.0495214462280273, "learning_rate": 4.199745329855823e-05, "loss": 0.4953, "step": 16826 }, { "epoch": 1.1293245193114325, "grad_norm": 1.0303970575332642, "learning_rate": 4.1986725287252634e-05, "loss": 0.5049, "step": 16828 }, { "epoch": 1.1294587429952014, "grad_norm": 0.956167459487915, "learning_rate": 4.19759976545585e-05, "loss": 0.531, "step": 16830 }, { "epoch": 1.1295929666789706, "grad_norm": 1.0434627532958984, "learning_rate": 4.19652704009827e-05, "loss": 0.5046, "step": 16832 }, { "epoch": 1.1297271903627395, "grad_norm": 1.7044645547866821, "learning_rate": 4.1954543527032076e-05, "loss": 0.3955, "step": 16834 }, { "epoch": 1.1298614140465084, "grad_norm": 1.1061593294143677, "learning_rate": 4.194381703321346e-05, "loss": 0.5206, "step": 16836 }, { "epoch": 1.1299956377302776, "grad_norm": 0.972310483455658, "learning_rate": 4.193309092003363e-05, "loss": 0.5309, "step": 16838 }, { "epoch": 1.1301298614140465, "grad_norm": 1.1154252290725708, "learning_rate": 4.1922365187999404e-05, "loss": 0.5935, "step": 16840 }, { "epoch": 1.1302640850978154, "grad_norm": 1.104824185371399, "learning_rate": 4.1911639837617526e-05, "loss": 0.5234, "step": 16842 }, { "epoch": 1.1303983087815845, "grad_norm": 1.0318411588668823, "learning_rate": 4.1900914869394775e-05, "loss": 0.4876, "step": 16844 }, { "epoch": 1.1305325324653535, "grad_norm": 0.9598281383514404, "learning_rate": 4.189019028383785e-05, "loss": 0.4643, "step": 16846 }, { "epoch": 1.1306667561491226, "grad_norm": 1.2466892004013062, "learning_rate": 4.187946608145351e-05, "loss": 0.5751, "step": 16848 }, { "epoch": 1.1308009798328915, "grad_norm": 1.0190337896347046, "learning_rate": 4.18687422627484e-05, "loss": 0.512, "step": 16850 }, { "epoch": 1.1309352035166604, "grad_norm": 1.062680721282959, "learning_rate": 4.185801882822927e-05, "loss": 0.524, "step": 16852 }, { "epoch": 1.1310694272004296, "grad_norm": 1.2178183794021606, "learning_rate": 4.184729577840271e-05, "loss": 0.5675, "step": 16854 }, { "epoch": 1.1312036508841985, "grad_norm": 1.0182420015335083, "learning_rate": 4.18365731137754e-05, "loss": 0.5095, "step": 16856 }, { "epoch": 1.1313378745679676, "grad_norm": 0.9893097877502441, "learning_rate": 4.1825850834853954e-05, "loss": 0.476, "step": 16858 }, { "epoch": 1.1314720982517366, "grad_norm": 1.0728018283843994, "learning_rate": 4.181512894214499e-05, "loss": 0.5199, "step": 16860 }, { "epoch": 1.1316063219355055, "grad_norm": 1.1170811653137207, "learning_rate": 4.180440743615508e-05, "loss": 0.547, "step": 16862 }, { "epoch": 1.1317405456192744, "grad_norm": 1.0509802103042603, "learning_rate": 4.1793686317390815e-05, "loss": 0.4813, "step": 16864 }, { "epoch": 1.1318747693030435, "grad_norm": 0.9971725344657898, "learning_rate": 4.178296558635873e-05, "loss": 0.4501, "step": 16866 }, { "epoch": 1.1320089929868125, "grad_norm": 0.957833468914032, "learning_rate": 4.1772245243565364e-05, "loss": 0.5182, "step": 16868 }, { "epoch": 1.1321432166705816, "grad_norm": 1.008810043334961, "learning_rate": 4.1761525289517215e-05, "loss": 0.4993, "step": 16870 }, { "epoch": 1.1322774403543505, "grad_norm": 0.941467821598053, "learning_rate": 4.175080572472082e-05, "loss": 0.5912, "step": 16872 }, { "epoch": 1.1324116640381194, "grad_norm": 1.0539593696594238, "learning_rate": 4.1740086549682606e-05, "loss": 0.4971, "step": 16874 }, { "epoch": 1.1325458877218886, "grad_norm": 1.0735893249511719, "learning_rate": 4.172936776490909e-05, "loss": 0.5694, "step": 16876 }, { "epoch": 1.1326801114056575, "grad_norm": 0.9888498783111572, "learning_rate": 4.171864937090665e-05, "loss": 0.4602, "step": 16878 }, { "epoch": 1.1328143350894266, "grad_norm": 0.9963198900222778, "learning_rate": 4.170793136818178e-05, "loss": 0.5226, "step": 16880 }, { "epoch": 1.1329485587731956, "grad_norm": 2.036478042602539, "learning_rate": 4.1697213757240814e-05, "loss": 0.4859, "step": 16882 }, { "epoch": 1.1330827824569645, "grad_norm": 0.9562026858329773, "learning_rate": 4.168649653859019e-05, "loss": 0.4906, "step": 16884 }, { "epoch": 1.1332170061407336, "grad_norm": 1.1929901838302612, "learning_rate": 4.1675779712736254e-05, "loss": 0.5306, "step": 16886 }, { "epoch": 1.1333512298245025, "grad_norm": 1.0257091522216797, "learning_rate": 4.1665063280185354e-05, "loss": 0.5154, "step": 16888 }, { "epoch": 1.1334854535082715, "grad_norm": 1.0640220642089844, "learning_rate": 4.165434724144383e-05, "loss": 0.4901, "step": 16890 }, { "epoch": 1.1336196771920406, "grad_norm": 1.0287139415740967, "learning_rate": 4.1643631597017987e-05, "loss": 0.5058, "step": 16892 }, { "epoch": 1.1337539008758095, "grad_norm": 1.0756909847259521, "learning_rate": 4.1632916347414124e-05, "loss": 0.4885, "step": 16894 }, { "epoch": 1.1338881245595784, "grad_norm": 1.0119127035140991, "learning_rate": 4.1622201493138495e-05, "loss": 0.4996, "step": 16896 }, { "epoch": 1.1340223482433476, "grad_norm": 1.1204429864883423, "learning_rate": 4.161148703469739e-05, "loss": 0.5635, "step": 16898 }, { "epoch": 1.1341565719271165, "grad_norm": 1.0327421426773071, "learning_rate": 4.160077297259701e-05, "loss": 0.4881, "step": 16900 }, { "epoch": 1.1342907956108856, "grad_norm": 1.0640689134597778, "learning_rate": 4.1590059307343596e-05, "loss": 0.5416, "step": 16902 }, { "epoch": 1.1344250192946546, "grad_norm": 1.1616973876953125, "learning_rate": 4.1579346039443346e-05, "loss": 0.4643, "step": 16904 }, { "epoch": 1.1345592429784235, "grad_norm": 1.0026849508285522, "learning_rate": 4.156863316940244e-05, "loss": 0.5394, "step": 16906 }, { "epoch": 1.1346934666621926, "grad_norm": 1.00323486328125, "learning_rate": 4.155792069772702e-05, "loss": 0.5543, "step": 16908 }, { "epoch": 1.1348276903459615, "grad_norm": 1.1087974309921265, "learning_rate": 4.1547208624923266e-05, "loss": 0.5363, "step": 16910 }, { "epoch": 1.1349619140297305, "grad_norm": 0.9458073377609253, "learning_rate": 4.153649695149726e-05, "loss": 0.5073, "step": 16912 }, { "epoch": 1.1350961377134996, "grad_norm": 1.2469216585159302, "learning_rate": 4.152578567795516e-05, "loss": 0.5397, "step": 16914 }, { "epoch": 1.1352303613972685, "grad_norm": 0.915696382522583, "learning_rate": 4.1515074804802995e-05, "loss": 0.5485, "step": 16916 }, { "epoch": 1.1353645850810374, "grad_norm": 0.8987810611724854, "learning_rate": 4.150436433254688e-05, "loss": 0.4696, "step": 16918 }, { "epoch": 1.1354988087648066, "grad_norm": 0.939405620098114, "learning_rate": 4.1493654261692834e-05, "loss": 0.494, "step": 16920 }, { "epoch": 1.1356330324485755, "grad_norm": 0.9830738306045532, "learning_rate": 4.148294459274691e-05, "loss": 0.4773, "step": 16922 }, { "epoch": 1.1357672561323446, "grad_norm": 1.1729371547698975, "learning_rate": 4.147223532621508e-05, "loss": 0.4847, "step": 16924 }, { "epoch": 1.1359014798161136, "grad_norm": 1.1334617137908936, "learning_rate": 4.146152646260339e-05, "loss": 0.5156, "step": 16926 }, { "epoch": 1.1360357034998825, "grad_norm": 1.2039645910263062, "learning_rate": 4.1450818002417766e-05, "loss": 0.5845, "step": 16928 }, { "epoch": 1.1361699271836516, "grad_norm": 1.044554591178894, "learning_rate": 4.1440109946164196e-05, "loss": 0.4928, "step": 16930 }, { "epoch": 1.1363041508674205, "grad_norm": 0.9057065844535828, "learning_rate": 4.142940229434858e-05, "loss": 0.46, "step": 16932 }, { "epoch": 1.1364383745511897, "grad_norm": 0.9855564832687378, "learning_rate": 4.141869504747687e-05, "loss": 0.5202, "step": 16934 }, { "epoch": 1.1365725982349586, "grad_norm": 0.9572040438652039, "learning_rate": 4.140798820605493e-05, "loss": 0.4946, "step": 16936 }, { "epoch": 1.1367068219187275, "grad_norm": 0.9390221238136292, "learning_rate": 4.139728177058867e-05, "loss": 0.5045, "step": 16938 }, { "epoch": 1.1368410456024964, "grad_norm": 0.9028573632240295, "learning_rate": 4.1386575741583904e-05, "loss": 0.5411, "step": 16940 }, { "epoch": 1.1369752692862656, "grad_norm": 1.138342261314392, "learning_rate": 4.1375870119546526e-05, "loss": 0.5993, "step": 16942 }, { "epoch": 1.1371094929700345, "grad_norm": 0.9852227568626404, "learning_rate": 4.1365164904982296e-05, "loss": 0.5267, "step": 16944 }, { "epoch": 1.1372437166538036, "grad_norm": 0.9474890232086182, "learning_rate": 4.135446009839707e-05, "loss": 0.5092, "step": 16946 }, { "epoch": 1.1373779403375726, "grad_norm": 0.9510831832885742, "learning_rate": 4.1343755700296584e-05, "loss": 0.5578, "step": 16948 }, { "epoch": 1.1375121640213415, "grad_norm": 1.0347990989685059, "learning_rate": 4.1333051711186635e-05, "loss": 0.555, "step": 16950 }, { "epoch": 1.1376463877051106, "grad_norm": 1.0550622940063477, "learning_rate": 4.132234813157294e-05, "loss": 0.5184, "step": 16952 }, { "epoch": 1.1377806113888795, "grad_norm": 1.0261003971099854, "learning_rate": 4.1311644961961244e-05, "loss": 0.523, "step": 16954 }, { "epoch": 1.1379148350726487, "grad_norm": 0.9454745054244995, "learning_rate": 4.1300942202857226e-05, "loss": 0.5207, "step": 16956 }, { "epoch": 1.1380490587564176, "grad_norm": 0.9956864714622498, "learning_rate": 4.129023985476659e-05, "loss": 0.564, "step": 16958 }, { "epoch": 1.1381832824401865, "grad_norm": 1.1226762533187866, "learning_rate": 4.127953791819499e-05, "loss": 0.4966, "step": 16960 }, { "epoch": 1.1383175061239557, "grad_norm": 1.004565954208374, "learning_rate": 4.126883639364808e-05, "loss": 0.4772, "step": 16962 }, { "epoch": 1.1384517298077246, "grad_norm": 1.0866649150848389, "learning_rate": 4.125813528163148e-05, "loss": 0.5378, "step": 16964 }, { "epoch": 1.1385859534914935, "grad_norm": 0.9746032357215881, "learning_rate": 4.124743458265079e-05, "loss": 0.4925, "step": 16966 }, { "epoch": 1.1387201771752626, "grad_norm": 0.9077035784721375, "learning_rate": 4.123673429721162e-05, "loss": 0.5254, "step": 16968 }, { "epoch": 1.1388544008590316, "grad_norm": 1.0778063535690308, "learning_rate": 4.12260344258195e-05, "loss": 0.6036, "step": 16970 }, { "epoch": 1.1389886245428005, "grad_norm": 1.1656830310821533, "learning_rate": 4.121533496898002e-05, "loss": 0.4911, "step": 16972 }, { "epoch": 1.1391228482265696, "grad_norm": 1.0001049041748047, "learning_rate": 4.120463592719867e-05, "loss": 0.4811, "step": 16974 }, { "epoch": 1.1392570719103385, "grad_norm": 0.9660854935646057, "learning_rate": 4.119393730098101e-05, "loss": 0.5281, "step": 16976 }, { "epoch": 1.1393912955941077, "grad_norm": 0.9911681413650513, "learning_rate": 4.1183239090832455e-05, "loss": 0.5248, "step": 16978 }, { "epoch": 1.1395255192778766, "grad_norm": 1.0164984464645386, "learning_rate": 4.117254129725854e-05, "loss": 0.5099, "step": 16980 }, { "epoch": 1.1396597429616455, "grad_norm": 1.0523251295089722, "learning_rate": 4.116184392076468e-05, "loss": 0.5631, "step": 16982 }, { "epoch": 1.1397939666454147, "grad_norm": 1.0661886930465698, "learning_rate": 4.1151146961856326e-05, "loss": 0.5161, "step": 16984 }, { "epoch": 1.1399281903291836, "grad_norm": 0.9028160572052002, "learning_rate": 4.114045042103887e-05, "loss": 0.4946, "step": 16986 }, { "epoch": 1.1400624140129525, "grad_norm": 1.2710407972335815, "learning_rate": 4.1129754298817714e-05, "loss": 0.5375, "step": 16988 }, { "epoch": 1.1401966376967216, "grad_norm": 1.012258529663086, "learning_rate": 4.1119058595698217e-05, "loss": 0.5062, "step": 16990 }, { "epoch": 1.1403308613804906, "grad_norm": 0.9147583842277527, "learning_rate": 4.110836331218575e-05, "loss": 0.5091, "step": 16992 }, { "epoch": 1.1404650850642595, "grad_norm": 1.1260439157485962, "learning_rate": 4.109766844878562e-05, "loss": 0.5854, "step": 16994 }, { "epoch": 1.1405993087480286, "grad_norm": 0.9493597149848938, "learning_rate": 4.108697400600316e-05, "loss": 0.4336, "step": 16996 }, { "epoch": 1.1407335324317975, "grad_norm": 0.9938174486160278, "learning_rate": 4.1076279984343636e-05, "loss": 0.5017, "step": 16998 }, { "epoch": 1.1408677561155667, "grad_norm": 1.0089324712753296, "learning_rate": 4.106558638431236e-05, "loss": 0.5049, "step": 17000 }, { "epoch": 1.1410019797993356, "grad_norm": 1.0204837322235107, "learning_rate": 4.105489320641452e-05, "loss": 0.5761, "step": 17002 }, { "epoch": 1.1411362034831045, "grad_norm": 1.0670700073242188, "learning_rate": 4.104420045115543e-05, "loss": 0.5781, "step": 17004 }, { "epoch": 1.1412704271668737, "grad_norm": 0.95860755443573, "learning_rate": 4.103350811904022e-05, "loss": 0.4447, "step": 17006 }, { "epoch": 1.1414046508506426, "grad_norm": 0.926114559173584, "learning_rate": 4.102281621057415e-05, "loss": 0.4874, "step": 17008 }, { "epoch": 1.1415388745344117, "grad_norm": 0.9847375154495239, "learning_rate": 4.1012124726262345e-05, "loss": 0.5116, "step": 17010 }, { "epoch": 1.1416730982181806, "grad_norm": 1.06489896774292, "learning_rate": 4.100143366660998e-05, "loss": 0.5412, "step": 17012 }, { "epoch": 1.1418073219019496, "grad_norm": 1.0089995861053467, "learning_rate": 4.099074303212218e-05, "loss": 0.523, "step": 17014 }, { "epoch": 1.1419415455857185, "grad_norm": 1.2394014596939087, "learning_rate": 4.098005282330406e-05, "loss": 0.6122, "step": 17016 }, { "epoch": 1.1420757692694876, "grad_norm": 1.0862969160079956, "learning_rate": 4.0969363040660685e-05, "loss": 0.5264, "step": 17018 }, { "epoch": 1.1422099929532565, "grad_norm": 1.0558875799179077, "learning_rate": 4.095867368469717e-05, "loss": 0.5508, "step": 17020 }, { "epoch": 1.1423442166370257, "grad_norm": 1.0416982173919678, "learning_rate": 4.094798475591853e-05, "loss": 0.512, "step": 17022 }, { "epoch": 1.1424784403207946, "grad_norm": 0.9815146327018738, "learning_rate": 4.093729625482983e-05, "loss": 0.5085, "step": 17024 }, { "epoch": 1.1426126640045635, "grad_norm": 0.936656653881073, "learning_rate": 4.092660818193603e-05, "loss": 0.512, "step": 17026 }, { "epoch": 1.1427468876883327, "grad_norm": 1.0042047500610352, "learning_rate": 4.091592053774219e-05, "loss": 0.4615, "step": 17028 }, { "epoch": 1.1428811113721016, "grad_norm": 1.1232362985610962, "learning_rate": 4.0905233322753224e-05, "loss": 0.5795, "step": 17030 }, { "epoch": 1.1430153350558707, "grad_norm": 1.0395944118499756, "learning_rate": 4.089454653747409e-05, "loss": 0.5091, "step": 17032 }, { "epoch": 1.1431495587396396, "grad_norm": 1.0014408826828003, "learning_rate": 4.088386018240974e-05, "loss": 0.506, "step": 17034 }, { "epoch": 1.1432837824234086, "grad_norm": 1.0939862728118896, "learning_rate": 4.087317425806504e-05, "loss": 0.5409, "step": 17036 }, { "epoch": 1.1434180061071777, "grad_norm": 1.2131550312042236, "learning_rate": 4.0862488764944956e-05, "loss": 0.5304, "step": 17038 }, { "epoch": 1.1435522297909466, "grad_norm": 1.020034909248352, "learning_rate": 4.085180370355427e-05, "loss": 0.5231, "step": 17040 }, { "epoch": 1.1436864534747155, "grad_norm": 1.0369129180908203, "learning_rate": 4.0841119074397895e-05, "loss": 0.592, "step": 17042 }, { "epoch": 1.1438206771584847, "grad_norm": 1.0240205526351929, "learning_rate": 4.083043487798062e-05, "loss": 0.5103, "step": 17044 }, { "epoch": 1.1439549008422536, "grad_norm": 0.9685449004173279, "learning_rate": 4.081975111480728e-05, "loss": 0.5502, "step": 17046 }, { "epoch": 1.1440891245260225, "grad_norm": 0.9529426693916321, "learning_rate": 4.080906778538263e-05, "loss": 0.4839, "step": 17048 }, { "epoch": 1.1442233482097917, "grad_norm": 0.9597352147102356, "learning_rate": 4.079838489021146e-05, "loss": 0.5054, "step": 17050 }, { "epoch": 1.1443575718935606, "grad_norm": 0.9930350184440613, "learning_rate": 4.078770242979851e-05, "loss": 0.5735, "step": 17052 }, { "epoch": 1.1444917955773297, "grad_norm": 0.8961820006370544, "learning_rate": 4.077702040464851e-05, "loss": 0.4948, "step": 17054 }, { "epoch": 1.1446260192610986, "grad_norm": 1.060964822769165, "learning_rate": 4.0766338815266156e-05, "loss": 0.4675, "step": 17056 }, { "epoch": 1.1447602429448676, "grad_norm": 1.0138031244277954, "learning_rate": 4.0755657662156144e-05, "loss": 0.4757, "step": 17058 }, { "epoch": 1.1448944666286367, "grad_norm": 0.8255618214607239, "learning_rate": 4.0744976945823116e-05, "loss": 0.5061, "step": 17060 }, { "epoch": 1.1450286903124056, "grad_norm": 1.083939790725708, "learning_rate": 4.073429666677174e-05, "loss": 0.5133, "step": 17062 }, { "epoch": 1.1451629139961745, "grad_norm": 1.0115509033203125, "learning_rate": 4.07236168255066e-05, "loss": 0.5213, "step": 17064 }, { "epoch": 1.1452971376799437, "grad_norm": 1.0404711961746216, "learning_rate": 4.0712937422532366e-05, "loss": 0.4725, "step": 17066 }, { "epoch": 1.1454313613637126, "grad_norm": 1.0284857749938965, "learning_rate": 4.070225845835354e-05, "loss": 0.5641, "step": 17068 }, { "epoch": 1.1455655850474815, "grad_norm": 0.8678731322288513, "learning_rate": 4.069157993347474e-05, "loss": 0.4719, "step": 17070 }, { "epoch": 1.1456998087312507, "grad_norm": 0.972980260848999, "learning_rate": 4.068090184840047e-05, "loss": 0.4771, "step": 17072 }, { "epoch": 1.1458340324150196, "grad_norm": 0.9753967523574829, "learning_rate": 4.0670224203635274e-05, "loss": 0.5157, "step": 17074 }, { "epoch": 1.1459682560987887, "grad_norm": 0.9964821934700012, "learning_rate": 4.065954699968363e-05, "loss": 0.4685, "step": 17076 }, { "epoch": 1.1461024797825576, "grad_norm": 1.0178292989730835, "learning_rate": 4.064887023705003e-05, "loss": 0.4768, "step": 17078 }, { "epoch": 1.1462367034663266, "grad_norm": 1.0333075523376465, "learning_rate": 4.0638193916238914e-05, "loss": 0.5122, "step": 17080 }, { "epoch": 1.1463709271500957, "grad_norm": 0.8620358109474182, "learning_rate": 4.0627518037754745e-05, "loss": 0.4573, "step": 17082 }, { "epoch": 1.1465051508338646, "grad_norm": 0.8973593711853027, "learning_rate": 4.0616842602101894e-05, "loss": 0.4458, "step": 17084 }, { "epoch": 1.1466393745176338, "grad_norm": 1.081467628479004, "learning_rate": 4.060616760978479e-05, "loss": 0.4875, "step": 17086 }, { "epoch": 1.1467735982014027, "grad_norm": 1.004448413848877, "learning_rate": 4.05954930613078e-05, "loss": 0.5093, "step": 17088 }, { "epoch": 1.1469078218851716, "grad_norm": 0.8252903819084167, "learning_rate": 4.058481895717527e-05, "loss": 0.5308, "step": 17090 }, { "epoch": 1.1470420455689405, "grad_norm": 0.9397380352020264, "learning_rate": 4.057414529789151e-05, "loss": 0.4919, "step": 17092 }, { "epoch": 1.1471762692527097, "grad_norm": 1.4505244493484497, "learning_rate": 4.056347208396088e-05, "loss": 0.53, "step": 17094 }, { "epoch": 1.1473104929364786, "grad_norm": 0.9114947319030762, "learning_rate": 4.05527993158876e-05, "loss": 0.4595, "step": 17096 }, { "epoch": 1.1474447166202477, "grad_norm": 1.0862253904342651, "learning_rate": 4.0542126994176005e-05, "loss": 0.4962, "step": 17098 }, { "epoch": 1.1475789403040166, "grad_norm": 1.0601911544799805, "learning_rate": 4.053145511933032e-05, "loss": 0.526, "step": 17100 }, { "epoch": 1.1477131639877856, "grad_norm": 0.9861023426055908, "learning_rate": 4.052078369185473e-05, "loss": 0.5379, "step": 17102 }, { "epoch": 1.1478473876715547, "grad_norm": 0.9312143325805664, "learning_rate": 4.051011271225349e-05, "loss": 0.4921, "step": 17104 }, { "epoch": 1.1479816113553236, "grad_norm": 1.0780442953109741, "learning_rate": 4.049944218103076e-05, "loss": 0.5544, "step": 17106 }, { "epoch": 1.1481158350390928, "grad_norm": 1.0807689428329468, "learning_rate": 4.048877209869071e-05, "loss": 0.5142, "step": 17108 }, { "epoch": 1.1482500587228617, "grad_norm": 1.092450737953186, "learning_rate": 4.047810246573746e-05, "loss": 0.5646, "step": 17110 }, { "epoch": 1.1483842824066306, "grad_norm": 1.3242136240005493, "learning_rate": 4.046743328267516e-05, "loss": 0.5701, "step": 17112 }, { "epoch": 1.1485185060903997, "grad_norm": 1.155165195465088, "learning_rate": 4.045676455000788e-05, "loss": 0.5041, "step": 17114 }, { "epoch": 1.1486527297741687, "grad_norm": 0.9011285901069641, "learning_rate": 4.0446096268239716e-05, "loss": 0.4813, "step": 17116 }, { "epoch": 1.1487869534579376, "grad_norm": 0.9024450778961182, "learning_rate": 4.043542843787472e-05, "loss": 0.5612, "step": 17118 }, { "epoch": 1.1489211771417067, "grad_norm": 0.931702196598053, "learning_rate": 4.0424761059416915e-05, "loss": 0.5496, "step": 17120 }, { "epoch": 1.1490554008254756, "grad_norm": 0.9446266293525696, "learning_rate": 4.0414094133370325e-05, "loss": 0.5196, "step": 17122 }, { "epoch": 1.1491896245092446, "grad_norm": 0.907971203327179, "learning_rate": 4.040342766023894e-05, "loss": 0.46, "step": 17124 }, { "epoch": 1.1493238481930137, "grad_norm": 1.0521533489227295, "learning_rate": 4.0392761640526705e-05, "loss": 0.4844, "step": 17126 }, { "epoch": 1.1494580718767826, "grad_norm": 0.9352225661277771, "learning_rate": 4.038209607473763e-05, "loss": 0.5457, "step": 17128 }, { "epoch": 1.1495922955605518, "grad_norm": 1.092270016670227, "learning_rate": 4.037143096337556e-05, "loss": 0.5504, "step": 17130 }, { "epoch": 1.1497265192443207, "grad_norm": 0.9602010250091553, "learning_rate": 4.0360766306944474e-05, "loss": 0.5331, "step": 17132 }, { "epoch": 1.1498607429280896, "grad_norm": 0.9188903570175171, "learning_rate": 4.035010210594821e-05, "loss": 0.4765, "step": 17134 }, { "epoch": 1.1499949666118587, "grad_norm": 1.1557133197784424, "learning_rate": 4.033943836089066e-05, "loss": 0.4902, "step": 17136 }, { "epoch": 1.1501291902956277, "grad_norm": 0.8777872323989868, "learning_rate": 4.032877507227564e-05, "loss": 0.4979, "step": 17138 }, { "epoch": 1.1502634139793966, "grad_norm": 0.9906134009361267, "learning_rate": 4.031811224060699e-05, "loss": 0.5023, "step": 17140 }, { "epoch": 1.1503976376631657, "grad_norm": 1.0566591024398804, "learning_rate": 4.030744986638848e-05, "loss": 0.456, "step": 17142 }, { "epoch": 1.1505318613469346, "grad_norm": 1.0089061260223389, "learning_rate": 4.029678795012392e-05, "loss": 0.4734, "step": 17144 }, { "epoch": 1.1506660850307036, "grad_norm": 0.9829335808753967, "learning_rate": 4.028612649231704e-05, "loss": 0.5027, "step": 17146 }, { "epoch": 1.1508003087144727, "grad_norm": 1.0287353992462158, "learning_rate": 4.027546549347159e-05, "loss": 0.5008, "step": 17148 }, { "epoch": 1.1509345323982416, "grad_norm": 1.08994722366333, "learning_rate": 4.026480495409127e-05, "loss": 0.5361, "step": 17150 }, { "epoch": 1.1510687560820108, "grad_norm": 1.085174322128296, "learning_rate": 4.025414487467977e-05, "loss": 0.5735, "step": 17152 }, { "epoch": 1.1512029797657797, "grad_norm": 1.1141557693481445, "learning_rate": 4.024348525574075e-05, "loss": 0.521, "step": 17154 }, { "epoch": 1.1513372034495486, "grad_norm": 1.0411978960037231, "learning_rate": 4.0232826097777895e-05, "loss": 0.4969, "step": 17156 }, { "epoch": 1.1514714271333177, "grad_norm": 1.0179568529129028, "learning_rate": 4.0222167401294766e-05, "loss": 0.4433, "step": 17158 }, { "epoch": 1.1516056508170867, "grad_norm": 0.9610757231712341, "learning_rate": 4.021150916679502e-05, "loss": 0.4584, "step": 17160 }, { "epoch": 1.1517398745008558, "grad_norm": 1.2858824729919434, "learning_rate": 4.020085139478221e-05, "loss": 0.5452, "step": 17162 }, { "epoch": 1.1518740981846247, "grad_norm": 1.0095821619033813, "learning_rate": 4.0190194085759915e-05, "loss": 0.476, "step": 17164 }, { "epoch": 1.1520083218683936, "grad_norm": 0.9941380023956299, "learning_rate": 4.017953724023165e-05, "loss": 0.475, "step": 17166 }, { "epoch": 1.1521425455521626, "grad_norm": 1.1222082376480103, "learning_rate": 4.016888085870094e-05, "loss": 0.4566, "step": 17168 }, { "epoch": 1.1522767692359317, "grad_norm": 1.0464144945144653, "learning_rate": 4.0158224941671285e-05, "loss": 0.5358, "step": 17170 }, { "epoch": 1.1524109929197006, "grad_norm": 0.8921653032302856, "learning_rate": 4.0147569489646135e-05, "loss": 0.5072, "step": 17172 }, { "epoch": 1.1525452166034698, "grad_norm": 0.8786313533782959, "learning_rate": 4.013691450312897e-05, "loss": 0.6054, "step": 17174 }, { "epoch": 1.1526794402872387, "grad_norm": 0.9894219636917114, "learning_rate": 4.012625998262318e-05, "loss": 0.4763, "step": 17176 }, { "epoch": 1.1528136639710076, "grad_norm": 1.0877069234848022, "learning_rate": 4.0115605928632205e-05, "loss": 0.4868, "step": 17178 }, { "epoch": 1.1529478876547767, "grad_norm": 1.1121222972869873, "learning_rate": 4.0104952341659405e-05, "loss": 0.5178, "step": 17180 }, { "epoch": 1.1530821113385457, "grad_norm": 1.0284626483917236, "learning_rate": 4.0094299222208154e-05, "loss": 0.5683, "step": 17182 }, { "epoch": 1.1532163350223148, "grad_norm": 0.9250473976135254, "learning_rate": 4.008364657078178e-05, "loss": 0.4824, "step": 17184 }, { "epoch": 1.1533505587060837, "grad_norm": 1.0787099599838257, "learning_rate": 4.007299438788362e-05, "loss": 0.524, "step": 17186 }, { "epoch": 1.1534847823898526, "grad_norm": 0.9786050319671631, "learning_rate": 4.0062342674016926e-05, "loss": 0.5332, "step": 17188 }, { "epoch": 1.1536190060736218, "grad_norm": 0.860023558139801, "learning_rate": 4.005169142968503e-05, "loss": 0.5021, "step": 17190 }, { "epoch": 1.1537532297573907, "grad_norm": 1.308913230895996, "learning_rate": 4.004104065539112e-05, "loss": 0.5735, "step": 17192 }, { "epoch": 1.1538874534411596, "grad_norm": 1.022002935409546, "learning_rate": 4.003039035163848e-05, "loss": 0.5111, "step": 17194 }, { "epoch": 1.1540216771249288, "grad_norm": 1.0221171379089355, "learning_rate": 4.0019740518930285e-05, "loss": 0.5244, "step": 17196 }, { "epoch": 1.1541559008086977, "grad_norm": 1.025030493736267, "learning_rate": 4.000909115776973e-05, "loss": 0.5237, "step": 17198 }, { "epoch": 1.1542901244924666, "grad_norm": 1.0764104127883911, "learning_rate": 3.999844226865996e-05, "loss": 0.5109, "step": 17200 }, { "epoch": 1.1544243481762357, "grad_norm": 1.0224589109420776, "learning_rate": 3.998779385210414e-05, "loss": 0.4532, "step": 17202 }, { "epoch": 1.1545585718600047, "grad_norm": 1.0146963596343994, "learning_rate": 3.9977145908605355e-05, "loss": 0.4632, "step": 17204 }, { "epoch": 1.1546927955437738, "grad_norm": 1.0280194282531738, "learning_rate": 3.9966498438666734e-05, "loss": 0.5198, "step": 17206 }, { "epoch": 1.1548270192275427, "grad_norm": 1.2473407983779907, "learning_rate": 3.995585144279132e-05, "loss": 0.6126, "step": 17208 }, { "epoch": 1.1549612429113116, "grad_norm": 1.049044132232666, "learning_rate": 3.994520492148218e-05, "loss": 0.5154, "step": 17210 }, { "epoch": 1.1550954665950808, "grad_norm": 0.9378153085708618, "learning_rate": 3.9934558875242337e-05, "loss": 0.528, "step": 17212 }, { "epoch": 1.1552296902788497, "grad_norm": 1.051458716392517, "learning_rate": 3.9923913304574794e-05, "loss": 0.5225, "step": 17214 }, { "epoch": 1.1553639139626186, "grad_norm": 0.990528404712677, "learning_rate": 3.991326820998251e-05, "loss": 0.5219, "step": 17216 }, { "epoch": 1.1554981376463878, "grad_norm": 1.0346277952194214, "learning_rate": 3.9902623591968514e-05, "loss": 0.4648, "step": 17218 }, { "epoch": 1.1556323613301567, "grad_norm": 0.9941197633743286, "learning_rate": 3.989197945103565e-05, "loss": 0.5467, "step": 17220 }, { "epoch": 1.1557665850139256, "grad_norm": 1.0439867973327637, "learning_rate": 3.9881335787686915e-05, "loss": 0.5672, "step": 17222 }, { "epoch": 1.1559008086976947, "grad_norm": 0.9574592709541321, "learning_rate": 3.987069260242515e-05, "loss": 0.506, "step": 17224 }, { "epoch": 1.1560350323814637, "grad_norm": 0.9048669934272766, "learning_rate": 3.986004989575325e-05, "loss": 0.4766, "step": 17226 }, { "epoch": 1.1561692560652328, "grad_norm": 1.0663636922836304, "learning_rate": 3.9849407668174044e-05, "loss": 0.541, "step": 17228 }, { "epoch": 1.1563034797490017, "grad_norm": 1.2869949340820312, "learning_rate": 3.9838765920190376e-05, "loss": 0.4894, "step": 17230 }, { "epoch": 1.1564377034327706, "grad_norm": 0.9784888029098511, "learning_rate": 3.9828124652305024e-05, "loss": 0.4924, "step": 17232 }, { "epoch": 1.1565719271165398, "grad_norm": 1.0746228694915771, "learning_rate": 3.98174838650208e-05, "loss": 0.5297, "step": 17234 }, { "epoch": 1.1567061508003087, "grad_norm": 0.8882965445518494, "learning_rate": 3.9806843558840427e-05, "loss": 0.4856, "step": 17236 }, { "epoch": 1.1568403744840778, "grad_norm": 1.1055036783218384, "learning_rate": 3.979620373426665e-05, "loss": 0.513, "step": 17238 }, { "epoch": 1.1569745981678468, "grad_norm": 1.1486135721206665, "learning_rate": 3.978556439180219e-05, "loss": 0.5003, "step": 17240 }, { "epoch": 1.1571088218516157, "grad_norm": 1.0955274105072021, "learning_rate": 3.9774925531949714e-05, "loss": 0.5072, "step": 17242 }, { "epoch": 1.1572430455353846, "grad_norm": 1.2183367013931274, "learning_rate": 3.9764287155211916e-05, "loss": 0.5716, "step": 17244 }, { "epoch": 1.1573772692191537, "grad_norm": 1.0250377655029297, "learning_rate": 3.975364926209141e-05, "loss": 0.5355, "step": 17246 }, { "epoch": 1.1575114929029227, "grad_norm": 1.3676515817642212, "learning_rate": 3.974301185309083e-05, "loss": 0.5002, "step": 17248 }, { "epoch": 1.1576457165866918, "grad_norm": 1.0352555513381958, "learning_rate": 3.973237492871276e-05, "loss": 0.5515, "step": 17250 }, { "epoch": 1.1577799402704607, "grad_norm": 1.0024343729019165, "learning_rate": 3.972173848945981e-05, "loss": 0.5135, "step": 17252 }, { "epoch": 1.1579141639542296, "grad_norm": 0.9455167055130005, "learning_rate": 3.971110253583448e-05, "loss": 0.54, "step": 17254 }, { "epoch": 1.1580483876379988, "grad_norm": 1.19954252243042, "learning_rate": 3.970046706833935e-05, "loss": 0.5259, "step": 17256 }, { "epoch": 1.1581826113217677, "grad_norm": 0.9135200381278992, "learning_rate": 3.968983208747686e-05, "loss": 0.4701, "step": 17258 }, { "epoch": 1.1583168350055368, "grad_norm": 1.0010595321655273, "learning_rate": 3.967919759374956e-05, "loss": 0.5032, "step": 17260 }, { "epoch": 1.1584510586893058, "grad_norm": 0.9441422820091248, "learning_rate": 3.9668563587659865e-05, "loss": 0.4644, "step": 17262 }, { "epoch": 1.1585852823730747, "grad_norm": 1.0468347072601318, "learning_rate": 3.9657930069710236e-05, "loss": 0.5695, "step": 17264 }, { "epoch": 1.1587195060568438, "grad_norm": 1.0343860387802124, "learning_rate": 3.9647297040403066e-05, "loss": 0.5416, "step": 17266 }, { "epoch": 1.1588537297406127, "grad_norm": 0.8644536137580872, "learning_rate": 3.963666450024075e-05, "loss": 0.4487, "step": 17268 }, { "epoch": 1.1589879534243817, "grad_norm": 0.938911497592926, "learning_rate": 3.962603244972566e-05, "loss": 0.4666, "step": 17270 }, { "epoch": 1.1591221771081508, "grad_norm": 0.8899960517883301, "learning_rate": 3.9615400889360146e-05, "loss": 0.4681, "step": 17272 }, { "epoch": 1.1592564007919197, "grad_norm": 1.0408364534378052, "learning_rate": 3.9604769819646515e-05, "loss": 0.4553, "step": 17274 }, { "epoch": 1.1593906244756886, "grad_norm": 1.0575523376464844, "learning_rate": 3.959413924108707e-05, "loss": 0.5335, "step": 17276 }, { "epoch": 1.1595248481594578, "grad_norm": 1.0249072313308716, "learning_rate": 3.958350915418407e-05, "loss": 0.512, "step": 17278 }, { "epoch": 1.1596590718432267, "grad_norm": 0.9576659202575684, "learning_rate": 3.95728795594398e-05, "loss": 0.5155, "step": 17280 }, { "epoch": 1.1597932955269958, "grad_norm": 1.0557490587234497, "learning_rate": 3.9562250457356445e-05, "loss": 0.5466, "step": 17282 }, { "epoch": 1.1599275192107648, "grad_norm": 0.9900712370872498, "learning_rate": 3.955162184843625e-05, "loss": 0.5618, "step": 17284 }, { "epoch": 1.1600617428945337, "grad_norm": 0.9188636541366577, "learning_rate": 3.9540993733181366e-05, "loss": 0.4764, "step": 17286 }, { "epoch": 1.1601959665783028, "grad_norm": 0.980354905128479, "learning_rate": 3.953036611209397e-05, "loss": 0.4352, "step": 17288 }, { "epoch": 1.1603301902620717, "grad_norm": 0.9077704548835754, "learning_rate": 3.951973898567617e-05, "loss": 0.467, "step": 17290 }, { "epoch": 1.1604644139458407, "grad_norm": 1.2130227088928223, "learning_rate": 3.950911235443012e-05, "loss": 0.5626, "step": 17292 }, { "epoch": 1.1605986376296098, "grad_norm": 1.041313648223877, "learning_rate": 3.949848621885786e-05, "loss": 0.4997, "step": 17294 }, { "epoch": 1.1607328613133787, "grad_norm": 1.0547500848770142, "learning_rate": 3.948786057946149e-05, "loss": 0.5183, "step": 17296 }, { "epoch": 1.1608670849971476, "grad_norm": 1.0261955261230469, "learning_rate": 3.9477235436743026e-05, "loss": 0.5048, "step": 17298 }, { "epoch": 1.1610013086809168, "grad_norm": 0.8961619734764099, "learning_rate": 3.9466610791204504e-05, "loss": 0.4825, "step": 17300 }, { "epoch": 1.1611355323646857, "grad_norm": 0.8076702356338501, "learning_rate": 3.94559866433479e-05, "loss": 0.4813, "step": 17302 }, { "epoch": 1.1612697560484548, "grad_norm": 0.9887005686759949, "learning_rate": 3.9445362993675204e-05, "loss": 0.525, "step": 17304 }, { "epoch": 1.1614039797322238, "grad_norm": 1.0277241468429565, "learning_rate": 3.9434739842688356e-05, "loss": 0.5315, "step": 17306 }, { "epoch": 1.1615382034159927, "grad_norm": 1.036078929901123, "learning_rate": 3.942411719088926e-05, "loss": 0.5651, "step": 17308 }, { "epoch": 1.1616724270997618, "grad_norm": 1.1413941383361816, "learning_rate": 3.941349503877984e-05, "loss": 0.5602, "step": 17310 }, { "epoch": 1.1618066507835307, "grad_norm": 1.0826722383499146, "learning_rate": 3.9402873386861944e-05, "loss": 0.5312, "step": 17312 }, { "epoch": 1.1619408744672999, "grad_norm": 1.1246724128723145, "learning_rate": 3.939225223563747e-05, "loss": 0.5123, "step": 17314 }, { "epoch": 1.1620750981510688, "grad_norm": 0.9715235829353333, "learning_rate": 3.938163158560819e-05, "loss": 0.4429, "step": 17316 }, { "epoch": 1.1622093218348377, "grad_norm": 0.9817221164703369, "learning_rate": 3.9371011437275965e-05, "loss": 0.4916, "step": 17318 }, { "epoch": 1.1623435455186066, "grad_norm": 0.9534182548522949, "learning_rate": 3.936039179114253e-05, "loss": 0.5323, "step": 17320 }, { "epoch": 1.1624777692023758, "grad_norm": 1.2011499404907227, "learning_rate": 3.9349772647709666e-05, "loss": 0.4955, "step": 17322 }, { "epoch": 1.1626119928861447, "grad_norm": 0.9890552163124084, "learning_rate": 3.9339154007479096e-05, "loss": 0.5058, "step": 17324 }, { "epoch": 1.1627462165699138, "grad_norm": 0.9952134490013123, "learning_rate": 3.9328535870952545e-05, "loss": 0.5537, "step": 17326 }, { "epoch": 1.1628804402536828, "grad_norm": 1.0444145202636719, "learning_rate": 3.931791823863168e-05, "loss": 0.5134, "step": 17328 }, { "epoch": 1.1630146639374517, "grad_norm": 1.0761719942092896, "learning_rate": 3.930730111101819e-05, "loss": 0.5787, "step": 17330 }, { "epoch": 1.1631488876212208, "grad_norm": 1.03164803981781, "learning_rate": 3.9296684488613675e-05, "loss": 0.5072, "step": 17332 }, { "epoch": 1.1632831113049897, "grad_norm": 0.9586412906646729, "learning_rate": 3.928606837191979e-05, "loss": 0.4944, "step": 17334 }, { "epoch": 1.1634173349887589, "grad_norm": 0.9396413564682007, "learning_rate": 3.9275452761438095e-05, "loss": 0.5199, "step": 17336 }, { "epoch": 1.1635515586725278, "grad_norm": 1.0381124019622803, "learning_rate": 3.926483765767017e-05, "loss": 0.5051, "step": 17338 }, { "epoch": 1.1636857823562967, "grad_norm": 1.0771284103393555, "learning_rate": 3.925422306111754e-05, "loss": 0.5278, "step": 17340 }, { "epoch": 1.1638200060400659, "grad_norm": 0.9608303904533386, "learning_rate": 3.924360897228177e-05, "loss": 0.5329, "step": 17342 }, { "epoch": 1.1639542297238348, "grad_norm": 1.1120320558547974, "learning_rate": 3.9232995391664296e-05, "loss": 0.484, "step": 17344 }, { "epoch": 1.1640884534076037, "grad_norm": 0.8794587254524231, "learning_rate": 3.922238231976665e-05, "loss": 0.4966, "step": 17346 }, { "epoch": 1.1642226770913728, "grad_norm": 1.0041494369506836, "learning_rate": 3.921176975709021e-05, "loss": 0.4705, "step": 17348 }, { "epoch": 1.1643569007751418, "grad_norm": 0.9785823225975037, "learning_rate": 3.920115770413646e-05, "loss": 0.6067, "step": 17350 }, { "epoch": 1.1644911244589107, "grad_norm": 1.7183648347854614, "learning_rate": 3.919054616140676e-05, "loss": 0.5145, "step": 17352 }, { "epoch": 1.1646253481426798, "grad_norm": 1.084672212600708, "learning_rate": 3.917993512940251e-05, "loss": 0.493, "step": 17354 }, { "epoch": 1.1647595718264487, "grad_norm": 0.9071468710899353, "learning_rate": 3.916932460862504e-05, "loss": 0.4254, "step": 17356 }, { "epoch": 1.1648937955102179, "grad_norm": 0.8470814824104309, "learning_rate": 3.915871459957569e-05, "loss": 0.5063, "step": 17358 }, { "epoch": 1.1650280191939868, "grad_norm": 0.9796807765960693, "learning_rate": 3.914810510275575e-05, "loss": 0.492, "step": 17360 }, { "epoch": 1.1651622428777557, "grad_norm": 0.9707605838775635, "learning_rate": 3.913749611866651e-05, "loss": 0.519, "step": 17362 }, { "epoch": 1.1652964665615249, "grad_norm": 0.9335168600082397, "learning_rate": 3.912688764780921e-05, "loss": 0.4455, "step": 17364 }, { "epoch": 1.1654306902452938, "grad_norm": 0.9438152313232422, "learning_rate": 3.91162796906851e-05, "loss": 0.4892, "step": 17366 }, { "epoch": 1.1655649139290627, "grad_norm": 1.0461525917053223, "learning_rate": 3.910567224779535e-05, "loss": 0.5111, "step": 17368 }, { "epoch": 1.1656991376128318, "grad_norm": 0.9843353629112244, "learning_rate": 3.90950653196412e-05, "loss": 0.4903, "step": 17370 }, { "epoch": 1.1658333612966008, "grad_norm": 1.0085928440093994, "learning_rate": 3.908445890672373e-05, "loss": 0.5351, "step": 17372 }, { "epoch": 1.1659675849803697, "grad_norm": 0.9849995970726013, "learning_rate": 3.907385300954414e-05, "loss": 0.4564, "step": 17374 }, { "epoch": 1.1661018086641388, "grad_norm": 1.2877445220947266, "learning_rate": 3.906324762860352e-05, "loss": 0.5493, "step": 17376 }, { "epoch": 1.1662360323479077, "grad_norm": 0.9882004261016846, "learning_rate": 3.9052642764402906e-05, "loss": 0.5426, "step": 17378 }, { "epoch": 1.1663702560316769, "grad_norm": 1.2059866189956665, "learning_rate": 3.904203841744343e-05, "loss": 0.5313, "step": 17380 }, { "epoch": 1.1665044797154458, "grad_norm": 1.0301743745803833, "learning_rate": 3.903143458822606e-05, "loss": 0.4965, "step": 17382 }, { "epoch": 1.1666387033992147, "grad_norm": 0.9556608200073242, "learning_rate": 3.9020831277251863e-05, "loss": 0.5347, "step": 17384 }, { "epoch": 1.1667729270829839, "grad_norm": 0.8305768370628357, "learning_rate": 3.9010228485021784e-05, "loss": 0.4965, "step": 17386 }, { "epoch": 1.1669071507667528, "grad_norm": 1.1022326946258545, "learning_rate": 3.8999626212036805e-05, "loss": 0.4865, "step": 17388 }, { "epoch": 1.167041374450522, "grad_norm": 0.9911981225013733, "learning_rate": 3.898902445879784e-05, "loss": 0.4622, "step": 17390 }, { "epoch": 1.1671755981342908, "grad_norm": 1.1780292987823486, "learning_rate": 3.897842322580584e-05, "loss": 0.5141, "step": 17392 }, { "epoch": 1.1673098218180598, "grad_norm": 1.2555780410766602, "learning_rate": 3.8967822513561656e-05, "loss": 0.5295, "step": 17394 }, { "epoch": 1.1674440455018287, "grad_norm": 1.005345106124878, "learning_rate": 3.895722232256618e-05, "loss": 0.5085, "step": 17396 }, { "epoch": 1.1675782691855978, "grad_norm": 0.9374826550483704, "learning_rate": 3.8946622653320216e-05, "loss": 0.4729, "step": 17398 }, { "epoch": 1.1677124928693667, "grad_norm": 1.1584421396255493, "learning_rate": 3.89360235063246e-05, "loss": 0.5027, "step": 17400 }, { "epoch": 1.1678467165531359, "grad_norm": 1.0123322010040283, "learning_rate": 3.8925424882080105e-05, "loss": 0.487, "step": 17402 }, { "epoch": 1.1679809402369048, "grad_norm": 1.1311192512512207, "learning_rate": 3.891482678108754e-05, "loss": 0.5161, "step": 17404 }, { "epoch": 1.1681151639206737, "grad_norm": 1.065680980682373, "learning_rate": 3.890422920384758e-05, "loss": 0.553, "step": 17406 }, { "epoch": 1.1682493876044429, "grad_norm": 0.8839694261550903, "learning_rate": 3.8893632150861004e-05, "loss": 0.4933, "step": 17408 }, { "epoch": 1.1683836112882118, "grad_norm": 0.9866513013839722, "learning_rate": 3.888303562262843e-05, "loss": 0.5221, "step": 17410 }, { "epoch": 1.168517834971981, "grad_norm": 1.0440144538879395, "learning_rate": 3.887243961965059e-05, "loss": 0.499, "step": 17412 }, { "epoch": 1.1686520586557498, "grad_norm": 0.9592545628547668, "learning_rate": 3.8861844142428083e-05, "loss": 0.4699, "step": 17414 }, { "epoch": 1.1687862823395188, "grad_norm": 0.9670283198356628, "learning_rate": 3.8851249191461544e-05, "loss": 0.4649, "step": 17416 }, { "epoch": 1.168920506023288, "grad_norm": 1.100111961364746, "learning_rate": 3.884065476725156e-05, "loss": 0.4961, "step": 17418 }, { "epoch": 1.1690547297070568, "grad_norm": 1.0785095691680908, "learning_rate": 3.8830060870298696e-05, "loss": 0.5124, "step": 17420 }, { "epoch": 1.1691889533908257, "grad_norm": 1.0409162044525146, "learning_rate": 3.881946750110348e-05, "loss": 0.5206, "step": 17422 }, { "epoch": 1.1693231770745949, "grad_norm": 1.0038281679153442, "learning_rate": 3.880887466016645e-05, "loss": 0.5282, "step": 17424 }, { "epoch": 1.1694574007583638, "grad_norm": 0.9651196002960205, "learning_rate": 3.879828234798808e-05, "loss": 0.4753, "step": 17426 }, { "epoch": 1.1695916244421327, "grad_norm": 1.3088198900222778, "learning_rate": 3.8787690565068846e-05, "loss": 0.547, "step": 17428 }, { "epoch": 1.1697258481259019, "grad_norm": 1.2230758666992188, "learning_rate": 3.877709931190917e-05, "loss": 0.4999, "step": 17430 }, { "epoch": 1.1698600718096708, "grad_norm": 0.9876248836517334, "learning_rate": 3.876650858900951e-05, "loss": 0.5295, "step": 17432 }, { "epoch": 1.16999429549344, "grad_norm": 1.0522669553756714, "learning_rate": 3.87559183968702e-05, "loss": 0.5396, "step": 17434 }, { "epoch": 1.1701285191772088, "grad_norm": 1.0726021528244019, "learning_rate": 3.8745328735991664e-05, "loss": 0.4984, "step": 17436 }, { "epoch": 1.1702627428609778, "grad_norm": 1.0583510398864746, "learning_rate": 3.8734739606874184e-05, "loss": 0.4875, "step": 17438 }, { "epoch": 1.170396966544747, "grad_norm": 1.0662593841552734, "learning_rate": 3.872415101001813e-05, "loss": 0.5118, "step": 17440 }, { "epoch": 1.1705311902285158, "grad_norm": 0.8635338544845581, "learning_rate": 3.871356294592377e-05, "loss": 0.4636, "step": 17442 }, { "epoch": 1.1706654139122847, "grad_norm": 0.9625066518783569, "learning_rate": 3.870297541509134e-05, "loss": 0.4783, "step": 17444 }, { "epoch": 1.1707996375960539, "grad_norm": 1.0663460493087769, "learning_rate": 3.869238841802113e-05, "loss": 0.5439, "step": 17446 }, { "epoch": 1.1709338612798228, "grad_norm": 1.0197993516921997, "learning_rate": 3.868180195521332e-05, "loss": 0.4795, "step": 17448 }, { "epoch": 1.1710680849635917, "grad_norm": 0.9224337935447693, "learning_rate": 3.867121602716812e-05, "loss": 0.494, "step": 17450 }, { "epoch": 1.1712023086473609, "grad_norm": 1.0983052253723145, "learning_rate": 3.8660630634385674e-05, "loss": 0.5593, "step": 17452 }, { "epoch": 1.1713365323311298, "grad_norm": 0.97585529088974, "learning_rate": 3.865004577736615e-05, "loss": 0.453, "step": 17454 }, { "epoch": 1.171470756014899, "grad_norm": 0.9972829818725586, "learning_rate": 3.8639461456609626e-05, "loss": 0.5571, "step": 17456 }, { "epoch": 1.1716049796986678, "grad_norm": 1.0324832201004028, "learning_rate": 3.862887767261623e-05, "loss": 0.5233, "step": 17458 }, { "epoch": 1.1717392033824368, "grad_norm": 1.1102116107940674, "learning_rate": 3.861829442588598e-05, "loss": 0.5954, "step": 17460 }, { "epoch": 1.171873427066206, "grad_norm": 1.0274536609649658, "learning_rate": 3.860771171691895e-05, "loss": 0.5099, "step": 17462 }, { "epoch": 1.1720076507499748, "grad_norm": 0.940142035484314, "learning_rate": 3.8597129546215135e-05, "loss": 0.5394, "step": 17464 }, { "epoch": 1.172141874433744, "grad_norm": 0.8878188729286194, "learning_rate": 3.858654791427453e-05, "loss": 0.4924, "step": 17466 }, { "epoch": 1.1722760981175129, "grad_norm": 0.9084573984146118, "learning_rate": 3.8575966821597075e-05, "loss": 0.4914, "step": 17468 }, { "epoch": 1.1724103218012818, "grad_norm": 1.108896017074585, "learning_rate": 3.856538626868276e-05, "loss": 0.5614, "step": 17470 }, { "epoch": 1.1725445454850507, "grad_norm": 3.371222496032715, "learning_rate": 3.855480625603142e-05, "loss": 0.5331, "step": 17472 }, { "epoch": 1.1726787691688199, "grad_norm": 0.939922571182251, "learning_rate": 3.8544226784143e-05, "loss": 0.5053, "step": 17474 }, { "epoch": 1.1728129928525888, "grad_norm": 0.8527662754058838, "learning_rate": 3.853364785351733e-05, "loss": 0.4852, "step": 17476 }, { "epoch": 1.172947216536358, "grad_norm": 0.8310046792030334, "learning_rate": 3.852306946465426e-05, "loss": 0.4621, "step": 17478 }, { "epoch": 1.1730814402201268, "grad_norm": 0.9874976277351379, "learning_rate": 3.851249161805358e-05, "loss": 0.5087, "step": 17480 }, { "epoch": 1.1732156639038958, "grad_norm": 1.1040178537368774, "learning_rate": 3.8501914314215095e-05, "loss": 0.4731, "step": 17482 }, { "epoch": 1.173349887587665, "grad_norm": 1.460074543952942, "learning_rate": 3.849133755363853e-05, "loss": 0.4691, "step": 17484 }, { "epoch": 1.1734841112714338, "grad_norm": 0.9203589558601379, "learning_rate": 3.8480761336823654e-05, "loss": 0.4678, "step": 17486 }, { "epoch": 1.173618334955203, "grad_norm": 1.0377877950668335, "learning_rate": 3.8470185664270134e-05, "loss": 0.5609, "step": 17488 }, { "epoch": 1.1737525586389719, "grad_norm": 1.0666682720184326, "learning_rate": 3.845961053647768e-05, "loss": 0.491, "step": 17490 }, { "epoch": 1.1738867823227408, "grad_norm": 0.9724449515342712, "learning_rate": 3.844903595394592e-05, "loss": 0.5039, "step": 17492 }, { "epoch": 1.17402100600651, "grad_norm": 1.1078293323516846, "learning_rate": 3.843846191717453e-05, "loss": 0.5721, "step": 17494 }, { "epoch": 1.1741552296902789, "grad_norm": 1.0953867435455322, "learning_rate": 3.8427888426663044e-05, "loss": 0.5078, "step": 17496 }, { "epoch": 1.1742894533740478, "grad_norm": 0.9513339400291443, "learning_rate": 3.841731548291111e-05, "loss": 0.5298, "step": 17498 }, { "epoch": 1.174423677057817, "grad_norm": 1.0575677156448364, "learning_rate": 3.840674308641821e-05, "loss": 0.5594, "step": 17500 }, { "epoch": 1.1745579007415858, "grad_norm": 1.0335451364517212, "learning_rate": 3.839617123768392e-05, "loss": 0.4774, "step": 17502 }, { "epoch": 1.1746921244253548, "grad_norm": 0.9178796410560608, "learning_rate": 3.8385599937207713e-05, "loss": 0.5096, "step": 17504 }, { "epoch": 1.174826348109124, "grad_norm": 1.166209101676941, "learning_rate": 3.837502918548907e-05, "loss": 0.4891, "step": 17506 }, { "epoch": 1.1749605717928928, "grad_norm": 0.899621307849884, "learning_rate": 3.836445898302745e-05, "loss": 0.4807, "step": 17508 }, { "epoch": 1.175094795476662, "grad_norm": 1.127547025680542, "learning_rate": 3.835388933032226e-05, "loss": 0.5472, "step": 17510 }, { "epoch": 1.1752290191604309, "grad_norm": 0.9545964002609253, "learning_rate": 3.8343320227872894e-05, "loss": 0.4765, "step": 17512 }, { "epoch": 1.1753632428441998, "grad_norm": 1.27057945728302, "learning_rate": 3.833275167617872e-05, "loss": 0.5113, "step": 17514 }, { "epoch": 1.175497466527969, "grad_norm": 0.8805884718894958, "learning_rate": 3.83221836757391e-05, "loss": 0.4793, "step": 17516 }, { "epoch": 1.1756316902117379, "grad_norm": 1.0999140739440918, "learning_rate": 3.831161622705332e-05, "loss": 0.5474, "step": 17518 }, { "epoch": 1.1757659138955068, "grad_norm": 0.852262020111084, "learning_rate": 3.83010493306207e-05, "loss": 0.4854, "step": 17520 }, { "epoch": 1.175900137579276, "grad_norm": 0.9464651346206665, "learning_rate": 3.829048298694049e-05, "loss": 0.524, "step": 17522 }, { "epoch": 1.1760343612630448, "grad_norm": 0.8718568682670593, "learning_rate": 3.827991719651194e-05, "loss": 0.4508, "step": 17524 }, { "epoch": 1.1761685849468138, "grad_norm": 0.8984642028808594, "learning_rate": 3.826935195983425e-05, "loss": 0.49, "step": 17526 }, { "epoch": 1.176302808630583, "grad_norm": 0.9936639070510864, "learning_rate": 3.8258787277406615e-05, "loss": 0.5017, "step": 17528 }, { "epoch": 1.1764370323143518, "grad_norm": 1.0150092840194702, "learning_rate": 3.824822314972818e-05, "loss": 0.5155, "step": 17530 }, { "epoch": 1.176571255998121, "grad_norm": 0.9806212782859802, "learning_rate": 3.823765957729813e-05, "loss": 0.4745, "step": 17532 }, { "epoch": 1.1767054796818899, "grad_norm": 0.9721527695655823, "learning_rate": 3.8227096560615486e-05, "loss": 0.4945, "step": 17534 }, { "epoch": 1.1768397033656588, "grad_norm": 0.9306116104125977, "learning_rate": 3.821653410017941e-05, "loss": 0.531, "step": 17536 }, { "epoch": 1.176973927049428, "grad_norm": 1.0064274072647095, "learning_rate": 3.820597219648891e-05, "loss": 0.5429, "step": 17538 }, { "epoch": 1.1771081507331969, "grad_norm": 1.1770319938659668, "learning_rate": 3.819541085004304e-05, "loss": 0.515, "step": 17540 }, { "epoch": 1.177242374416966, "grad_norm": 0.8967007994651794, "learning_rate": 3.81848500613408e-05, "loss": 0.4666, "step": 17542 }, { "epoch": 1.177376598100735, "grad_norm": 1.0089564323425293, "learning_rate": 3.817428983088116e-05, "loss": 0.4986, "step": 17544 }, { "epoch": 1.1775108217845038, "grad_norm": 0.8636226654052734, "learning_rate": 3.8163730159163065e-05, "loss": 0.4436, "step": 17546 }, { "epoch": 1.1776450454682728, "grad_norm": 1.1310228109359741, "learning_rate": 3.8153171046685454e-05, "loss": 0.5249, "step": 17548 }, { "epoch": 1.177779269152042, "grad_norm": 0.976316511631012, "learning_rate": 3.8142612493947205e-05, "loss": 0.4639, "step": 17550 }, { "epoch": 1.1779134928358108, "grad_norm": 1.0009357929229736, "learning_rate": 3.8132054501447206e-05, "loss": 0.5191, "step": 17552 }, { "epoch": 1.17804771651958, "grad_norm": 0.9338767528533936, "learning_rate": 3.8121497069684285e-05, "loss": 0.5434, "step": 17554 }, { "epoch": 1.1781819402033489, "grad_norm": 0.9037238359451294, "learning_rate": 3.8110940199157284e-05, "loss": 0.4996, "step": 17556 }, { "epoch": 1.1783161638871178, "grad_norm": 1.027163028717041, "learning_rate": 3.8100383890364956e-05, "loss": 0.5148, "step": 17558 }, { "epoch": 1.178450387570887, "grad_norm": 1.1842451095581055, "learning_rate": 3.808982814380612e-05, "loss": 0.5143, "step": 17560 }, { "epoch": 1.1785846112546559, "grad_norm": 1.0291099548339844, "learning_rate": 3.8079272959979454e-05, "loss": 0.5255, "step": 17562 }, { "epoch": 1.178718834938425, "grad_norm": 0.9508962035179138, "learning_rate": 3.8068718339383726e-05, "loss": 0.4743, "step": 17564 }, { "epoch": 1.178853058622194, "grad_norm": 1.0064043998718262, "learning_rate": 3.805816428251757e-05, "loss": 0.496, "step": 17566 }, { "epoch": 1.1789872823059628, "grad_norm": 0.9309611320495605, "learning_rate": 3.8047610789879696e-05, "loss": 0.511, "step": 17568 }, { "epoch": 1.179121505989732, "grad_norm": 1.0128785371780396, "learning_rate": 3.8037057861968684e-05, "loss": 0.5054, "step": 17570 }, { "epoch": 1.179255729673501, "grad_norm": 1.043404221534729, "learning_rate": 3.8026505499283184e-05, "loss": 0.5251, "step": 17572 }, { "epoch": 1.1793899533572698, "grad_norm": 1.0299290418624878, "learning_rate": 3.801595370232174e-05, "loss": 0.5318, "step": 17574 }, { "epoch": 1.179524177041039, "grad_norm": 0.9969633221626282, "learning_rate": 3.800540247158293e-05, "loss": 0.4817, "step": 17576 }, { "epoch": 1.1796584007248079, "grad_norm": 1.067553997039795, "learning_rate": 3.7994851807565254e-05, "loss": 0.4643, "step": 17578 }, { "epoch": 1.1797926244085768, "grad_norm": 1.044662356376648, "learning_rate": 3.798430171076723e-05, "loss": 0.5075, "step": 17580 }, { "epoch": 1.179926848092346, "grad_norm": 1.7358472347259521, "learning_rate": 3.7973752181687335e-05, "loss": 0.5194, "step": 17582 }, { "epoch": 1.1800610717761149, "grad_norm": 1.1128405332565308, "learning_rate": 3.796320322082398e-05, "loss": 0.5602, "step": 17584 }, { "epoch": 1.180195295459884, "grad_norm": 1.165319800376892, "learning_rate": 3.7952654828675616e-05, "loss": 0.5301, "step": 17586 }, { "epoch": 1.180329519143653, "grad_norm": 1.0974600315093994, "learning_rate": 3.794210700574061e-05, "loss": 0.6221, "step": 17588 }, { "epoch": 1.1804637428274218, "grad_norm": 0.8745746612548828, "learning_rate": 3.7931559752517354e-05, "loss": 0.5553, "step": 17590 }, { "epoch": 1.180597966511191, "grad_norm": 1.0914452075958252, "learning_rate": 3.792101306950414e-05, "loss": 0.5834, "step": 17592 }, { "epoch": 1.18073219019496, "grad_norm": 0.9152393937110901, "learning_rate": 3.7910466957199336e-05, "loss": 0.4696, "step": 17594 }, { "epoch": 1.1808664138787288, "grad_norm": 1.0173288583755493, "learning_rate": 3.789992141610117e-05, "loss": 0.4779, "step": 17596 }, { "epoch": 1.181000637562498, "grad_norm": 1.014234185218811, "learning_rate": 3.788937644670794e-05, "loss": 0.5314, "step": 17598 }, { "epoch": 1.1811348612462669, "grad_norm": 1.069540023803711, "learning_rate": 3.7878832049517846e-05, "loss": 0.526, "step": 17600 }, { "epoch": 1.1812690849300358, "grad_norm": 1.0426692962646484, "learning_rate": 3.786828822502912e-05, "loss": 0.4956, "step": 17602 }, { "epoch": 1.181403308613805, "grad_norm": 1.066277265548706, "learning_rate": 3.78577449737399e-05, "loss": 0.5578, "step": 17604 }, { "epoch": 1.1815375322975739, "grad_norm": 0.8797937631607056, "learning_rate": 3.784720229614838e-05, "loss": 0.5221, "step": 17606 }, { "epoch": 1.181671755981343, "grad_norm": 1.0879156589508057, "learning_rate": 3.783666019275263e-05, "loss": 0.5703, "step": 17608 }, { "epoch": 1.181805979665112, "grad_norm": 0.9450921416282654, "learning_rate": 3.782611866405078e-05, "loss": 0.4624, "step": 17610 }, { "epoch": 1.1819402033488808, "grad_norm": 1.0539926290512085, "learning_rate": 3.7815577710540874e-05, "loss": 0.5211, "step": 17612 }, { "epoch": 1.18207442703265, "grad_norm": 0.9046851992607117, "learning_rate": 3.7805037332720975e-05, "loss": 0.5124, "step": 17614 }, { "epoch": 1.182208650716419, "grad_norm": 1.0149006843566895, "learning_rate": 3.7794497531089066e-05, "loss": 0.4943, "step": 17616 }, { "epoch": 1.182342874400188, "grad_norm": 0.942730188369751, "learning_rate": 3.778395830614317e-05, "loss": 0.4962, "step": 17618 }, { "epoch": 1.182477098083957, "grad_norm": 1.0554370880126953, "learning_rate": 3.7773419658381185e-05, "loss": 0.4686, "step": 17620 }, { "epoch": 1.1826113217677259, "grad_norm": 0.9436022043228149, "learning_rate": 3.7762881588301115e-05, "loss": 0.472, "step": 17622 }, { "epoch": 1.1827455454514948, "grad_norm": 0.913763165473938, "learning_rate": 3.7752344096400794e-05, "loss": 0.5218, "step": 17624 }, { "epoch": 1.182879769135264, "grad_norm": 1.0532349348068237, "learning_rate": 3.774180718317816e-05, "loss": 0.4568, "step": 17626 }, { "epoch": 1.1830139928190329, "grad_norm": 1.3988685607910156, "learning_rate": 3.7731270849131014e-05, "loss": 0.4813, "step": 17628 }, { "epoch": 1.183148216502802, "grad_norm": 0.9777106642723083, "learning_rate": 3.772073509475721e-05, "loss": 0.5098, "step": 17630 }, { "epoch": 1.183282440186571, "grad_norm": 1.157128930091858, "learning_rate": 3.7710199920554513e-05, "loss": 0.5584, "step": 17632 }, { "epoch": 1.1834166638703398, "grad_norm": 1.0951553583145142, "learning_rate": 3.769966532702071e-05, "loss": 0.4952, "step": 17634 }, { "epoch": 1.183550887554109, "grad_norm": 1.0618690252304077, "learning_rate": 3.7689131314653525e-05, "loss": 0.4437, "step": 17636 }, { "epoch": 1.183685111237878, "grad_norm": 0.9891539216041565, "learning_rate": 3.7678597883950696e-05, "loss": 0.4713, "step": 17638 }, { "epoch": 1.183819334921647, "grad_norm": 1.0724852085113525, "learning_rate": 3.7668065035409864e-05, "loss": 0.528, "step": 17640 }, { "epoch": 1.183953558605416, "grad_norm": 0.9540714621543884, "learning_rate": 3.765753276952873e-05, "loss": 0.5084, "step": 17642 }, { "epoch": 1.1840877822891849, "grad_norm": 1.026107907295227, "learning_rate": 3.764700108680489e-05, "loss": 0.6074, "step": 17644 }, { "epoch": 1.184222005972954, "grad_norm": 1.0301669836044312, "learning_rate": 3.763646998773596e-05, "loss": 0.4662, "step": 17646 }, { "epoch": 1.184356229656723, "grad_norm": 0.8550382852554321, "learning_rate": 3.762593947281952e-05, "loss": 0.415, "step": 17648 }, { "epoch": 1.1844904533404919, "grad_norm": 1.0265635251998901, "learning_rate": 3.7615409542553094e-05, "loss": 0.5387, "step": 17650 }, { "epoch": 1.184624677024261, "grad_norm": 0.789578914642334, "learning_rate": 3.760488019743422e-05, "loss": 0.459, "step": 17652 }, { "epoch": 1.18475890070803, "grad_norm": 2.789158344268799, "learning_rate": 3.759435143796037e-05, "loss": 0.5348, "step": 17654 }, { "epoch": 1.1848931243917988, "grad_norm": 1.1224685907363892, "learning_rate": 3.7583823264629045e-05, "loss": 0.5749, "step": 17656 }, { "epoch": 1.185027348075568, "grad_norm": 1.0600457191467285, "learning_rate": 3.7573295677937633e-05, "loss": 0.5241, "step": 17658 }, { "epoch": 1.185161571759337, "grad_norm": 0.9870001077651978, "learning_rate": 3.7562768678383576e-05, "loss": 0.4698, "step": 17660 }, { "epoch": 1.185295795443106, "grad_norm": 1.0906633138656616, "learning_rate": 3.755224226646423e-05, "loss": 0.4884, "step": 17662 }, { "epoch": 1.185430019126875, "grad_norm": 1.013982892036438, "learning_rate": 3.754171644267698e-05, "loss": 0.5416, "step": 17664 }, { "epoch": 1.1855642428106439, "grad_norm": 1.0187809467315674, "learning_rate": 3.7531191207519114e-05, "loss": 0.5265, "step": 17666 }, { "epoch": 1.185698466494413, "grad_norm": 1.1359308958053589, "learning_rate": 3.752066656148796e-05, "loss": 0.5757, "step": 17668 }, { "epoch": 1.185832690178182, "grad_norm": 0.9246326088905334, "learning_rate": 3.7510142505080746e-05, "loss": 0.5185, "step": 17670 }, { "epoch": 1.1859669138619509, "grad_norm": 1.0252164602279663, "learning_rate": 3.749961903879477e-05, "loss": 0.5504, "step": 17672 }, { "epoch": 1.18610113754572, "grad_norm": 1.0328184366226196, "learning_rate": 3.748909616312718e-05, "loss": 0.4842, "step": 17674 }, { "epoch": 1.186235361229489, "grad_norm": 0.8675329685211182, "learning_rate": 3.747857387857522e-05, "loss": 0.4546, "step": 17676 }, { "epoch": 1.1863695849132578, "grad_norm": 0.9986277222633362, "learning_rate": 3.7468052185636e-05, "loss": 0.5553, "step": 17678 }, { "epoch": 1.186503808597027, "grad_norm": 1.0309573411941528, "learning_rate": 3.74575310848067e-05, "loss": 0.5489, "step": 17680 }, { "epoch": 1.186638032280796, "grad_norm": 1.0841816663742065, "learning_rate": 3.744701057658436e-05, "loss": 0.4927, "step": 17682 }, { "epoch": 1.186772255964565, "grad_norm": 1.0273849964141846, "learning_rate": 3.743649066146612e-05, "loss": 0.5874, "step": 17684 }, { "epoch": 1.186906479648334, "grad_norm": 1.7728848457336426, "learning_rate": 3.742597133994896e-05, "loss": 0.5942, "step": 17686 }, { "epoch": 1.1870407033321029, "grad_norm": 0.9678778052330017, "learning_rate": 3.741545261252994e-05, "loss": 0.4759, "step": 17688 }, { "epoch": 1.187174927015872, "grad_norm": 0.9566680788993835, "learning_rate": 3.740493447970604e-05, "loss": 0.5046, "step": 17690 }, { "epoch": 1.187309150699641, "grad_norm": 1.0579038858413696, "learning_rate": 3.739441694197422e-05, "loss": 0.516, "step": 17692 }, { "epoch": 1.18744337438341, "grad_norm": 0.9175028800964355, "learning_rate": 3.7383899999831406e-05, "loss": 0.4448, "step": 17694 }, { "epoch": 1.187577598067179, "grad_norm": 0.9152856469154358, "learning_rate": 3.737338365377452e-05, "loss": 0.4691, "step": 17696 }, { "epoch": 1.187711821750948, "grad_norm": 1.0249311923980713, "learning_rate": 3.7362867904300415e-05, "loss": 0.5694, "step": 17698 }, { "epoch": 1.1878460454347168, "grad_norm": 1.0089964866638184, "learning_rate": 3.7352352751905964e-05, "loss": 0.4709, "step": 17700 }, { "epoch": 1.187980269118486, "grad_norm": 1.1326613426208496, "learning_rate": 3.7341838197087964e-05, "loss": 0.539, "step": 17702 }, { "epoch": 1.188114492802255, "grad_norm": 0.9482781887054443, "learning_rate": 3.7331324240343226e-05, "loss": 0.5831, "step": 17704 }, { "epoch": 1.188248716486024, "grad_norm": 0.9407302141189575, "learning_rate": 3.7320810882168495e-05, "loss": 0.5088, "step": 17706 }, { "epoch": 1.188382940169793, "grad_norm": 1.0773608684539795, "learning_rate": 3.7310298123060527e-05, "loss": 0.6095, "step": 17708 }, { "epoch": 1.1885171638535619, "grad_norm": 1.024487853050232, "learning_rate": 3.7299785963516e-05, "loss": 0.5277, "step": 17710 }, { "epoch": 1.188651387537331, "grad_norm": 0.9741911888122559, "learning_rate": 3.728927440403165e-05, "loss": 0.5577, "step": 17712 }, { "epoch": 1.1887856112211, "grad_norm": 0.9415025115013123, "learning_rate": 3.727876344510405e-05, "loss": 0.4607, "step": 17714 }, { "epoch": 1.188919834904869, "grad_norm": 1.1595509052276611, "learning_rate": 3.726825308722989e-05, "loss": 0.5347, "step": 17716 }, { "epoch": 1.189054058588638, "grad_norm": 1.1408499479293823, "learning_rate": 3.725774333090574e-05, "loss": 0.4462, "step": 17718 }, { "epoch": 1.189188282272407, "grad_norm": 1.05015230178833, "learning_rate": 3.724723417662813e-05, "loss": 0.5859, "step": 17720 }, { "epoch": 1.189322505956176, "grad_norm": 1.0466431379318237, "learning_rate": 3.723672562489366e-05, "loss": 0.4653, "step": 17722 }, { "epoch": 1.189456729639945, "grad_norm": 1.0187463760375977, "learning_rate": 3.7226217676198775e-05, "loss": 0.555, "step": 17724 }, { "epoch": 1.189590953323714, "grad_norm": 1.1777077913284302, "learning_rate": 3.721571033104001e-05, "loss": 0.5346, "step": 17726 }, { "epoch": 1.189725177007483, "grad_norm": 0.9326265454292297, "learning_rate": 3.720520358991378e-05, "loss": 0.44, "step": 17728 }, { "epoch": 1.189859400691252, "grad_norm": 1.1017358303070068, "learning_rate": 3.719469745331653e-05, "loss": 0.5862, "step": 17730 }, { "epoch": 1.1899936243750209, "grad_norm": 1.0286736488342285, "learning_rate": 3.718419192174464e-05, "loss": 0.4303, "step": 17732 }, { "epoch": 1.19012784805879, "grad_norm": 0.9467357993125916, "learning_rate": 3.717368699569448e-05, "loss": 0.5062, "step": 17734 }, { "epoch": 1.190262071742559, "grad_norm": 0.9172195196151733, "learning_rate": 3.716318267566238e-05, "loss": 0.4467, "step": 17736 }, { "epoch": 1.190396295426328, "grad_norm": 1.2418681383132935, "learning_rate": 3.715267896214467e-05, "loss": 0.5656, "step": 17738 }, { "epoch": 1.190530519110097, "grad_norm": 0.925208330154419, "learning_rate": 3.7142175855637606e-05, "loss": 0.4698, "step": 17740 }, { "epoch": 1.190664742793866, "grad_norm": 0.93855881690979, "learning_rate": 3.7131673356637464e-05, "loss": 0.4955, "step": 17742 }, { "epoch": 1.190798966477635, "grad_norm": 1.0195592641830444, "learning_rate": 3.7121171465640425e-05, "loss": 0.5211, "step": 17744 }, { "epoch": 1.190933190161404, "grad_norm": 0.9680092930793762, "learning_rate": 3.7110670183142735e-05, "loss": 0.5084, "step": 17746 }, { "epoch": 1.191067413845173, "grad_norm": 0.9840260744094849, "learning_rate": 3.7100169509640517e-05, "loss": 0.5218, "step": 17748 }, { "epoch": 1.191201637528942, "grad_norm": 1.0281223058700562, "learning_rate": 3.708966944562993e-05, "loss": 0.5306, "step": 17750 }, { "epoch": 1.191335861212711, "grad_norm": 0.9090124368667603, "learning_rate": 3.707916999160706e-05, "loss": 0.4871, "step": 17752 }, { "epoch": 1.1914700848964799, "grad_norm": 1.0749856233596802, "learning_rate": 3.706867114806803e-05, "loss": 0.4986, "step": 17754 }, { "epoch": 1.191604308580249, "grad_norm": 0.9622864127159119, "learning_rate": 3.705817291550884e-05, "loss": 0.5304, "step": 17756 }, { "epoch": 1.191738532264018, "grad_norm": 1.1090136766433716, "learning_rate": 3.7047675294425535e-05, "loss": 0.574, "step": 17758 }, { "epoch": 1.191872755947787, "grad_norm": 1.0562835931777954, "learning_rate": 3.70371782853141e-05, "loss": 0.4782, "step": 17760 }, { "epoch": 1.192006979631556, "grad_norm": 1.0050835609436035, "learning_rate": 3.702668188867051e-05, "loss": 0.5709, "step": 17762 }, { "epoch": 1.192141203315325, "grad_norm": 0.9055975675582886, "learning_rate": 3.701618610499068e-05, "loss": 0.4767, "step": 17764 }, { "epoch": 1.192275426999094, "grad_norm": 1.0100618600845337, "learning_rate": 3.7005690934770534e-05, "loss": 0.4838, "step": 17766 }, { "epoch": 1.192409650682863, "grad_norm": 1.1675429344177246, "learning_rate": 3.6995196378505926e-05, "loss": 0.5449, "step": 17768 }, { "epoch": 1.192543874366632, "grad_norm": 1.0675374269485474, "learning_rate": 3.6984702436692726e-05, "loss": 0.5021, "step": 17770 }, { "epoch": 1.192678098050401, "grad_norm": 1.0601750612258911, "learning_rate": 3.6974209109826726e-05, "loss": 0.5486, "step": 17772 }, { "epoch": 1.19281232173417, "grad_norm": 1.0038232803344727, "learning_rate": 3.696371639840376e-05, "loss": 0.4954, "step": 17774 }, { "epoch": 1.1929465454179389, "grad_norm": 1.0992913246154785, "learning_rate": 3.695322430291953e-05, "loss": 0.5561, "step": 17776 }, { "epoch": 1.193080769101708, "grad_norm": 0.9731093049049377, "learning_rate": 3.694273282386981e-05, "loss": 0.4909, "step": 17778 }, { "epoch": 1.193214992785477, "grad_norm": 1.0533447265625, "learning_rate": 3.693224196175029e-05, "loss": 0.5425, "step": 17780 }, { "epoch": 1.193349216469246, "grad_norm": 0.852382481098175, "learning_rate": 3.692175171705664e-05, "loss": 0.5027, "step": 17782 }, { "epoch": 1.193483440153015, "grad_norm": 1.2404472827911377, "learning_rate": 3.6911262090284506e-05, "loss": 0.5075, "step": 17784 }, { "epoch": 1.193617663836784, "grad_norm": 1.0792241096496582, "learning_rate": 3.6900773081929496e-05, "loss": 0.4673, "step": 17786 }, { "epoch": 1.193751887520553, "grad_norm": 0.8949421048164368, "learning_rate": 3.689028469248721e-05, "loss": 0.4906, "step": 17788 }, { "epoch": 1.193886111204322, "grad_norm": 1.0198348760604858, "learning_rate": 3.687979692245318e-05, "loss": 0.5639, "step": 17790 }, { "epoch": 1.1940203348880911, "grad_norm": 1.091529369354248, "learning_rate": 3.6869309772322955e-05, "loss": 0.5408, "step": 17792 }, { "epoch": 1.19415455857186, "grad_norm": 0.9910798072814941, "learning_rate": 3.685882324259201e-05, "loss": 0.5494, "step": 17794 }, { "epoch": 1.194288782255629, "grad_norm": 1.0622038841247559, "learning_rate": 3.684833733375584e-05, "loss": 0.4839, "step": 17796 }, { "epoch": 1.194423005939398, "grad_norm": 0.9485304355621338, "learning_rate": 3.683785204630986e-05, "loss": 0.5038, "step": 17798 }, { "epoch": 1.194557229623167, "grad_norm": 1.4749153852462769, "learning_rate": 3.68273673807495e-05, "loss": 0.5175, "step": 17800 }, { "epoch": 1.194691453306936, "grad_norm": 1.138503909111023, "learning_rate": 3.681688333757011e-05, "loss": 0.5002, "step": 17802 }, { "epoch": 1.194825676990705, "grad_norm": 0.9076254963874817, "learning_rate": 3.6806399917267074e-05, "loss": 0.5075, "step": 17804 }, { "epoch": 1.194959900674474, "grad_norm": 1.083914875984192, "learning_rate": 3.679591712033568e-05, "loss": 0.4778, "step": 17806 }, { "epoch": 1.195094124358243, "grad_norm": 1.0164463520050049, "learning_rate": 3.678543494727126e-05, "loss": 0.5803, "step": 17808 }, { "epoch": 1.195228348042012, "grad_norm": 0.9743123650550842, "learning_rate": 3.677495339856903e-05, "loss": 0.5415, "step": 17810 }, { "epoch": 1.195362571725781, "grad_norm": 1.047502875328064, "learning_rate": 3.676447247472428e-05, "loss": 0.5183, "step": 17812 }, { "epoch": 1.1954967954095501, "grad_norm": 1.0990514755249023, "learning_rate": 3.6753992176232136e-05, "loss": 0.4921, "step": 17814 }, { "epoch": 1.195631019093319, "grad_norm": 0.9854269623756409, "learning_rate": 3.6743512503587844e-05, "loss": 0.4958, "step": 17816 }, { "epoch": 1.195765242777088, "grad_norm": 1.0102423429489136, "learning_rate": 3.673303345728651e-05, "loss": 0.4884, "step": 17818 }, { "epoch": 1.195899466460857, "grad_norm": 1.0501770973205566, "learning_rate": 3.672255503782326e-05, "loss": 0.5247, "step": 17820 }, { "epoch": 1.196033690144626, "grad_norm": 0.9007217288017273, "learning_rate": 3.671207724569317e-05, "loss": 0.5295, "step": 17822 }, { "epoch": 1.196167913828395, "grad_norm": 0.9345261454582214, "learning_rate": 3.670160008139131e-05, "loss": 0.472, "step": 17824 }, { "epoch": 1.196302137512164, "grad_norm": 1.0468041896820068, "learning_rate": 3.669112354541269e-05, "loss": 0.4921, "step": 17826 }, { "epoch": 1.196436361195933, "grad_norm": 1.0646734237670898, "learning_rate": 3.668064763825231e-05, "loss": 0.5565, "step": 17828 }, { "epoch": 1.196570584879702, "grad_norm": 0.9364396929740906, "learning_rate": 3.6670172360405136e-05, "loss": 0.4596, "step": 17830 }, { "epoch": 1.196704808563471, "grad_norm": 0.9761340022087097, "learning_rate": 3.6659697712366116e-05, "loss": 0.523, "step": 17832 }, { "epoch": 1.19683903224724, "grad_norm": 0.9941263198852539, "learning_rate": 3.664922369463012e-05, "loss": 0.4505, "step": 17834 }, { "epoch": 1.1969732559310091, "grad_norm": 0.9420185089111328, "learning_rate": 3.663875030769209e-05, "loss": 0.4849, "step": 17836 }, { "epoch": 1.197107479614778, "grad_norm": 1.0919946432113647, "learning_rate": 3.66282775520468e-05, "loss": 0.4925, "step": 17838 }, { "epoch": 1.197241703298547, "grad_norm": 1.6442017555236816, "learning_rate": 3.6617805428189135e-05, "loss": 0.5326, "step": 17840 }, { "epoch": 1.197375926982316, "grad_norm": 0.9553593993186951, "learning_rate": 3.660733393661381e-05, "loss": 0.5323, "step": 17842 }, { "epoch": 1.197510150666085, "grad_norm": 0.9864370226860046, "learning_rate": 3.6596863077815644e-05, "loss": 0.5602, "step": 17844 }, { "epoch": 1.197644374349854, "grad_norm": 1.0713117122650146, "learning_rate": 3.658639285228934e-05, "loss": 0.6159, "step": 17846 }, { "epoch": 1.197778598033623, "grad_norm": 0.8509141802787781, "learning_rate": 3.65759232605296e-05, "loss": 0.4617, "step": 17848 }, { "epoch": 1.197912821717392, "grad_norm": 0.9234373569488525, "learning_rate": 3.656545430303108e-05, "loss": 0.4666, "step": 17850 }, { "epoch": 1.198047045401161, "grad_norm": 0.9886850118637085, "learning_rate": 3.655498598028844e-05, "loss": 0.5268, "step": 17852 }, { "epoch": 1.19818126908493, "grad_norm": 0.9016915559768677, "learning_rate": 3.654451829279626e-05, "loss": 0.4801, "step": 17854 }, { "epoch": 1.198315492768699, "grad_norm": 1.187814474105835, "learning_rate": 3.6534051241049137e-05, "loss": 0.5369, "step": 17856 }, { "epoch": 1.1984497164524681, "grad_norm": 0.9405223727226257, "learning_rate": 3.652358482554162e-05, "loss": 0.5279, "step": 17858 }, { "epoch": 1.198583940136237, "grad_norm": 0.956626296043396, "learning_rate": 3.6513119046768206e-05, "loss": 0.5491, "step": 17860 }, { "epoch": 1.198718163820006, "grad_norm": 1.0439075231552124, "learning_rate": 3.65026539052234e-05, "loss": 0.5631, "step": 17862 }, { "epoch": 1.198852387503775, "grad_norm": 0.9437693357467651, "learning_rate": 3.6492189401401656e-05, "loss": 0.5267, "step": 17864 }, { "epoch": 1.198986611187544, "grad_norm": 1.1598464250564575, "learning_rate": 3.6481725535797415e-05, "loss": 0.5814, "step": 17866 }, { "epoch": 1.1991208348713132, "grad_norm": 0.9224652051925659, "learning_rate": 3.647126230890503e-05, "loss": 0.5349, "step": 17868 }, { "epoch": 1.199255058555082, "grad_norm": 0.9480599164962769, "learning_rate": 3.646079972121893e-05, "loss": 0.5187, "step": 17870 }, { "epoch": 1.199389282238851, "grad_norm": 1.0072935819625854, "learning_rate": 3.645033777323339e-05, "loss": 0.4771, "step": 17872 }, { "epoch": 1.1995235059226201, "grad_norm": 0.9206082224845886, "learning_rate": 3.643987646544278e-05, "loss": 0.568, "step": 17874 }, { "epoch": 1.199657729606389, "grad_norm": 0.9697580337524414, "learning_rate": 3.642941579834131e-05, "loss": 0.4322, "step": 17876 }, { "epoch": 1.199791953290158, "grad_norm": 0.9863641262054443, "learning_rate": 3.641895577242327e-05, "loss": 0.5506, "step": 17878 }, { "epoch": 1.1999261769739271, "grad_norm": 0.8394425511360168, "learning_rate": 3.640849638818286e-05, "loss": 0.4906, "step": 17880 }, { "epoch": 1.200060400657696, "grad_norm": 0.9852147698402405, "learning_rate": 3.639803764611428e-05, "loss": 0.4999, "step": 17882 }, { "epoch": 1.200194624341465, "grad_norm": 1.240363597869873, "learning_rate": 3.6387579546711656e-05, "loss": 0.5641, "step": 17884 }, { "epoch": 1.200328848025234, "grad_norm": 1.0349351167678833, "learning_rate": 3.637712209046915e-05, "loss": 0.4971, "step": 17886 }, { "epoch": 1.200463071709003, "grad_norm": 0.9062631726264954, "learning_rate": 3.636666527788083e-05, "loss": 0.4782, "step": 17888 }, { "epoch": 1.2005972953927722, "grad_norm": 1.0077486038208008, "learning_rate": 3.635620910944077e-05, "loss": 0.4889, "step": 17890 }, { "epoch": 1.200731519076541, "grad_norm": 1.0355712175369263, "learning_rate": 3.6345753585642996e-05, "loss": 0.5339, "step": 17892 }, { "epoch": 1.20086574276031, "grad_norm": 0.9321998357772827, "learning_rate": 3.6335298706981517e-05, "loss": 0.5161, "step": 17894 }, { "epoch": 1.2009999664440791, "grad_norm": 0.967764675617218, "learning_rate": 3.632484447395029e-05, "loss": 0.6395, "step": 17896 }, { "epoch": 1.201134190127848, "grad_norm": 1.0610249042510986, "learning_rate": 3.63143908870433e-05, "loss": 0.4735, "step": 17898 }, { "epoch": 1.201268413811617, "grad_norm": 0.9627799391746521, "learning_rate": 3.63039379467544e-05, "loss": 0.518, "step": 17900 }, { "epoch": 1.2014026374953861, "grad_norm": 0.9245826005935669, "learning_rate": 3.629348565357752e-05, "loss": 0.4635, "step": 17902 }, { "epoch": 1.201536861179155, "grad_norm": 0.9800781011581421, "learning_rate": 3.6283034008006465e-05, "loss": 0.4852, "step": 17904 }, { "epoch": 1.201671084862924, "grad_norm": 1.0018424987792969, "learning_rate": 3.627258301053511e-05, "loss": 0.5053, "step": 17906 }, { "epoch": 1.201805308546693, "grad_norm": 1.0354385375976562, "learning_rate": 3.626213266165719e-05, "loss": 0.5597, "step": 17908 }, { "epoch": 1.201939532230462, "grad_norm": 2.0732288360595703, "learning_rate": 3.6251682961866506e-05, "loss": 0.5155, "step": 17910 }, { "epoch": 1.2020737559142312, "grad_norm": 0.9796667098999023, "learning_rate": 3.6241233911656755e-05, "loss": 0.4653, "step": 17912 }, { "epoch": 1.202207979598, "grad_norm": 1.0895826816558838, "learning_rate": 3.6230785511521656e-05, "loss": 0.4678, "step": 17914 }, { "epoch": 1.202342203281769, "grad_norm": 1.074872612953186, "learning_rate": 3.6220337761954855e-05, "loss": 0.4601, "step": 17916 }, { "epoch": 1.2024764269655381, "grad_norm": 1.085161805152893, "learning_rate": 3.620989066345001e-05, "loss": 0.5764, "step": 17918 }, { "epoch": 1.202610650649307, "grad_norm": 1.0547800064086914, "learning_rate": 3.61994442165007e-05, "loss": 0.4965, "step": 17920 }, { "epoch": 1.202744874333076, "grad_norm": 0.9630693793296814, "learning_rate": 3.618899842160053e-05, "loss": 0.4983, "step": 17922 }, { "epoch": 1.2028790980168451, "grad_norm": 0.9729456901550293, "learning_rate": 3.617855327924302e-05, "loss": 0.5085, "step": 17924 }, { "epoch": 1.203013321700614, "grad_norm": 1.0706628561019897, "learning_rate": 3.616810878992168e-05, "loss": 0.5246, "step": 17926 }, { "epoch": 1.203147545384383, "grad_norm": 0.9929572939872742, "learning_rate": 3.6157664954130014e-05, "loss": 0.4808, "step": 17928 }, { "epoch": 1.203281769068152, "grad_norm": 1.0386592149734497, "learning_rate": 3.6147221772361446e-05, "loss": 0.5013, "step": 17930 }, { "epoch": 1.203415992751921, "grad_norm": 1.1223738193511963, "learning_rate": 3.613677924510942e-05, "loss": 0.4968, "step": 17932 }, { "epoch": 1.2035502164356902, "grad_norm": 1.0150068998336792, "learning_rate": 3.612633737286729e-05, "loss": 0.5536, "step": 17934 }, { "epoch": 1.203684440119459, "grad_norm": 0.9773173332214355, "learning_rate": 3.611589615612847e-05, "loss": 0.5046, "step": 17936 }, { "epoch": 1.203818663803228, "grad_norm": 0.9851709604263306, "learning_rate": 3.610545559538623e-05, "loss": 0.5024, "step": 17938 }, { "epoch": 1.2039528874869971, "grad_norm": 0.9655515551567078, "learning_rate": 3.6095015691133904e-05, "loss": 0.5187, "step": 17940 }, { "epoch": 1.204087111170766, "grad_norm": 1.0774941444396973, "learning_rate": 3.608457644386474e-05, "loss": 0.5763, "step": 17942 }, { "epoch": 1.2042213348545352, "grad_norm": 1.0457144975662231, "learning_rate": 3.6074137854071976e-05, "loss": 0.5097, "step": 17944 }, { "epoch": 1.2043555585383041, "grad_norm": 0.9518623352050781, "learning_rate": 3.606369992224882e-05, "loss": 0.4483, "step": 17946 }, { "epoch": 1.204489782222073, "grad_norm": 0.9342052936553955, "learning_rate": 3.605326264888843e-05, "loss": 0.5088, "step": 17948 }, { "epoch": 1.2046240059058422, "grad_norm": 1.091081142425537, "learning_rate": 3.6042826034483956e-05, "loss": 0.5486, "step": 17950 }, { "epoch": 1.204758229589611, "grad_norm": 1.0078307390213013, "learning_rate": 3.603239007952851e-05, "loss": 0.5193, "step": 17952 }, { "epoch": 1.20489245327338, "grad_norm": 0.965714156627655, "learning_rate": 3.602195478451516e-05, "loss": 0.519, "step": 17954 }, { "epoch": 1.2050266769571492, "grad_norm": 0.8684780597686768, "learning_rate": 3.6011520149936975e-05, "loss": 0.4679, "step": 17956 }, { "epoch": 1.205160900640918, "grad_norm": 0.9268240928649902, "learning_rate": 3.600108617628692e-05, "loss": 0.5211, "step": 17958 }, { "epoch": 1.205295124324687, "grad_norm": 1.1090248823165894, "learning_rate": 3.599065286405806e-05, "loss": 0.5528, "step": 17960 }, { "epoch": 1.2054293480084561, "grad_norm": 1.0047954320907593, "learning_rate": 3.598022021374327e-05, "loss": 0.496, "step": 17962 }, { "epoch": 1.205563571692225, "grad_norm": 1.0637043714523315, "learning_rate": 3.596978822583554e-05, "loss": 0.4953, "step": 17964 }, { "epoch": 1.2056977953759942, "grad_norm": 1.0383027791976929, "learning_rate": 3.595935690082769e-05, "loss": 0.5041, "step": 17966 }, { "epoch": 1.2058320190597631, "grad_norm": 0.9820025563240051, "learning_rate": 3.594892623921264e-05, "loss": 0.5287, "step": 17968 }, { "epoch": 1.205966242743532, "grad_norm": 1.118740200996399, "learning_rate": 3.593849624148319e-05, "loss": 0.5667, "step": 17970 }, { "epoch": 1.2061004664273012, "grad_norm": 1.0724904537200928, "learning_rate": 3.5928066908132144e-05, "loss": 0.5043, "step": 17972 }, { "epoch": 1.20623469011107, "grad_norm": 1.0753625631332397, "learning_rate": 3.591763823965226e-05, "loss": 0.5786, "step": 17974 }, { "epoch": 1.206368913794839, "grad_norm": 1.0068565607070923, "learning_rate": 3.590721023653628e-05, "loss": 0.5206, "step": 17976 }, { "epoch": 1.2065031374786082, "grad_norm": 0.9540721774101257, "learning_rate": 3.5896782899276905e-05, "loss": 0.4794, "step": 17978 }, { "epoch": 1.206637361162377, "grad_norm": 1.0614107847213745, "learning_rate": 3.58863562283668e-05, "loss": 0.5097, "step": 17980 }, { "epoch": 1.206771584846146, "grad_norm": 1.1896748542785645, "learning_rate": 3.5875930224298616e-05, "loss": 0.4741, "step": 17982 }, { "epoch": 1.2069058085299151, "grad_norm": 1.0033884048461914, "learning_rate": 3.586550488756496e-05, "loss": 0.5514, "step": 17984 }, { "epoch": 1.207040032213684, "grad_norm": 1.0701922178268433, "learning_rate": 3.585508021865838e-05, "loss": 0.54, "step": 17986 }, { "epoch": 1.2071742558974532, "grad_norm": 0.9692116379737854, "learning_rate": 3.584465621807148e-05, "loss": 0.5118, "step": 17988 }, { "epoch": 1.2073084795812221, "grad_norm": 0.9265748262405396, "learning_rate": 3.58342328862967e-05, "loss": 0.5253, "step": 17990 }, { "epoch": 1.207442703264991, "grad_norm": 1.0179463624954224, "learning_rate": 3.58238102238266e-05, "loss": 0.5352, "step": 17992 }, { "epoch": 1.2075769269487602, "grad_norm": 1.0018413066864014, "learning_rate": 3.5813388231153576e-05, "loss": 0.5197, "step": 17994 }, { "epoch": 1.207711150632529, "grad_norm": 0.993140697479248, "learning_rate": 3.5802966908770044e-05, "loss": 0.4556, "step": 17996 }, { "epoch": 1.207845374316298, "grad_norm": 0.9674150347709656, "learning_rate": 3.579254625716844e-05, "loss": 0.4942, "step": 17998 }, { "epoch": 1.2079795980000672, "grad_norm": 1.0136640071868896, "learning_rate": 3.5782126276841045e-05, "loss": 0.5252, "step": 18000 }, { "epoch": 1.208113821683836, "grad_norm": 1.0575333833694458, "learning_rate": 3.577170696828026e-05, "loss": 0.4999, "step": 18002 }, { "epoch": 1.208248045367605, "grad_norm": 0.998807430267334, "learning_rate": 3.576128833197832e-05, "loss": 0.478, "step": 18004 }, { "epoch": 1.2083822690513741, "grad_norm": 1.1003563404083252, "learning_rate": 3.5750870368427526e-05, "loss": 0.5106, "step": 18006 }, { "epoch": 1.208516492735143, "grad_norm": 0.9529297947883606, "learning_rate": 3.574045307812007e-05, "loss": 0.4817, "step": 18008 }, { "epoch": 1.2086507164189122, "grad_norm": 0.9703645706176758, "learning_rate": 3.573003646154819e-05, "loss": 0.5092, "step": 18010 }, { "epoch": 1.2087849401026811, "grad_norm": 0.9716821312904358, "learning_rate": 3.5719620519204e-05, "loss": 0.4638, "step": 18012 }, { "epoch": 1.20891916378645, "grad_norm": 1.0448511838912964, "learning_rate": 3.570920525157968e-05, "loss": 0.5714, "step": 18014 }, { "epoch": 1.2090533874702192, "grad_norm": 1.0114126205444336, "learning_rate": 3.569879065916729e-05, "loss": 0.5992, "step": 18016 }, { "epoch": 1.209187611153988, "grad_norm": 0.994085431098938, "learning_rate": 3.568837674245894e-05, "loss": 0.4827, "step": 18018 }, { "epoch": 1.2093218348377572, "grad_norm": 0.9882687330245972, "learning_rate": 3.5677963501946634e-05, "loss": 0.5636, "step": 18020 }, { "epoch": 1.2094560585215262, "grad_norm": 1.0266276597976685, "learning_rate": 3.5667550938122396e-05, "loss": 0.5164, "step": 18022 }, { "epoch": 1.209590282205295, "grad_norm": 1.452270746231079, "learning_rate": 3.565713905147817e-05, "loss": 0.5519, "step": 18024 }, { "epoch": 1.2097245058890642, "grad_norm": 0.9671774506568909, "learning_rate": 3.5646727842505955e-05, "loss": 0.5172, "step": 18026 }, { "epoch": 1.2098587295728331, "grad_norm": 1.0094542503356934, "learning_rate": 3.56363173116976e-05, "loss": 0.5207, "step": 18028 }, { "epoch": 1.209992953256602, "grad_norm": 1.1328052282333374, "learning_rate": 3.562590745954501e-05, "loss": 0.5359, "step": 18030 }, { "epoch": 1.2101271769403712, "grad_norm": 1.0051718950271606, "learning_rate": 3.561549828654003e-05, "loss": 0.5685, "step": 18032 }, { "epoch": 1.2102614006241401, "grad_norm": 1.0846080780029297, "learning_rate": 3.5605089793174485e-05, "loss": 0.5447, "step": 18034 }, { "epoch": 1.210395624307909, "grad_norm": 1.0940159559249878, "learning_rate": 3.559468197994012e-05, "loss": 0.469, "step": 18036 }, { "epoch": 1.2105298479916782, "grad_norm": 1.0636924505233765, "learning_rate": 3.558427484732872e-05, "loss": 0.4979, "step": 18038 }, { "epoch": 1.210664071675447, "grad_norm": 1.1106228828430176, "learning_rate": 3.5573868395831975e-05, "loss": 0.4764, "step": 18040 }, { "epoch": 1.2107982953592162, "grad_norm": 1.08779776096344, "learning_rate": 3.556346262594159e-05, "loss": 0.5086, "step": 18042 }, { "epoch": 1.2109325190429852, "grad_norm": 0.8930120468139648, "learning_rate": 3.5553057538149196e-05, "loss": 0.5473, "step": 18044 }, { "epoch": 1.211066742726754, "grad_norm": 3.5846095085144043, "learning_rate": 3.5542653132946435e-05, "loss": 0.5397, "step": 18046 }, { "epoch": 1.2112009664105232, "grad_norm": 1.1256413459777832, "learning_rate": 3.553224941082487e-05, "loss": 0.587, "step": 18048 }, { "epoch": 1.2113351900942921, "grad_norm": 1.0590184926986694, "learning_rate": 3.552184637227609e-05, "loss": 0.5121, "step": 18050 }, { "epoch": 1.211469413778061, "grad_norm": 0.9384714961051941, "learning_rate": 3.5511444017791575e-05, "loss": 0.5786, "step": 18052 }, { "epoch": 1.2116036374618302, "grad_norm": 1.056459665298462, "learning_rate": 3.550104234786287e-05, "loss": 0.5226, "step": 18054 }, { "epoch": 1.2117378611455991, "grad_norm": 1.0418684482574463, "learning_rate": 3.5490641362981374e-05, "loss": 0.5002, "step": 18056 }, { "epoch": 1.211872084829368, "grad_norm": 1.3959248065948486, "learning_rate": 3.5480241063638566e-05, "loss": 0.5453, "step": 18058 }, { "epoch": 1.2120063085131372, "grad_norm": 0.9421146512031555, "learning_rate": 3.546984145032582e-05, "loss": 0.4287, "step": 18060 }, { "epoch": 1.212140532196906, "grad_norm": 0.9701901078224182, "learning_rate": 3.545944252353446e-05, "loss": 0.4897, "step": 18062 }, { "epoch": 1.2122747558806752, "grad_norm": 1.0220611095428467, "learning_rate": 3.5449044283755874e-05, "loss": 0.5086, "step": 18064 }, { "epoch": 1.2124089795644442, "grad_norm": 1.0941497087478638, "learning_rate": 3.543864673148132e-05, "loss": 0.4985, "step": 18066 }, { "epoch": 1.212543203248213, "grad_norm": 0.967139720916748, "learning_rate": 3.542824986720209e-05, "loss": 0.5037, "step": 18068 }, { "epoch": 1.2126774269319822, "grad_norm": 1.2793251276016235, "learning_rate": 3.541785369140938e-05, "loss": 0.5264, "step": 18070 }, { "epoch": 1.2128116506157511, "grad_norm": 1.0796058177947998, "learning_rate": 3.5407458204594426e-05, "loss": 0.5904, "step": 18072 }, { "epoch": 1.21294587429952, "grad_norm": 1.0878667831420898, "learning_rate": 3.5397063407248365e-05, "loss": 0.5369, "step": 18074 }, { "epoch": 1.2130800979832892, "grad_norm": 1.0877381563186646, "learning_rate": 3.5386669299862355e-05, "loss": 0.5289, "step": 18076 }, { "epoch": 1.2132143216670581, "grad_norm": 1.190150260925293, "learning_rate": 3.5376275882927466e-05, "loss": 0.5311, "step": 18078 }, { "epoch": 1.213348545350827, "grad_norm": 0.9739012718200684, "learning_rate": 3.5365883156934795e-05, "loss": 0.4665, "step": 18080 }, { "epoch": 1.2134827690345962, "grad_norm": 0.9805138111114502, "learning_rate": 3.535549112237537e-05, "loss": 0.505, "step": 18082 }, { "epoch": 1.213616992718365, "grad_norm": 0.8651003241539001, "learning_rate": 3.534509977974019e-05, "loss": 0.4986, "step": 18084 }, { "epoch": 1.2137512164021342, "grad_norm": 1.2091014385223389, "learning_rate": 3.533470912952022e-05, "loss": 0.5289, "step": 18086 }, { "epoch": 1.2138854400859032, "grad_norm": 0.9878976941108704, "learning_rate": 3.532431917220642e-05, "loss": 0.4601, "step": 18088 }, { "epoch": 1.214019663769672, "grad_norm": 0.9306507706642151, "learning_rate": 3.5313929908289665e-05, "loss": 0.5849, "step": 18090 }, { "epoch": 1.2141538874534412, "grad_norm": 0.9268865585327148, "learning_rate": 3.5303541338260856e-05, "loss": 0.479, "step": 18092 }, { "epoch": 1.2142881111372101, "grad_norm": 1.228854775428772, "learning_rate": 3.529315346261081e-05, "loss": 0.509, "step": 18094 }, { "epoch": 1.2144223348209793, "grad_norm": 0.8898868560791016, "learning_rate": 3.528276628183035e-05, "loss": 0.4212, "step": 18096 }, { "epoch": 1.2145565585047482, "grad_norm": 0.9570024013519287, "learning_rate": 3.527237979641024e-05, "loss": 0.502, "step": 18098 }, { "epoch": 1.2146907821885171, "grad_norm": 1.054985523223877, "learning_rate": 3.526199400684124e-05, "loss": 0.4765, "step": 18100 }, { "epoch": 1.2148250058722863, "grad_norm": 1.0664186477661133, "learning_rate": 3.5251608913614014e-05, "loss": 0.5318, "step": 18102 }, { "epoch": 1.2149592295560552, "grad_norm": 0.9964617490768433, "learning_rate": 3.5241224517219286e-05, "loss": 0.5151, "step": 18104 }, { "epoch": 1.215093453239824, "grad_norm": 0.9645223021507263, "learning_rate": 3.523084081814767e-05, "loss": 0.5289, "step": 18106 }, { "epoch": 1.2152276769235932, "grad_norm": 1.088595986366272, "learning_rate": 3.5220457816889784e-05, "loss": 0.4541, "step": 18108 }, { "epoch": 1.2153619006073622, "grad_norm": 1.0221766233444214, "learning_rate": 3.5210075513936206e-05, "loss": 0.5183, "step": 18110 }, { "epoch": 1.215496124291131, "grad_norm": 0.8846173882484436, "learning_rate": 3.519969390977748e-05, "loss": 0.4248, "step": 18112 }, { "epoch": 1.2156303479749002, "grad_norm": 1.0733497142791748, "learning_rate": 3.518931300490409e-05, "loss": 0.5247, "step": 18114 }, { "epoch": 1.2157645716586691, "grad_norm": 1.0206555128097534, "learning_rate": 3.517893279980656e-05, "loss": 0.5308, "step": 18116 }, { "epoch": 1.2158987953424383, "grad_norm": 1.1100720167160034, "learning_rate": 3.516855329497529e-05, "loss": 0.4776, "step": 18118 }, { "epoch": 1.2160330190262072, "grad_norm": 0.9295620918273926, "learning_rate": 3.5158174490900744e-05, "loss": 0.4824, "step": 18120 }, { "epoch": 1.2161672427099761, "grad_norm": 1.0154778957366943, "learning_rate": 3.514779638807324e-05, "loss": 0.5329, "step": 18122 }, { "epoch": 1.2163014663937453, "grad_norm": 1.0279161930084229, "learning_rate": 3.513741898698317e-05, "loss": 0.4524, "step": 18124 }, { "epoch": 1.2164356900775142, "grad_norm": 1.1146999597549438, "learning_rate": 3.512704228812082e-05, "loss": 0.4705, "step": 18126 }, { "epoch": 1.216569913761283, "grad_norm": 1.072011113166809, "learning_rate": 3.5116666291976474e-05, "loss": 0.5498, "step": 18128 }, { "epoch": 1.2167041374450522, "grad_norm": 1.0384671688079834, "learning_rate": 3.510629099904038e-05, "loss": 0.4668, "step": 18130 }, { "epoch": 1.2168383611288212, "grad_norm": 0.9664222598075867, "learning_rate": 3.509591640980274e-05, "loss": 0.4974, "step": 18132 }, { "epoch": 1.21697258481259, "grad_norm": 1.051636815071106, "learning_rate": 3.508554252475376e-05, "loss": 0.5368, "step": 18134 }, { "epoch": 1.2171068084963592, "grad_norm": 0.8387489914894104, "learning_rate": 3.507516934438355e-05, "loss": 0.4544, "step": 18136 }, { "epoch": 1.2172410321801281, "grad_norm": 1.011088490486145, "learning_rate": 3.506479686918226e-05, "loss": 0.5168, "step": 18138 }, { "epoch": 1.2173752558638973, "grad_norm": 0.9723619222640991, "learning_rate": 3.505442509963993e-05, "loss": 0.4749, "step": 18140 }, { "epoch": 1.2175094795476662, "grad_norm": 0.9728646874427795, "learning_rate": 3.504405403624664e-05, "loss": 0.494, "step": 18142 }, { "epoch": 1.2176437032314351, "grad_norm": 1.091109275817871, "learning_rate": 3.503368367949237e-05, "loss": 0.5355, "step": 18144 }, { "epoch": 1.2177779269152043, "grad_norm": 1.0038992166519165, "learning_rate": 3.502331402986713e-05, "loss": 0.5109, "step": 18146 }, { "epoch": 1.2179121505989732, "grad_norm": 1.120638132095337, "learning_rate": 3.501294508786084e-05, "loss": 0.4706, "step": 18148 }, { "epoch": 1.218046374282742, "grad_norm": 0.9986687898635864, "learning_rate": 3.500257685396344e-05, "loss": 0.5479, "step": 18150 }, { "epoch": 1.2181805979665112, "grad_norm": 1.1052039861679077, "learning_rate": 3.4992209328664774e-05, "loss": 0.5399, "step": 18152 }, { "epoch": 1.2183148216502802, "grad_norm": 0.9729425311088562, "learning_rate": 3.4981842512454716e-05, "loss": 0.505, "step": 18154 }, { "epoch": 1.218449045334049, "grad_norm": 1.1042462587356567, "learning_rate": 3.497147640582306e-05, "loss": 0.4806, "step": 18156 }, { "epoch": 1.2185832690178182, "grad_norm": 1.1737509965896606, "learning_rate": 3.4961111009259605e-05, "loss": 0.5463, "step": 18158 }, { "epoch": 1.2187174927015871, "grad_norm": 0.889025092124939, "learning_rate": 3.495074632325407e-05, "loss": 0.5417, "step": 18160 }, { "epoch": 1.2188517163853563, "grad_norm": 1.0269781351089478, "learning_rate": 3.4940382348296186e-05, "loss": 0.5851, "step": 18162 }, { "epoch": 1.2189859400691252, "grad_norm": 0.9341825842857361, "learning_rate": 3.493001908487561e-05, "loss": 0.5085, "step": 18164 }, { "epoch": 1.2191201637528941, "grad_norm": 0.9956706166267395, "learning_rate": 3.491965653348202e-05, "loss": 0.5383, "step": 18166 }, { "epoch": 1.2192543874366633, "grad_norm": 1.0724912881851196, "learning_rate": 3.490929469460499e-05, "loss": 0.5229, "step": 18168 }, { "epoch": 1.2193886111204322, "grad_norm": 0.9348374605178833, "learning_rate": 3.4898933568734115e-05, "loss": 0.492, "step": 18170 }, { "epoch": 1.2195228348042013, "grad_norm": 0.9576309323310852, "learning_rate": 3.488857315635893e-05, "loss": 0.4348, "step": 18172 }, { "epoch": 1.2196570584879702, "grad_norm": 1.3052387237548828, "learning_rate": 3.487821345796895e-05, "loss": 0.4717, "step": 18174 }, { "epoch": 1.2197912821717392, "grad_norm": 1.044317364692688, "learning_rate": 3.4867854474053644e-05, "loss": 0.4311, "step": 18176 }, { "epoch": 1.2199255058555083, "grad_norm": 1.0891656875610352, "learning_rate": 3.4857496205102474e-05, "loss": 0.5037, "step": 18178 }, { "epoch": 1.2200597295392772, "grad_norm": 1.0032423734664917, "learning_rate": 3.4847138651604805e-05, "loss": 0.5302, "step": 18180 }, { "epoch": 1.2201939532230461, "grad_norm": 0.8056291937828064, "learning_rate": 3.483678181405006e-05, "loss": 0.4807, "step": 18182 }, { "epoch": 1.2203281769068153, "grad_norm": 1.104245662689209, "learning_rate": 3.4826425692927545e-05, "loss": 0.5011, "step": 18184 }, { "epoch": 1.2204624005905842, "grad_norm": 0.9866969585418701, "learning_rate": 3.481607028872659e-05, "loss": 0.5141, "step": 18186 }, { "epoch": 1.2205966242743531, "grad_norm": 1.0253126621246338, "learning_rate": 3.480571560193645e-05, "loss": 0.5778, "step": 18188 }, { "epoch": 1.2207308479581223, "grad_norm": 0.9861544370651245, "learning_rate": 3.479536163304637e-05, "loss": 0.4452, "step": 18190 }, { "epoch": 1.2208650716418912, "grad_norm": 0.9421694874763489, "learning_rate": 3.4785008382545546e-05, "loss": 0.5036, "step": 18192 }, { "epoch": 1.2209992953256603, "grad_norm": 1.0168511867523193, "learning_rate": 3.4774655850923174e-05, "loss": 0.5127, "step": 18194 }, { "epoch": 1.2211335190094292, "grad_norm": 1.1280962228775024, "learning_rate": 3.476430403866836e-05, "loss": 0.5206, "step": 18196 }, { "epoch": 1.2212677426931982, "grad_norm": 0.909195065498352, "learning_rate": 3.475395294627023e-05, "loss": 0.4774, "step": 18198 }, { "epoch": 1.2214019663769673, "grad_norm": 1.0700228214263916, "learning_rate": 3.474360257421784e-05, "loss": 0.4393, "step": 18200 }, { "epoch": 1.2215361900607362, "grad_norm": 1.0085256099700928, "learning_rate": 3.4733252923000226e-05, "loss": 0.4778, "step": 18202 }, { "epoch": 1.2216704137445051, "grad_norm": 0.963640570640564, "learning_rate": 3.4722903993106395e-05, "loss": 0.5338, "step": 18204 }, { "epoch": 1.2218046374282743, "grad_norm": 1.1290463209152222, "learning_rate": 3.4712555785025305e-05, "loss": 0.5166, "step": 18206 }, { "epoch": 1.2219388611120432, "grad_norm": 0.9944272637367249, "learning_rate": 3.4702208299245895e-05, "loss": 0.4689, "step": 18208 }, { "epoch": 1.2220730847958121, "grad_norm": 0.8961957693099976, "learning_rate": 3.469186153625704e-05, "loss": 0.5015, "step": 18210 }, { "epoch": 1.2222073084795813, "grad_norm": 0.9712990522384644, "learning_rate": 3.468151549654766e-05, "loss": 0.4926, "step": 18212 }, { "epoch": 1.2223415321633502, "grad_norm": 0.9702814817428589, "learning_rate": 3.467117018060652e-05, "loss": 0.4775, "step": 18214 }, { "epoch": 1.2224757558471193, "grad_norm": 0.9661044478416443, "learning_rate": 3.466082558892247e-05, "loss": 0.5065, "step": 18216 }, { "epoch": 1.2226099795308882, "grad_norm": 0.9817913174629211, "learning_rate": 3.465048172198423e-05, "loss": 0.4714, "step": 18218 }, { "epoch": 1.2227442032146572, "grad_norm": 1.1519465446472168, "learning_rate": 3.464013858028056e-05, "loss": 0.5162, "step": 18220 }, { "epoch": 1.2228784268984263, "grad_norm": 0.944974958896637, "learning_rate": 3.462979616430012e-05, "loss": 0.5193, "step": 18222 }, { "epoch": 1.2230126505821952, "grad_norm": 1.0148048400878906, "learning_rate": 3.461945447453161e-05, "loss": 0.4923, "step": 18224 }, { "epoch": 1.2231468742659641, "grad_norm": 1.0852816104888916, "learning_rate": 3.4609113511463616e-05, "loss": 0.4819, "step": 18226 }, { "epoch": 1.2232810979497333, "grad_norm": 1.0011029243469238, "learning_rate": 3.459877327558475e-05, "loss": 0.5228, "step": 18228 }, { "epoch": 1.2234153216335022, "grad_norm": 1.0051935911178589, "learning_rate": 3.458843376738355e-05, "loss": 0.5644, "step": 18230 }, { "epoch": 1.2235495453172711, "grad_norm": 1.0228527784347534, "learning_rate": 3.457809498734857e-05, "loss": 0.478, "step": 18232 }, { "epoch": 1.2236837690010403, "grad_norm": 0.8186175227165222, "learning_rate": 3.456775693596825e-05, "loss": 0.4832, "step": 18234 }, { "epoch": 1.2238179926848092, "grad_norm": 0.9312829971313477, "learning_rate": 3.455741961373109e-05, "loss": 0.52, "step": 18236 }, { "epoch": 1.2239522163685783, "grad_norm": 1.4396902322769165, "learning_rate": 3.454708302112547e-05, "loss": 0.5446, "step": 18238 }, { "epoch": 1.2240864400523472, "grad_norm": 0.9807689785957336, "learning_rate": 3.453674715863982e-05, "loss": 0.4864, "step": 18240 }, { "epoch": 1.2242206637361162, "grad_norm": 0.9718871116638184, "learning_rate": 3.452641202676242e-05, "loss": 0.4939, "step": 18242 }, { "epoch": 1.2243548874198853, "grad_norm": 1.0906612873077393, "learning_rate": 3.4516077625981644e-05, "loss": 0.5018, "step": 18244 }, { "epoch": 1.2244891111036542, "grad_norm": 1.0393626689910889, "learning_rate": 3.450574395678575e-05, "loss": 0.5002, "step": 18246 }, { "epoch": 1.2246233347874234, "grad_norm": 1.05021071434021, "learning_rate": 3.4495411019662994e-05, "loss": 0.4996, "step": 18248 }, { "epoch": 1.2247575584711923, "grad_norm": 1.0882809162139893, "learning_rate": 3.4485078815101566e-05, "loss": 0.5134, "step": 18250 }, { "epoch": 1.2248917821549612, "grad_norm": 1.0747679471969604, "learning_rate": 3.447474734358967e-05, "loss": 0.4203, "step": 18252 }, { "epoch": 1.2250260058387303, "grad_norm": 1.0003612041473389, "learning_rate": 3.4464416605615424e-05, "loss": 0.4996, "step": 18254 }, { "epoch": 1.2251602295224993, "grad_norm": 1.214212417602539, "learning_rate": 3.4454086601666945e-05, "loss": 0.5272, "step": 18256 }, { "epoch": 1.2252944532062682, "grad_norm": 1.1573606729507446, "learning_rate": 3.44437573322323e-05, "loss": 0.4412, "step": 18258 }, { "epoch": 1.2254286768900373, "grad_norm": 0.9980739951133728, "learning_rate": 3.443342879779954e-05, "loss": 0.5256, "step": 18260 }, { "epoch": 1.2255629005738062, "grad_norm": 1.5467917919158936, "learning_rate": 3.442310099885665e-05, "loss": 0.482, "step": 18262 }, { "epoch": 1.2256971242575752, "grad_norm": 2.0626678466796875, "learning_rate": 3.441277393589162e-05, "loss": 0.528, "step": 18264 }, { "epoch": 1.2258313479413443, "grad_norm": 0.8996275663375854, "learning_rate": 3.440244760939236e-05, "loss": 0.4628, "step": 18266 }, { "epoch": 1.2259655716251132, "grad_norm": 0.7829926609992981, "learning_rate": 3.439212201984679e-05, "loss": 0.5047, "step": 18268 }, { "epoch": 1.2260997953088824, "grad_norm": 1.0295792818069458, "learning_rate": 3.438179716774275e-05, "loss": 0.5633, "step": 18270 }, { "epoch": 1.2262340189926513, "grad_norm": 0.9000246524810791, "learning_rate": 3.437147305356807e-05, "loss": 0.4807, "step": 18272 }, { "epoch": 1.2263682426764202, "grad_norm": 0.8799797892570496, "learning_rate": 3.436114967781059e-05, "loss": 0.5576, "step": 18274 }, { "epoch": 1.2265024663601893, "grad_norm": 0.9025986790657043, "learning_rate": 3.4350827040958e-05, "loss": 0.4866, "step": 18276 }, { "epoch": 1.2266366900439583, "grad_norm": 1.0311486721038818, "learning_rate": 3.434050514349809e-05, "loss": 0.5643, "step": 18278 }, { "epoch": 1.2267709137277272, "grad_norm": 0.9919482469558716, "learning_rate": 3.43301839859185e-05, "loss": 0.4589, "step": 18280 }, { "epoch": 1.2269051374114963, "grad_norm": 1.0341503620147705, "learning_rate": 3.431986356870691e-05, "loss": 0.4944, "step": 18282 }, { "epoch": 1.2270393610952652, "grad_norm": 0.968529224395752, "learning_rate": 3.430954389235092e-05, "loss": 0.4862, "step": 18284 }, { "epoch": 1.2271735847790342, "grad_norm": 1.0049059391021729, "learning_rate": 3.429922495733815e-05, "loss": 0.5348, "step": 18286 }, { "epoch": 1.2273078084628033, "grad_norm": 0.8700787425041199, "learning_rate": 3.4288906764156106e-05, "loss": 0.4803, "step": 18288 }, { "epoch": 1.2274420321465722, "grad_norm": 1.1416560411453247, "learning_rate": 3.4278589313292345e-05, "loss": 0.5415, "step": 18290 }, { "epoch": 1.2275762558303414, "grad_norm": 1.0105998516082764, "learning_rate": 3.4268272605234296e-05, "loss": 0.4932, "step": 18292 }, { "epoch": 1.2277104795141103, "grad_norm": 0.9765534400939941, "learning_rate": 3.425795664046946e-05, "loss": 0.5506, "step": 18294 }, { "epoch": 1.2278447031978792, "grad_norm": 1.060732364654541, "learning_rate": 3.42476414194852e-05, "loss": 0.5005, "step": 18296 }, { "epoch": 1.2279789268816483, "grad_norm": 1.036497712135315, "learning_rate": 3.423732694276891e-05, "loss": 0.5465, "step": 18298 }, { "epoch": 1.2281131505654173, "grad_norm": 1.1618542671203613, "learning_rate": 3.422701321080791e-05, "loss": 0.5404, "step": 18300 }, { "epoch": 1.2282473742491862, "grad_norm": 0.8722122311592102, "learning_rate": 3.421670022408956e-05, "loss": 0.49, "step": 18302 }, { "epoch": 1.2283815979329553, "grad_norm": 0.9689675569534302, "learning_rate": 3.420638798310105e-05, "loss": 0.497, "step": 18304 }, { "epoch": 1.2285158216167242, "grad_norm": 1.1279895305633545, "learning_rate": 3.419607648832968e-05, "loss": 0.5396, "step": 18306 }, { "epoch": 1.2286500453004932, "grad_norm": 0.9110646843910217, "learning_rate": 3.4185765740262586e-05, "loss": 0.485, "step": 18308 }, { "epoch": 1.2287842689842623, "grad_norm": 0.9667492508888245, "learning_rate": 3.417545573938699e-05, "loss": 0.5364, "step": 18310 }, { "epoch": 1.2289184926680312, "grad_norm": 1.0547741651535034, "learning_rate": 3.416514648618998e-05, "loss": 0.4854, "step": 18312 }, { "epoch": 1.2290527163518004, "grad_norm": 1.0077714920043945, "learning_rate": 3.415483798115867e-05, "loss": 0.4669, "step": 18314 }, { "epoch": 1.2291869400355693, "grad_norm": 1.0353007316589355, "learning_rate": 3.4144530224780095e-05, "loss": 0.551, "step": 18316 }, { "epoch": 1.2293211637193382, "grad_norm": 1.0020809173583984, "learning_rate": 3.41342232175413e-05, "loss": 0.5244, "step": 18318 }, { "epoch": 1.2294553874031073, "grad_norm": 1.1620763540267944, "learning_rate": 3.412391695992925e-05, "loss": 0.5181, "step": 18320 }, { "epoch": 1.2295896110868763, "grad_norm": 1.1190643310546875, "learning_rate": 3.4113611452430914e-05, "loss": 0.5102, "step": 18322 }, { "epoch": 1.2297238347706454, "grad_norm": 1.0470566749572754, "learning_rate": 3.410330669553319e-05, "loss": 0.5045, "step": 18324 }, { "epoch": 1.2298580584544143, "grad_norm": 1.0887559652328491, "learning_rate": 3.409300268972298e-05, "loss": 0.4794, "step": 18326 }, { "epoch": 1.2299922821381832, "grad_norm": 1.129022240638733, "learning_rate": 3.4082699435487084e-05, "loss": 0.5229, "step": 18328 }, { "epoch": 1.2301265058219524, "grad_norm": 0.9262421727180481, "learning_rate": 3.4072396933312365e-05, "loss": 0.5555, "step": 18330 }, { "epoch": 1.2302607295057213, "grad_norm": 1.0439683198928833, "learning_rate": 3.406209518368555e-05, "loss": 0.517, "step": 18332 }, { "epoch": 1.2303949531894902, "grad_norm": 1.0444456338882446, "learning_rate": 3.405179418709342e-05, "loss": 0.5042, "step": 18334 }, { "epoch": 1.2305291768732594, "grad_norm": 1.017127275466919, "learning_rate": 3.4041493944022655e-05, "loss": 0.4798, "step": 18336 }, { "epoch": 1.2306634005570283, "grad_norm": 1.0182819366455078, "learning_rate": 3.40311944549599e-05, "loss": 0.5153, "step": 18338 }, { "epoch": 1.2307976242407972, "grad_norm": 0.9072675704956055, "learning_rate": 3.402089572039183e-05, "loss": 0.4493, "step": 18340 }, { "epoch": 1.2309318479245663, "grad_norm": 0.9794762134552002, "learning_rate": 3.401059774080498e-05, "loss": 0.467, "step": 18342 }, { "epoch": 1.2310660716083353, "grad_norm": 1.4422526359558105, "learning_rate": 3.4000300516685965e-05, "loss": 0.5094, "step": 18344 }, { "epoch": 1.2312002952921044, "grad_norm": 1.0380927324295044, "learning_rate": 3.399000404852129e-05, "loss": 0.5298, "step": 18346 }, { "epoch": 1.2313345189758733, "grad_norm": 1.0192526578903198, "learning_rate": 3.397970833679744e-05, "loss": 0.4584, "step": 18348 }, { "epoch": 1.2314687426596422, "grad_norm": 1.0050543546676636, "learning_rate": 3.396941338200087e-05, "loss": 0.4855, "step": 18350 }, { "epoch": 1.2316029663434114, "grad_norm": 1.0599147081375122, "learning_rate": 3.3959119184618e-05, "loss": 0.5053, "step": 18352 }, { "epoch": 1.2317371900271803, "grad_norm": 0.9676487445831299, "learning_rate": 3.394882574513519e-05, "loss": 0.4528, "step": 18354 }, { "epoch": 1.2318714137109492, "grad_norm": 1.147121548652649, "learning_rate": 3.393853306403881e-05, "loss": 0.5288, "step": 18356 }, { "epoch": 1.2320056373947184, "grad_norm": 0.8826063871383667, "learning_rate": 3.392824114181516e-05, "loss": 0.5134, "step": 18358 }, { "epoch": 1.2321398610784873, "grad_norm": 1.051510214805603, "learning_rate": 3.391794997895051e-05, "loss": 0.4756, "step": 18360 }, { "epoch": 1.2322740847622562, "grad_norm": 0.9662967920303345, "learning_rate": 3.3907659575931096e-05, "loss": 0.4593, "step": 18362 }, { "epoch": 1.2324083084460253, "grad_norm": 1.0281141996383667, "learning_rate": 3.389736993324314e-05, "loss": 0.5, "step": 18364 }, { "epoch": 1.2325425321297943, "grad_norm": 0.966301441192627, "learning_rate": 3.388708105137276e-05, "loss": 0.477, "step": 18366 }, { "epoch": 1.2326767558135634, "grad_norm": 0.9663693308830261, "learning_rate": 3.3876792930806144e-05, "loss": 0.5562, "step": 18368 }, { "epoch": 1.2328109794973323, "grad_norm": 0.892840564250946, "learning_rate": 3.3866505572029334e-05, "loss": 0.4397, "step": 18370 }, { "epoch": 1.2329452031811012, "grad_norm": 1.1237777471542358, "learning_rate": 3.3856218975528434e-05, "loss": 0.5463, "step": 18372 }, { "epoch": 1.2330794268648704, "grad_norm": 0.975803554058075, "learning_rate": 3.384593314178942e-05, "loss": 0.4984, "step": 18374 }, { "epoch": 1.2332136505486393, "grad_norm": 0.9467883110046387, "learning_rate": 3.383564807129832e-05, "loss": 0.5388, "step": 18376 }, { "epoch": 1.2333478742324082, "grad_norm": 1.12064528465271, "learning_rate": 3.382536376454104e-05, "loss": 0.5373, "step": 18378 }, { "epoch": 1.2334820979161774, "grad_norm": 0.9428995251655579, "learning_rate": 3.3815080222003533e-05, "loss": 0.4736, "step": 18380 }, { "epoch": 1.2336163215999463, "grad_norm": 0.9213509559631348, "learning_rate": 3.3804797444171654e-05, "loss": 0.4904, "step": 18382 }, { "epoch": 1.2337505452837152, "grad_norm": 1.0931910276412964, "learning_rate": 3.379451543153126e-05, "loss": 0.5332, "step": 18384 }, { "epoch": 1.2338847689674843, "grad_norm": 1.0226064920425415, "learning_rate": 3.378423418456813e-05, "loss": 0.5754, "step": 18386 }, { "epoch": 1.2340189926512533, "grad_norm": 1.019049882888794, "learning_rate": 3.3773953703768055e-05, "loss": 0.4714, "step": 18388 }, { "epoch": 1.2341532163350224, "grad_norm": 1.0632864236831665, "learning_rate": 3.376367398961674e-05, "loss": 0.5205, "step": 18390 }, { "epoch": 1.2342874400187913, "grad_norm": 1.2359843254089355, "learning_rate": 3.375339504259994e-05, "loss": 0.506, "step": 18392 }, { "epoch": 1.2344216637025602, "grad_norm": 1.0137488842010498, "learning_rate": 3.3743116863203236e-05, "loss": 0.5279, "step": 18394 }, { "epoch": 1.2345558873863294, "grad_norm": 0.8961823582649231, "learning_rate": 3.3732839451912335e-05, "loss": 0.4639, "step": 18396 }, { "epoch": 1.2346901110700983, "grad_norm": 0.9982473850250244, "learning_rate": 3.372256280921274e-05, "loss": 0.5054, "step": 18398 }, { "epoch": 1.2348243347538674, "grad_norm": 1.088832974433899, "learning_rate": 3.371228693559007e-05, "loss": 0.5486, "step": 18400 }, { "epoch": 1.2349585584376364, "grad_norm": 0.926771342754364, "learning_rate": 3.3702011831529805e-05, "loss": 0.5207, "step": 18402 }, { "epoch": 1.2350927821214053, "grad_norm": 1.06533944606781, "learning_rate": 3.369173749751744e-05, "loss": 0.4763, "step": 18404 }, { "epoch": 1.2352270058051744, "grad_norm": 1.282995343208313, "learning_rate": 3.3681463934038415e-05, "loss": 0.4892, "step": 18406 }, { "epoch": 1.2353612294889433, "grad_norm": 0.9002024531364441, "learning_rate": 3.3671191141578114e-05, "loss": 0.4841, "step": 18408 }, { "epoch": 1.2354954531727123, "grad_norm": 1.3077336549758911, "learning_rate": 3.366091912062194e-05, "loss": 0.4384, "step": 18410 }, { "epoch": 1.2356296768564814, "grad_norm": 0.9488261938095093, "learning_rate": 3.365064787165518e-05, "loss": 0.4489, "step": 18412 }, { "epoch": 1.2357639005402503, "grad_norm": 1.1497156620025635, "learning_rate": 3.3640377395163185e-05, "loss": 0.5817, "step": 18414 }, { "epoch": 1.2358981242240192, "grad_norm": 1.0375531911849976, "learning_rate": 3.3630107691631174e-05, "loss": 0.567, "step": 18416 }, { "epoch": 1.2360323479077884, "grad_norm": 0.9430070519447327, "learning_rate": 3.361983876154439e-05, "loss": 0.4803, "step": 18418 }, { "epoch": 1.2361665715915573, "grad_norm": 0.9468973278999329, "learning_rate": 3.360957060538801e-05, "loss": 0.4333, "step": 18420 }, { "epoch": 1.2363007952753264, "grad_norm": 0.9819739460945129, "learning_rate": 3.35993032236472e-05, "loss": 0.5326, "step": 18422 }, { "epoch": 1.2364350189590954, "grad_norm": 0.9785555601119995, "learning_rate": 3.358903661680703e-05, "loss": 0.4747, "step": 18424 }, { "epoch": 1.2365692426428643, "grad_norm": 0.9492508769035339, "learning_rate": 3.3578770785352654e-05, "loss": 0.44, "step": 18426 }, { "epoch": 1.2367034663266334, "grad_norm": 0.9501255750656128, "learning_rate": 3.356850572976903e-05, "loss": 0.5529, "step": 18428 }, { "epoch": 1.2368376900104023, "grad_norm": 1.0544824600219727, "learning_rate": 3.3558241450541235e-05, "loss": 0.4882, "step": 18430 }, { "epoch": 1.2369719136941713, "grad_norm": 0.9859625101089478, "learning_rate": 3.354797794815416e-05, "loss": 0.5196, "step": 18432 }, { "epoch": 1.2371061373779404, "grad_norm": 0.9515497088432312, "learning_rate": 3.35377152230928e-05, "loss": 0.4984, "step": 18434 }, { "epoch": 1.2372403610617093, "grad_norm": 1.6617851257324219, "learning_rate": 3.352745327584202e-05, "loss": 0.5168, "step": 18436 }, { "epoch": 1.2373745847454782, "grad_norm": 0.9807457327842712, "learning_rate": 3.3517192106886684e-05, "loss": 0.5494, "step": 18438 }, { "epoch": 1.2375088084292474, "grad_norm": 1.1575415134429932, "learning_rate": 3.3506931716711596e-05, "loss": 0.4799, "step": 18440 }, { "epoch": 1.2376430321130163, "grad_norm": 1.086361289024353, "learning_rate": 3.3496672105801575e-05, "loss": 0.4879, "step": 18442 }, { "epoch": 1.2377772557967854, "grad_norm": 1.0192978382110596, "learning_rate": 3.3486413274641327e-05, "loss": 0.5117, "step": 18444 }, { "epoch": 1.2379114794805544, "grad_norm": 1.3816787004470825, "learning_rate": 3.347615522371559e-05, "loss": 0.519, "step": 18446 }, { "epoch": 1.2380457031643233, "grad_norm": 1.0699265003204346, "learning_rate": 3.3465897953509006e-05, "loss": 0.5066, "step": 18448 }, { "epoch": 1.2381799268480924, "grad_norm": 1.1121770143508911, "learning_rate": 3.345564146450625e-05, "loss": 0.5127, "step": 18450 }, { "epoch": 1.2383141505318613, "grad_norm": 1.0726310014724731, "learning_rate": 3.3445385757191885e-05, "loss": 0.508, "step": 18452 }, { "epoch": 1.2384483742156303, "grad_norm": 1.135878562927246, "learning_rate": 3.3435130832050517e-05, "loss": 0.5387, "step": 18454 }, { "epoch": 1.2385825978993994, "grad_norm": 0.9350636005401611, "learning_rate": 3.342487668956661e-05, "loss": 0.5215, "step": 18456 }, { "epoch": 1.2387168215831683, "grad_norm": 1.0614064931869507, "learning_rate": 3.341462333022472e-05, "loss": 0.5485, "step": 18458 }, { "epoch": 1.2388510452669372, "grad_norm": 1.1100685596466064, "learning_rate": 3.340437075450923e-05, "loss": 0.5167, "step": 18460 }, { "epoch": 1.2389852689507064, "grad_norm": 1.136499285697937, "learning_rate": 3.339411896290462e-05, "loss": 0.553, "step": 18462 }, { "epoch": 1.2391194926344753, "grad_norm": 1.1006940603256226, "learning_rate": 3.338386795589521e-05, "loss": 0.5592, "step": 18464 }, { "epoch": 1.2392537163182444, "grad_norm": 0.9916535019874573, "learning_rate": 3.337361773396538e-05, "loss": 0.4779, "step": 18466 }, { "epoch": 1.2393879400020134, "grad_norm": 1.1026976108551025, "learning_rate": 3.336336829759941e-05, "loss": 0.5057, "step": 18468 }, { "epoch": 1.2395221636857823, "grad_norm": 1.1068087816238403, "learning_rate": 3.3353119647281575e-05, "loss": 0.4868, "step": 18470 }, { "epoch": 1.2396563873695514, "grad_norm": 1.0291463136672974, "learning_rate": 3.334287178349611e-05, "loss": 0.5074, "step": 18472 }, { "epoch": 1.2397906110533203, "grad_norm": 0.9090981483459473, "learning_rate": 3.333262470672719e-05, "loss": 0.4529, "step": 18474 }, { "epoch": 1.2399248347370895, "grad_norm": 1.090463638305664, "learning_rate": 3.332237841745898e-05, "loss": 0.5083, "step": 18476 }, { "epoch": 1.2400590584208584, "grad_norm": 1.8354182243347168, "learning_rate": 3.3312132916175586e-05, "loss": 0.4737, "step": 18478 }, { "epoch": 1.2401932821046273, "grad_norm": 1.238089919090271, "learning_rate": 3.330188820336111e-05, "loss": 0.599, "step": 18480 }, { "epoch": 1.2403275057883965, "grad_norm": 1.2819405794143677, "learning_rate": 3.329164427949957e-05, "loss": 0.5218, "step": 18482 }, { "epoch": 1.2404617294721654, "grad_norm": 0.9651596546173096, "learning_rate": 3.3281401145075e-05, "loss": 0.5906, "step": 18484 }, { "epoch": 1.2405959531559343, "grad_norm": 0.9273151159286499, "learning_rate": 3.3271158800571326e-05, "loss": 0.5921, "step": 18486 }, { "epoch": 1.2407301768397034, "grad_norm": 1.0414173603057861, "learning_rate": 3.3260917246472525e-05, "loss": 0.5502, "step": 18488 }, { "epoch": 1.2408644005234724, "grad_norm": 1.1283290386199951, "learning_rate": 3.325067648326244e-05, "loss": 0.5242, "step": 18490 }, { "epoch": 1.2409986242072413, "grad_norm": 0.974013090133667, "learning_rate": 3.324043651142499e-05, "loss": 0.4687, "step": 18492 }, { "epoch": 1.2411328478910104, "grad_norm": 0.985163688659668, "learning_rate": 3.323019733144392e-05, "loss": 0.5497, "step": 18494 }, { "epoch": 1.2412670715747793, "grad_norm": 0.972523033618927, "learning_rate": 3.3219958943803076e-05, "loss": 0.5479, "step": 18496 }, { "epoch": 1.2414012952585485, "grad_norm": 0.9983100891113281, "learning_rate": 3.3209721348986166e-05, "loss": 0.4886, "step": 18498 }, { "epoch": 1.2415355189423174, "grad_norm": 1.0459272861480713, "learning_rate": 3.3199484547476915e-05, "loss": 0.5291, "step": 18500 }, { "epoch": 1.2416697426260863, "grad_norm": 1.0113303661346436, "learning_rate": 3.3189248539758974e-05, "loss": 0.462, "step": 18502 }, { "epoch": 1.2418039663098555, "grad_norm": 0.879380464553833, "learning_rate": 3.317901332631599e-05, "loss": 0.4726, "step": 18504 }, { "epoch": 1.2419381899936244, "grad_norm": 1.1066372394561768, "learning_rate": 3.3168778907631534e-05, "loss": 0.4816, "step": 18506 }, { "epoch": 1.2420724136773933, "grad_norm": 1.1736201047897339, "learning_rate": 3.3158545284189204e-05, "loss": 0.5234, "step": 18508 }, { "epoch": 1.2422066373611624, "grad_norm": 1.0361710786819458, "learning_rate": 3.314831245647247e-05, "loss": 0.486, "step": 18510 }, { "epoch": 1.2423408610449314, "grad_norm": 1.0722993612289429, "learning_rate": 3.3138080424964845e-05, "loss": 0.5198, "step": 18512 }, { "epoch": 1.2424750847287003, "grad_norm": 1.145302176475525, "learning_rate": 3.312784919014974e-05, "loss": 0.4976, "step": 18514 }, { "epoch": 1.2426093084124694, "grad_norm": 0.9962978363037109, "learning_rate": 3.311761875251062e-05, "loss": 0.5858, "step": 18516 }, { "epoch": 1.2427435320962383, "grad_norm": 1.1119589805603027, "learning_rate": 3.3107389112530784e-05, "loss": 0.5064, "step": 18518 }, { "epoch": 1.2428777557800075, "grad_norm": 1.1276476383209229, "learning_rate": 3.309716027069362e-05, "loss": 0.5459, "step": 18520 }, { "epoch": 1.2430119794637764, "grad_norm": 0.9385380744934082, "learning_rate": 3.308693222748237e-05, "loss": 0.485, "step": 18522 }, { "epoch": 1.2431462031475453, "grad_norm": 1.0502797365188599, "learning_rate": 3.3076704983380334e-05, "loss": 0.5509, "step": 18524 }, { "epoch": 1.2432804268313145, "grad_norm": 0.9831997156143188, "learning_rate": 3.30664785388707e-05, "loss": 0.5499, "step": 18526 }, { "epoch": 1.2434146505150834, "grad_norm": 0.9537842273712158, "learning_rate": 3.305625289443667e-05, "loss": 0.4943, "step": 18528 }, { "epoch": 1.2435488741988523, "grad_norm": 0.9327796697616577, "learning_rate": 3.304602805056135e-05, "loss": 0.5067, "step": 18530 }, { "epoch": 1.2436830978826214, "grad_norm": 0.793362021446228, "learning_rate": 3.3035804007727885e-05, "loss": 0.4489, "step": 18532 }, { "epoch": 1.2438173215663904, "grad_norm": 0.9797016382217407, "learning_rate": 3.302558076641931e-05, "loss": 0.4644, "step": 18534 }, { "epoch": 1.2439515452501593, "grad_norm": 1.111334204673767, "learning_rate": 3.301535832711867e-05, "loss": 0.5596, "step": 18536 }, { "epoch": 1.2440857689339284, "grad_norm": 1.0660854578018188, "learning_rate": 3.300513669030895e-05, "loss": 0.5534, "step": 18538 }, { "epoch": 1.2442199926176973, "grad_norm": 1.0276641845703125, "learning_rate": 3.299491585647311e-05, "loss": 0.4905, "step": 18540 }, { "epoch": 1.2443542163014665, "grad_norm": 1.1686550378799438, "learning_rate": 3.298469582609405e-05, "loss": 0.5589, "step": 18542 }, { "epoch": 1.2444884399852354, "grad_norm": 1.0222535133361816, "learning_rate": 3.2974476599654646e-05, "loss": 0.5572, "step": 18544 }, { "epoch": 1.2446226636690043, "grad_norm": 1.0813424587249756, "learning_rate": 3.296425817763776e-05, "loss": 0.5217, "step": 18546 }, { "epoch": 1.2447568873527735, "grad_norm": 1.1131865978240967, "learning_rate": 3.295404056052616e-05, "loss": 0.5526, "step": 18548 }, { "epoch": 1.2448911110365424, "grad_norm": 0.9508737921714783, "learning_rate": 3.2943823748802635e-05, "loss": 0.4891, "step": 18550 }, { "epoch": 1.2450253347203115, "grad_norm": 1.0114701986312866, "learning_rate": 3.2933607742949876e-05, "loss": 0.5116, "step": 18552 }, { "epoch": 1.2451595584040804, "grad_norm": 1.0921638011932373, "learning_rate": 3.292339254345063e-05, "loss": 0.4719, "step": 18554 }, { "epoch": 1.2452937820878494, "grad_norm": 1.1859817504882812, "learning_rate": 3.2913178150787465e-05, "loss": 0.546, "step": 18556 }, { "epoch": 1.2454280057716185, "grad_norm": 1.0289673805236816, "learning_rate": 3.290296456544306e-05, "loss": 0.4925, "step": 18558 }, { "epoch": 1.2455622294553874, "grad_norm": 0.9060097932815552, "learning_rate": 3.289275178789995e-05, "loss": 0.5271, "step": 18560 }, { "epoch": 1.2456964531391563, "grad_norm": 0.955970048904419, "learning_rate": 3.288253981864068e-05, "loss": 0.5207, "step": 18562 }, { "epoch": 1.2458306768229255, "grad_norm": 0.9810526967048645, "learning_rate": 3.287232865814774e-05, "loss": 0.4588, "step": 18564 }, { "epoch": 1.2459649005066944, "grad_norm": 0.9905154705047607, "learning_rate": 3.28621183069036e-05, "loss": 0.5414, "step": 18566 }, { "epoch": 1.2460991241904633, "grad_norm": 1.0671234130859375, "learning_rate": 3.285190876539066e-05, "loss": 0.5407, "step": 18568 }, { "epoch": 1.2462333478742325, "grad_norm": 1.082667350769043, "learning_rate": 3.284170003409132e-05, "loss": 0.5051, "step": 18570 }, { "epoch": 1.2463675715580014, "grad_norm": 1.0688824653625488, "learning_rate": 3.2831492113487904e-05, "loss": 0.5403, "step": 18572 }, { "epoch": 1.2465017952417705, "grad_norm": 0.9907695055007935, "learning_rate": 3.282128500406273e-05, "loss": 0.5392, "step": 18574 }, { "epoch": 1.2466360189255394, "grad_norm": 0.8497958183288574, "learning_rate": 3.281107870629806e-05, "loss": 0.4663, "step": 18576 }, { "epoch": 1.2467702426093084, "grad_norm": 0.9073302745819092, "learning_rate": 3.280087322067612e-05, "loss": 0.5011, "step": 18578 }, { "epoch": 1.2469044662930775, "grad_norm": 1.1064773797988892, "learning_rate": 3.279066854767908e-05, "loss": 0.5251, "step": 18580 }, { "epoch": 1.2470386899768464, "grad_norm": 1.1607712507247925, "learning_rate": 3.2780464687789144e-05, "loss": 0.4841, "step": 18582 }, { "epoch": 1.2471729136606153, "grad_norm": 1.0058317184448242, "learning_rate": 3.277026164148836e-05, "loss": 0.5717, "step": 18584 }, { "epoch": 1.2473071373443845, "grad_norm": 1.129913091659546, "learning_rate": 3.2760059409258855e-05, "loss": 0.5153, "step": 18586 }, { "epoch": 1.2474413610281534, "grad_norm": 0.946969211101532, "learning_rate": 3.2749857991582635e-05, "loss": 0.4985, "step": 18588 }, { "epoch": 1.2475755847119223, "grad_norm": 1.0105154514312744, "learning_rate": 3.273965738894171e-05, "loss": 0.5492, "step": 18590 }, { "epoch": 1.2477098083956915, "grad_norm": 0.8665843605995178, "learning_rate": 3.272945760181802e-05, "loss": 0.442, "step": 18592 }, { "epoch": 1.2478440320794604, "grad_norm": 1.0189673900604248, "learning_rate": 3.271925863069351e-05, "loss": 0.511, "step": 18594 }, { "epoch": 1.2479782557632295, "grad_norm": 1.064059853553772, "learning_rate": 3.270906047605003e-05, "loss": 0.4635, "step": 18596 }, { "epoch": 1.2481124794469984, "grad_norm": 0.9782739281654358, "learning_rate": 3.269886313836947e-05, "loss": 0.4735, "step": 18598 }, { "epoch": 1.2482467031307674, "grad_norm": 0.9779257774353027, "learning_rate": 3.268866661813358e-05, "loss": 0.4897, "step": 18600 }, { "epoch": 1.2483809268145365, "grad_norm": 0.9636147022247314, "learning_rate": 3.2678470915824166e-05, "loss": 0.4963, "step": 18602 }, { "epoch": 1.2485151504983054, "grad_norm": 1.0301083326339722, "learning_rate": 3.266827603192292e-05, "loss": 0.5127, "step": 18604 }, { "epoch": 1.2486493741820743, "grad_norm": 0.9167841076850891, "learning_rate": 3.265808196691158e-05, "loss": 0.4677, "step": 18606 }, { "epoch": 1.2487835978658435, "grad_norm": 1.0668751001358032, "learning_rate": 3.264788872127173e-05, "loss": 0.5269, "step": 18608 }, { "epoch": 1.2489178215496124, "grad_norm": 0.9439718127250671, "learning_rate": 3.2637696295485055e-05, "loss": 0.5401, "step": 18610 }, { "epoch": 1.2490520452333813, "grad_norm": 0.9846333861351013, "learning_rate": 3.262750469003307e-05, "loss": 0.4892, "step": 18612 }, { "epoch": 1.2491862689171505, "grad_norm": 1.0669738054275513, "learning_rate": 3.261731390539732e-05, "loss": 0.5529, "step": 18614 }, { "epoch": 1.2493204926009194, "grad_norm": 0.9945108294487, "learning_rate": 3.260712394205933e-05, "loss": 0.473, "step": 18616 }, { "epoch": 1.2494547162846885, "grad_norm": 0.9489200711250305, "learning_rate": 3.25969348005005e-05, "loss": 0.5114, "step": 18618 }, { "epoch": 1.2495889399684574, "grad_norm": 0.9898130893707275, "learning_rate": 3.2586746481202294e-05, "loss": 0.5466, "step": 18620 }, { "epoch": 1.2497231636522264, "grad_norm": 1.1003575325012207, "learning_rate": 3.2576558984646075e-05, "loss": 0.5652, "step": 18622 }, { "epoch": 1.2498573873359955, "grad_norm": 0.9520372152328491, "learning_rate": 3.256637231131319e-05, "loss": 0.5103, "step": 18624 }, { "epoch": 1.2499916110197644, "grad_norm": 0.9153604507446289, "learning_rate": 3.255618646168491e-05, "loss": 0.533, "step": 18626 }, { "epoch": 1.2501258347035336, "grad_norm": 1.1091843843460083, "learning_rate": 3.2546001436242547e-05, "loss": 0.5138, "step": 18628 }, { "epoch": 1.2502600583873025, "grad_norm": 0.919145941734314, "learning_rate": 3.253581723546727e-05, "loss": 0.4221, "step": 18630 }, { "epoch": 1.2503942820710714, "grad_norm": 1.1235380172729492, "learning_rate": 3.25256338598403e-05, "loss": 0.5764, "step": 18632 }, { "epoch": 1.2505285057548403, "grad_norm": 1.2690304517745972, "learning_rate": 3.251545130984276e-05, "loss": 0.4321, "step": 18634 }, { "epoch": 1.2506627294386095, "grad_norm": 1.2638576030731201, "learning_rate": 3.250526958595578e-05, "loss": 0.533, "step": 18636 }, { "epoch": 1.2507969531223784, "grad_norm": 0.9857358932495117, "learning_rate": 3.249508868866039e-05, "loss": 0.4782, "step": 18638 }, { "epoch": 1.2509311768061475, "grad_norm": 1.0567487478256226, "learning_rate": 3.248490861843765e-05, "loss": 0.5068, "step": 18640 }, { "epoch": 1.2510654004899164, "grad_norm": 1.0033124685287476, "learning_rate": 3.247472937576852e-05, "loss": 0.5441, "step": 18642 }, { "epoch": 1.2511996241736854, "grad_norm": 1.047810435295105, "learning_rate": 3.246455096113399e-05, "loss": 0.5456, "step": 18644 }, { "epoch": 1.2513338478574545, "grad_norm": 1.0675344467163086, "learning_rate": 3.245437337501491e-05, "loss": 0.5262, "step": 18646 }, { "epoch": 1.2514680715412234, "grad_norm": 1.068942666053772, "learning_rate": 3.244419661789222e-05, "loss": 0.5746, "step": 18648 }, { "epoch": 1.2516022952249926, "grad_norm": 0.9902501106262207, "learning_rate": 3.24340206902467e-05, "loss": 0.4986, "step": 18650 }, { "epoch": 1.2517365189087615, "grad_norm": 0.9147794246673584, "learning_rate": 3.242384559255917e-05, "loss": 0.445, "step": 18652 }, { "epoch": 1.2518707425925304, "grad_norm": 1.0337930917739868, "learning_rate": 3.241367132531036e-05, "loss": 0.525, "step": 18654 }, { "epoch": 1.2520049662762995, "grad_norm": 0.9941875338554382, "learning_rate": 3.2403497888981015e-05, "loss": 0.478, "step": 18656 }, { "epoch": 1.2521391899600685, "grad_norm": 1.101161241531372, "learning_rate": 3.2393325284051776e-05, "loss": 0.5018, "step": 18658 }, { "epoch": 1.2522734136438374, "grad_norm": 0.95537269115448, "learning_rate": 3.238315351100332e-05, "loss": 0.5611, "step": 18660 }, { "epoch": 1.2524076373276065, "grad_norm": 1.0222139358520508, "learning_rate": 3.237298257031619e-05, "loss": 0.5171, "step": 18662 }, { "epoch": 1.2525418610113754, "grad_norm": 0.8889300227165222, "learning_rate": 3.236281246247099e-05, "loss": 0.4837, "step": 18664 }, { "epoch": 1.2526760846951444, "grad_norm": 1.0732251405715942, "learning_rate": 3.235264318794821e-05, "loss": 0.5226, "step": 18666 }, { "epoch": 1.2528103083789135, "grad_norm": 1.0490918159484863, "learning_rate": 3.234247474722835e-05, "loss": 0.4917, "step": 18668 }, { "epoch": 1.2529445320626824, "grad_norm": 1.081900954246521, "learning_rate": 3.2332307140791805e-05, "loss": 0.5014, "step": 18670 }, { "epoch": 1.2530787557464516, "grad_norm": 0.9502929449081421, "learning_rate": 3.2322140369119045e-05, "loss": 0.5165, "step": 18672 }, { "epoch": 1.2532129794302205, "grad_norm": 1.0244060754776, "learning_rate": 3.231197443269036e-05, "loss": 0.4838, "step": 18674 }, { "epoch": 1.2533472031139894, "grad_norm": 1.018920660018921, "learning_rate": 3.230180933198612e-05, "loss": 0.4737, "step": 18676 }, { "epoch": 1.2534814267977585, "grad_norm": 2.8826301097869873, "learning_rate": 3.2291645067486595e-05, "loss": 0.559, "step": 18678 }, { "epoch": 1.2536156504815275, "grad_norm": 1.0019567012786865, "learning_rate": 3.2281481639672e-05, "loss": 0.4896, "step": 18680 }, { "epoch": 1.2537498741652966, "grad_norm": 1.0178946256637573, "learning_rate": 3.227131904902257e-05, "loss": 0.4947, "step": 18682 }, { "epoch": 1.2538840978490655, "grad_norm": 0.9037721157073975, "learning_rate": 3.226115729601843e-05, "loss": 0.4808, "step": 18684 }, { "epoch": 1.2540183215328344, "grad_norm": 1.0834360122680664, "learning_rate": 3.225099638113974e-05, "loss": 0.547, "step": 18686 }, { "epoch": 1.2541525452166034, "grad_norm": 0.911413311958313, "learning_rate": 3.224083630486656e-05, "loss": 0.5445, "step": 18688 }, { "epoch": 1.2542867689003725, "grad_norm": 0.995908796787262, "learning_rate": 3.223067706767896e-05, "loss": 0.5148, "step": 18690 }, { "epoch": 1.2544209925841414, "grad_norm": 1.0455437898635864, "learning_rate": 3.2220518670056916e-05, "loss": 0.5453, "step": 18692 }, { "epoch": 1.2545552162679106, "grad_norm": 0.9652132987976074, "learning_rate": 3.221036111248041e-05, "loss": 0.4771, "step": 18694 }, { "epoch": 1.2546894399516795, "grad_norm": 2.6709609031677246, "learning_rate": 3.2200204395429344e-05, "loss": 0.5278, "step": 18696 }, { "epoch": 1.2548236636354484, "grad_norm": 0.9316728711128235, "learning_rate": 3.2190048519383635e-05, "loss": 0.4673, "step": 18698 }, { "epoch": 1.2549578873192175, "grad_norm": 0.9168113470077515, "learning_rate": 3.2179893484823106e-05, "loss": 0.4929, "step": 18700 }, { "epoch": 1.2550921110029865, "grad_norm": 1.005332589149475, "learning_rate": 3.216973929222757e-05, "loss": 0.4858, "step": 18702 }, { "epoch": 1.2552263346867556, "grad_norm": 1.124756097793579, "learning_rate": 3.2159585942076775e-05, "loss": 0.5492, "step": 18704 }, { "epoch": 1.2553605583705245, "grad_norm": 1.0058296918869019, "learning_rate": 3.214943343485049e-05, "loss": 0.4753, "step": 18706 }, { "epoch": 1.2554947820542934, "grad_norm": 1.2010509967803955, "learning_rate": 3.213928177102834e-05, "loss": 0.4831, "step": 18708 }, { "epoch": 1.2556290057380624, "grad_norm": 1.0022845268249512, "learning_rate": 3.2129130951090036e-05, "loss": 0.4934, "step": 18710 }, { "epoch": 1.2557632294218315, "grad_norm": 1.1425693035125732, "learning_rate": 3.2118980975515134e-05, "loss": 0.4571, "step": 18712 }, { "epoch": 1.2558974531056004, "grad_norm": 0.9820336699485779, "learning_rate": 3.210883184478323e-05, "loss": 0.4552, "step": 18714 }, { "epoch": 1.2560316767893696, "grad_norm": 0.9208889603614807, "learning_rate": 3.2098683559373835e-05, "loss": 0.5166, "step": 18716 }, { "epoch": 1.2561659004731385, "grad_norm": 0.9718737602233887, "learning_rate": 3.208853611976644e-05, "loss": 0.4928, "step": 18718 }, { "epoch": 1.2563001241569074, "grad_norm": 1.2219953536987305, "learning_rate": 3.2078389526440495e-05, "loss": 0.6289, "step": 18720 }, { "epoch": 1.2564343478406765, "grad_norm": 1.119890570640564, "learning_rate": 3.206824377987541e-05, "loss": 0.5022, "step": 18722 }, { "epoch": 1.2565685715244455, "grad_norm": 1.0400066375732422, "learning_rate": 3.205809888055053e-05, "loss": 0.5057, "step": 18724 }, { "epoch": 1.2567027952082146, "grad_norm": 0.9680889844894409, "learning_rate": 3.204795482894521e-05, "loss": 0.4882, "step": 18726 }, { "epoch": 1.2568370188919835, "grad_norm": 1.02334463596344, "learning_rate": 3.203781162553871e-05, "loss": 0.4791, "step": 18728 }, { "epoch": 1.2569712425757524, "grad_norm": 1.0177057981491089, "learning_rate": 3.20276692708103e-05, "loss": 0.4628, "step": 18730 }, { "epoch": 1.2571054662595216, "grad_norm": 0.9529568552970886, "learning_rate": 3.201752776523917e-05, "loss": 0.4753, "step": 18732 }, { "epoch": 1.2572396899432905, "grad_norm": 1.061122179031372, "learning_rate": 3.20073871093045e-05, "loss": 0.4968, "step": 18734 }, { "epoch": 1.2573739136270594, "grad_norm": 1.1048305034637451, "learning_rate": 3.199724730348539e-05, "loss": 0.48, "step": 18736 }, { "epoch": 1.2575081373108286, "grad_norm": 1.2485487461090088, "learning_rate": 3.198710834826096e-05, "loss": 0.5333, "step": 18738 }, { "epoch": 1.2576423609945975, "grad_norm": 0.9257562756538391, "learning_rate": 3.1976970244110234e-05, "loss": 0.4765, "step": 18740 }, { "epoch": 1.2577765846783664, "grad_norm": 1.0418267250061035, "learning_rate": 3.196683299151223e-05, "loss": 0.4768, "step": 18742 }, { "epoch": 1.2579108083621355, "grad_norm": 1.0285526514053345, "learning_rate": 3.19566965909459e-05, "loss": 0.5373, "step": 18744 }, { "epoch": 1.2580450320459045, "grad_norm": 0.8784838914871216, "learning_rate": 3.1946561042890174e-05, "loss": 0.4501, "step": 18746 }, { "epoch": 1.2581792557296736, "grad_norm": 0.9794188737869263, "learning_rate": 3.1936426347823955e-05, "loss": 0.4713, "step": 18748 }, { "epoch": 1.2583134794134425, "grad_norm": 0.976508378982544, "learning_rate": 3.1926292506226054e-05, "loss": 0.5099, "step": 18750 }, { "epoch": 1.2584477030972114, "grad_norm": 0.9098379015922546, "learning_rate": 3.1916159518575294e-05, "loss": 0.5241, "step": 18752 }, { "epoch": 1.2585819267809806, "grad_norm": 1.3595097064971924, "learning_rate": 3.190602738535043e-05, "loss": 0.5093, "step": 18754 }, { "epoch": 1.2587161504647495, "grad_norm": 1.1624239683151245, "learning_rate": 3.189589610703021e-05, "loss": 0.4938, "step": 18756 }, { "epoch": 1.2588503741485186, "grad_norm": 1.031873106956482, "learning_rate": 3.188576568409328e-05, "loss": 0.5355, "step": 18758 }, { "epoch": 1.2589845978322876, "grad_norm": 1.002721905708313, "learning_rate": 3.187563611701832e-05, "loss": 0.4826, "step": 18760 }, { "epoch": 1.2591188215160565, "grad_norm": 1.8391629457473755, "learning_rate": 3.18655074062839e-05, "loss": 0.4936, "step": 18762 }, { "epoch": 1.2592530451998254, "grad_norm": 1.0190123319625854, "learning_rate": 3.1855379552368605e-05, "loss": 0.612, "step": 18764 }, { "epoch": 1.2593872688835945, "grad_norm": 0.82330721616745, "learning_rate": 3.184525255575092e-05, "loss": 0.4679, "step": 18766 }, { "epoch": 1.2595214925673635, "grad_norm": 1.0846850872039795, "learning_rate": 3.183512641690939e-05, "loss": 0.4938, "step": 18768 }, { "epoch": 1.2596557162511326, "grad_norm": 1.0195902585983276, "learning_rate": 3.182500113632238e-05, "loss": 0.506, "step": 18770 }, { "epoch": 1.2597899399349015, "grad_norm": 1.0187736749649048, "learning_rate": 3.181487671446836e-05, "loss": 0.5466, "step": 18772 }, { "epoch": 1.2599241636186704, "grad_norm": 1.1696752309799194, "learning_rate": 3.180475315182563e-05, "loss": 0.4986, "step": 18774 }, { "epoch": 1.2600583873024396, "grad_norm": 0.9804502725601196, "learning_rate": 3.1794630448872545e-05, "loss": 0.4734, "step": 18776 }, { "epoch": 1.2601926109862085, "grad_norm": 1.367415189743042, "learning_rate": 3.1784508606087374e-05, "loss": 0.4764, "step": 18778 }, { "epoch": 1.2603268346699776, "grad_norm": 0.9473814368247986, "learning_rate": 3.177438762394835e-05, "loss": 0.5156, "step": 18780 }, { "epoch": 1.2604610583537466, "grad_norm": 1.252681851387024, "learning_rate": 3.1764267502933666e-05, "loss": 0.4782, "step": 18782 }, { "epoch": 1.2605952820375155, "grad_norm": 1.0670397281646729, "learning_rate": 3.17541482435215e-05, "loss": 0.5068, "step": 18784 }, { "epoch": 1.2607295057212844, "grad_norm": 0.9958670139312744, "learning_rate": 3.1744029846189925e-05, "loss": 0.5183, "step": 18786 }, { "epoch": 1.2608637294050535, "grad_norm": 0.9645728468894958, "learning_rate": 3.173391231141706e-05, "loss": 0.4881, "step": 18788 }, { "epoch": 1.2609979530888225, "grad_norm": 1.0699021816253662, "learning_rate": 3.172379563968092e-05, "loss": 0.5038, "step": 18790 }, { "epoch": 1.2611321767725916, "grad_norm": 0.9757247567176819, "learning_rate": 3.171367983145949e-05, "loss": 0.5218, "step": 18792 }, { "epoch": 1.2612664004563605, "grad_norm": 0.9883769154548645, "learning_rate": 3.170356488723072e-05, "loss": 0.4885, "step": 18794 }, { "epoch": 1.2614006241401294, "grad_norm": 0.9620486497879028, "learning_rate": 3.169345080747256e-05, "loss": 0.5286, "step": 18796 }, { "epoch": 1.2615348478238986, "grad_norm": 1.0763846635818481, "learning_rate": 3.168333759266282e-05, "loss": 0.5452, "step": 18798 }, { "epoch": 1.2616690715076675, "grad_norm": 1.1380316019058228, "learning_rate": 3.167322524327938e-05, "loss": 0.5408, "step": 18800 }, { "epoch": 1.2618032951914366, "grad_norm": 1.1167099475860596, "learning_rate": 3.166311375979999e-05, "loss": 0.4681, "step": 18802 }, { "epoch": 1.2619375188752056, "grad_norm": 0.990744948387146, "learning_rate": 3.1653003142702444e-05, "loss": 0.5582, "step": 18804 }, { "epoch": 1.2620717425589745, "grad_norm": 0.7896711230278015, "learning_rate": 3.1642893392464404e-05, "loss": 0.4751, "step": 18806 }, { "epoch": 1.2622059662427434, "grad_norm": 1.1622854471206665, "learning_rate": 3.1632784509563565e-05, "loss": 0.501, "step": 18808 }, { "epoch": 1.2623401899265125, "grad_norm": 0.9174784421920776, "learning_rate": 3.162267649447752e-05, "loss": 0.5044, "step": 18810 }, { "epoch": 1.2624744136102815, "grad_norm": 1.0127310752868652, "learning_rate": 3.1612569347683895e-05, "loss": 0.4747, "step": 18812 }, { "epoch": 1.2626086372940506, "grad_norm": 1.002046823501587, "learning_rate": 3.160246306966019e-05, "loss": 0.5464, "step": 18814 }, { "epoch": 1.2627428609778195, "grad_norm": 0.9624472856521606, "learning_rate": 3.1592357660883954e-05, "loss": 0.466, "step": 18816 }, { "epoch": 1.2628770846615884, "grad_norm": 1.0694067478179932, "learning_rate": 3.158225312183261e-05, "loss": 0.5229, "step": 18818 }, { "epoch": 1.2630113083453576, "grad_norm": 0.891897439956665, "learning_rate": 3.157214945298358e-05, "loss": 0.4503, "step": 18820 }, { "epoch": 1.2631455320291265, "grad_norm": 0.9390667676925659, "learning_rate": 3.156204665481426e-05, "loss": 0.4689, "step": 18822 }, { "epoch": 1.2632797557128956, "grad_norm": 1.0218350887298584, "learning_rate": 3.155194472780197e-05, "loss": 0.5263, "step": 18824 }, { "epoch": 1.2634139793966646, "grad_norm": 1.1957632303237915, "learning_rate": 3.154184367242403e-05, "loss": 0.5302, "step": 18826 }, { "epoch": 1.2635482030804335, "grad_norm": 1.0991414785385132, "learning_rate": 3.1531743489157664e-05, "loss": 0.5263, "step": 18828 }, { "epoch": 1.2636824267642026, "grad_norm": 1.0328730344772339, "learning_rate": 3.152164417848012e-05, "loss": 0.5064, "step": 18830 }, { "epoch": 1.2638166504479715, "grad_norm": 1.0690430402755737, "learning_rate": 3.151154574086853e-05, "loss": 0.4402, "step": 18832 }, { "epoch": 1.2639508741317407, "grad_norm": 1.054821491241455, "learning_rate": 3.150144817680008e-05, "loss": 0.4937, "step": 18834 }, { "epoch": 1.2640850978155096, "grad_norm": 1.02252197265625, "learning_rate": 3.149135148675179e-05, "loss": 0.5199, "step": 18836 }, { "epoch": 1.2642193214992785, "grad_norm": 0.9816807508468628, "learning_rate": 3.148125567120076e-05, "loss": 0.488, "step": 18838 }, { "epoch": 1.2643535451830474, "grad_norm": 0.902094841003418, "learning_rate": 3.147116073062399e-05, "loss": 0.4548, "step": 18840 }, { "epoch": 1.2644877688668166, "grad_norm": 1.0778688192367554, "learning_rate": 3.146106666549844e-05, "loss": 0.5177, "step": 18842 }, { "epoch": 1.2646219925505855, "grad_norm": 0.8925482630729675, "learning_rate": 3.1450973476301026e-05, "loss": 0.4492, "step": 18844 }, { "epoch": 1.2647562162343546, "grad_norm": 0.9562098383903503, "learning_rate": 3.144088116350866e-05, "loss": 0.4942, "step": 18846 }, { "epoch": 1.2648904399181236, "grad_norm": 0.9487055540084839, "learning_rate": 3.1430789727598145e-05, "loss": 0.5028, "step": 18848 }, { "epoch": 1.2650246636018925, "grad_norm": 1.0395872592926025, "learning_rate": 3.142069916904631e-05, "loss": 0.4787, "step": 18850 }, { "epoch": 1.2651588872856616, "grad_norm": 0.9606418013572693, "learning_rate": 3.14106094883299e-05, "loss": 0.5069, "step": 18852 }, { "epoch": 1.2652931109694305, "grad_norm": 1.210218071937561, "learning_rate": 3.140052068592564e-05, "loss": 0.5159, "step": 18854 }, { "epoch": 1.2654273346531997, "grad_norm": 1.0063384771347046, "learning_rate": 3.139043276231019e-05, "loss": 0.4747, "step": 18856 }, { "epoch": 1.2655615583369686, "grad_norm": 0.8223646283149719, "learning_rate": 3.138034571796022e-05, "loss": 0.4376, "step": 18858 }, { "epoch": 1.2656957820207375, "grad_norm": 1.190323829650879, "learning_rate": 3.1370259553352274e-05, "loss": 0.5213, "step": 18860 }, { "epoch": 1.2658300057045064, "grad_norm": 1.1596482992172241, "learning_rate": 3.1360174268962973e-05, "loss": 0.4962, "step": 18862 }, { "epoch": 1.2659642293882756, "grad_norm": 0.9268326163291931, "learning_rate": 3.135008986526874e-05, "loss": 0.4507, "step": 18864 }, { "epoch": 1.2660984530720445, "grad_norm": 1.0648741722106934, "learning_rate": 3.1340006342746107e-05, "loss": 0.5527, "step": 18866 }, { "epoch": 1.2662326767558136, "grad_norm": 1.0832303762435913, "learning_rate": 3.132992370187148e-05, "loss": 0.5718, "step": 18868 }, { "epoch": 1.2663669004395826, "grad_norm": 1.0149818658828735, "learning_rate": 3.131984194312125e-05, "loss": 0.5047, "step": 18870 }, { "epoch": 1.2665011241233515, "grad_norm": 0.9199078679084778, "learning_rate": 3.130976106697174e-05, "loss": 0.4241, "step": 18872 }, { "epoch": 1.2666353478071206, "grad_norm": 1.0143941640853882, "learning_rate": 3.1299681073899284e-05, "loss": 0.4366, "step": 18874 }, { "epoch": 1.2667695714908895, "grad_norm": 1.0422862768173218, "learning_rate": 3.128960196438011e-05, "loss": 0.4941, "step": 18876 }, { "epoch": 1.2669037951746587, "grad_norm": 1.0042037963867188, "learning_rate": 3.127952373889046e-05, "loss": 0.4664, "step": 18878 }, { "epoch": 1.2670380188584276, "grad_norm": 2.3638975620269775, "learning_rate": 3.12694463979065e-05, "loss": 0.4863, "step": 18880 }, { "epoch": 1.2671722425421965, "grad_norm": 1.1389439105987549, "learning_rate": 3.1259369941904374e-05, "loss": 0.5407, "step": 18882 }, { "epoch": 1.2673064662259654, "grad_norm": 1.0771986246109009, "learning_rate": 3.124929437136016e-05, "loss": 0.4671, "step": 18884 }, { "epoch": 1.2674406899097346, "grad_norm": 0.9763202667236328, "learning_rate": 3.123921968674991e-05, "loss": 0.5225, "step": 18886 }, { "epoch": 1.2675749135935035, "grad_norm": 0.9527340531349182, "learning_rate": 3.122914588854964e-05, "loss": 0.4885, "step": 18888 }, { "epoch": 1.2677091372772726, "grad_norm": 1.0093185901641846, "learning_rate": 3.1219072977235305e-05, "loss": 0.5397, "step": 18890 }, { "epoch": 1.2678433609610416, "grad_norm": 0.91331547498703, "learning_rate": 3.120900095328286e-05, "loss": 0.5074, "step": 18892 }, { "epoch": 1.2679775846448105, "grad_norm": 1.0820149183273315, "learning_rate": 3.1198929817168154e-05, "loss": 0.5128, "step": 18894 }, { "epoch": 1.2681118083285796, "grad_norm": 1.0016778707504272, "learning_rate": 3.118885956936706e-05, "loss": 0.5063, "step": 18896 }, { "epoch": 1.2682460320123485, "grad_norm": 0.9243509769439697, "learning_rate": 3.117879021035534e-05, "loss": 0.5134, "step": 18898 }, { "epoch": 1.2683802556961177, "grad_norm": 1.0297565460205078, "learning_rate": 3.116872174060878e-05, "loss": 0.5065, "step": 18900 }, { "epoch": 1.2685144793798866, "grad_norm": 1.028468132019043, "learning_rate": 3.1158654160603096e-05, "loss": 0.5198, "step": 18902 }, { "epoch": 1.2686487030636555, "grad_norm": 1.0754987001419067, "learning_rate": 3.114858747081395e-05, "loss": 0.5308, "step": 18904 }, { "epoch": 1.2687829267474247, "grad_norm": 0.9786719083786011, "learning_rate": 3.113852167171697e-05, "loss": 0.4841, "step": 18906 }, { "epoch": 1.2689171504311936, "grad_norm": 0.9417799115180969, "learning_rate": 3.112845676378776e-05, "loss": 0.4579, "step": 18908 }, { "epoch": 1.2690513741149627, "grad_norm": 1.0206085443496704, "learning_rate": 3.1118392747501854e-05, "loss": 0.5222, "step": 18910 }, { "epoch": 1.2691855977987316, "grad_norm": 0.9408379197120667, "learning_rate": 3.1108329623334766e-05, "loss": 0.494, "step": 18912 }, { "epoch": 1.2693198214825006, "grad_norm": 0.9950227737426758, "learning_rate": 3.109826739176195e-05, "loss": 0.4814, "step": 18914 }, { "epoch": 1.2694540451662695, "grad_norm": 1.0771654844284058, "learning_rate": 3.108820605325883e-05, "loss": 0.5509, "step": 18916 }, { "epoch": 1.2695882688500386, "grad_norm": 1.1590389013290405, "learning_rate": 3.1078145608300776e-05, "loss": 0.5334, "step": 18918 }, { "epoch": 1.2697224925338075, "grad_norm": 0.8306449055671692, "learning_rate": 3.1068086057363155e-05, "loss": 0.4865, "step": 18920 }, { "epoch": 1.2698567162175767, "grad_norm": 1.0557630062103271, "learning_rate": 3.105802740092122e-05, "loss": 0.507, "step": 18922 }, { "epoch": 1.2699909399013456, "grad_norm": 1.056467890739441, "learning_rate": 3.104796963945027e-05, "loss": 0.5766, "step": 18924 }, { "epoch": 1.2701251635851145, "grad_norm": 1.1036617755889893, "learning_rate": 3.103791277342545e-05, "loss": 0.489, "step": 18926 }, { "epoch": 1.2702593872688837, "grad_norm": 1.1211748123168945, "learning_rate": 3.102785680332199e-05, "loss": 0.5598, "step": 18928 }, { "epoch": 1.2703936109526526, "grad_norm": 1.1051607131958008, "learning_rate": 3.101780172961497e-05, "loss": 0.5387, "step": 18930 }, { "epoch": 1.2705278346364217, "grad_norm": 1.058781623840332, "learning_rate": 3.1007747552779515e-05, "loss": 0.5104, "step": 18932 }, { "epoch": 1.2706620583201906, "grad_norm": 0.948963463306427, "learning_rate": 3.099769427329062e-05, "loss": 0.5391, "step": 18934 }, { "epoch": 1.2707962820039596, "grad_norm": 1.0071995258331299, "learning_rate": 3.098764189162332e-05, "loss": 0.4627, "step": 18936 }, { "epoch": 1.2709305056877285, "grad_norm": 1.0042026042938232, "learning_rate": 3.0977590408252546e-05, "loss": 0.5054, "step": 18938 }, { "epoch": 1.2710647293714976, "grad_norm": 0.9338868856430054, "learning_rate": 3.096753982365323e-05, "loss": 0.5491, "step": 18940 }, { "epoch": 1.2711989530552665, "grad_norm": 1.0394976139068604, "learning_rate": 3.0957490138300215e-05, "loss": 0.5156, "step": 18942 }, { "epoch": 1.2713331767390357, "grad_norm": 0.9292700886726379, "learning_rate": 3.094744135266836e-05, "loss": 0.4393, "step": 18944 }, { "epoch": 1.2714674004228046, "grad_norm": 1.1042059659957886, "learning_rate": 3.093739346723242e-05, "loss": 0.4668, "step": 18946 }, { "epoch": 1.2716016241065735, "grad_norm": 1.0075424909591675, "learning_rate": 3.092734648246718e-05, "loss": 0.5124, "step": 18948 }, { "epoch": 1.2717358477903427, "grad_norm": 1.0175364017486572, "learning_rate": 3.091730039884728e-05, "loss": 0.5205, "step": 18950 }, { "epoch": 1.2718700714741116, "grad_norm": 1.0737347602844238, "learning_rate": 3.090725521684744e-05, "loss": 0.4697, "step": 18952 }, { "epoch": 1.2720042951578807, "grad_norm": 0.8391603231430054, "learning_rate": 3.089721093694224e-05, "loss": 0.4156, "step": 18954 }, { "epoch": 1.2721385188416496, "grad_norm": 0.9844167232513428, "learning_rate": 3.088716755960624e-05, "loss": 0.5035, "step": 18956 }, { "epoch": 1.2722727425254186, "grad_norm": 1.1497169733047485, "learning_rate": 3.087712508531402e-05, "loss": 0.5162, "step": 18958 }, { "epoch": 1.2724069662091875, "grad_norm": 1.0359950065612793, "learning_rate": 3.0867083514540004e-05, "loss": 0.5119, "step": 18960 }, { "epoch": 1.2725411898929566, "grad_norm": 1.0207481384277344, "learning_rate": 3.085704284775869e-05, "loss": 0.457, "step": 18962 }, { "epoch": 1.2726754135767255, "grad_norm": 1.1166799068450928, "learning_rate": 3.084700308544445e-05, "loss": 0.5362, "step": 18964 }, { "epoch": 1.2728096372604947, "grad_norm": 1.011577844619751, "learning_rate": 3.083696422807166e-05, "loss": 0.5093, "step": 18966 }, { "epoch": 1.2729438609442636, "grad_norm": 0.7684579491615295, "learning_rate": 3.082692627611462e-05, "loss": 0.4695, "step": 18968 }, { "epoch": 1.2730780846280325, "grad_norm": 1.02020263671875, "learning_rate": 3.081688923004762e-05, "loss": 0.5184, "step": 18970 }, { "epoch": 1.2732123083118017, "grad_norm": 1.0614569187164307, "learning_rate": 3.080685309034487e-05, "loss": 0.5676, "step": 18972 }, { "epoch": 1.2733465319955706, "grad_norm": 1.0228722095489502, "learning_rate": 3.079681785748057e-05, "loss": 0.5098, "step": 18974 }, { "epoch": 1.2734807556793397, "grad_norm": 0.8926354050636292, "learning_rate": 3.078678353192887e-05, "loss": 0.4811, "step": 18976 }, { "epoch": 1.2736149793631086, "grad_norm": 0.8983010053634644, "learning_rate": 3.0776750114163876e-05, "loss": 0.4778, "step": 18978 }, { "epoch": 1.2737492030468776, "grad_norm": 1.0334417819976807, "learning_rate": 3.076671760465961e-05, "loss": 0.5023, "step": 18980 }, { "epoch": 1.2738834267306467, "grad_norm": 1.8472157716751099, "learning_rate": 3.0756686003890156e-05, "loss": 0.5451, "step": 18982 }, { "epoch": 1.2740176504144156, "grad_norm": 0.9782081842422485, "learning_rate": 3.074665531232941e-05, "loss": 0.4861, "step": 18984 }, { "epoch": 1.2741518740981848, "grad_norm": 0.9048913717269897, "learning_rate": 3.073662553045138e-05, "loss": 0.4748, "step": 18986 }, { "epoch": 1.2742860977819537, "grad_norm": 1.0437411069869995, "learning_rate": 3.072659665872988e-05, "loss": 0.4714, "step": 18988 }, { "epoch": 1.2744203214657226, "grad_norm": 0.9941788911819458, "learning_rate": 3.07165686976388e-05, "loss": 0.5814, "step": 18990 }, { "epoch": 1.2745545451494915, "grad_norm": 1.147536277770996, "learning_rate": 3.070654164765193e-05, "loss": 0.5174, "step": 18992 }, { "epoch": 1.2746887688332607, "grad_norm": 1.0783060789108276, "learning_rate": 3.0696515509243026e-05, "loss": 0.487, "step": 18994 }, { "epoch": 1.2748229925170296, "grad_norm": 0.9300075173377991, "learning_rate": 3.068649028288581e-05, "loss": 0.5002, "step": 18996 }, { "epoch": 1.2749572162007987, "grad_norm": 1.0361125469207764, "learning_rate": 3.0676465969053946e-05, "loss": 0.5958, "step": 18998 }, { "epoch": 1.2750914398845676, "grad_norm": 1.1151272058486938, "learning_rate": 3.0666442568221065e-05, "loss": 0.5107, "step": 19000 }, { "epoch": 1.2752256635683366, "grad_norm": 0.9330003261566162, "learning_rate": 3.065642008086076e-05, "loss": 0.4949, "step": 19002 }, { "epoch": 1.2753598872521057, "grad_norm": 0.9569537043571472, "learning_rate": 3.064639850744657e-05, "loss": 0.5493, "step": 19004 }, { "epoch": 1.2754941109358746, "grad_norm": 1.061761736869812, "learning_rate": 3.063637784845199e-05, "loss": 0.5463, "step": 19006 }, { "epoch": 1.2756283346196438, "grad_norm": 1.055582880973816, "learning_rate": 3.0626358104350475e-05, "loss": 0.4809, "step": 19008 }, { "epoch": 1.2757625583034127, "grad_norm": 1.039762020111084, "learning_rate": 3.0616339275615455e-05, "loss": 0.5304, "step": 19010 }, { "epoch": 1.2758967819871816, "grad_norm": 1.0182931423187256, "learning_rate": 3.0606321362720267e-05, "loss": 0.4901, "step": 19012 }, { "epoch": 1.2760310056709505, "grad_norm": 0.9698779582977295, "learning_rate": 3.059630436613829e-05, "loss": 0.4836, "step": 19014 }, { "epoch": 1.2761652293547197, "grad_norm": 0.9517755508422852, "learning_rate": 3.058628828634274e-05, "loss": 0.482, "step": 19016 }, { "epoch": 1.2762994530384886, "grad_norm": 1.0961393117904663, "learning_rate": 3.057627312380692e-05, "loss": 0.5077, "step": 19018 }, { "epoch": 1.2764336767222577, "grad_norm": 1.094814419746399, "learning_rate": 3.0566258879003986e-05, "loss": 0.5784, "step": 19020 }, { "epoch": 1.2765679004060266, "grad_norm": 1.0417190790176392, "learning_rate": 3.055624555240711e-05, "loss": 0.4749, "step": 19022 }, { "epoch": 1.2767021240897956, "grad_norm": 1.007354497909546, "learning_rate": 3.0546233144489414e-05, "loss": 0.4932, "step": 19024 }, { "epoch": 1.2768363477735647, "grad_norm": 1.0553150177001953, "learning_rate": 3.053622165572392e-05, "loss": 0.4851, "step": 19026 }, { "epoch": 1.2769705714573336, "grad_norm": 0.872083306312561, "learning_rate": 3.05262110865837e-05, "loss": 0.4629, "step": 19028 }, { "epoch": 1.2771047951411028, "grad_norm": 0.9458115100860596, "learning_rate": 3.0516201437541693e-05, "loss": 0.5173, "step": 19030 }, { "epoch": 1.2772390188248717, "grad_norm": 1.1611723899841309, "learning_rate": 3.0506192709070868e-05, "loss": 0.5351, "step": 19032 }, { "epoch": 1.2773732425086406, "grad_norm": 1.0064661502838135, "learning_rate": 3.04961849016441e-05, "loss": 0.4785, "step": 19034 }, { "epoch": 1.2775074661924095, "grad_norm": 1.0834848880767822, "learning_rate": 3.0486178015734246e-05, "loss": 0.5189, "step": 19036 }, { "epoch": 1.2776416898761787, "grad_norm": 1.0904122591018677, "learning_rate": 3.04761720518141e-05, "loss": 0.559, "step": 19038 }, { "epoch": 1.2777759135599476, "grad_norm": 0.9285237193107605, "learning_rate": 3.046616701035645e-05, "loss": 0.4663, "step": 19040 }, { "epoch": 1.2779101372437167, "grad_norm": 1.1051727533340454, "learning_rate": 3.0456162891833978e-05, "loss": 0.4522, "step": 19042 }, { "epoch": 1.2780443609274856, "grad_norm": 0.9298257231712341, "learning_rate": 3.044615969671939e-05, "loss": 0.5165, "step": 19044 }, { "epoch": 1.2781785846112546, "grad_norm": 0.9940264225006104, "learning_rate": 3.043615742548529e-05, "loss": 0.4633, "step": 19046 }, { "epoch": 1.2783128082950237, "grad_norm": 1.0981922149658203, "learning_rate": 3.0426156078604305e-05, "loss": 0.4873, "step": 19048 }, { "epoch": 1.2784470319787926, "grad_norm": 1.0183321237564087, "learning_rate": 3.0416155656548928e-05, "loss": 0.5701, "step": 19050 }, { "epoch": 1.2785812556625618, "grad_norm": 1.0703001022338867, "learning_rate": 3.040615615979171e-05, "loss": 0.5107, "step": 19052 }, { "epoch": 1.2787154793463307, "grad_norm": 1.0372713804244995, "learning_rate": 3.0396157588805068e-05, "loss": 0.4989, "step": 19054 }, { "epoch": 1.2788497030300996, "grad_norm": 0.9407340288162231, "learning_rate": 3.0386159944061444e-05, "loss": 0.4823, "step": 19056 }, { "epoch": 1.2789839267138687, "grad_norm": 1.0311121940612793, "learning_rate": 3.0376163226033183e-05, "loss": 0.485, "step": 19058 }, { "epoch": 1.2791181503976377, "grad_norm": 1.3180968761444092, "learning_rate": 3.0366167435192627e-05, "loss": 0.5411, "step": 19060 }, { "epoch": 1.2792523740814068, "grad_norm": 1.1690349578857422, "learning_rate": 3.035617257201204e-05, "loss": 0.4954, "step": 19062 }, { "epoch": 1.2793865977651757, "grad_norm": 1.024618148803711, "learning_rate": 3.034617863696369e-05, "loss": 0.5332, "step": 19064 }, { "epoch": 1.2795208214489446, "grad_norm": 1.085595726966858, "learning_rate": 3.0336185630519737e-05, "loss": 0.5139, "step": 19066 }, { "epoch": 1.2796550451327136, "grad_norm": 0.8518025875091553, "learning_rate": 3.032619355315236e-05, "loss": 0.4813, "step": 19068 }, { "epoch": 1.2797892688164827, "grad_norm": 1.0916320085525513, "learning_rate": 3.0316202405333626e-05, "loss": 0.5583, "step": 19070 }, { "epoch": 1.2799234925002516, "grad_norm": 1.020693063735962, "learning_rate": 3.0306212187535653e-05, "loss": 0.4914, "step": 19072 }, { "epoch": 1.2800577161840208, "grad_norm": 1.0341733694076538, "learning_rate": 3.0296222900230397e-05, "loss": 0.4996, "step": 19074 }, { "epoch": 1.2801919398677897, "grad_norm": 1.2896177768707275, "learning_rate": 3.0286234543889892e-05, "loss": 0.6399, "step": 19076 }, { "epoch": 1.2803261635515586, "grad_norm": 0.9945932626724243, "learning_rate": 3.0276247118986013e-05, "loss": 0.5082, "step": 19078 }, { "epoch": 1.2804603872353277, "grad_norm": 0.8634927272796631, "learning_rate": 3.026626062599069e-05, "loss": 0.4289, "step": 19080 }, { "epoch": 1.2805946109190967, "grad_norm": 1.334445834159851, "learning_rate": 3.025627506537574e-05, "loss": 0.5118, "step": 19082 }, { "epoch": 1.2807288346028658, "grad_norm": 0.8973318934440613, "learning_rate": 3.0246290437612974e-05, "loss": 0.4608, "step": 19084 }, { "epoch": 1.2808630582866347, "grad_norm": 1.1655054092407227, "learning_rate": 3.0236306743174135e-05, "loss": 0.5179, "step": 19086 }, { "epoch": 1.2809972819704036, "grad_norm": 1.0608400106430054, "learning_rate": 3.0226323982530946e-05, "loss": 0.4694, "step": 19088 }, { "epoch": 1.2811315056541726, "grad_norm": 1.2491742372512817, "learning_rate": 3.0216342156155063e-05, "loss": 0.5309, "step": 19090 }, { "epoch": 1.2812657293379417, "grad_norm": 1.1221964359283447, "learning_rate": 3.0206361264518106e-05, "loss": 0.5641, "step": 19092 }, { "epoch": 1.2813999530217106, "grad_norm": 0.9746096730232239, "learning_rate": 3.019638130809167e-05, "loss": 0.4834, "step": 19094 }, { "epoch": 1.2815341767054798, "grad_norm": 1.0384272336959839, "learning_rate": 3.018640228734726e-05, "loss": 0.5042, "step": 19096 }, { "epoch": 1.2816684003892487, "grad_norm": 0.9357865452766418, "learning_rate": 3.0176424202756388e-05, "loss": 0.4443, "step": 19098 }, { "epoch": 1.2818026240730176, "grad_norm": 0.948830783367157, "learning_rate": 3.016644705479048e-05, "loss": 0.5083, "step": 19100 }, { "epoch": 1.2819368477567867, "grad_norm": 0.9100587368011475, "learning_rate": 3.0156470843920965e-05, "loss": 0.4692, "step": 19102 }, { "epoch": 1.2820710714405557, "grad_norm": 1.0029624700546265, "learning_rate": 3.014649557061917e-05, "loss": 0.4883, "step": 19104 }, { "epoch": 1.2822052951243248, "grad_norm": 1.0393444299697876, "learning_rate": 3.0136521235356425e-05, "loss": 0.5089, "step": 19106 }, { "epoch": 1.2823395188080937, "grad_norm": 1.058200716972351, "learning_rate": 3.0126547838603976e-05, "loss": 0.5405, "step": 19108 }, { "epoch": 1.2824737424918626, "grad_norm": 1.0188109874725342, "learning_rate": 3.0116575380833088e-05, "loss": 0.5337, "step": 19110 }, { "epoch": 1.2826079661756316, "grad_norm": 1.1087276935577393, "learning_rate": 3.0106603862514882e-05, "loss": 0.4425, "step": 19112 }, { "epoch": 1.2827421898594007, "grad_norm": 0.9219624400138855, "learning_rate": 3.0096633284120556e-05, "loss": 0.4702, "step": 19114 }, { "epoch": 1.2828764135431696, "grad_norm": 0.9535489082336426, "learning_rate": 3.008666364612115e-05, "loss": 0.4899, "step": 19116 }, { "epoch": 1.2830106372269388, "grad_norm": 0.8020244240760803, "learning_rate": 3.0076694948987738e-05, "loss": 0.5104, "step": 19118 }, { "epoch": 1.2831448609107077, "grad_norm": 0.9598153233528137, "learning_rate": 3.0066727193191307e-05, "loss": 0.4854, "step": 19120 }, { "epoch": 1.2832790845944766, "grad_norm": 1.084541916847229, "learning_rate": 3.0056760379202824e-05, "loss": 0.5471, "step": 19122 }, { "epoch": 1.2834133082782457, "grad_norm": 1.1065452098846436, "learning_rate": 3.0046794507493192e-05, "loss": 0.4721, "step": 19124 }, { "epoch": 1.2835475319620147, "grad_norm": 0.9663994312286377, "learning_rate": 3.0036829578533287e-05, "loss": 0.4458, "step": 19126 }, { "epoch": 1.2836817556457838, "grad_norm": 0.9064396619796753, "learning_rate": 3.0026865592793928e-05, "loss": 0.4571, "step": 19128 }, { "epoch": 1.2838159793295527, "grad_norm": 1.063199520111084, "learning_rate": 3.0016902550745897e-05, "loss": 0.5314, "step": 19130 }, { "epoch": 1.2839502030133216, "grad_norm": 0.9132474064826965, "learning_rate": 3.0006940452859916e-05, "loss": 0.5467, "step": 19132 }, { "epoch": 1.2840844266970908, "grad_norm": 1.0902132987976074, "learning_rate": 2.9996979299606697e-05, "loss": 0.5334, "step": 19134 }, { "epoch": 1.2842186503808597, "grad_norm": 0.9519376158714294, "learning_rate": 2.9987019091456846e-05, "loss": 0.4686, "step": 19136 }, { "epoch": 1.2843528740646288, "grad_norm": 1.072738766670227, "learning_rate": 2.9977059828881027e-05, "loss": 0.5224, "step": 19138 }, { "epoch": 1.2844870977483978, "grad_norm": 1.204823613166809, "learning_rate": 2.996710151234972e-05, "loss": 0.6014, "step": 19140 }, { "epoch": 1.2846213214321667, "grad_norm": 1.0488771200180054, "learning_rate": 2.9957144142333494e-05, "loss": 0.4384, "step": 19142 }, { "epoch": 1.2847555451159356, "grad_norm": 0.9874991774559021, "learning_rate": 2.9947187719302792e-05, "loss": 0.4718, "step": 19144 }, { "epoch": 1.2848897687997047, "grad_norm": 1.00027334690094, "learning_rate": 2.993723224372804e-05, "loss": 0.5419, "step": 19146 }, { "epoch": 1.2850239924834737, "grad_norm": 1.0105254650115967, "learning_rate": 2.9927277716079605e-05, "loss": 0.5151, "step": 19148 }, { "epoch": 1.2851582161672428, "grad_norm": 0.9950233697891235, "learning_rate": 2.991732413682784e-05, "loss": 0.487, "step": 19150 }, { "epoch": 1.2852924398510117, "grad_norm": 1.175166130065918, "learning_rate": 2.9907371506443003e-05, "loss": 0.5578, "step": 19152 }, { "epoch": 1.2854266635347806, "grad_norm": 1.0153224468231201, "learning_rate": 2.9897419825395367e-05, "loss": 0.4839, "step": 19154 }, { "epoch": 1.2855608872185498, "grad_norm": 1.0609861612319946, "learning_rate": 2.9887469094155108e-05, "loss": 0.5828, "step": 19156 }, { "epoch": 1.2856951109023187, "grad_norm": 1.1728836297988892, "learning_rate": 2.9877519313192386e-05, "loss": 0.5589, "step": 19158 }, { "epoch": 1.2858293345860878, "grad_norm": 1.1383951902389526, "learning_rate": 2.9867570482977316e-05, "loss": 0.5306, "step": 19160 }, { "epoch": 1.2859635582698568, "grad_norm": 1.0976660251617432, "learning_rate": 2.9857622603979933e-05, "loss": 0.5036, "step": 19162 }, { "epoch": 1.2860977819536257, "grad_norm": 0.9632569551467896, "learning_rate": 2.9847675676670285e-05, "loss": 0.5693, "step": 19164 }, { "epoch": 1.2862320056373946, "grad_norm": 1.1621909141540527, "learning_rate": 2.9837729701518325e-05, "loss": 0.4986, "step": 19166 }, { "epoch": 1.2863662293211637, "grad_norm": 1.0787842273712158, "learning_rate": 2.9827784678994003e-05, "loss": 0.4265, "step": 19168 }, { "epoch": 1.2865004530049327, "grad_norm": 1.0132031440734863, "learning_rate": 2.9817840609567166e-05, "loss": 0.5004, "step": 19170 }, { "epoch": 1.2866346766887018, "grad_norm": 1.047211766242981, "learning_rate": 2.9807897493707703e-05, "loss": 0.4887, "step": 19172 }, { "epoch": 1.2867689003724707, "grad_norm": 1.1641210317611694, "learning_rate": 2.9797955331885346e-05, "loss": 0.5006, "step": 19174 }, { "epoch": 1.2869031240562396, "grad_norm": 1.0389865636825562, "learning_rate": 2.9788014124569895e-05, "loss": 0.527, "step": 19176 }, { "epoch": 1.2870373477400088, "grad_norm": 1.0260602235794067, "learning_rate": 2.977807387223102e-05, "loss": 0.4668, "step": 19178 }, { "epoch": 1.2871715714237777, "grad_norm": 1.4412453174591064, "learning_rate": 2.9768134575338402e-05, "loss": 0.5256, "step": 19180 }, { "epoch": 1.2873057951075468, "grad_norm": 0.9133895635604858, "learning_rate": 2.975819623436163e-05, "loss": 0.5167, "step": 19182 }, { "epoch": 1.2874400187913158, "grad_norm": 0.9934803247451782, "learning_rate": 2.9748258849770293e-05, "loss": 0.5346, "step": 19184 }, { "epoch": 1.2875742424750847, "grad_norm": 0.9523311257362366, "learning_rate": 2.9738322422033895e-05, "loss": 0.4901, "step": 19186 }, { "epoch": 1.2877084661588536, "grad_norm": 1.0600826740264893, "learning_rate": 2.9728386951621923e-05, "loss": 0.522, "step": 19188 }, { "epoch": 1.2878426898426227, "grad_norm": 1.0519115924835205, "learning_rate": 2.9718452439003796e-05, "loss": 0.4582, "step": 19190 }, { "epoch": 1.2879769135263917, "grad_norm": 1.139541745185852, "learning_rate": 2.9708518884648923e-05, "loss": 0.5583, "step": 19192 }, { "epoch": 1.2881111372101608, "grad_norm": 1.0064599514007568, "learning_rate": 2.969858628902662e-05, "loss": 0.5957, "step": 19194 }, { "epoch": 1.2882453608939297, "grad_norm": 1.003114938735962, "learning_rate": 2.9688654652606207e-05, "loss": 0.4909, "step": 19196 }, { "epoch": 1.2883795845776986, "grad_norm": 0.9284659028053284, "learning_rate": 2.967872397585689e-05, "loss": 0.5212, "step": 19198 }, { "epoch": 1.2885138082614678, "grad_norm": 1.0107605457305908, "learning_rate": 2.9668794259247945e-05, "loss": 0.5099, "step": 19200 }, { "epoch": 1.2886480319452367, "grad_norm": 0.9981799721717834, "learning_rate": 2.9658865503248463e-05, "loss": 0.5337, "step": 19202 }, { "epoch": 1.2887822556290058, "grad_norm": 1.0402655601501465, "learning_rate": 2.96489377083276e-05, "loss": 0.5138, "step": 19204 }, { "epoch": 1.2889164793127748, "grad_norm": 0.9867530465126038, "learning_rate": 2.963901087495441e-05, "loss": 0.5181, "step": 19206 }, { "epoch": 1.2890507029965437, "grad_norm": 0.9896501898765564, "learning_rate": 2.9629085003597918e-05, "loss": 0.5025, "step": 19208 }, { "epoch": 1.2891849266803128, "grad_norm": 0.958095371723175, "learning_rate": 2.9619160094727093e-05, "loss": 0.502, "step": 19210 }, { "epoch": 1.2893191503640817, "grad_norm": 1.1133953332901, "learning_rate": 2.960923614881089e-05, "loss": 0.5428, "step": 19212 }, { "epoch": 1.2894533740478509, "grad_norm": 1.0463382005691528, "learning_rate": 2.9599313166318177e-05, "loss": 0.5092, "step": 19214 }, { "epoch": 1.2895875977316198, "grad_norm": 0.9142300486564636, "learning_rate": 2.9589391147717803e-05, "loss": 0.4992, "step": 19216 }, { "epoch": 1.2897218214153887, "grad_norm": 1.1417711973190308, "learning_rate": 2.9579470093478558e-05, "loss": 0.4988, "step": 19218 }, { "epoch": 1.2898560450991576, "grad_norm": 1.090346097946167, "learning_rate": 2.9569550004069203e-05, "loss": 0.4902, "step": 19220 }, { "epoch": 1.2899902687829268, "grad_norm": 0.9615224599838257, "learning_rate": 2.9559630879958422e-05, "loss": 0.426, "step": 19222 }, { "epoch": 1.2901244924666957, "grad_norm": 1.0067687034606934, "learning_rate": 2.9549712721614902e-05, "loss": 0.5547, "step": 19224 }, { "epoch": 1.2902587161504648, "grad_norm": 1.2727751731872559, "learning_rate": 2.953979552950722e-05, "loss": 0.546, "step": 19226 }, { "epoch": 1.2903929398342338, "grad_norm": 1.0097273588180542, "learning_rate": 2.9529879304103997e-05, "loss": 0.4805, "step": 19228 }, { "epoch": 1.2905271635180027, "grad_norm": 1.0178848505020142, "learning_rate": 2.9519964045873716e-05, "loss": 0.5131, "step": 19230 }, { "epoch": 1.2906613872017718, "grad_norm": 1.038375973701477, "learning_rate": 2.951004975528484e-05, "loss": 0.4862, "step": 19232 }, { "epoch": 1.2907956108855407, "grad_norm": 0.9703474044799805, "learning_rate": 2.9500136432805848e-05, "loss": 0.5114, "step": 19234 }, { "epoch": 1.2909298345693099, "grad_norm": 1.1284089088439941, "learning_rate": 2.949022407890507e-05, "loss": 0.5053, "step": 19236 }, { "epoch": 1.2910640582530788, "grad_norm": 1.1331610679626465, "learning_rate": 2.9480312694050905e-05, "loss": 0.4618, "step": 19238 }, { "epoch": 1.2911982819368477, "grad_norm": 1.259121060371399, "learning_rate": 2.9470402278711584e-05, "loss": 0.5041, "step": 19240 }, { "epoch": 1.2913325056206166, "grad_norm": 0.8875788450241089, "learning_rate": 2.9460492833355407e-05, "loss": 0.4756, "step": 19242 }, { "epoch": 1.2914667293043858, "grad_norm": 1.0083508491516113, "learning_rate": 2.945058435845054e-05, "loss": 0.5571, "step": 19244 }, { "epoch": 1.2916009529881547, "grad_norm": 1.2376141548156738, "learning_rate": 2.9440676854465165e-05, "loss": 0.5199, "step": 19246 }, { "epoch": 1.2917351766719238, "grad_norm": 0.9780316948890686, "learning_rate": 2.9430770321867374e-05, "loss": 0.501, "step": 19248 }, { "epoch": 1.2918694003556928, "grad_norm": 0.9901347756385803, "learning_rate": 2.942086476112525e-05, "loss": 0.4607, "step": 19250 }, { "epoch": 1.2920036240394617, "grad_norm": 0.9821428656578064, "learning_rate": 2.941096017270678e-05, "loss": 0.5326, "step": 19252 }, { "epoch": 1.2921378477232308, "grad_norm": 0.9778178334236145, "learning_rate": 2.9401056557079975e-05, "loss": 0.4725, "step": 19254 }, { "epoch": 1.2922720714069997, "grad_norm": 1.065484881401062, "learning_rate": 2.9391153914712722e-05, "loss": 0.5688, "step": 19256 }, { "epoch": 1.2924062950907689, "grad_norm": 1.1318145990371704, "learning_rate": 2.938125224607294e-05, "loss": 0.5338, "step": 19258 }, { "epoch": 1.2925405187745378, "grad_norm": 1.0310784578323364, "learning_rate": 2.937135155162842e-05, "loss": 0.5378, "step": 19260 }, { "epoch": 1.2926747424583067, "grad_norm": 1.0071526765823364, "learning_rate": 2.9361451831847004e-05, "loss": 0.4794, "step": 19262 }, { "epoch": 1.2928089661420756, "grad_norm": 1.001133918762207, "learning_rate": 2.935155308719637e-05, "loss": 0.5911, "step": 19264 }, { "epoch": 1.2929431898258448, "grad_norm": 1.0991449356079102, "learning_rate": 2.9341655318144278e-05, "loss": 0.505, "step": 19266 }, { "epoch": 1.2930774135096137, "grad_norm": 1.0803444385528564, "learning_rate": 2.9331758525158338e-05, "loss": 0.5172, "step": 19268 }, { "epoch": 1.2932116371933828, "grad_norm": 1.080772876739502, "learning_rate": 2.9321862708706172e-05, "loss": 0.5521, "step": 19270 }, { "epoch": 1.2933458608771518, "grad_norm": 1.0210713148117065, "learning_rate": 2.9311967869255324e-05, "loss": 0.533, "step": 19272 }, { "epoch": 1.2934800845609207, "grad_norm": 1.0075637102127075, "learning_rate": 2.9302074007273317e-05, "loss": 0.5399, "step": 19274 }, { "epoch": 1.2936143082446898, "grad_norm": 1.0783876180648804, "learning_rate": 2.9292181123227612e-05, "loss": 0.5131, "step": 19276 }, { "epoch": 1.2937485319284587, "grad_norm": 1.3644986152648926, "learning_rate": 2.9282289217585633e-05, "loss": 0.4527, "step": 19278 }, { "epoch": 1.2938827556122279, "grad_norm": 0.9839290380477905, "learning_rate": 2.927239829081474e-05, "loss": 0.502, "step": 19280 }, { "epoch": 1.2940169792959968, "grad_norm": 0.9688754677772522, "learning_rate": 2.9262508343382276e-05, "loss": 0.4621, "step": 19282 }, { "epoch": 1.2941512029797657, "grad_norm": 1.1157327890396118, "learning_rate": 2.9252619375755508e-05, "loss": 0.447, "step": 19284 }, { "epoch": 1.2942854266635349, "grad_norm": 0.9797788858413696, "learning_rate": 2.9242731388401685e-05, "loss": 0.473, "step": 19286 }, { "epoch": 1.2944196503473038, "grad_norm": 1.1115068197250366, "learning_rate": 2.9232844381787967e-05, "loss": 0.5085, "step": 19288 }, { "epoch": 1.294553874031073, "grad_norm": 0.9349199533462524, "learning_rate": 2.9222958356381547e-05, "loss": 0.561, "step": 19290 }, { "epoch": 1.2946880977148418, "grad_norm": 1.0068470239639282, "learning_rate": 2.9213073312649452e-05, "loss": 0.4471, "step": 19292 }, { "epoch": 1.2948223213986108, "grad_norm": 0.9416777491569519, "learning_rate": 2.9203189251058792e-05, "loss": 0.4911, "step": 19294 }, { "epoch": 1.2949565450823797, "grad_norm": 1.0156686305999756, "learning_rate": 2.9193306172076553e-05, "loss": 0.5114, "step": 19296 }, { "epoch": 1.2950907687661488, "grad_norm": 1.1254180669784546, "learning_rate": 2.9183424076169653e-05, "loss": 0.4796, "step": 19298 }, { "epoch": 1.2952249924499177, "grad_norm": 0.9957659840583801, "learning_rate": 2.9173542963805058e-05, "loss": 0.4872, "step": 19300 }, { "epoch": 1.2953592161336869, "grad_norm": 0.9471935629844666, "learning_rate": 2.916366283544959e-05, "loss": 0.4351, "step": 19302 }, { "epoch": 1.2954934398174558, "grad_norm": 1.0207605361938477, "learning_rate": 2.915378369157009e-05, "loss": 0.5233, "step": 19304 }, { "epoch": 1.2956276635012247, "grad_norm": 1.0082496404647827, "learning_rate": 2.91439055326333e-05, "loss": 0.4257, "step": 19306 }, { "epoch": 1.2957618871849939, "grad_norm": 1.0640394687652588, "learning_rate": 2.913402835910598e-05, "loss": 0.5308, "step": 19308 }, { "epoch": 1.2958961108687628, "grad_norm": 1.0955893993377686, "learning_rate": 2.9124152171454766e-05, "loss": 0.5193, "step": 19310 }, { "epoch": 1.296030334552532, "grad_norm": 1.0826829671859741, "learning_rate": 2.9114276970146355e-05, "loss": 0.5126, "step": 19312 }, { "epoch": 1.2961645582363008, "grad_norm": 3.258791923522949, "learning_rate": 2.910440275564724e-05, "loss": 0.4691, "step": 19314 }, { "epoch": 1.2962987819200698, "grad_norm": 0.9478439688682556, "learning_rate": 2.9094529528424032e-05, "loss": 0.5007, "step": 19316 }, { "epoch": 1.2964330056038387, "grad_norm": 1.0688092708587646, "learning_rate": 2.9084657288943174e-05, "loss": 0.523, "step": 19318 }, { "epoch": 1.2965672292876078, "grad_norm": 1.0586588382720947, "learning_rate": 2.9074786037671153e-05, "loss": 0.5582, "step": 19320 }, { "epoch": 1.2967014529713767, "grad_norm": 1.011034369468689, "learning_rate": 2.9064915775074342e-05, "loss": 0.5154, "step": 19322 }, { "epoch": 1.2968356766551459, "grad_norm": 1.160041093826294, "learning_rate": 2.905504650161909e-05, "loss": 0.4864, "step": 19324 }, { "epoch": 1.2969699003389148, "grad_norm": 1.1135926246643066, "learning_rate": 2.9045178217771684e-05, "loss": 0.5105, "step": 19326 }, { "epoch": 1.2971041240226837, "grad_norm": 2.92923903465271, "learning_rate": 2.9035310923998427e-05, "loss": 0.4478, "step": 19328 }, { "epoch": 1.2972383477064529, "grad_norm": 0.9935290813446045, "learning_rate": 2.9025444620765502e-05, "loss": 0.4581, "step": 19330 }, { "epoch": 1.2973725713902218, "grad_norm": 1.0143234729766846, "learning_rate": 2.901557930853907e-05, "loss": 0.4711, "step": 19332 }, { "epoch": 1.297506795073991, "grad_norm": 0.9148914217948914, "learning_rate": 2.9005714987785236e-05, "loss": 0.5094, "step": 19334 }, { "epoch": 1.2976410187577598, "grad_norm": 0.9044680595397949, "learning_rate": 2.8995851658970102e-05, "loss": 0.4913, "step": 19336 }, { "epoch": 1.2977752424415288, "grad_norm": 1.0622960329055786, "learning_rate": 2.898598932255966e-05, "loss": 0.5493, "step": 19338 }, { "epoch": 1.2979094661252977, "grad_norm": 1.059946060180664, "learning_rate": 2.8976127979019934e-05, "loss": 0.5537, "step": 19340 }, { "epoch": 1.2980436898090668, "grad_norm": 1.0593196153640747, "learning_rate": 2.896626762881678e-05, "loss": 0.5263, "step": 19342 }, { "epoch": 1.2981779134928357, "grad_norm": 0.8505914807319641, "learning_rate": 2.8956408272416148e-05, "loss": 0.4574, "step": 19344 }, { "epoch": 1.2983121371766049, "grad_norm": 0.9502744674682617, "learning_rate": 2.8946549910283817e-05, "loss": 0.531, "step": 19346 }, { "epoch": 1.2984463608603738, "grad_norm": 1.0109035968780518, "learning_rate": 2.8936692542885617e-05, "loss": 0.5213, "step": 19348 }, { "epoch": 1.2985805845441427, "grad_norm": 1.015946626663208, "learning_rate": 2.8926836170687284e-05, "loss": 0.4924, "step": 19350 }, { "epoch": 1.2987148082279119, "grad_norm": 1.1477079391479492, "learning_rate": 2.8916980794154503e-05, "loss": 0.526, "step": 19352 }, { "epoch": 1.2988490319116808, "grad_norm": 1.1128402948379517, "learning_rate": 2.8907126413752895e-05, "loss": 0.5568, "step": 19354 }, { "epoch": 1.29898325559545, "grad_norm": 1.0364487171173096, "learning_rate": 2.889727302994811e-05, "loss": 0.6144, "step": 19356 }, { "epoch": 1.2991174792792188, "grad_norm": 0.9741474986076355, "learning_rate": 2.8887420643205682e-05, "loss": 0.4971, "step": 19358 }, { "epoch": 1.2992517029629878, "grad_norm": 1.0670655965805054, "learning_rate": 2.887756925399111e-05, "loss": 0.4898, "step": 19360 }, { "epoch": 1.299385926646757, "grad_norm": 0.8790115714073181, "learning_rate": 2.8867718862769834e-05, "loss": 0.4672, "step": 19362 }, { "epoch": 1.2995201503305258, "grad_norm": 1.0886361598968506, "learning_rate": 2.8857869470007302e-05, "loss": 0.4906, "step": 19364 }, { "epoch": 1.299654374014295, "grad_norm": 0.8426548838615417, "learning_rate": 2.8848021076168875e-05, "loss": 0.4661, "step": 19366 }, { "epoch": 1.2997885976980639, "grad_norm": 1.0217941999435425, "learning_rate": 2.883817368171985e-05, "loss": 0.504, "step": 19368 }, { "epoch": 1.2999228213818328, "grad_norm": 1.404261827468872, "learning_rate": 2.882832728712551e-05, "loss": 0.554, "step": 19370 }, { "epoch": 1.3000570450656017, "grad_norm": 0.9625460505485535, "learning_rate": 2.881848189285105e-05, "loss": 0.5089, "step": 19372 }, { "epoch": 1.3001912687493709, "grad_norm": 0.9874593615531921, "learning_rate": 2.880863749936169e-05, "loss": 0.5137, "step": 19374 }, { "epoch": 1.3003254924331398, "grad_norm": 1.07978093624115, "learning_rate": 2.879879410712252e-05, "loss": 0.4855, "step": 19376 }, { "epoch": 1.300459716116909, "grad_norm": 0.9420575499534607, "learning_rate": 2.8788951716598656e-05, "loss": 0.4991, "step": 19378 }, { "epoch": 1.3005939398006778, "grad_norm": 0.9697789549827576, "learning_rate": 2.877911032825511e-05, "loss": 0.4976, "step": 19380 }, { "epoch": 1.3007281634844468, "grad_norm": 1.011218547821045, "learning_rate": 2.8769269942556875e-05, "loss": 0.5941, "step": 19382 }, { "epoch": 1.300862387168216, "grad_norm": 1.0524111986160278, "learning_rate": 2.875943055996887e-05, "loss": 0.5196, "step": 19384 }, { "epoch": 1.3009966108519848, "grad_norm": 1.0156795978546143, "learning_rate": 2.874959218095602e-05, "loss": 0.4749, "step": 19386 }, { "epoch": 1.301130834535754, "grad_norm": 0.8876588940620422, "learning_rate": 2.873975480598315e-05, "loss": 0.4607, "step": 19388 }, { "epoch": 1.3012650582195229, "grad_norm": 0.9444760680198669, "learning_rate": 2.8729918435515058e-05, "loss": 0.4715, "step": 19390 }, { "epoch": 1.3013992819032918, "grad_norm": 1.092552661895752, "learning_rate": 2.872008307001648e-05, "loss": 0.4848, "step": 19392 }, { "epoch": 1.3015335055870607, "grad_norm": 1.0454710721969604, "learning_rate": 2.8710248709952147e-05, "loss": 0.5212, "step": 19394 }, { "epoch": 1.3016677292708299, "grad_norm": 1.009968876838684, "learning_rate": 2.8700415355786704e-05, "loss": 0.5196, "step": 19396 }, { "epoch": 1.3018019529545988, "grad_norm": 0.9652559161186218, "learning_rate": 2.8690583007984745e-05, "loss": 0.5146, "step": 19398 }, { "epoch": 1.301936176638368, "grad_norm": 0.9651011228561401, "learning_rate": 2.868075166701082e-05, "loss": 0.4049, "step": 19400 }, { "epoch": 1.3020704003221368, "grad_norm": 1.0508171319961548, "learning_rate": 2.867092133332947e-05, "loss": 0.5036, "step": 19402 }, { "epoch": 1.3022046240059058, "grad_norm": 0.9191226959228516, "learning_rate": 2.8661092007405132e-05, "loss": 0.4603, "step": 19404 }, { "epoch": 1.302338847689675, "grad_norm": 0.8831053972244263, "learning_rate": 2.8651263689702256e-05, "loss": 0.4623, "step": 19406 }, { "epoch": 1.3024730713734438, "grad_norm": 1.0167165994644165, "learning_rate": 2.8641436380685184e-05, "loss": 0.4737, "step": 19408 }, { "epoch": 1.302607295057213, "grad_norm": 0.9504034519195557, "learning_rate": 2.863161008081825e-05, "loss": 0.4663, "step": 19410 }, { "epoch": 1.3027415187409819, "grad_norm": 1.011612892150879, "learning_rate": 2.8621784790565696e-05, "loss": 0.4308, "step": 19412 }, { "epoch": 1.3028757424247508, "grad_norm": 1.1726253032684326, "learning_rate": 2.8611960510391795e-05, "loss": 0.5226, "step": 19414 }, { "epoch": 1.3030099661085197, "grad_norm": 1.0991902351379395, "learning_rate": 2.8602137240760695e-05, "loss": 0.5698, "step": 19416 }, { "epoch": 1.3031441897922889, "grad_norm": 0.9891123175621033, "learning_rate": 2.859231498213654e-05, "loss": 0.4947, "step": 19418 }, { "epoch": 1.3032784134760578, "grad_norm": 0.8932777643203735, "learning_rate": 2.8582493734983384e-05, "loss": 0.511, "step": 19420 }, { "epoch": 1.303412637159827, "grad_norm": 0.9994890689849854, "learning_rate": 2.8572673499765307e-05, "loss": 0.4924, "step": 19422 }, { "epoch": 1.3035468608435958, "grad_norm": 1.1127361059188843, "learning_rate": 2.856285427694627e-05, "loss": 0.5761, "step": 19424 }, { "epoch": 1.3036810845273648, "grad_norm": 0.9759175777435303, "learning_rate": 2.8553036066990214e-05, "loss": 0.4683, "step": 19426 }, { "epoch": 1.303815308211134, "grad_norm": 1.0085070133209229, "learning_rate": 2.854321887036101e-05, "loss": 0.4728, "step": 19428 }, { "epoch": 1.3039495318949028, "grad_norm": 1.080127239227295, "learning_rate": 2.8533402687522538e-05, "loss": 0.4812, "step": 19430 }, { "epoch": 1.304083755578672, "grad_norm": 1.0481758117675781, "learning_rate": 2.8523587518938554e-05, "loss": 0.4898, "step": 19432 }, { "epoch": 1.3042179792624409, "grad_norm": 0.9496031403541565, "learning_rate": 2.851377336507286e-05, "loss": 0.4965, "step": 19434 }, { "epoch": 1.3043522029462098, "grad_norm": 1.1129904985427856, "learning_rate": 2.8503960226389136e-05, "loss": 0.5361, "step": 19436 }, { "epoch": 1.304486426629979, "grad_norm": 0.9533309936523438, "learning_rate": 2.8494148103350983e-05, "loss": 0.4836, "step": 19438 }, { "epoch": 1.3046206503137479, "grad_norm": 1.0634078979492188, "learning_rate": 2.8484336996422057e-05, "loss": 0.6061, "step": 19440 }, { "epoch": 1.304754873997517, "grad_norm": 1.0369467735290527, "learning_rate": 2.847452690606589e-05, "loss": 0.5382, "step": 19442 }, { "epoch": 1.304889097681286, "grad_norm": 0.9479460716247559, "learning_rate": 2.8464717832746014e-05, "loss": 0.5047, "step": 19444 }, { "epoch": 1.3050233213650548, "grad_norm": 1.0582891702651978, "learning_rate": 2.8454909776925865e-05, "loss": 0.5063, "step": 19446 }, { "epoch": 1.3051575450488238, "grad_norm": 0.9738956689834595, "learning_rate": 2.8445102739068873e-05, "loss": 0.4998, "step": 19448 }, { "epoch": 1.305291768732593, "grad_norm": 1.0738046169281006, "learning_rate": 2.8435296719638366e-05, "loss": 0.5153, "step": 19450 }, { "epoch": 1.3054259924163618, "grad_norm": 0.8883415460586548, "learning_rate": 2.8425491719097707e-05, "loss": 0.4196, "step": 19452 }, { "epoch": 1.305560216100131, "grad_norm": 1.0990755558013916, "learning_rate": 2.8415687737910134e-05, "loss": 0.4818, "step": 19454 }, { "epoch": 1.3056944397838999, "grad_norm": 1.189955234527588, "learning_rate": 2.8405884776538876e-05, "loss": 0.5433, "step": 19456 }, { "epoch": 1.3058286634676688, "grad_norm": 1.0909309387207031, "learning_rate": 2.839608283544708e-05, "loss": 0.5031, "step": 19458 }, { "epoch": 1.305962887151438, "grad_norm": 0.9963827729225159, "learning_rate": 2.8386281915097907e-05, "loss": 0.4778, "step": 19460 }, { "epoch": 1.3060971108352069, "grad_norm": 0.9648716449737549, "learning_rate": 2.8376482015954387e-05, "loss": 0.437, "step": 19462 }, { "epoch": 1.306231334518976, "grad_norm": 1.0067102909088135, "learning_rate": 2.836668313847962e-05, "loss": 0.5187, "step": 19464 }, { "epoch": 1.306365558202745, "grad_norm": 1.1235438585281372, "learning_rate": 2.8356885283136485e-05, "loss": 0.5222, "step": 19466 }, { "epoch": 1.3064997818865138, "grad_norm": 1.0012166500091553, "learning_rate": 2.8347088450387986e-05, "loss": 0.4864, "step": 19468 }, { "epoch": 1.3066340055702828, "grad_norm": 1.0444602966308594, "learning_rate": 2.833729264069696e-05, "loss": 0.447, "step": 19470 }, { "epoch": 1.306768229254052, "grad_norm": 0.9028139114379883, "learning_rate": 2.8327497854526276e-05, "loss": 0.4706, "step": 19472 }, { "epoch": 1.3069024529378208, "grad_norm": 0.9326144456863403, "learning_rate": 2.8317704092338703e-05, "loss": 0.3933, "step": 19474 }, { "epoch": 1.30703667662159, "grad_norm": 0.9908169507980347, "learning_rate": 2.8307911354596978e-05, "loss": 0.5367, "step": 19476 }, { "epoch": 1.3071709003053589, "grad_norm": 0.9125736355781555, "learning_rate": 2.8298119641763763e-05, "loss": 0.4801, "step": 19478 }, { "epoch": 1.3073051239891278, "grad_norm": 0.9635627269744873, "learning_rate": 2.828832895430174e-05, "loss": 0.5259, "step": 19480 }, { "epoch": 1.307439347672897, "grad_norm": 1.1029691696166992, "learning_rate": 2.827853929267348e-05, "loss": 0.4381, "step": 19482 }, { "epoch": 1.3075735713566659, "grad_norm": 0.9907110333442688, "learning_rate": 2.8268750657341524e-05, "loss": 0.5017, "step": 19484 }, { "epoch": 1.307707795040435, "grad_norm": 0.8806270360946655, "learning_rate": 2.825896304876835e-05, "loss": 0.4635, "step": 19486 }, { "epoch": 1.307842018724204, "grad_norm": 1.082257628440857, "learning_rate": 2.8249176467416438e-05, "loss": 0.5017, "step": 19488 }, { "epoch": 1.3079762424079728, "grad_norm": 0.9463380575180054, "learning_rate": 2.8239390913748144e-05, "loss": 0.483, "step": 19490 }, { "epoch": 1.3081104660917418, "grad_norm": 0.9221293926239014, "learning_rate": 2.822960638822588e-05, "loss": 0.4758, "step": 19492 }, { "epoch": 1.308244689775511, "grad_norm": 1.0206997394561768, "learning_rate": 2.8219822891311863e-05, "loss": 0.4737, "step": 19494 }, { "epoch": 1.3083789134592798, "grad_norm": 1.2292686700820923, "learning_rate": 2.8210040423468408e-05, "loss": 0.5189, "step": 19496 }, { "epoch": 1.308513137143049, "grad_norm": 0.9695619940757751, "learning_rate": 2.820025898515768e-05, "loss": 0.4991, "step": 19498 }, { "epoch": 1.3086473608268179, "grad_norm": 1.0157265663146973, "learning_rate": 2.8190478576841862e-05, "loss": 0.4968, "step": 19500 }, { "epoch": 1.3087815845105868, "grad_norm": 1.0575307607650757, "learning_rate": 2.8180699198983062e-05, "loss": 0.5299, "step": 19502 }, { "epoch": 1.308915808194356, "grad_norm": 1.0781140327453613, "learning_rate": 2.8170920852043286e-05, "loss": 0.5328, "step": 19504 }, { "epoch": 1.3090500318781249, "grad_norm": 1.0960808992385864, "learning_rate": 2.8161143536484592e-05, "loss": 0.5449, "step": 19506 }, { "epoch": 1.309184255561894, "grad_norm": 1.0679954290390015, "learning_rate": 2.81513672527689e-05, "loss": 0.4821, "step": 19508 }, { "epoch": 1.309318479245663, "grad_norm": 0.8787037134170532, "learning_rate": 2.8141592001358163e-05, "loss": 0.4442, "step": 19510 }, { "epoch": 1.3094527029294318, "grad_norm": 0.9036421179771423, "learning_rate": 2.813181778271422e-05, "loss": 0.4954, "step": 19512 }, { "epoch": 1.309586926613201, "grad_norm": 1.6593456268310547, "learning_rate": 2.8122044597298886e-05, "loss": 0.5353, "step": 19514 }, { "epoch": 1.30972115029697, "grad_norm": 0.9559675455093384, "learning_rate": 2.8112272445573905e-05, "loss": 0.5036, "step": 19516 }, { "epoch": 1.309855373980739, "grad_norm": 1.2117623090744019, "learning_rate": 2.810250132800103e-05, "loss": 0.51, "step": 19518 }, { "epoch": 1.309989597664508, "grad_norm": 1.0451873540878296, "learning_rate": 2.8092731245041903e-05, "loss": 0.5105, "step": 19520 }, { "epoch": 1.3101238213482769, "grad_norm": 0.9599100351333618, "learning_rate": 2.8082962197158148e-05, "loss": 0.4935, "step": 19522 }, { "epoch": 1.3102580450320458, "grad_norm": 0.9579646587371826, "learning_rate": 2.8073194184811314e-05, "loss": 0.5055, "step": 19524 }, { "epoch": 1.310392268715815, "grad_norm": 0.9336437582969666, "learning_rate": 2.8063427208462957e-05, "loss": 0.5059, "step": 19526 }, { "epoch": 1.3105264923995839, "grad_norm": 0.8615583777427673, "learning_rate": 2.8053661268574505e-05, "loss": 0.4374, "step": 19528 }, { "epoch": 1.310660716083353, "grad_norm": 1.0524922609329224, "learning_rate": 2.8043896365607447e-05, "loss": 0.5419, "step": 19530 }, { "epoch": 1.310794939767122, "grad_norm": 0.9826512336730957, "learning_rate": 2.803413250002307e-05, "loss": 0.513, "step": 19532 }, { "epoch": 1.3109291634508908, "grad_norm": 0.9838429093360901, "learning_rate": 2.8024369672282756e-05, "loss": 0.4585, "step": 19534 }, { "epoch": 1.31106338713466, "grad_norm": 1.0887281894683838, "learning_rate": 2.8014607882847743e-05, "loss": 0.4766, "step": 19536 }, { "epoch": 1.311197610818429, "grad_norm": 1.03814697265625, "learning_rate": 2.800484713217929e-05, "loss": 0.5152, "step": 19538 }, { "epoch": 1.311331834502198, "grad_norm": 1.0641275644302368, "learning_rate": 2.7995087420738565e-05, "loss": 0.4898, "step": 19540 }, { "epoch": 1.311466058185967, "grad_norm": 2.263648509979248, "learning_rate": 2.7985328748986682e-05, "loss": 0.5257, "step": 19542 }, { "epoch": 1.3116002818697359, "grad_norm": 0.9365549683570862, "learning_rate": 2.7975571117384713e-05, "loss": 0.4544, "step": 19544 }, { "epoch": 1.3117345055535048, "grad_norm": 0.9181808829307556, "learning_rate": 2.7965814526393718e-05, "loss": 0.4883, "step": 19546 }, { "epoch": 1.311868729237274, "grad_norm": 1.0153909921646118, "learning_rate": 2.795605897647466e-05, "loss": 0.5232, "step": 19548 }, { "epoch": 1.3120029529210429, "grad_norm": 1.0753653049468994, "learning_rate": 2.7946304468088463e-05, "loss": 0.4537, "step": 19550 }, { "epoch": 1.312137176604812, "grad_norm": 1.0972576141357422, "learning_rate": 2.7936551001695992e-05, "loss": 0.5292, "step": 19552 }, { "epoch": 1.312271400288581, "grad_norm": 0.9783473014831543, "learning_rate": 2.792679857775813e-05, "loss": 0.4966, "step": 19554 }, { "epoch": 1.3124056239723498, "grad_norm": 0.9188259840011597, "learning_rate": 2.7917047196735602e-05, "loss": 0.5125, "step": 19556 }, { "epoch": 1.312539847656119, "grad_norm": 1.1747207641601562, "learning_rate": 2.790729685908919e-05, "loss": 0.526, "step": 19558 }, { "epoch": 1.312674071339888, "grad_norm": 1.0269527435302734, "learning_rate": 2.7897547565279557e-05, "loss": 0.557, "step": 19560 }, { "epoch": 1.312808295023657, "grad_norm": 1.1212568283081055, "learning_rate": 2.788779931576734e-05, "loss": 0.5312, "step": 19562 }, { "epoch": 1.312942518707426, "grad_norm": 0.9822105169296265, "learning_rate": 2.787805211101311e-05, "loss": 0.5002, "step": 19564 }, { "epoch": 1.3130767423911949, "grad_norm": 1.115074872970581, "learning_rate": 2.7868305951477425e-05, "loss": 0.5395, "step": 19566 }, { "epoch": 1.3132109660749638, "grad_norm": 1.0336101055145264, "learning_rate": 2.7858560837620773e-05, "loss": 0.4518, "step": 19568 }, { "epoch": 1.313345189758733, "grad_norm": 0.9578868746757507, "learning_rate": 2.7848816769903574e-05, "loss": 0.536, "step": 19570 }, { "epoch": 1.3134794134425019, "grad_norm": 1.0108757019042969, "learning_rate": 2.783907374878623e-05, "loss": 0.495, "step": 19572 }, { "epoch": 1.313613637126271, "grad_norm": 0.9814471006393433, "learning_rate": 2.7829331774729056e-05, "loss": 0.4853, "step": 19574 }, { "epoch": 1.31374786081004, "grad_norm": 0.9435616731643677, "learning_rate": 2.781959084819238e-05, "loss": 0.4488, "step": 19576 }, { "epoch": 1.3138820844938088, "grad_norm": 1.0283944606781006, "learning_rate": 2.780985096963641e-05, "loss": 0.4549, "step": 19578 }, { "epoch": 1.314016308177578, "grad_norm": 0.9773117303848267, "learning_rate": 2.780011213952135e-05, "loss": 0.5221, "step": 19580 }, { "epoch": 1.314150531861347, "grad_norm": 1.0373390913009644, "learning_rate": 2.7790374358307327e-05, "loss": 0.4952, "step": 19582 }, { "epoch": 1.314284755545116, "grad_norm": 1.02969491481781, "learning_rate": 2.7780637626454452e-05, "loss": 0.5134, "step": 19584 }, { "epoch": 1.314418979228885, "grad_norm": 1.0415964126586914, "learning_rate": 2.7770901944422744e-05, "loss": 0.4908, "step": 19586 }, { "epoch": 1.3145532029126539, "grad_norm": 0.9986392855644226, "learning_rate": 2.7761167312672242e-05, "loss": 0.5128, "step": 19588 }, { "epoch": 1.314687426596423, "grad_norm": 1.0504610538482666, "learning_rate": 2.775143373166281e-05, "loss": 0.4669, "step": 19590 }, { "epoch": 1.314821650280192, "grad_norm": 1.1573312282562256, "learning_rate": 2.7741701201854414e-05, "loss": 0.4936, "step": 19592 }, { "epoch": 1.314955873963961, "grad_norm": 1.3695416450500488, "learning_rate": 2.773196972370684e-05, "loss": 0.4605, "step": 19594 }, { "epoch": 1.31509009764773, "grad_norm": 1.1023151874542236, "learning_rate": 2.772223929767993e-05, "loss": 0.5058, "step": 19596 }, { "epoch": 1.315224321331499, "grad_norm": 0.9775272607803345, "learning_rate": 2.771250992423341e-05, "loss": 0.4713, "step": 19598 }, { "epoch": 1.3153585450152678, "grad_norm": 1.0285531282424927, "learning_rate": 2.7702781603826965e-05, "loss": 0.5197, "step": 19600 }, { "epoch": 1.315492768699037, "grad_norm": 1.1167120933532715, "learning_rate": 2.7693054336920228e-05, "loss": 0.5114, "step": 19602 }, { "epoch": 1.315626992382806, "grad_norm": 0.9397265315055847, "learning_rate": 2.7683328123972823e-05, "loss": 0.4821, "step": 19604 }, { "epoch": 1.315761216066575, "grad_norm": 0.9107526540756226, "learning_rate": 2.7673602965444285e-05, "loss": 0.5193, "step": 19606 }, { "epoch": 1.315895439750344, "grad_norm": 0.9846282005310059, "learning_rate": 2.76638788617941e-05, "loss": 0.5116, "step": 19608 }, { "epoch": 1.3160296634341129, "grad_norm": 0.9590386152267456, "learning_rate": 2.76541558134817e-05, "loss": 0.4889, "step": 19610 }, { "epoch": 1.316163887117882, "grad_norm": 0.9941418766975403, "learning_rate": 2.764443382096652e-05, "loss": 0.5371, "step": 19612 }, { "epoch": 1.316298110801651, "grad_norm": 1.1155186891555786, "learning_rate": 2.7634712884707852e-05, "loss": 0.4739, "step": 19614 }, { "epoch": 1.31643233448542, "grad_norm": 1.1311336755752563, "learning_rate": 2.7624993005165066e-05, "loss": 0.5331, "step": 19616 }, { "epoch": 1.316566558169189, "grad_norm": 0.9965952038764954, "learning_rate": 2.7615274182797325e-05, "loss": 0.478, "step": 19618 }, { "epoch": 1.316700781852958, "grad_norm": 0.9219123721122742, "learning_rate": 2.7605556418063877e-05, "loss": 0.4746, "step": 19620 }, { "epoch": 1.3168350055367268, "grad_norm": 1.1114145517349243, "learning_rate": 2.759583971142383e-05, "loss": 0.5478, "step": 19622 }, { "epoch": 1.316969229220496, "grad_norm": 0.9328547120094299, "learning_rate": 2.758612406333633e-05, "loss": 0.506, "step": 19624 }, { "epoch": 1.317103452904265, "grad_norm": 0.8412764072418213, "learning_rate": 2.7576409474260378e-05, "loss": 0.5242, "step": 19626 }, { "epoch": 1.317237676588034, "grad_norm": 0.966641902923584, "learning_rate": 2.7566695944654997e-05, "loss": 0.4647, "step": 19628 }, { "epoch": 1.317371900271803, "grad_norm": 0.9684114456176758, "learning_rate": 2.7556983474979093e-05, "loss": 0.4965, "step": 19630 }, { "epoch": 1.3175061239555719, "grad_norm": 0.9587264060974121, "learning_rate": 2.754727206569161e-05, "loss": 0.5264, "step": 19632 }, { "epoch": 1.317640347639341, "grad_norm": 1.021653652191162, "learning_rate": 2.753756171725137e-05, "loss": 0.508, "step": 19634 }, { "epoch": 1.31777457132311, "grad_norm": 0.9952699542045593, "learning_rate": 2.7527852430117167e-05, "loss": 0.4698, "step": 19636 }, { "epoch": 1.317908795006879, "grad_norm": 0.9360623359680176, "learning_rate": 2.7518144204747732e-05, "loss": 0.4765, "step": 19638 }, { "epoch": 1.318043018690648, "grad_norm": 0.9587939381599426, "learning_rate": 2.7508437041601786e-05, "loss": 0.4329, "step": 19640 }, { "epoch": 1.318177242374417, "grad_norm": 1.0178338289260864, "learning_rate": 2.749873094113797e-05, "loss": 0.5465, "step": 19642 }, { "epoch": 1.3183114660581858, "grad_norm": 1.1787763833999634, "learning_rate": 2.748902590381487e-05, "loss": 0.4357, "step": 19644 }, { "epoch": 1.318445689741955, "grad_norm": 1.0718886852264404, "learning_rate": 2.7479321930091023e-05, "loss": 0.4981, "step": 19646 }, { "epoch": 1.318579913425724, "grad_norm": 1.336592435836792, "learning_rate": 2.7469619020424913e-05, "loss": 0.453, "step": 19648 }, { "epoch": 1.318714137109493, "grad_norm": 1.0587416887283325, "learning_rate": 2.7459917175275018e-05, "loss": 0.4984, "step": 19650 }, { "epoch": 1.318848360793262, "grad_norm": 1.0716010332107544, "learning_rate": 2.7450216395099705e-05, "loss": 0.5693, "step": 19652 }, { "epoch": 1.3189825844770309, "grad_norm": 0.9708546996116638, "learning_rate": 2.7440516680357354e-05, "loss": 0.5115, "step": 19654 }, { "epoch": 1.3191168081608, "grad_norm": 1.0340770483016968, "learning_rate": 2.7430818031506188e-05, "loss": 0.494, "step": 19656 }, { "epoch": 1.319251031844569, "grad_norm": 1.0604135990142822, "learning_rate": 2.742112044900451e-05, "loss": 0.5184, "step": 19658 }, { "epoch": 1.319385255528338, "grad_norm": 0.9252541065216064, "learning_rate": 2.7411423933310476e-05, "loss": 0.4121, "step": 19660 }, { "epoch": 1.319519479212107, "grad_norm": 1.0229439735412598, "learning_rate": 2.740172848488226e-05, "loss": 0.4907, "step": 19662 }, { "epoch": 1.319653702895876, "grad_norm": 0.9297716021537781, "learning_rate": 2.739203410417794e-05, "loss": 0.5034, "step": 19664 }, { "epoch": 1.319787926579645, "grad_norm": 0.9669817090034485, "learning_rate": 2.738234079165555e-05, "loss": 0.5169, "step": 19666 }, { "epoch": 1.319922150263414, "grad_norm": 1.0199397802352905, "learning_rate": 2.737264854777306e-05, "loss": 0.4526, "step": 19668 }, { "epoch": 1.3200563739471831, "grad_norm": 0.8246620893478394, "learning_rate": 2.7362957372988452e-05, "loss": 0.4251, "step": 19670 }, { "epoch": 1.320190597630952, "grad_norm": 1.0211527347564697, "learning_rate": 2.7353267267759587e-05, "loss": 0.5057, "step": 19672 }, { "epoch": 1.320324821314721, "grad_norm": 1.0440993309020996, "learning_rate": 2.734357823254432e-05, "loss": 0.5367, "step": 19674 }, { "epoch": 1.3204590449984899, "grad_norm": 1.2122056484222412, "learning_rate": 2.7333890267800412e-05, "loss": 0.5195, "step": 19676 }, { "epoch": 1.320593268682259, "grad_norm": 1.0193716287612915, "learning_rate": 2.7324203373985626e-05, "loss": 0.5795, "step": 19678 }, { "epoch": 1.320727492366028, "grad_norm": 0.9539496898651123, "learning_rate": 2.7314517551557627e-05, "loss": 0.4247, "step": 19680 }, { "epoch": 1.320861716049797, "grad_norm": 1.0349043607711792, "learning_rate": 2.7304832800974105e-05, "loss": 0.5161, "step": 19682 }, { "epoch": 1.320995939733566, "grad_norm": 0.9099202156066895, "learning_rate": 2.7295149122692566e-05, "loss": 0.4728, "step": 19684 }, { "epoch": 1.321130163417335, "grad_norm": 0.9715734720230103, "learning_rate": 2.7285466517170605e-05, "loss": 0.5016, "step": 19686 }, { "epoch": 1.321264387101104, "grad_norm": 1.1370961666107178, "learning_rate": 2.727578498486566e-05, "loss": 0.4813, "step": 19688 }, { "epoch": 1.321398610784873, "grad_norm": 1.0140807628631592, "learning_rate": 2.7266104526235215e-05, "loss": 0.5288, "step": 19690 }, { "epoch": 1.3215328344686421, "grad_norm": 1.0584005117416382, "learning_rate": 2.725642514173662e-05, "loss": 0.4736, "step": 19692 }, { "epoch": 1.321667058152411, "grad_norm": 0.9137961268424988, "learning_rate": 2.724674683182722e-05, "loss": 0.4887, "step": 19694 }, { "epoch": 1.32180128183618, "grad_norm": 1.1081452369689941, "learning_rate": 2.7237069596964266e-05, "loss": 0.4869, "step": 19696 }, { "epoch": 1.3219355055199489, "grad_norm": 1.1327177286148071, "learning_rate": 2.722739343760503e-05, "loss": 0.5325, "step": 19698 }, { "epoch": 1.322069729203718, "grad_norm": 0.9909605979919434, "learning_rate": 2.721771835420668e-05, "loss": 0.5747, "step": 19700 }, { "epoch": 1.322203952887487, "grad_norm": 1.0612339973449707, "learning_rate": 2.7208044347226335e-05, "loss": 0.4779, "step": 19702 }, { "epoch": 1.322338176571256, "grad_norm": 0.8930952548980713, "learning_rate": 2.719837141712106e-05, "loss": 0.4119, "step": 19704 }, { "epoch": 1.322472400255025, "grad_norm": 0.9589643478393555, "learning_rate": 2.718869956434791e-05, "loss": 0.4433, "step": 19706 }, { "epoch": 1.322606623938794, "grad_norm": 1.1247379779815674, "learning_rate": 2.717902878936386e-05, "loss": 0.5554, "step": 19708 }, { "epoch": 1.322740847622563, "grad_norm": 1.1248613595962524, "learning_rate": 2.7169359092625813e-05, "loss": 0.5025, "step": 19710 }, { "epoch": 1.322875071306332, "grad_norm": 1.054959774017334, "learning_rate": 2.715969047459066e-05, "loss": 0.5498, "step": 19712 }, { "epoch": 1.3230092949901011, "grad_norm": 1.0124552249908447, "learning_rate": 2.7150022935715196e-05, "loss": 0.5236, "step": 19714 }, { "epoch": 1.32314351867387, "grad_norm": 1.1516460180282593, "learning_rate": 2.714035647645624e-05, "loss": 0.5068, "step": 19716 }, { "epoch": 1.323277742357639, "grad_norm": 0.8048339486122131, "learning_rate": 2.7130691097270468e-05, "loss": 0.4566, "step": 19718 }, { "epoch": 1.3234119660414079, "grad_norm": 0.946758508682251, "learning_rate": 2.7121026798614583e-05, "loss": 0.4247, "step": 19720 }, { "epoch": 1.323546189725177, "grad_norm": 1.184579849243164, "learning_rate": 2.7111363580945202e-05, "loss": 0.542, "step": 19722 }, { "epoch": 1.323680413408946, "grad_norm": 1.1211533546447754, "learning_rate": 2.710170144471888e-05, "loss": 0.4784, "step": 19724 }, { "epoch": 1.323814637092715, "grad_norm": 1.029905080795288, "learning_rate": 2.7092040390392115e-05, "loss": 0.5168, "step": 19726 }, { "epoch": 1.323948860776484, "grad_norm": 0.9693290591239929, "learning_rate": 2.7082380418421417e-05, "loss": 0.5317, "step": 19728 }, { "epoch": 1.324083084460253, "grad_norm": 1.0033167600631714, "learning_rate": 2.7072721529263177e-05, "loss": 0.4559, "step": 19730 }, { "epoch": 1.324217308144022, "grad_norm": 1.0830321311950684, "learning_rate": 2.706306372337376e-05, "loss": 0.4772, "step": 19732 }, { "epoch": 1.324351531827791, "grad_norm": 1.0198765993118286, "learning_rate": 2.7053407001209465e-05, "loss": 0.5029, "step": 19734 }, { "epoch": 1.3244857555115601, "grad_norm": 1.114503264427185, "learning_rate": 2.7043751363226575e-05, "loss": 0.4737, "step": 19736 }, { "epoch": 1.324619979195329, "grad_norm": 1.0865391492843628, "learning_rate": 2.703409680988128e-05, "loss": 0.4728, "step": 19738 }, { "epoch": 1.324754202879098, "grad_norm": 1.020459771156311, "learning_rate": 2.702444334162979e-05, "loss": 0.4673, "step": 19740 }, { "epoch": 1.324888426562867, "grad_norm": 0.9951459765434265, "learning_rate": 2.7014790958928138e-05, "loss": 0.4764, "step": 19742 }, { "epoch": 1.325022650246636, "grad_norm": 0.8151239156723022, "learning_rate": 2.7005139662232425e-05, "loss": 0.4571, "step": 19744 }, { "epoch": 1.3251568739304052, "grad_norm": 1.0674978494644165, "learning_rate": 2.699548945199863e-05, "loss": 0.5049, "step": 19746 }, { "epoch": 1.325291097614174, "grad_norm": 1.6327846050262451, "learning_rate": 2.6985840328682737e-05, "loss": 0.5086, "step": 19748 }, { "epoch": 1.325425321297943, "grad_norm": 1.055503487586975, "learning_rate": 2.6976192292740637e-05, "loss": 0.5274, "step": 19750 }, { "epoch": 1.325559544981712, "grad_norm": 0.9699060916900635, "learning_rate": 2.696654534462818e-05, "loss": 0.5392, "step": 19752 }, { "epoch": 1.325693768665481, "grad_norm": 0.9717655181884766, "learning_rate": 2.6956899484801134e-05, "loss": 0.5691, "step": 19754 }, { "epoch": 1.32582799234925, "grad_norm": 0.9502808451652527, "learning_rate": 2.6947254713715304e-05, "loss": 0.4602, "step": 19756 }, { "epoch": 1.3259622160330191, "grad_norm": 0.837290346622467, "learning_rate": 2.693761103182635e-05, "loss": 0.4719, "step": 19758 }, { "epoch": 1.326096439716788, "grad_norm": 0.9364441633224487, "learning_rate": 2.692796843958993e-05, "loss": 0.4502, "step": 19760 }, { "epoch": 1.326230663400557, "grad_norm": 1.1107641458511353, "learning_rate": 2.691832693746161e-05, "loss": 0.565, "step": 19762 }, { "epoch": 1.326364887084326, "grad_norm": 1.2745716571807861, "learning_rate": 2.6908686525896977e-05, "loss": 0.4906, "step": 19764 }, { "epoch": 1.326499110768095, "grad_norm": 1.05853271484375, "learning_rate": 2.689904720535147e-05, "loss": 0.5468, "step": 19766 }, { "epoch": 1.3266333344518642, "grad_norm": 1.0703691244125366, "learning_rate": 2.68894089762806e-05, "loss": 0.4954, "step": 19768 }, { "epoch": 1.326767558135633, "grad_norm": 0.9572019577026367, "learning_rate": 2.6879771839139678e-05, "loss": 0.4597, "step": 19770 }, { "epoch": 1.326901781819402, "grad_norm": 1.1199136972427368, "learning_rate": 2.6870135794384084e-05, "loss": 0.6664, "step": 19772 }, { "epoch": 1.327036005503171, "grad_norm": 1.0304343700408936, "learning_rate": 2.686050084246907e-05, "loss": 0.4733, "step": 19774 }, { "epoch": 1.32717022918694, "grad_norm": 0.8788327574729919, "learning_rate": 2.6850866983849915e-05, "loss": 0.4735, "step": 19776 }, { "epoch": 1.327304452870709, "grad_norm": 0.9124723672866821, "learning_rate": 2.684123421898179e-05, "loss": 0.5136, "step": 19778 }, { "epoch": 1.3274386765544781, "grad_norm": 1.0493173599243164, "learning_rate": 2.6831602548319773e-05, "loss": 0.4791, "step": 19780 }, { "epoch": 1.327572900238247, "grad_norm": 1.0095930099487305, "learning_rate": 2.6821971972318992e-05, "loss": 0.4779, "step": 19782 }, { "epoch": 1.327707123922016, "grad_norm": 2.1133108139038086, "learning_rate": 2.6812342491434444e-05, "loss": 0.4133, "step": 19784 }, { "epoch": 1.327841347605785, "grad_norm": 2.8399405479431152, "learning_rate": 2.680271410612113e-05, "loss": 0.5151, "step": 19786 }, { "epoch": 1.327975571289554, "grad_norm": 0.8534837961196899, "learning_rate": 2.6793086816833967e-05, "loss": 0.4798, "step": 19788 }, { "epoch": 1.3281097949733232, "grad_norm": 1.0487769842147827, "learning_rate": 2.6783460624027813e-05, "loss": 0.5263, "step": 19790 }, { "epoch": 1.328244018657092, "grad_norm": 0.954577624797821, "learning_rate": 2.6773835528157464e-05, "loss": 0.4473, "step": 19792 }, { "epoch": 1.328378242340861, "grad_norm": 1.010227084159851, "learning_rate": 2.676421152967774e-05, "loss": 0.5166, "step": 19794 }, { "epoch": 1.32851246602463, "grad_norm": 1.0502123832702637, "learning_rate": 2.675458862904333e-05, "loss": 0.4875, "step": 19796 }, { "epoch": 1.328646689708399, "grad_norm": 1.1167939901351929, "learning_rate": 2.6744966826708906e-05, "loss": 0.5347, "step": 19798 }, { "epoch": 1.328780913392168, "grad_norm": 1.131696343421936, "learning_rate": 2.673534612312904e-05, "loss": 0.5007, "step": 19800 }, { "epoch": 1.3289151370759371, "grad_norm": 0.9861233234405518, "learning_rate": 2.6725726518758344e-05, "loss": 0.5568, "step": 19802 }, { "epoch": 1.329049360759706, "grad_norm": 0.9725451469421387, "learning_rate": 2.6716108014051282e-05, "loss": 0.5044, "step": 19804 }, { "epoch": 1.329183584443475, "grad_norm": 0.9981563091278076, "learning_rate": 2.670649060946237e-05, "loss": 0.5231, "step": 19806 }, { "epoch": 1.329317808127244, "grad_norm": 1.1433606147766113, "learning_rate": 2.6696874305445936e-05, "loss": 0.5069, "step": 19808 }, { "epoch": 1.329452031811013, "grad_norm": 1.113306999206543, "learning_rate": 2.6687259102456386e-05, "loss": 0.4659, "step": 19810 }, { "epoch": 1.3295862554947822, "grad_norm": 1.0019900798797607, "learning_rate": 2.6677645000947982e-05, "loss": 0.4865, "step": 19812 }, { "epoch": 1.329720479178551, "grad_norm": 0.9664418697357178, "learning_rate": 2.6668032001375002e-05, "loss": 0.4992, "step": 19814 }, { "epoch": 1.32985470286232, "grad_norm": 1.1979312896728516, "learning_rate": 2.665842010419164e-05, "loss": 0.4842, "step": 19816 }, { "epoch": 1.3299889265460891, "grad_norm": 1.013890027999878, "learning_rate": 2.6648809309852017e-05, "loss": 0.5108, "step": 19818 }, { "epoch": 1.330123150229858, "grad_norm": 0.8482784032821655, "learning_rate": 2.6639199618810228e-05, "loss": 0.4415, "step": 19820 }, { "epoch": 1.3302573739136272, "grad_norm": 1.0452470779418945, "learning_rate": 2.6629591031520334e-05, "loss": 0.5919, "step": 19822 }, { "epoch": 1.3303915975973961, "grad_norm": 1.0971609354019165, "learning_rate": 2.6619983548436313e-05, "loss": 0.4799, "step": 19824 }, { "epoch": 1.330525821281165, "grad_norm": 1.0110511779785156, "learning_rate": 2.661037717001209e-05, "loss": 0.5498, "step": 19826 }, { "epoch": 1.330660044964934, "grad_norm": 1.140190601348877, "learning_rate": 2.660077189670153e-05, "loss": 0.4567, "step": 19828 }, { "epoch": 1.330794268648703, "grad_norm": 1.0503569841384888, "learning_rate": 2.659116772895851e-05, "loss": 0.5217, "step": 19830 }, { "epoch": 1.330928492332472, "grad_norm": 1.0813215970993042, "learning_rate": 2.6581564667236758e-05, "loss": 0.4438, "step": 19832 }, { "epoch": 1.3310627160162412, "grad_norm": 0.8762857913970947, "learning_rate": 2.6571962711990073e-05, "loss": 0.5069, "step": 19834 }, { "epoch": 1.33119693970001, "grad_norm": 1.0669820308685303, "learning_rate": 2.6562361863672037e-05, "loss": 0.4461, "step": 19836 }, { "epoch": 1.331331163383779, "grad_norm": 1.0361627340316772, "learning_rate": 2.655276212273633e-05, "loss": 0.489, "step": 19838 }, { "epoch": 1.3314653870675481, "grad_norm": 1.078757405281067, "learning_rate": 2.654316348963649e-05, "loss": 0.4468, "step": 19840 }, { "epoch": 1.331599610751317, "grad_norm": 1.0235754251480103, "learning_rate": 2.653356596482607e-05, "loss": 0.5024, "step": 19842 }, { "epoch": 1.3317338344350862, "grad_norm": 1.0501896142959595, "learning_rate": 2.652396954875851e-05, "loss": 0.5146, "step": 19844 }, { "epoch": 1.3318680581188551, "grad_norm": 0.8992530107498169, "learning_rate": 2.651437424188723e-05, "loss": 0.5311, "step": 19846 }, { "epoch": 1.332002281802624, "grad_norm": 1.1598973274230957, "learning_rate": 2.6504780044665578e-05, "loss": 0.5134, "step": 19848 }, { "epoch": 1.332136505486393, "grad_norm": 0.9470548629760742, "learning_rate": 2.6495186957546852e-05, "loss": 0.5206, "step": 19850 }, { "epoch": 1.332270729170162, "grad_norm": 1.2048908472061157, "learning_rate": 2.648559498098434e-05, "loss": 0.5537, "step": 19852 }, { "epoch": 1.332404952853931, "grad_norm": 0.9466072916984558, "learning_rate": 2.647600411543123e-05, "loss": 0.4738, "step": 19854 }, { "epoch": 1.3325391765377002, "grad_norm": 1.1737912893295288, "learning_rate": 2.6466414361340663e-05, "loss": 0.5499, "step": 19856 }, { "epoch": 1.332673400221469, "grad_norm": 1.0548076629638672, "learning_rate": 2.6456825719165723e-05, "loss": 0.5891, "step": 19858 }, { "epoch": 1.332807623905238, "grad_norm": 0.9503657817840576, "learning_rate": 2.644723818935949e-05, "loss": 0.5022, "step": 19860 }, { "epoch": 1.3329418475890071, "grad_norm": 1.0045098066329956, "learning_rate": 2.6437651772374937e-05, "loss": 0.4418, "step": 19862 }, { "epoch": 1.333076071272776, "grad_norm": 0.9361532926559448, "learning_rate": 2.6428066468665013e-05, "loss": 0.4746, "step": 19864 }, { "epoch": 1.3332102949565452, "grad_norm": 1.0541296005249023, "learning_rate": 2.6418482278682566e-05, "loss": 0.4461, "step": 19866 }, { "epoch": 1.3333445186403141, "grad_norm": 0.9788280129432678, "learning_rate": 2.640889920288049e-05, "loss": 0.5003, "step": 19868 }, { "epoch": 1.333478742324083, "grad_norm": 1.0545464754104614, "learning_rate": 2.6399317241711507e-05, "loss": 0.5072, "step": 19870 }, { "epoch": 1.333612966007852, "grad_norm": 1.0186614990234375, "learning_rate": 2.63897363956284e-05, "loss": 0.4442, "step": 19872 }, { "epoch": 1.333747189691621, "grad_norm": 1.0518426895141602, "learning_rate": 2.638015666508382e-05, "loss": 0.4598, "step": 19874 }, { "epoch": 1.33388141337539, "grad_norm": 1.012442946434021, "learning_rate": 2.6370578050530392e-05, "loss": 0.544, "step": 19876 }, { "epoch": 1.3340156370591592, "grad_norm": 1.0759896039962769, "learning_rate": 2.636100055242067e-05, "loss": 0.4876, "step": 19878 }, { "epoch": 1.334149860742928, "grad_norm": 1.1002321243286133, "learning_rate": 2.63514241712072e-05, "loss": 0.5496, "step": 19880 }, { "epoch": 1.334284084426697, "grad_norm": 1.0845890045166016, "learning_rate": 2.6341848907342436e-05, "loss": 0.5002, "step": 19882 }, { "epoch": 1.3344183081104661, "grad_norm": 0.9589855074882507, "learning_rate": 2.6332274761278797e-05, "loss": 0.5202, "step": 19884 }, { "epoch": 1.334552531794235, "grad_norm": 0.9928118586540222, "learning_rate": 2.6322701733468608e-05, "loss": 0.4507, "step": 19886 }, { "epoch": 1.3346867554780042, "grad_norm": 0.9842401146888733, "learning_rate": 2.6313129824364224e-05, "loss": 0.555, "step": 19888 }, { "epoch": 1.3348209791617731, "grad_norm": 1.1226009130477905, "learning_rate": 2.630355903441788e-05, "loss": 0.5192, "step": 19890 }, { "epoch": 1.334955202845542, "grad_norm": 1.1191109418869019, "learning_rate": 2.6293989364081774e-05, "loss": 0.5041, "step": 19892 }, { "epoch": 1.3350894265293112, "grad_norm": 0.9476625323295593, "learning_rate": 2.628442081380803e-05, "loss": 0.4581, "step": 19894 }, { "epoch": 1.33522365021308, "grad_norm": 0.902819812297821, "learning_rate": 2.6274853384048793e-05, "loss": 0.4951, "step": 19896 }, { "epoch": 1.3353578738968492, "grad_norm": 0.9704957604408264, "learning_rate": 2.6265287075256057e-05, "loss": 0.4322, "step": 19898 }, { "epoch": 1.3354920975806182, "grad_norm": 1.0102123022079468, "learning_rate": 2.6255721887881857e-05, "loss": 0.4701, "step": 19900 }, { "epoch": 1.335626321264387, "grad_norm": 0.9383575916290283, "learning_rate": 2.6246157822378104e-05, "loss": 0.4717, "step": 19902 }, { "epoch": 1.335760544948156, "grad_norm": 1.066762924194336, "learning_rate": 2.6236594879196685e-05, "loss": 0.4914, "step": 19904 }, { "epoch": 1.3358947686319251, "grad_norm": 0.9062620401382446, "learning_rate": 2.6227033058789408e-05, "loss": 0.5052, "step": 19906 }, { "epoch": 1.336028992315694, "grad_norm": 1.233633041381836, "learning_rate": 2.6217472361608094e-05, "loss": 0.4843, "step": 19908 }, { "epoch": 1.3361632159994632, "grad_norm": 1.0275330543518066, "learning_rate": 2.6207912788104438e-05, "loss": 0.5502, "step": 19910 }, { "epoch": 1.3362974396832321, "grad_norm": 0.9860764145851135, "learning_rate": 2.6198354338730123e-05, "loss": 0.4641, "step": 19912 }, { "epoch": 1.336431663367001, "grad_norm": 0.9409505128860474, "learning_rate": 2.6188797013936762e-05, "loss": 0.5787, "step": 19914 }, { "epoch": 1.3365658870507702, "grad_norm": 1.118374228477478, "learning_rate": 2.6179240814175897e-05, "loss": 0.4895, "step": 19916 }, { "epoch": 1.336700110734539, "grad_norm": 0.9939571022987366, "learning_rate": 2.6169685739899085e-05, "loss": 0.5625, "step": 19918 }, { "epoch": 1.3368343344183082, "grad_norm": 0.8455119132995605, "learning_rate": 2.6160131791557763e-05, "loss": 0.4674, "step": 19920 }, { "epoch": 1.3369685581020772, "grad_norm": 1.0985572338104248, "learning_rate": 2.6150578969603334e-05, "loss": 0.5249, "step": 19922 }, { "epoch": 1.337102781785846, "grad_norm": 1.0769293308258057, "learning_rate": 2.614102727448713e-05, "loss": 0.5309, "step": 19924 }, { "epoch": 1.337237005469615, "grad_norm": 1.0132557153701782, "learning_rate": 2.6131476706660485e-05, "loss": 0.4845, "step": 19926 }, { "epoch": 1.3373712291533841, "grad_norm": 1.0289278030395508, "learning_rate": 2.612192726657462e-05, "loss": 0.4253, "step": 19928 }, { "epoch": 1.337505452837153, "grad_norm": 1.099208950996399, "learning_rate": 2.6112378954680773e-05, "loss": 0.4906, "step": 19930 }, { "epoch": 1.3376396765209222, "grad_norm": 0.9392862319946289, "learning_rate": 2.6102831771430003e-05, "loss": 0.4784, "step": 19932 }, { "epoch": 1.3377739002046911, "grad_norm": 1.1693406105041504, "learning_rate": 2.6093285717273465e-05, "loss": 0.4967, "step": 19934 }, { "epoch": 1.33790812388846, "grad_norm": 1.182614803314209, "learning_rate": 2.6083740792662148e-05, "loss": 0.5382, "step": 19936 }, { "epoch": 1.3380423475722292, "grad_norm": 1.1358389854431152, "learning_rate": 2.607419699804706e-05, "loss": 0.587, "step": 19938 }, { "epoch": 1.338176571255998, "grad_norm": 0.9437689185142517, "learning_rate": 2.6064654333879123e-05, "loss": 0.5281, "step": 19940 }, { "epoch": 1.3383107949397672, "grad_norm": 1.1665679216384888, "learning_rate": 2.6055112800609206e-05, "loss": 0.5308, "step": 19942 }, { "epoch": 1.3384450186235362, "grad_norm": 1.0980030298233032, "learning_rate": 2.6045572398688095e-05, "loss": 0.4539, "step": 19944 }, { "epoch": 1.338579242307305, "grad_norm": 1.079601764678955, "learning_rate": 2.6036033128566608e-05, "loss": 0.5316, "step": 19946 }, { "epoch": 1.338713465991074, "grad_norm": 1.1160887479782104, "learning_rate": 2.6026494990695426e-05, "loss": 0.5248, "step": 19948 }, { "epoch": 1.3388476896748431, "grad_norm": 1.0072522163391113, "learning_rate": 2.6016957985525227e-05, "loss": 0.4889, "step": 19950 }, { "epoch": 1.338981913358612, "grad_norm": 0.8797785639762878, "learning_rate": 2.6007422113506573e-05, "loss": 0.5538, "step": 19952 }, { "epoch": 1.3391161370423812, "grad_norm": 0.938903272151947, "learning_rate": 2.599788737509007e-05, "loss": 0.5266, "step": 19954 }, { "epoch": 1.3392503607261501, "grad_norm": 0.9888512492179871, "learning_rate": 2.5988353770726166e-05, "loss": 0.4562, "step": 19956 }, { "epoch": 1.339384584409919, "grad_norm": 0.9281738996505737, "learning_rate": 2.5978821300865368e-05, "loss": 0.4834, "step": 19958 }, { "epoch": 1.3395188080936882, "grad_norm": 0.9474917650222778, "learning_rate": 2.5969289965957988e-05, "loss": 0.382, "step": 19960 }, { "epoch": 1.339653031777457, "grad_norm": 1.0445213317871094, "learning_rate": 2.5959759766454428e-05, "loss": 0.4968, "step": 19962 }, { "epoch": 1.3397872554612262, "grad_norm": 1.082736611366272, "learning_rate": 2.5950230702804923e-05, "loss": 0.4946, "step": 19964 }, { "epoch": 1.3399214791449952, "grad_norm": 1.0125893354415894, "learning_rate": 2.5940702775459747e-05, "loss": 0.4805, "step": 19966 }, { "epoch": 1.340055702828764, "grad_norm": 0.8601479530334473, "learning_rate": 2.593117598486905e-05, "loss": 0.5045, "step": 19968 }, { "epoch": 1.3401899265125332, "grad_norm": 0.9454371929168701, "learning_rate": 2.5921650331482962e-05, "loss": 0.4602, "step": 19970 }, { "epoch": 1.3403241501963021, "grad_norm": 1.1329702138900757, "learning_rate": 2.591212581575153e-05, "loss": 0.4194, "step": 19972 }, { "epoch": 1.3404583738800713, "grad_norm": 0.9190551042556763, "learning_rate": 2.59026024381248e-05, "loss": 0.491, "step": 19974 }, { "epoch": 1.3405925975638402, "grad_norm": 0.9530711770057678, "learning_rate": 2.589308019905273e-05, "loss": 0.5468, "step": 19976 }, { "epoch": 1.3407268212476091, "grad_norm": 1.036069393157959, "learning_rate": 2.5883559098985204e-05, "loss": 0.5485, "step": 19978 }, { "epoch": 1.340861044931378, "grad_norm": 0.9575065970420837, "learning_rate": 2.5874039138372075e-05, "loss": 0.4476, "step": 19980 }, { "epoch": 1.3409952686151472, "grad_norm": 0.9909329414367676, "learning_rate": 2.586452031766317e-05, "loss": 0.5144, "step": 19982 }, { "epoch": 1.341129492298916, "grad_norm": 1.0757123231887817, "learning_rate": 2.5855002637308224e-05, "loss": 0.5254, "step": 19984 }, { "epoch": 1.3412637159826852, "grad_norm": 0.9575532674789429, "learning_rate": 2.584548609775692e-05, "loss": 0.4818, "step": 19986 }, { "epoch": 1.3413979396664542, "grad_norm": 0.9985975623130798, "learning_rate": 2.5835970699458906e-05, "loss": 0.4669, "step": 19988 }, { "epoch": 1.341532163350223, "grad_norm": 0.9868609309196472, "learning_rate": 2.582645644286374e-05, "loss": 0.4987, "step": 19990 }, { "epoch": 1.3416663870339922, "grad_norm": 0.9942282438278198, "learning_rate": 2.5816943328420984e-05, "loss": 0.5489, "step": 19992 }, { "epoch": 1.3418006107177611, "grad_norm": 1.0065810680389404, "learning_rate": 2.5807431356580092e-05, "loss": 0.4594, "step": 19994 }, { "epoch": 1.3419348344015303, "grad_norm": 1.054410457611084, "learning_rate": 2.5797920527790526e-05, "loss": 0.5011, "step": 19996 }, { "epoch": 1.3420690580852992, "grad_norm": 0.9025766253471375, "learning_rate": 2.5788410842501588e-05, "loss": 0.5257, "step": 19998 }, { "epoch": 1.3422032817690681, "grad_norm": 0.9478522539138794, "learning_rate": 2.577890230116265e-05, "loss": 0.4517, "step": 20000 }, { "epoch": 1.342337505452837, "grad_norm": 0.9698958396911621, "learning_rate": 2.5769394904222926e-05, "loss": 0.4975, "step": 20002 }, { "epoch": 1.3424717291366062, "grad_norm": 1.0644009113311768, "learning_rate": 2.575988865213167e-05, "loss": 0.4529, "step": 20004 }, { "epoch": 1.342605952820375, "grad_norm": 1.110205888748169, "learning_rate": 2.5750383545338002e-05, "loss": 0.5225, "step": 20006 }, { "epoch": 1.3427401765041442, "grad_norm": 1.094616174697876, "learning_rate": 2.5740879584291034e-05, "loss": 0.4981, "step": 20008 }, { "epoch": 1.3428744001879132, "grad_norm": 0.9325581789016724, "learning_rate": 2.5731376769439776e-05, "loss": 0.5181, "step": 20010 }, { "epoch": 1.343008623871682, "grad_norm": 1.0330641269683838, "learning_rate": 2.572187510123326e-05, "loss": 0.499, "step": 20012 }, { "epoch": 1.3431428475554512, "grad_norm": 1.054914951324463, "learning_rate": 2.5712374580120414e-05, "loss": 0.5062, "step": 20014 }, { "epoch": 1.3432770712392201, "grad_norm": 0.9782353043556213, "learning_rate": 2.5702875206550105e-05, "loss": 0.5673, "step": 20016 }, { "epoch": 1.3434112949229893, "grad_norm": 1.109312891960144, "learning_rate": 2.5693376980971145e-05, "loss": 0.4858, "step": 20018 }, { "epoch": 1.3435455186067582, "grad_norm": 1.1009002923965454, "learning_rate": 2.5683879903832347e-05, "loss": 0.4806, "step": 20020 }, { "epoch": 1.3436797422905271, "grad_norm": 1.0390616655349731, "learning_rate": 2.567438397558239e-05, "loss": 0.5129, "step": 20022 }, { "epoch": 1.343813965974296, "grad_norm": 0.9547844529151917, "learning_rate": 2.5664889196669973e-05, "loss": 0.4629, "step": 20024 }, { "epoch": 1.3439481896580652, "grad_norm": 1.1426942348480225, "learning_rate": 2.5655395567543684e-05, "loss": 0.5165, "step": 20026 }, { "epoch": 1.344082413341834, "grad_norm": 1.0668246746063232, "learning_rate": 2.564590308865209e-05, "loss": 0.4642, "step": 20028 }, { "epoch": 1.3442166370256032, "grad_norm": 1.1261327266693115, "learning_rate": 2.5636411760443657e-05, "loss": 0.5446, "step": 20030 }, { "epoch": 1.3443508607093722, "grad_norm": 1.0175567865371704, "learning_rate": 2.5626921583366886e-05, "loss": 0.476, "step": 20032 }, { "epoch": 1.344485084393141, "grad_norm": 0.9739302396774292, "learning_rate": 2.561743255787014e-05, "loss": 0.5262, "step": 20034 }, { "epoch": 1.3446193080769102, "grad_norm": 0.9953427314758301, "learning_rate": 2.5607944684401764e-05, "loss": 0.4766, "step": 20036 }, { "epoch": 1.3447535317606791, "grad_norm": 0.981567919254303, "learning_rate": 2.559845796341001e-05, "loss": 0.4691, "step": 20038 }, { "epoch": 1.3448877554444483, "grad_norm": 1.0105043649673462, "learning_rate": 2.5588972395343157e-05, "loss": 0.4665, "step": 20040 }, { "epoch": 1.3450219791282172, "grad_norm": 1.0410128831863403, "learning_rate": 2.557948798064935e-05, "loss": 0.5859, "step": 20042 }, { "epoch": 1.3451562028119861, "grad_norm": 1.1123363971710205, "learning_rate": 2.557000471977673e-05, "loss": 0.4977, "step": 20044 }, { "epoch": 1.3452904264957553, "grad_norm": 1.0192078351974487, "learning_rate": 2.5560522613173314e-05, "loss": 0.446, "step": 20046 }, { "epoch": 1.3454246501795242, "grad_norm": 1.1729594469070435, "learning_rate": 2.555104166128717e-05, "loss": 0.476, "step": 20048 }, { "epoch": 1.3455588738632933, "grad_norm": 1.0416300296783447, "learning_rate": 2.554156186456621e-05, "loss": 0.5405, "step": 20050 }, { "epoch": 1.3456930975470622, "grad_norm": 1.0445678234100342, "learning_rate": 2.553208322345838e-05, "loss": 0.4817, "step": 20052 }, { "epoch": 1.3458273212308312, "grad_norm": 1.0217093229293823, "learning_rate": 2.552260573841152e-05, "loss": 0.5277, "step": 20054 }, { "epoch": 1.3459615449146, "grad_norm": 1.1795514822006226, "learning_rate": 2.5513129409873372e-05, "loss": 0.4549, "step": 20056 }, { "epoch": 1.3460957685983692, "grad_norm": 0.9328517317771912, "learning_rate": 2.550365423829172e-05, "loss": 0.4461, "step": 20058 }, { "epoch": 1.3462299922821381, "grad_norm": 1.020693302154541, "learning_rate": 2.5494180224114217e-05, "loss": 0.5468, "step": 20060 }, { "epoch": 1.3463642159659073, "grad_norm": 0.8491286039352417, "learning_rate": 2.548470736778853e-05, "loss": 0.4652, "step": 20062 }, { "epoch": 1.3464984396496762, "grad_norm": 0.8926975727081299, "learning_rate": 2.547523566976222e-05, "loss": 0.513, "step": 20064 }, { "epoch": 1.3466326633334451, "grad_norm": 1.0468785762786865, "learning_rate": 2.546576513048279e-05, "loss": 0.5602, "step": 20066 }, { "epoch": 1.3467668870172143, "grad_norm": 1.0157579183578491, "learning_rate": 2.5456295750397698e-05, "loss": 0.4753, "step": 20068 }, { "epoch": 1.3469011107009832, "grad_norm": 0.9623618721961975, "learning_rate": 2.5446827529954385e-05, "loss": 0.5044, "step": 20070 }, { "epoch": 1.3470353343847523, "grad_norm": 1.1392110586166382, "learning_rate": 2.543736046960019e-05, "loss": 0.4971, "step": 20072 }, { "epoch": 1.3471695580685212, "grad_norm": 1.296471357345581, "learning_rate": 2.5427894569782412e-05, "loss": 0.4338, "step": 20074 }, { "epoch": 1.3473037817522902, "grad_norm": 1.1017106771469116, "learning_rate": 2.5418429830948276e-05, "loss": 0.4988, "step": 20076 }, { "epoch": 1.347438005436059, "grad_norm": 1.0007191896438599, "learning_rate": 2.540896625354502e-05, "loss": 0.5375, "step": 20078 }, { "epoch": 1.3475722291198282, "grad_norm": 1.0098400115966797, "learning_rate": 2.5399503838019722e-05, "loss": 0.4733, "step": 20080 }, { "epoch": 1.3477064528035971, "grad_norm": 0.9135158061981201, "learning_rate": 2.539004258481954e-05, "loss": 0.4741, "step": 20082 }, { "epoch": 1.3478406764873663, "grad_norm": 0.9549738764762878, "learning_rate": 2.5380582494391413e-05, "loss": 0.5202, "step": 20084 }, { "epoch": 1.3479749001711352, "grad_norm": 1.1197456121444702, "learning_rate": 2.5371123567182363e-05, "loss": 0.4726, "step": 20086 }, { "epoch": 1.3481091238549041, "grad_norm": 0.9651446342468262, "learning_rate": 2.5361665803639282e-05, "loss": 0.4997, "step": 20088 }, { "epoch": 1.3482433475386733, "grad_norm": 1.2065728902816772, "learning_rate": 2.5352209204209065e-05, "loss": 0.5219, "step": 20090 }, { "epoch": 1.3483775712224422, "grad_norm": 1.1290955543518066, "learning_rate": 2.534275376933849e-05, "loss": 0.5273, "step": 20092 }, { "epoch": 1.3485117949062113, "grad_norm": 1.0234849452972412, "learning_rate": 2.533329949947431e-05, "loss": 0.4433, "step": 20094 }, { "epoch": 1.3486460185899802, "grad_norm": 0.9947075247764587, "learning_rate": 2.532384639506321e-05, "loss": 0.4801, "step": 20096 }, { "epoch": 1.3487802422737492, "grad_norm": 1.044743299484253, "learning_rate": 2.531439445655186e-05, "loss": 0.4686, "step": 20098 }, { "epoch": 1.348914465957518, "grad_norm": 1.1770745515823364, "learning_rate": 2.530494368438683e-05, "loss": 0.4632, "step": 20100 }, { "epoch": 1.3490486896412872, "grad_norm": 1.1431256532669067, "learning_rate": 2.5295494079014647e-05, "loss": 0.5243, "step": 20102 }, { "epoch": 1.3491829133250561, "grad_norm": 1.1603831052780151, "learning_rate": 2.528604564088178e-05, "loss": 0.5558, "step": 20104 }, { "epoch": 1.3493171370088253, "grad_norm": 1.0677740573883057, "learning_rate": 2.5276598370434666e-05, "loss": 0.4388, "step": 20106 }, { "epoch": 1.3494513606925942, "grad_norm": 0.952692449092865, "learning_rate": 2.5267152268119653e-05, "loss": 0.5032, "step": 20108 }, { "epoch": 1.3495855843763631, "grad_norm": 1.0014994144439697, "learning_rate": 2.5257707334383095e-05, "loss": 0.531, "step": 20110 }, { "epoch": 1.3497198080601323, "grad_norm": 0.9802749156951904, "learning_rate": 2.524826356967117e-05, "loss": 0.4783, "step": 20112 }, { "epoch": 1.3498540317439012, "grad_norm": 1.054930329322815, "learning_rate": 2.5238820974430143e-05, "loss": 0.4785, "step": 20114 }, { "epoch": 1.3499882554276703, "grad_norm": 1.3713618516921997, "learning_rate": 2.5229379549106115e-05, "loss": 0.5262, "step": 20116 }, { "epoch": 1.3501224791114392, "grad_norm": 1.056983232498169, "learning_rate": 2.521993929414521e-05, "loss": 0.5007, "step": 20118 }, { "epoch": 1.3502567027952082, "grad_norm": 0.9778099060058594, "learning_rate": 2.521050020999347e-05, "loss": 0.4966, "step": 20120 }, { "epoch": 1.3503909264789773, "grad_norm": 1.049523115158081, "learning_rate": 2.5201062297096807e-05, "loss": 0.5833, "step": 20122 }, { "epoch": 1.3505251501627462, "grad_norm": 0.9236955046653748, "learning_rate": 2.51916255559012e-05, "loss": 0.5228, "step": 20124 }, { "epoch": 1.3506593738465154, "grad_norm": 1.0097215175628662, "learning_rate": 2.518218998685249e-05, "loss": 0.5528, "step": 20126 }, { "epoch": 1.3507935975302843, "grad_norm": 0.9391131401062012, "learning_rate": 2.5172755590396517e-05, "loss": 0.4982, "step": 20128 }, { "epoch": 1.3509278212140532, "grad_norm": 0.8988783955574036, "learning_rate": 2.5163322366979024e-05, "loss": 0.5132, "step": 20130 }, { "epoch": 1.3510620448978221, "grad_norm": 0.9671209454536438, "learning_rate": 2.5153890317045702e-05, "loss": 0.4777, "step": 20132 }, { "epoch": 1.3511962685815913, "grad_norm": 1.1700220108032227, "learning_rate": 2.5144459441042196e-05, "loss": 0.5518, "step": 20134 }, { "epoch": 1.3513304922653602, "grad_norm": 1.0199360847473145, "learning_rate": 2.5135029739414116e-05, "loss": 0.4867, "step": 20136 }, { "epoch": 1.3514647159491293, "grad_norm": 0.9349581003189087, "learning_rate": 2.5125601212606992e-05, "loss": 0.4588, "step": 20138 }, { "epoch": 1.3515989396328982, "grad_norm": 1.1096220016479492, "learning_rate": 2.5116173861066296e-05, "loss": 0.458, "step": 20140 }, { "epoch": 1.3517331633166672, "grad_norm": 0.8157988786697388, "learning_rate": 2.510674768523743e-05, "loss": 0.4541, "step": 20142 }, { "epoch": 1.3518673870004363, "grad_norm": 1.1094350814819336, "learning_rate": 2.509732268556581e-05, "loss": 0.4659, "step": 20144 }, { "epoch": 1.3520016106842052, "grad_norm": 1.0682333707809448, "learning_rate": 2.5087898862496705e-05, "loss": 0.4778, "step": 20146 }, { "epoch": 1.3521358343679744, "grad_norm": 0.9747878313064575, "learning_rate": 2.5078476216475422e-05, "loss": 0.4584, "step": 20148 }, { "epoch": 1.3522700580517433, "grad_norm": 1.086549162864685, "learning_rate": 2.5069054747947102e-05, "loss": 0.5574, "step": 20150 }, { "epoch": 1.3524042817355122, "grad_norm": 1.0562546253204346, "learning_rate": 2.5059634457356933e-05, "loss": 0.4842, "step": 20152 }, { "epoch": 1.3525385054192811, "grad_norm": 1.0438446998596191, "learning_rate": 2.5050215345149975e-05, "loss": 0.5118, "step": 20154 }, { "epoch": 1.3526727291030503, "grad_norm": 1.0002363920211792, "learning_rate": 2.5040797411771306e-05, "loss": 0.4894, "step": 20156 }, { "epoch": 1.3528069527868192, "grad_norm": 1.0864993333816528, "learning_rate": 2.503138065766587e-05, "loss": 0.476, "step": 20158 }, { "epoch": 1.3529411764705883, "grad_norm": 1.0022969245910645, "learning_rate": 2.502196508327861e-05, "loss": 0.4776, "step": 20160 }, { "epoch": 1.3530754001543572, "grad_norm": 1.067216396331787, "learning_rate": 2.5012550689054355e-05, "loss": 0.4521, "step": 20162 }, { "epoch": 1.3532096238381262, "grad_norm": 0.8796282410621643, "learning_rate": 2.5003137475437966e-05, "loss": 0.454, "step": 20164 }, { "epoch": 1.3533438475218953, "grad_norm": 1.033448576927185, "learning_rate": 2.499372544287417e-05, "loss": 0.5156, "step": 20166 }, { "epoch": 1.3534780712056642, "grad_norm": 1.0945485830307007, "learning_rate": 2.4984314591807682e-05, "loss": 0.4867, "step": 20168 }, { "epoch": 1.3536122948894334, "grad_norm": 1.0850149393081665, "learning_rate": 2.4974904922683118e-05, "loss": 0.5128, "step": 20170 }, { "epoch": 1.3537465185732023, "grad_norm": 1.0059179067611694, "learning_rate": 2.4965496435945106e-05, "loss": 0.462, "step": 20172 }, { "epoch": 1.3538807422569712, "grad_norm": 1.287937045097351, "learning_rate": 2.495608913203814e-05, "loss": 0.4408, "step": 20174 }, { "epoch": 1.3540149659407401, "grad_norm": 1.053192377090454, "learning_rate": 2.4946683011406752e-05, "loss": 0.5038, "step": 20176 }, { "epoch": 1.3541491896245093, "grad_norm": 1.0209565162658691, "learning_rate": 2.493727807449529e-05, "loss": 0.4562, "step": 20178 }, { "epoch": 1.3542834133082782, "grad_norm": 1.0276302099227905, "learning_rate": 2.4927874321748173e-05, "loss": 0.4979, "step": 20180 }, { "epoch": 1.3544176369920473, "grad_norm": 0.7771276831626892, "learning_rate": 2.4918471753609675e-05, "loss": 0.4448, "step": 20182 }, { "epoch": 1.3545518606758162, "grad_norm": 0.8976522088050842, "learning_rate": 2.4909070370524083e-05, "loss": 0.4867, "step": 20184 }, { "epoch": 1.3546860843595852, "grad_norm": 0.9804712533950806, "learning_rate": 2.489967017293558e-05, "loss": 0.5131, "step": 20186 }, { "epoch": 1.3548203080433543, "grad_norm": 1.0005247592926025, "learning_rate": 2.4890271161288302e-05, "loss": 0.4038, "step": 20188 }, { "epoch": 1.3549545317271232, "grad_norm": 1.1950335502624512, "learning_rate": 2.4880873336026332e-05, "loss": 0.5083, "step": 20190 }, { "epoch": 1.3550887554108924, "grad_norm": 1.0356762409210205, "learning_rate": 2.4871476697593687e-05, "loss": 0.4998, "step": 20192 }, { "epoch": 1.3552229790946613, "grad_norm": 1.0134941339492798, "learning_rate": 2.4862081246434378e-05, "loss": 0.5141, "step": 20194 }, { "epoch": 1.3553572027784302, "grad_norm": 1.083375096321106, "learning_rate": 2.485268698299229e-05, "loss": 0.5193, "step": 20196 }, { "epoch": 1.3554914264621993, "grad_norm": 0.9247611165046692, "learning_rate": 2.4843293907711295e-05, "loss": 0.4712, "step": 20198 }, { "epoch": 1.3556256501459683, "grad_norm": 1.007331371307373, "learning_rate": 2.4833902021035177e-05, "loss": 0.4537, "step": 20200 }, { "epoch": 1.3557598738297374, "grad_norm": 0.9057896733283997, "learning_rate": 2.482451132340772e-05, "loss": 0.48, "step": 20202 }, { "epoch": 1.3558940975135063, "grad_norm": 0.9654235243797302, "learning_rate": 2.4815121815272573e-05, "loss": 0.5029, "step": 20204 }, { "epoch": 1.3560283211972752, "grad_norm": 0.9667559862136841, "learning_rate": 2.480573349707344e-05, "loss": 0.463, "step": 20206 }, { "epoch": 1.3561625448810442, "grad_norm": 1.0404314994812012, "learning_rate": 2.479634636925382e-05, "loss": 0.4783, "step": 20208 }, { "epoch": 1.3562967685648133, "grad_norm": 1.0094547271728516, "learning_rate": 2.478696043225729e-05, "loss": 0.4635, "step": 20210 }, { "epoch": 1.3564309922485822, "grad_norm": 1.0778117179870605, "learning_rate": 2.477757568652728e-05, "loss": 0.4946, "step": 20212 }, { "epoch": 1.3565652159323514, "grad_norm": 0.9366231560707092, "learning_rate": 2.4768192132507245e-05, "loss": 0.5015, "step": 20214 }, { "epoch": 1.3566994396161203, "grad_norm": 1.0686755180358887, "learning_rate": 2.475880977064051e-05, "loss": 0.473, "step": 20216 }, { "epoch": 1.3568336632998892, "grad_norm": 0.9536647200584412, "learning_rate": 2.4749428601370384e-05, "loss": 0.4982, "step": 20218 }, { "epoch": 1.3569678869836583, "grad_norm": 0.9279668927192688, "learning_rate": 2.4740048625140078e-05, "loss": 0.4641, "step": 20220 }, { "epoch": 1.3571021106674273, "grad_norm": 1.2315804958343506, "learning_rate": 2.473066984239282e-05, "loss": 0.4984, "step": 20222 }, { "epoch": 1.3572363343511964, "grad_norm": 0.9835600256919861, "learning_rate": 2.472129225357173e-05, "loss": 0.4598, "step": 20224 }, { "epoch": 1.3573705580349653, "grad_norm": 0.9124851226806641, "learning_rate": 2.471191585911987e-05, "loss": 0.494, "step": 20226 }, { "epoch": 1.3575047817187342, "grad_norm": 1.033633828163147, "learning_rate": 2.4702540659480238e-05, "loss": 0.472, "step": 20228 }, { "epoch": 1.3576390054025032, "grad_norm": 0.9878078699111938, "learning_rate": 2.469316665509584e-05, "loss": 0.4974, "step": 20230 }, { "epoch": 1.3577732290862723, "grad_norm": 0.9498612284660339, "learning_rate": 2.4683793846409524e-05, "loss": 0.4954, "step": 20232 }, { "epoch": 1.3579074527700412, "grad_norm": 1.0005344152450562, "learning_rate": 2.4674422233864213e-05, "loss": 0.4534, "step": 20234 }, { "epoch": 1.3580416764538104, "grad_norm": 1.2731600999832153, "learning_rate": 2.466505181790261e-05, "loss": 0.4823, "step": 20236 }, { "epoch": 1.3581759001375793, "grad_norm": 0.9503934979438782, "learning_rate": 2.465568259896751e-05, "loss": 0.5024, "step": 20238 }, { "epoch": 1.3583101238213482, "grad_norm": 0.9364110827445984, "learning_rate": 2.464631457750155e-05, "loss": 0.4845, "step": 20240 }, { "epoch": 1.3584443475051173, "grad_norm": 1.1178056001663208, "learning_rate": 2.4636947753947387e-05, "loss": 0.4674, "step": 20242 }, { "epoch": 1.3585785711888863, "grad_norm": 1.00879967212677, "learning_rate": 2.4627582128747567e-05, "loss": 0.4675, "step": 20244 }, { "epoch": 1.3587127948726554, "grad_norm": 0.9615558981895447, "learning_rate": 2.4618217702344603e-05, "loss": 0.481, "step": 20246 }, { "epoch": 1.3588470185564243, "grad_norm": 1.1916038990020752, "learning_rate": 2.460885447518092e-05, "loss": 0.5446, "step": 20248 }, { "epoch": 1.3589812422401932, "grad_norm": 1.065394401550293, "learning_rate": 2.4599492447698958e-05, "loss": 0.5024, "step": 20250 }, { "epoch": 1.3591154659239622, "grad_norm": 1.1207654476165771, "learning_rate": 2.4590131620341023e-05, "loss": 0.4775, "step": 20252 }, { "epoch": 1.3592496896077313, "grad_norm": 0.9586560726165771, "learning_rate": 2.4580771993549407e-05, "loss": 0.4822, "step": 20254 }, { "epoch": 1.3593839132915002, "grad_norm": 0.9638738036155701, "learning_rate": 2.4571413567766305e-05, "loss": 0.4469, "step": 20256 }, { "epoch": 1.3595181369752694, "grad_norm": 1.106136679649353, "learning_rate": 2.4562056343433933e-05, "loss": 0.5251, "step": 20258 }, { "epoch": 1.3596523606590383, "grad_norm": 1.168971061706543, "learning_rate": 2.4552700320994377e-05, "loss": 0.4528, "step": 20260 }, { "epoch": 1.3597865843428072, "grad_norm": 1.069299578666687, "learning_rate": 2.4543345500889687e-05, "loss": 0.5055, "step": 20262 }, { "epoch": 1.3599208080265763, "grad_norm": 1.0085757970809937, "learning_rate": 2.4533991883561868e-05, "loss": 0.4194, "step": 20264 }, { "epoch": 1.3600550317103453, "grad_norm": 0.9998415112495422, "learning_rate": 2.4524639469452838e-05, "loss": 0.4773, "step": 20266 }, { "epoch": 1.3601892553941144, "grad_norm": 1.1032497882843018, "learning_rate": 2.451528825900451e-05, "loss": 0.5662, "step": 20268 }, { "epoch": 1.3603234790778833, "grad_norm": 0.9551237225532532, "learning_rate": 2.4505938252658682e-05, "loss": 0.5417, "step": 20270 }, { "epoch": 1.3604577027616522, "grad_norm": 1.0504215955734253, "learning_rate": 2.449658945085718e-05, "loss": 0.4674, "step": 20272 }, { "epoch": 1.3605919264454214, "grad_norm": 1.1610937118530273, "learning_rate": 2.4487241854041637e-05, "loss": 0.5049, "step": 20274 }, { "epoch": 1.3607261501291903, "grad_norm": 0.9728044271469116, "learning_rate": 2.4477895462653764e-05, "loss": 0.4661, "step": 20276 }, { "epoch": 1.3608603738129594, "grad_norm": 0.8766085505485535, "learning_rate": 2.4468550277135128e-05, "loss": 0.4712, "step": 20278 }, { "epoch": 1.3609945974967284, "grad_norm": 1.001575231552124, "learning_rate": 2.4459206297927306e-05, "loss": 0.4446, "step": 20280 }, { "epoch": 1.3611288211804973, "grad_norm": 1.0977261066436768, "learning_rate": 2.4449863525471763e-05, "loss": 0.4841, "step": 20282 }, { "epoch": 1.3612630448642662, "grad_norm": 1.1366521120071411, "learning_rate": 2.4440521960209924e-05, "loss": 0.5129, "step": 20284 }, { "epoch": 1.3613972685480353, "grad_norm": 1.0719273090362549, "learning_rate": 2.4431181602583137e-05, "loss": 0.5287, "step": 20286 }, { "epoch": 1.3615314922318043, "grad_norm": 0.9917060732841492, "learning_rate": 2.442184245303277e-05, "loss": 0.4884, "step": 20288 }, { "epoch": 1.3616657159155734, "grad_norm": 1.0406906604766846, "learning_rate": 2.441250451200004e-05, "loss": 0.4561, "step": 20290 }, { "epoch": 1.3617999395993423, "grad_norm": 0.9567008018493652, "learning_rate": 2.4403167779926162e-05, "loss": 0.5076, "step": 20292 }, { "epoch": 1.3619341632831112, "grad_norm": 0.8407172560691833, "learning_rate": 2.4393832257252252e-05, "loss": 0.4622, "step": 20294 }, { "epoch": 1.3620683869668804, "grad_norm": 0.9681487679481506, "learning_rate": 2.438449794441943e-05, "loss": 0.4714, "step": 20296 }, { "epoch": 1.3622026106506493, "grad_norm": 0.9833024144172668, "learning_rate": 2.4375164841868697e-05, "loss": 0.4349, "step": 20298 }, { "epoch": 1.3623368343344184, "grad_norm": 1.0088200569152832, "learning_rate": 2.4365832950041072e-05, "loss": 0.443, "step": 20300 }, { "epoch": 1.3624710580181874, "grad_norm": 1.1553938388824463, "learning_rate": 2.4356502269377397e-05, "loss": 0.5802, "step": 20302 }, { "epoch": 1.3626052817019563, "grad_norm": 0.9181938171386719, "learning_rate": 2.434717280031858e-05, "loss": 0.4581, "step": 20304 }, { "epoch": 1.3627395053857252, "grad_norm": 1.0648936033248901, "learning_rate": 2.4337844543305387e-05, "loss": 0.5063, "step": 20306 }, { "epoch": 1.3628737290694943, "grad_norm": 1.046919822692871, "learning_rate": 2.4328517498778596e-05, "loss": 0.5244, "step": 20308 }, { "epoch": 1.3630079527532633, "grad_norm": 1.006034016609192, "learning_rate": 2.431919166717887e-05, "loss": 0.4525, "step": 20310 }, { "epoch": 1.3631421764370324, "grad_norm": 0.9807382225990295, "learning_rate": 2.4309867048946838e-05, "loss": 0.4511, "step": 20312 }, { "epoch": 1.3632764001208013, "grad_norm": 1.055794596672058, "learning_rate": 2.4300543644523056e-05, "loss": 0.4722, "step": 20314 }, { "epoch": 1.3634106238045702, "grad_norm": 1.0888453722000122, "learning_rate": 2.429122145434807e-05, "loss": 0.4709, "step": 20316 }, { "epoch": 1.3635448474883394, "grad_norm": 1.0583890676498413, "learning_rate": 2.4281900478862306e-05, "loss": 0.467, "step": 20318 }, { "epoch": 1.3636790711721083, "grad_norm": 1.1082299947738647, "learning_rate": 2.427258071850618e-05, "loss": 0.533, "step": 20320 }, { "epoch": 1.3638132948558774, "grad_norm": 0.9727580547332764, "learning_rate": 2.4263262173720008e-05, "loss": 0.5385, "step": 20322 }, { "epoch": 1.3639475185396464, "grad_norm": 1.0671035051345825, "learning_rate": 2.4253944844944105e-05, "loss": 0.485, "step": 20324 }, { "epoch": 1.3640817422234153, "grad_norm": 0.9661836624145508, "learning_rate": 2.424462873261868e-05, "loss": 0.4603, "step": 20326 }, { "epoch": 1.3642159659071842, "grad_norm": 0.9973102807998657, "learning_rate": 2.42353138371839e-05, "loss": 0.5054, "step": 20328 }, { "epoch": 1.3643501895909533, "grad_norm": 1.0182770490646362, "learning_rate": 2.422600015907988e-05, "loss": 0.4084, "step": 20330 }, { "epoch": 1.3644844132747223, "grad_norm": 0.9462610483169556, "learning_rate": 2.421668769874665e-05, "loss": 0.4351, "step": 20332 }, { "epoch": 1.3646186369584914, "grad_norm": 1.179429292678833, "learning_rate": 2.4207376456624243e-05, "loss": 0.5364, "step": 20334 }, { "epoch": 1.3647528606422603, "grad_norm": 1.2717273235321045, "learning_rate": 2.419806643315256e-05, "loss": 0.5074, "step": 20336 }, { "epoch": 1.3648870843260292, "grad_norm": 1.2078745365142822, "learning_rate": 2.4188757628771524e-05, "loss": 0.4458, "step": 20338 }, { "epoch": 1.3650213080097984, "grad_norm": 0.9432024955749512, "learning_rate": 2.4179450043920933e-05, "loss": 0.4646, "step": 20340 }, { "epoch": 1.3651555316935673, "grad_norm": 1.0819599628448486, "learning_rate": 2.4170143679040552e-05, "loss": 0.5303, "step": 20342 }, { "epoch": 1.3652897553773364, "grad_norm": 1.0137512683868408, "learning_rate": 2.4160838534570074e-05, "loss": 0.4371, "step": 20344 }, { "epoch": 1.3654239790611054, "grad_norm": 1.136479139328003, "learning_rate": 2.4151534610949178e-05, "loss": 0.4898, "step": 20346 }, { "epoch": 1.3655582027448743, "grad_norm": 0.8670485615730286, "learning_rate": 2.4142231908617445e-05, "loss": 0.477, "step": 20348 }, { "epoch": 1.3656924264286434, "grad_norm": 1.0853685140609741, "learning_rate": 2.413293042801441e-05, "loss": 0.534, "step": 20350 }, { "epoch": 1.3658266501124123, "grad_norm": 1.0450760126113892, "learning_rate": 2.4123630169579526e-05, "loss": 0.4518, "step": 20352 }, { "epoch": 1.3659608737961815, "grad_norm": 1.060826063156128, "learning_rate": 2.4114331133752254e-05, "loss": 0.4867, "step": 20354 }, { "epoch": 1.3660950974799504, "grad_norm": 1.0384961366653442, "learning_rate": 2.410503332097193e-05, "loss": 0.4835, "step": 20356 }, { "epoch": 1.3662293211637193, "grad_norm": 1.0962611436843872, "learning_rate": 2.4095736731677863e-05, "loss": 0.5329, "step": 20358 }, { "epoch": 1.3663635448474882, "grad_norm": 1.1338086128234863, "learning_rate": 2.4086441366309277e-05, "loss": 0.4969, "step": 20360 }, { "epoch": 1.3664977685312574, "grad_norm": 1.0492395162582397, "learning_rate": 2.4077147225305403e-05, "loss": 0.4232, "step": 20362 }, { "epoch": 1.3666319922150263, "grad_norm": 0.9534843564033508, "learning_rate": 2.4067854309105326e-05, "loss": 0.4584, "step": 20364 }, { "epoch": 1.3667662158987954, "grad_norm": 1.0854427814483643, "learning_rate": 2.4058562618148155e-05, "loss": 0.4982, "step": 20366 }, { "epoch": 1.3669004395825644, "grad_norm": 1.072067379951477, "learning_rate": 2.4049272152872893e-05, "loss": 0.511, "step": 20368 }, { "epoch": 1.3670346632663333, "grad_norm": 1.0549293756484985, "learning_rate": 2.403998291371849e-05, "loss": 0.5146, "step": 20370 }, { "epoch": 1.3671688869501024, "grad_norm": 1.0453256368637085, "learning_rate": 2.4030694901123825e-05, "loss": 0.5173, "step": 20372 }, { "epoch": 1.3673031106338713, "grad_norm": 1.1417920589447021, "learning_rate": 2.402140811552778e-05, "loss": 0.4684, "step": 20374 }, { "epoch": 1.3674373343176405, "grad_norm": 0.9018839001655579, "learning_rate": 2.401212255736912e-05, "loss": 0.4742, "step": 20376 }, { "epoch": 1.3675715580014094, "grad_norm": 1.0004795789718628, "learning_rate": 2.400283822708656e-05, "loss": 0.5078, "step": 20378 }, { "epoch": 1.3677057816851783, "grad_norm": 0.8867558836936951, "learning_rate": 2.399355512511876e-05, "loss": 0.4474, "step": 20380 }, { "epoch": 1.3678400053689472, "grad_norm": 1.1225106716156006, "learning_rate": 2.3984273251904356e-05, "loss": 0.4629, "step": 20382 }, { "epoch": 1.3679742290527164, "grad_norm": 0.963327169418335, "learning_rate": 2.3974992607881887e-05, "loss": 0.4835, "step": 20384 }, { "epoch": 1.3681084527364853, "grad_norm": 1.0399508476257324, "learning_rate": 2.3965713193489836e-05, "loss": 0.4706, "step": 20386 }, { "epoch": 1.3682426764202544, "grad_norm": 1.0206793546676636, "learning_rate": 2.3956435009166627e-05, "loss": 0.4832, "step": 20388 }, { "epoch": 1.3683769001040234, "grad_norm": 1.1822404861450195, "learning_rate": 2.394715805535067e-05, "loss": 0.5164, "step": 20390 }, { "epoch": 1.3685111237877923, "grad_norm": 1.8236244916915894, "learning_rate": 2.3937882332480243e-05, "loss": 0.4687, "step": 20392 }, { "epoch": 1.3686453474715614, "grad_norm": 1.1084983348846436, "learning_rate": 2.392860784099365e-05, "loss": 0.5216, "step": 20394 }, { "epoch": 1.3687795711553303, "grad_norm": 1.001785159111023, "learning_rate": 2.3919334581329082e-05, "loss": 0.5059, "step": 20396 }, { "epoch": 1.3689137948390995, "grad_norm": 1.0605645179748535, "learning_rate": 2.391006255392464e-05, "loss": 0.5077, "step": 20398 }, { "epoch": 1.3690480185228684, "grad_norm": 0.8846589922904968, "learning_rate": 2.390079175921845e-05, "loss": 0.4692, "step": 20400 }, { "epoch": 1.3691822422066373, "grad_norm": 1.018807053565979, "learning_rate": 2.3891522197648508e-05, "loss": 0.498, "step": 20402 }, { "epoch": 1.3693164658904062, "grad_norm": 1.0309909582138062, "learning_rate": 2.3882253869652828e-05, "loss": 0.5042, "step": 20404 }, { "epoch": 1.3694506895741754, "grad_norm": 1.1537896394729614, "learning_rate": 2.3872986775669288e-05, "loss": 0.4564, "step": 20406 }, { "epoch": 1.3695849132579443, "grad_norm": 0.9856048822402954, "learning_rate": 2.386372091613575e-05, "loss": 0.4525, "step": 20408 }, { "epoch": 1.3697191369417134, "grad_norm": 1.110657811164856, "learning_rate": 2.3854456291489987e-05, "loss": 0.4641, "step": 20410 }, { "epoch": 1.3698533606254824, "grad_norm": 0.9858686327934265, "learning_rate": 2.3845192902169766e-05, "loss": 0.4832, "step": 20412 }, { "epoch": 1.3699875843092513, "grad_norm": 1.1257051229476929, "learning_rate": 2.383593074861275e-05, "loss": 0.4913, "step": 20414 }, { "epoch": 1.3701218079930204, "grad_norm": 0.9121658802032471, "learning_rate": 2.3826669831256554e-05, "loss": 0.4668, "step": 20416 }, { "epoch": 1.3702560316767893, "grad_norm": 1.0087146759033203, "learning_rate": 2.3817410150538728e-05, "loss": 0.5393, "step": 20418 }, { "epoch": 1.3703902553605585, "grad_norm": 1.0078567266464233, "learning_rate": 2.3808151706896804e-05, "loss": 0.5155, "step": 20420 }, { "epoch": 1.3705244790443274, "grad_norm": 1.0872535705566406, "learning_rate": 2.3798894500768183e-05, "loss": 0.4565, "step": 20422 }, { "epoch": 1.3706587027280963, "grad_norm": 1.031997561454773, "learning_rate": 2.378963853259032e-05, "loss": 0.4866, "step": 20424 }, { "epoch": 1.3707929264118655, "grad_norm": 1.018782615661621, "learning_rate": 2.3780383802800453e-05, "loss": 0.4596, "step": 20426 }, { "epoch": 1.3709271500956344, "grad_norm": 1.0133177042007446, "learning_rate": 2.3771130311835914e-05, "loss": 0.4945, "step": 20428 }, { "epoch": 1.3710613737794035, "grad_norm": 1.014116883277893, "learning_rate": 2.3761878060133868e-05, "loss": 0.4044, "step": 20430 }, { "epoch": 1.3711955974631724, "grad_norm": 1.083739995956421, "learning_rate": 2.3752627048131516e-05, "loss": 0.4973, "step": 20432 }, { "epoch": 1.3713298211469414, "grad_norm": 1.0590863227844238, "learning_rate": 2.3743377276265916e-05, "loss": 0.4977, "step": 20434 }, { "epoch": 1.3714640448307103, "grad_norm": 1.184765100479126, "learning_rate": 2.373412874497411e-05, "loss": 0.4348, "step": 20436 }, { "epoch": 1.3715982685144794, "grad_norm": 1.0506863594055176, "learning_rate": 2.3724881454693048e-05, "loss": 0.4796, "step": 20438 }, { "epoch": 1.3717324921982483, "grad_norm": 1.010758399963379, "learning_rate": 2.3715635405859692e-05, "loss": 0.4563, "step": 20440 }, { "epoch": 1.3718667158820175, "grad_norm": 0.9689465761184692, "learning_rate": 2.3706390598910878e-05, "loss": 0.4726, "step": 20442 }, { "epoch": 1.3720009395657864, "grad_norm": 0.9940888285636902, "learning_rate": 2.3697147034283407e-05, "loss": 0.4485, "step": 20444 }, { "epoch": 1.3721351632495553, "grad_norm": 0.9998622536659241, "learning_rate": 2.3687904712413996e-05, "loss": 0.4643, "step": 20446 }, { "epoch": 1.3722693869333245, "grad_norm": 1.1568114757537842, "learning_rate": 2.367866363373936e-05, "loss": 0.5178, "step": 20448 }, { "epoch": 1.3724036106170934, "grad_norm": 0.9747192859649658, "learning_rate": 2.3669423798696094e-05, "loss": 0.5501, "step": 20450 }, { "epoch": 1.3725378343008625, "grad_norm": 0.9922822117805481, "learning_rate": 2.3660185207720813e-05, "loss": 0.4787, "step": 20452 }, { "epoch": 1.3726720579846314, "grad_norm": 0.961625337600708, "learning_rate": 2.365094786124995e-05, "loss": 0.4688, "step": 20454 }, { "epoch": 1.3728062816684004, "grad_norm": 1.0886564254760742, "learning_rate": 2.3641711759720003e-05, "loss": 0.4928, "step": 20456 }, { "epoch": 1.3729405053521693, "grad_norm": 1.1044477224349976, "learning_rate": 2.363247690356733e-05, "loss": 0.4975, "step": 20458 }, { "epoch": 1.3730747290359384, "grad_norm": 1.0163726806640625, "learning_rate": 2.3623243293228287e-05, "loss": 0.5184, "step": 20460 }, { "epoch": 1.3732089527197073, "grad_norm": 0.9921759963035583, "learning_rate": 2.3614010929139136e-05, "loss": 0.4297, "step": 20462 }, { "epoch": 1.3733431764034765, "grad_norm": 1.009995937347412, "learning_rate": 2.360477981173608e-05, "loss": 0.4722, "step": 20464 }, { "epoch": 1.3734774000872454, "grad_norm": 0.9856315851211548, "learning_rate": 2.3595549941455276e-05, "loss": 0.5022, "step": 20466 }, { "epoch": 1.3736116237710143, "grad_norm": 0.8814576864242554, "learning_rate": 2.358632131873279e-05, "loss": 0.4634, "step": 20468 }, { "epoch": 1.3737458474547835, "grad_norm": 1.0409995317459106, "learning_rate": 2.3577093944004707e-05, "loss": 0.5505, "step": 20470 }, { "epoch": 1.3738800711385524, "grad_norm": 1.1030895709991455, "learning_rate": 2.3567867817706974e-05, "loss": 0.5467, "step": 20472 }, { "epoch": 1.3740142948223215, "grad_norm": 0.9939549565315247, "learning_rate": 2.355864294027551e-05, "loss": 0.4681, "step": 20474 }, { "epoch": 1.3741485185060904, "grad_norm": 1.004815697669983, "learning_rate": 2.3549419312146153e-05, "loss": 0.4403, "step": 20476 }, { "epoch": 1.3742827421898594, "grad_norm": 1.0654927492141724, "learning_rate": 2.3540196933754743e-05, "loss": 0.4571, "step": 20478 }, { "epoch": 1.3744169658736283, "grad_norm": 1.0886633396148682, "learning_rate": 2.3530975805536993e-05, "loss": 0.4856, "step": 20480 }, { "epoch": 1.3745511895573974, "grad_norm": 1.0018386840820312, "learning_rate": 2.352175592792859e-05, "loss": 0.4564, "step": 20482 }, { "epoch": 1.3746854132411663, "grad_norm": 0.9411597847938538, "learning_rate": 2.3512537301365134e-05, "loss": 0.4969, "step": 20484 }, { "epoch": 1.3748196369249355, "grad_norm": 0.9695761203765869, "learning_rate": 2.3503319926282218e-05, "loss": 0.4395, "step": 20486 }, { "epoch": 1.3749538606087044, "grad_norm": 1.0185626745224, "learning_rate": 2.349410380311532e-05, "loss": 0.5126, "step": 20488 }, { "epoch": 1.3750880842924733, "grad_norm": 1.1450790166854858, "learning_rate": 2.348488893229991e-05, "loss": 0.5268, "step": 20490 }, { "epoch": 1.3752223079762425, "grad_norm": 1.0380408763885498, "learning_rate": 2.3475675314271363e-05, "loss": 0.4382, "step": 20492 }, { "epoch": 1.3753565316600114, "grad_norm": 1.0181218385696411, "learning_rate": 2.3466462949464996e-05, "loss": 0.5113, "step": 20494 }, { "epoch": 1.3754907553437805, "grad_norm": 0.9403045773506165, "learning_rate": 2.345725183831606e-05, "loss": 0.5087, "step": 20496 }, { "epoch": 1.3756249790275494, "grad_norm": 1.0722616910934448, "learning_rate": 2.34480419812598e-05, "loss": 0.4588, "step": 20498 }, { "epoch": 1.3757592027113184, "grad_norm": 0.9313488006591797, "learning_rate": 2.3438833378731338e-05, "loss": 0.4626, "step": 20500 }, { "epoch": 1.3758934263950875, "grad_norm": 1.1240397691726685, "learning_rate": 2.342962603116578e-05, "loss": 0.5289, "step": 20502 }, { "epoch": 1.3760276500788564, "grad_norm": 1.1276453733444214, "learning_rate": 2.342041993899812e-05, "loss": 0.5325, "step": 20504 }, { "epoch": 1.3761618737626256, "grad_norm": 1.1331275701522827, "learning_rate": 2.341121510266337e-05, "loss": 0.4912, "step": 20506 }, { "epoch": 1.3762960974463945, "grad_norm": 0.7722598910331726, "learning_rate": 2.340201152259642e-05, "loss": 0.4556, "step": 20508 }, { "epoch": 1.3764303211301634, "grad_norm": 1.0412757396697998, "learning_rate": 2.3392809199232123e-05, "loss": 0.4943, "step": 20510 }, { "epoch": 1.3765645448139323, "grad_norm": 0.9970412254333496, "learning_rate": 2.3383608133005253e-05, "loss": 0.4934, "step": 20512 }, { "epoch": 1.3766987684977015, "grad_norm": 0.9749853014945984, "learning_rate": 2.337440832435058e-05, "loss": 0.4486, "step": 20514 }, { "epoch": 1.3768329921814704, "grad_norm": 1.1657603979110718, "learning_rate": 2.3365209773702736e-05, "loss": 0.5406, "step": 20516 }, { "epoch": 1.3769672158652395, "grad_norm": 0.8795127868652344, "learning_rate": 2.335601248149637e-05, "loss": 0.4899, "step": 20518 }, { "epoch": 1.3771014395490084, "grad_norm": 0.8475943207740784, "learning_rate": 2.3346816448166025e-05, "loss": 0.47, "step": 20520 }, { "epoch": 1.3772356632327774, "grad_norm": 1.1030148267745972, "learning_rate": 2.3337621674146193e-05, "loss": 0.5703, "step": 20522 }, { "epoch": 1.3773698869165465, "grad_norm": 1.0147314071655273, "learning_rate": 2.332842815987129e-05, "loss": 0.5035, "step": 20524 }, { "epoch": 1.3775041106003154, "grad_norm": 1.1151841878890991, "learning_rate": 2.3319235905775727e-05, "loss": 0.4712, "step": 20526 }, { "epoch": 1.3776383342840846, "grad_norm": 0.8732739090919495, "learning_rate": 2.33100449122938e-05, "loss": 0.5501, "step": 20528 }, { "epoch": 1.3777725579678535, "grad_norm": 1.0400519371032715, "learning_rate": 2.3300855179859765e-05, "loss": 0.4494, "step": 20530 }, { "epoch": 1.3779067816516224, "grad_norm": 1.0130256414413452, "learning_rate": 2.3291666708907826e-05, "loss": 0.4888, "step": 20532 }, { "epoch": 1.3780410053353913, "grad_norm": 0.9467426538467407, "learning_rate": 2.3282479499872095e-05, "loss": 0.4929, "step": 20534 }, { "epoch": 1.3781752290191605, "grad_norm": 0.9923089742660522, "learning_rate": 2.3273293553186688e-05, "loss": 0.5216, "step": 20536 }, { "epoch": 1.3783094527029294, "grad_norm": 1.347734808921814, "learning_rate": 2.3264108869285606e-05, "loss": 0.5083, "step": 20538 }, { "epoch": 1.3784436763866985, "grad_norm": 0.9640908241271973, "learning_rate": 2.3254925448602806e-05, "loss": 0.5134, "step": 20540 }, { "epoch": 1.3785779000704674, "grad_norm": 1.0231845378875732, "learning_rate": 2.3245743291572164e-05, "loss": 0.4919, "step": 20542 }, { "epoch": 1.3787121237542364, "grad_norm": 1.092300534248352, "learning_rate": 2.3236562398627566e-05, "loss": 0.5318, "step": 20544 }, { "epoch": 1.3788463474380055, "grad_norm": 1.0548770427703857, "learning_rate": 2.3227382770202744e-05, "loss": 0.4472, "step": 20546 }, { "epoch": 1.3789805711217744, "grad_norm": 0.9882622957229614, "learning_rate": 2.3218204406731474e-05, "loss": 0.518, "step": 20548 }, { "epoch": 1.3791147948055436, "grad_norm": 0.9383532404899597, "learning_rate": 2.3209027308647352e-05, "loss": 0.4538, "step": 20550 }, { "epoch": 1.3792490184893125, "grad_norm": 1.078097939491272, "learning_rate": 2.3199851476384022e-05, "loss": 0.535, "step": 20552 }, { "epoch": 1.3793832421730814, "grad_norm": 1.1034539937973022, "learning_rate": 2.319067691037499e-05, "loss": 0.4945, "step": 20554 }, { "epoch": 1.3795174658568503, "grad_norm": 0.9837169051170349, "learning_rate": 2.318150361105378e-05, "loss": 0.5057, "step": 20556 }, { "epoch": 1.3796516895406195, "grad_norm": 0.9580581188201904, "learning_rate": 2.3172331578853785e-05, "loss": 0.4469, "step": 20558 }, { "epoch": 1.3797859132243884, "grad_norm": 1.0174978971481323, "learning_rate": 2.3163160814208367e-05, "loss": 0.5027, "step": 20560 }, { "epoch": 1.3799201369081575, "grad_norm": 1.1509947776794434, "learning_rate": 2.315399131755081e-05, "loss": 0.4928, "step": 20562 }, { "epoch": 1.3800543605919264, "grad_norm": 0.9773992300033569, "learning_rate": 2.3144823089314388e-05, "loss": 0.4518, "step": 20564 }, { "epoch": 1.3801885842756954, "grad_norm": 0.983257532119751, "learning_rate": 2.313565612993227e-05, "loss": 0.5034, "step": 20566 }, { "epoch": 1.3803228079594645, "grad_norm": 1.0288691520690918, "learning_rate": 2.3126490439837578e-05, "loss": 0.5071, "step": 20568 }, { "epoch": 1.3804570316432334, "grad_norm": 1.0755268335342407, "learning_rate": 2.311732601946334e-05, "loss": 0.4987, "step": 20570 }, { "epoch": 1.3805912553270026, "grad_norm": 1.0262552499771118, "learning_rate": 2.310816286924261e-05, "loss": 0.5021, "step": 20572 }, { "epoch": 1.3807254790107715, "grad_norm": 1.126383662223816, "learning_rate": 2.3099000989608288e-05, "loss": 0.4603, "step": 20574 }, { "epoch": 1.3808597026945404, "grad_norm": 1.0954885482788086, "learning_rate": 2.3089840380993305e-05, "loss": 0.4856, "step": 20576 }, { "epoch": 1.3809939263783095, "grad_norm": 1.0396054983139038, "learning_rate": 2.308068104383041e-05, "loss": 0.4617, "step": 20578 }, { "epoch": 1.3811281500620785, "grad_norm": 1.0048110485076904, "learning_rate": 2.3071522978552418e-05, "loss": 0.4569, "step": 20580 }, { "epoch": 1.3812623737458474, "grad_norm": 0.8580957651138306, "learning_rate": 2.3062366185591995e-05, "loss": 0.4575, "step": 20582 }, { "epoch": 1.3813965974296165, "grad_norm": 0.9834871888160706, "learning_rate": 2.3053210665381818e-05, "loss": 0.4724, "step": 20584 }, { "epoch": 1.3815308211133854, "grad_norm": 1.0720840692520142, "learning_rate": 2.304405641835445e-05, "loss": 0.5053, "step": 20586 }, { "epoch": 1.3816650447971544, "grad_norm": 0.9133841395378113, "learning_rate": 2.3034903444942406e-05, "loss": 0.4478, "step": 20588 }, { "epoch": 1.3817992684809235, "grad_norm": 1.0490212440490723, "learning_rate": 2.3025751745578134e-05, "loss": 0.5149, "step": 20590 }, { "epoch": 1.3819334921646924, "grad_norm": 1.0795177221298218, "learning_rate": 2.3016601320694065e-05, "loss": 0.5253, "step": 20592 }, { "epoch": 1.3820677158484616, "grad_norm": 0.8143784999847412, "learning_rate": 2.3007452170722528e-05, "loss": 0.4588, "step": 20594 }, { "epoch": 1.3822019395322305, "grad_norm": 1.0718694925308228, "learning_rate": 2.299830429609579e-05, "loss": 0.4899, "step": 20596 }, { "epoch": 1.3823361632159994, "grad_norm": 1.0476452112197876, "learning_rate": 2.2989157697246066e-05, "loss": 0.5106, "step": 20598 }, { "epoch": 1.3824703868997685, "grad_norm": 0.9038621187210083, "learning_rate": 2.2980012374605542e-05, "loss": 0.4656, "step": 20600 }, { "epoch": 1.3826046105835375, "grad_norm": 1.0184696912765503, "learning_rate": 2.29708683286063e-05, "loss": 0.5373, "step": 20602 }, { "epoch": 1.3827388342673066, "grad_norm": 0.9636785387992859, "learning_rate": 2.296172555968038e-05, "loss": 0.5063, "step": 20604 }, { "epoch": 1.3828730579510755, "grad_norm": 0.9802426099777222, "learning_rate": 2.2952584068259757e-05, "loss": 0.4836, "step": 20606 }, { "epoch": 1.3830072816348444, "grad_norm": 1.0625511407852173, "learning_rate": 2.2943443854776335e-05, "loss": 0.486, "step": 20608 }, { "epoch": 1.3831415053186134, "grad_norm": 1.0863850116729736, "learning_rate": 2.2934304919662003e-05, "loss": 0.4964, "step": 20610 }, { "epoch": 1.3832757290023825, "grad_norm": 0.8596870303153992, "learning_rate": 2.2925167263348525e-05, "loss": 0.4592, "step": 20612 }, { "epoch": 1.3834099526861514, "grad_norm": 0.9515671133995056, "learning_rate": 2.2916030886267687e-05, "loss": 0.4575, "step": 20614 }, { "epoch": 1.3835441763699206, "grad_norm": 1.1263680458068848, "learning_rate": 2.2906895788851097e-05, "loss": 0.5034, "step": 20616 }, { "epoch": 1.3836784000536895, "grad_norm": 1.0509198904037476, "learning_rate": 2.2897761971530425e-05, "loss": 0.4805, "step": 20618 }, { "epoch": 1.3838126237374584, "grad_norm": 0.9480360746383667, "learning_rate": 2.288862943473718e-05, "loss": 0.5098, "step": 20620 }, { "epoch": 1.3839468474212275, "grad_norm": 0.9155513048171997, "learning_rate": 2.2879498178902904e-05, "loss": 0.4373, "step": 20622 }, { "epoch": 1.3840810711049965, "grad_norm": 0.9924333095550537, "learning_rate": 2.2870368204459008e-05, "loss": 0.4545, "step": 20624 }, { "epoch": 1.3842152947887656, "grad_norm": 1.1303173303604126, "learning_rate": 2.2861239511836867e-05, "loss": 0.4909, "step": 20626 }, { "epoch": 1.3843495184725345, "grad_norm": 0.9560405611991882, "learning_rate": 2.2852112101467778e-05, "loss": 0.4618, "step": 20628 }, { "epoch": 1.3844837421563034, "grad_norm": 1.2171663045883179, "learning_rate": 2.2842985973783026e-05, "loss": 0.4482, "step": 20630 }, { "epoch": 1.3846179658400724, "grad_norm": 0.9853360056877136, "learning_rate": 2.283386112921378e-05, "loss": 0.515, "step": 20632 }, { "epoch": 1.3847521895238415, "grad_norm": 0.9441184997558594, "learning_rate": 2.2824737568191178e-05, "loss": 0.5091, "step": 20634 }, { "epoch": 1.3848864132076104, "grad_norm": 1.2589091062545776, "learning_rate": 2.281561529114627e-05, "loss": 0.5033, "step": 20636 }, { "epoch": 1.3850206368913796, "grad_norm": 0.9849038124084473, "learning_rate": 2.28064942985101e-05, "loss": 0.5216, "step": 20638 }, { "epoch": 1.3851548605751485, "grad_norm": 0.9936666488647461, "learning_rate": 2.279737459071359e-05, "loss": 0.4874, "step": 20640 }, { "epoch": 1.3852890842589174, "grad_norm": 1.0241390466690063, "learning_rate": 2.2788256168187665e-05, "loss": 0.4866, "step": 20642 }, { "epoch": 1.3854233079426865, "grad_norm": 0.9956193566322327, "learning_rate": 2.2779139031363096e-05, "loss": 0.4914, "step": 20644 }, { "epoch": 1.3855575316264555, "grad_norm": 1.0614651441574097, "learning_rate": 2.2770023180670698e-05, "loss": 0.5134, "step": 20646 }, { "epoch": 1.3856917553102246, "grad_norm": 0.9981801509857178, "learning_rate": 2.2760908616541136e-05, "loss": 0.515, "step": 20648 }, { "epoch": 1.3858259789939935, "grad_norm": 0.9139376878738403, "learning_rate": 2.2751795339405103e-05, "loss": 0.4629, "step": 20650 }, { "epoch": 1.3859602026777624, "grad_norm": 0.9527412056922913, "learning_rate": 2.274268334969316e-05, "loss": 0.4909, "step": 20652 }, { "epoch": 1.3860944263615316, "grad_norm": 1.0902884006500244, "learning_rate": 2.2733572647835828e-05, "loss": 0.5259, "step": 20654 }, { "epoch": 1.3862286500453005, "grad_norm": 0.9523167014122009, "learning_rate": 2.2724463234263548e-05, "loss": 0.4923, "step": 20656 }, { "epoch": 1.3863628737290694, "grad_norm": 1.0305126905441284, "learning_rate": 2.2715355109406766e-05, "loss": 0.4921, "step": 20658 }, { "epoch": 1.3864970974128386, "grad_norm": 1.0861053466796875, "learning_rate": 2.27062482736958e-05, "loss": 0.5463, "step": 20660 }, { "epoch": 1.3866313210966075, "grad_norm": 1.0559502840042114, "learning_rate": 2.2697142727560937e-05, "loss": 0.5093, "step": 20662 }, { "epoch": 1.3867655447803764, "grad_norm": 0.9299308657646179, "learning_rate": 2.2688038471432373e-05, "loss": 0.564, "step": 20664 }, { "epoch": 1.3868997684641455, "grad_norm": 1.0282251834869385, "learning_rate": 2.26789355057403e-05, "loss": 0.4818, "step": 20666 }, { "epoch": 1.3870339921479145, "grad_norm": 0.9290526509284973, "learning_rate": 2.2669833830914782e-05, "loss": 0.4723, "step": 20668 }, { "epoch": 1.3871682158316836, "grad_norm": 0.9008547067642212, "learning_rate": 2.2660733447385894e-05, "loss": 0.4255, "step": 20670 }, { "epoch": 1.3873024395154525, "grad_norm": 0.8995382189750671, "learning_rate": 2.2651634355583606e-05, "loss": 0.5014, "step": 20672 }, { "epoch": 1.3874366631992214, "grad_norm": 1.4763998985290527, "learning_rate": 2.2642536555937783e-05, "loss": 0.493, "step": 20674 }, { "epoch": 1.3875708868829906, "grad_norm": 1.1050307750701904, "learning_rate": 2.2633440048878323e-05, "loss": 0.4752, "step": 20676 }, { "epoch": 1.3877051105667595, "grad_norm": 0.9923198819160461, "learning_rate": 2.2624344834834992e-05, "loss": 0.4421, "step": 20678 }, { "epoch": 1.3878393342505286, "grad_norm": 1.04277765750885, "learning_rate": 2.2615250914237558e-05, "loss": 0.5183, "step": 20680 }, { "epoch": 1.3879735579342976, "grad_norm": 1.0581557750701904, "learning_rate": 2.260615828751566e-05, "loss": 0.4809, "step": 20682 }, { "epoch": 1.3881077816180665, "grad_norm": 1.04483962059021, "learning_rate": 2.2597066955098923e-05, "loss": 0.4696, "step": 20684 }, { "epoch": 1.3882420053018354, "grad_norm": 0.9544757604598999, "learning_rate": 2.2587976917416864e-05, "loss": 0.45, "step": 20686 }, { "epoch": 1.3883762289856045, "grad_norm": 1.074668049812317, "learning_rate": 2.2578888174899014e-05, "loss": 0.5675, "step": 20688 }, { "epoch": 1.3885104526693735, "grad_norm": 0.9719125032424927, "learning_rate": 2.2569800727974777e-05, "loss": 0.4975, "step": 20690 }, { "epoch": 1.3886446763531426, "grad_norm": 0.8937525749206543, "learning_rate": 2.2560714577073512e-05, "loss": 0.4317, "step": 20692 }, { "epoch": 1.3887789000369115, "grad_norm": 1.0742558240890503, "learning_rate": 2.255162972262451e-05, "loss": 0.457, "step": 20694 }, { "epoch": 1.3889131237206804, "grad_norm": 0.9840453267097473, "learning_rate": 2.254254616505705e-05, "loss": 0.5375, "step": 20696 }, { "epoch": 1.3890473474044496, "grad_norm": 1.455519676208496, "learning_rate": 2.2533463904800268e-05, "loss": 0.4567, "step": 20698 }, { "epoch": 1.3891815710882185, "grad_norm": 1.0274455547332764, "learning_rate": 2.2524382942283345e-05, "loss": 0.5004, "step": 20700 }, { "epoch": 1.3893157947719876, "grad_norm": 1.2613234519958496, "learning_rate": 2.2515303277935257e-05, "loss": 0.5372, "step": 20702 }, { "epoch": 1.3894500184557566, "grad_norm": 0.9007759690284729, "learning_rate": 2.250622491218507e-05, "loss": 0.4797, "step": 20704 }, { "epoch": 1.3895842421395255, "grad_norm": 1.0045149326324463, "learning_rate": 2.2497147845461668e-05, "loss": 0.4741, "step": 20706 }, { "epoch": 1.3897184658232944, "grad_norm": 1.0059312582015991, "learning_rate": 2.2488072078193968e-05, "loss": 0.4911, "step": 20708 }, { "epoch": 1.3898526895070635, "grad_norm": 0.9289558529853821, "learning_rate": 2.2478997610810764e-05, "loss": 0.4067, "step": 20710 }, { "epoch": 1.3899869131908325, "grad_norm": 1.004197120666504, "learning_rate": 2.2469924443740804e-05, "loss": 0.5426, "step": 20712 }, { "epoch": 1.3901211368746016, "grad_norm": 1.1435902118682861, "learning_rate": 2.2460852577412754e-05, "loss": 0.5566, "step": 20714 }, { "epoch": 1.3902553605583705, "grad_norm": 1.0519390106201172, "learning_rate": 2.2451782012255294e-05, "loss": 0.4759, "step": 20716 }, { "epoch": 1.3903895842421394, "grad_norm": 0.9619807600975037, "learning_rate": 2.244271274869696e-05, "loss": 0.4767, "step": 20718 }, { "epoch": 1.3905238079259086, "grad_norm": 1.1317015886306763, "learning_rate": 2.243364478716626e-05, "loss": 0.4815, "step": 20720 }, { "epoch": 1.3906580316096775, "grad_norm": 1.0051021575927734, "learning_rate": 2.2424578128091617e-05, "loss": 0.5831, "step": 20722 }, { "epoch": 1.3907922552934466, "grad_norm": 1.0220969915390015, "learning_rate": 2.241551277190145e-05, "loss": 0.4742, "step": 20724 }, { "epoch": 1.3909264789772156, "grad_norm": 0.9811525344848633, "learning_rate": 2.2406448719024044e-05, "loss": 0.5178, "step": 20726 }, { "epoch": 1.3910607026609845, "grad_norm": 1.10579514503479, "learning_rate": 2.2397385969887713e-05, "loss": 0.5329, "step": 20728 }, { "epoch": 1.3911949263447536, "grad_norm": 1.003940463066101, "learning_rate": 2.2388324524920577e-05, "loss": 0.4809, "step": 20730 }, { "epoch": 1.3913291500285225, "grad_norm": 0.9447029829025269, "learning_rate": 2.2379264384550836e-05, "loss": 0.472, "step": 20732 }, { "epoch": 1.3914633737122915, "grad_norm": 0.9570050835609436, "learning_rate": 2.237020554920652e-05, "loss": 0.4414, "step": 20734 }, { "epoch": 1.3915975973960606, "grad_norm": 0.9190820455551147, "learning_rate": 2.236114801931567e-05, "loss": 0.4616, "step": 20736 }, { "epoch": 1.3917318210798295, "grad_norm": 1.0760704278945923, "learning_rate": 2.2352091795306252e-05, "loss": 0.4619, "step": 20738 }, { "epoch": 1.3918660447635984, "grad_norm": 0.981931209564209, "learning_rate": 2.2343036877606094e-05, "loss": 0.5765, "step": 20740 }, { "epoch": 1.3920002684473676, "grad_norm": 0.9680532813072205, "learning_rate": 2.233398326664307e-05, "loss": 0.4447, "step": 20742 }, { "epoch": 1.3921344921311365, "grad_norm": 1.0856574773788452, "learning_rate": 2.232493096284492e-05, "loss": 0.4907, "step": 20744 }, { "epoch": 1.3922687158149056, "grad_norm": 1.0441486835479736, "learning_rate": 2.231587996663938e-05, "loss": 0.5341, "step": 20746 }, { "epoch": 1.3924029394986746, "grad_norm": 1.130039095878601, "learning_rate": 2.2306830278454072e-05, "loss": 0.5279, "step": 20748 }, { "epoch": 1.3925371631824435, "grad_norm": 0.9704619646072388, "learning_rate": 2.229778189871658e-05, "loss": 0.4829, "step": 20750 }, { "epoch": 1.3926713868662126, "grad_norm": 1.6788395643234253, "learning_rate": 2.2288734827854395e-05, "loss": 0.4921, "step": 20752 }, { "epoch": 1.3928056105499815, "grad_norm": 0.9546937346458435, "learning_rate": 2.2279689066295022e-05, "loss": 0.4837, "step": 20754 }, { "epoch": 1.3929398342337507, "grad_norm": 1.1602784395217896, "learning_rate": 2.227064461446583e-05, "loss": 0.553, "step": 20756 }, { "epoch": 1.3930740579175196, "grad_norm": 1.0249425172805786, "learning_rate": 2.2261601472794147e-05, "loss": 0.5028, "step": 20758 }, { "epoch": 1.3932082816012885, "grad_norm": 1.065847396850586, "learning_rate": 2.2252559641707237e-05, "loss": 0.4242, "step": 20760 }, { "epoch": 1.3933425052850574, "grad_norm": 0.9325951933860779, "learning_rate": 2.2243519121632332e-05, "loss": 0.4736, "step": 20762 }, { "epoch": 1.3934767289688266, "grad_norm": 1.0492970943450928, "learning_rate": 2.2234479912996557e-05, "loss": 0.5372, "step": 20764 }, { "epoch": 1.3936109526525955, "grad_norm": 1.1325078010559082, "learning_rate": 2.2225442016227044e-05, "loss": 0.5958, "step": 20766 }, { "epoch": 1.3937451763363646, "grad_norm": 1.083055853843689, "learning_rate": 2.2216405431750735e-05, "loss": 0.5259, "step": 20768 }, { "epoch": 1.3938794000201336, "grad_norm": 0.9872101545333862, "learning_rate": 2.2207370159994662e-05, "loss": 0.5425, "step": 20770 }, { "epoch": 1.3940136237039025, "grad_norm": 0.9818466305732727, "learning_rate": 2.2198336201385674e-05, "loss": 0.4414, "step": 20772 }, { "epoch": 1.3941478473876716, "grad_norm": 0.9530790448188782, "learning_rate": 2.218930355635065e-05, "loss": 0.4587, "step": 20774 }, { "epoch": 1.3942820710714405, "grad_norm": 1.1509958505630493, "learning_rate": 2.2180272225316346e-05, "loss": 0.4853, "step": 20776 }, { "epoch": 1.3944162947552097, "grad_norm": 0.9824808239936829, "learning_rate": 2.2171242208709474e-05, "loss": 0.464, "step": 20778 }, { "epoch": 1.3945505184389786, "grad_norm": 1.0254404544830322, "learning_rate": 2.216221350695667e-05, "loss": 0.516, "step": 20780 }, { "epoch": 1.3946847421227475, "grad_norm": 1.113036870956421, "learning_rate": 2.2153186120484546e-05, "loss": 0.5511, "step": 20782 }, { "epoch": 1.3948189658065164, "grad_norm": 1.0598375797271729, "learning_rate": 2.2144160049719632e-05, "loss": 0.496, "step": 20784 }, { "epoch": 1.3949531894902856, "grad_norm": 1.0639839172363281, "learning_rate": 2.2135135295088373e-05, "loss": 0.5056, "step": 20786 }, { "epoch": 1.3950874131740545, "grad_norm": 1.0893691778182983, "learning_rate": 2.2126111857017157e-05, "loss": 0.5265, "step": 20788 }, { "epoch": 1.3952216368578236, "grad_norm": 1.1033680438995361, "learning_rate": 2.2117089735932367e-05, "loss": 0.5766, "step": 20790 }, { "epoch": 1.3953558605415926, "grad_norm": 1.1397067308425903, "learning_rate": 2.2108068932260235e-05, "loss": 0.5154, "step": 20792 }, { "epoch": 1.3954900842253615, "grad_norm": 1.0490998029708862, "learning_rate": 2.2099049446427038e-05, "loss": 0.5106, "step": 20794 }, { "epoch": 1.3956243079091306, "grad_norm": 1.205463171005249, "learning_rate": 2.2090031278858854e-05, "loss": 0.506, "step": 20796 }, { "epoch": 1.3957585315928995, "grad_norm": 1.0334906578063965, "learning_rate": 2.208101442998182e-05, "loss": 0.4647, "step": 20798 }, { "epoch": 1.3958927552766687, "grad_norm": 1.0384498834609985, "learning_rate": 2.2071998900221947e-05, "loss": 0.486, "step": 20800 }, { "epoch": 1.3960269789604376, "grad_norm": 1.1685583591461182, "learning_rate": 2.2062984690005224e-05, "loss": 0.4676, "step": 20802 }, { "epoch": 1.3961612026442065, "grad_norm": 0.9502678513526917, "learning_rate": 2.2053971799757535e-05, "loss": 0.486, "step": 20804 }, { "epoch": 1.3962954263279757, "grad_norm": 0.962885320186615, "learning_rate": 2.2044960229904722e-05, "loss": 0.5062, "step": 20806 }, { "epoch": 1.3964296500117446, "grad_norm": 1.1103259325027466, "learning_rate": 2.2035949980872573e-05, "loss": 0.5132, "step": 20808 }, { "epoch": 1.3965638736955135, "grad_norm": 1.0677655935287476, "learning_rate": 2.2026941053086773e-05, "loss": 0.5633, "step": 20810 }, { "epoch": 1.3966980973792826, "grad_norm": 1.0549362897872925, "learning_rate": 2.2017933446973027e-05, "loss": 0.4588, "step": 20812 }, { "epoch": 1.3968323210630516, "grad_norm": 0.9541140198707581, "learning_rate": 2.20089271629569e-05, "loss": 0.4411, "step": 20814 }, { "epoch": 1.3969665447468205, "grad_norm": 0.9641833305358887, "learning_rate": 2.199992220146392e-05, "loss": 0.488, "step": 20816 }, { "epoch": 1.3971007684305896, "grad_norm": 1.637166142463684, "learning_rate": 2.199091856291954e-05, "loss": 0.4531, "step": 20818 }, { "epoch": 1.3972349921143585, "grad_norm": 1.167707920074463, "learning_rate": 2.1981916247749196e-05, "loss": 0.4867, "step": 20820 }, { "epoch": 1.3973692157981277, "grad_norm": 0.9572454690933228, "learning_rate": 2.1972915256378217e-05, "loss": 0.4576, "step": 20822 }, { "epoch": 1.3975034394818966, "grad_norm": 1.1723190546035767, "learning_rate": 2.1963915589231876e-05, "loss": 0.4989, "step": 20824 }, { "epoch": 1.3976376631656655, "grad_norm": 0.9696754217147827, "learning_rate": 2.195491724673538e-05, "loss": 0.4452, "step": 20826 }, { "epoch": 1.3977718868494347, "grad_norm": 1.079972505569458, "learning_rate": 2.194592022931391e-05, "loss": 0.498, "step": 20828 }, { "epoch": 1.3979061105332036, "grad_norm": 0.9583383202552795, "learning_rate": 2.1936924537392527e-05, "loss": 0.4416, "step": 20830 }, { "epoch": 1.3980403342169727, "grad_norm": 1.1776984930038452, "learning_rate": 2.1927930171396293e-05, "loss": 0.5538, "step": 20832 }, { "epoch": 1.3981745579007416, "grad_norm": 1.0068161487579346, "learning_rate": 2.1918937131750157e-05, "loss": 0.5406, "step": 20834 }, { "epoch": 1.3983087815845106, "grad_norm": 1.1186065673828125, "learning_rate": 2.1909945418879017e-05, "loss": 0.4743, "step": 20836 }, { "epoch": 1.3984430052682795, "grad_norm": 1.0533246994018555, "learning_rate": 2.190095503320771e-05, "loss": 0.5541, "step": 20838 }, { "epoch": 1.3985772289520486, "grad_norm": 0.8219181299209595, "learning_rate": 2.1891965975161037e-05, "loss": 0.4716, "step": 20840 }, { "epoch": 1.3987114526358175, "grad_norm": 1.1751853227615356, "learning_rate": 2.1882978245163705e-05, "loss": 0.4849, "step": 20842 }, { "epoch": 1.3988456763195867, "grad_norm": 0.9777979254722595, "learning_rate": 2.187399184364035e-05, "loss": 0.4917, "step": 20844 }, { "epoch": 1.3989799000033556, "grad_norm": 1.089253306388855, "learning_rate": 2.1865006771015567e-05, "loss": 0.4869, "step": 20846 }, { "epoch": 1.3991141236871245, "grad_norm": 0.8422405123710632, "learning_rate": 2.18560230277139e-05, "loss": 0.4439, "step": 20848 }, { "epoch": 1.3992483473708937, "grad_norm": 1.0621827840805054, "learning_rate": 2.1847040614159804e-05, "loss": 0.527, "step": 20850 }, { "epoch": 1.3993825710546626, "grad_norm": 1.1477723121643066, "learning_rate": 2.1838059530777677e-05, "loss": 0.476, "step": 20852 }, { "epoch": 1.3995167947384317, "grad_norm": 0.860033392906189, "learning_rate": 2.1829079777991844e-05, "loss": 0.481, "step": 20854 }, { "epoch": 1.3996510184222006, "grad_norm": 0.8550235033035278, "learning_rate": 2.1820101356226612e-05, "loss": 0.466, "step": 20856 }, { "epoch": 1.3997852421059696, "grad_norm": 0.9906622767448425, "learning_rate": 2.1811124265906157e-05, "loss": 0.4855, "step": 20858 }, { "epoch": 1.3999194657897385, "grad_norm": 0.8928864002227783, "learning_rate": 2.180214850745467e-05, "loss": 0.5128, "step": 20860 }, { "epoch": 1.4000536894735076, "grad_norm": 0.9970691800117493, "learning_rate": 2.179317408129622e-05, "loss": 0.5149, "step": 20862 }, { "epoch": 1.4001879131572765, "grad_norm": 0.9386903047561646, "learning_rate": 2.1784200987854825e-05, "loss": 0.4421, "step": 20864 }, { "epoch": 1.4003221368410457, "grad_norm": 0.8827176094055176, "learning_rate": 2.1775229227554433e-05, "loss": 0.4363, "step": 20866 }, { "epoch": 1.4004563605248146, "grad_norm": 0.9380989670753479, "learning_rate": 2.1766258800818972e-05, "loss": 0.5722, "step": 20868 }, { "epoch": 1.4005905842085835, "grad_norm": 1.1577523946762085, "learning_rate": 2.1757289708072272e-05, "loss": 0.4723, "step": 20870 }, { "epoch": 1.4007248078923527, "grad_norm": 0.931845486164093, "learning_rate": 2.1748321949738088e-05, "loss": 0.4809, "step": 20872 }, { "epoch": 1.4008590315761216, "grad_norm": 1.0608102083206177, "learning_rate": 2.1739355526240124e-05, "loss": 0.5016, "step": 20874 }, { "epoch": 1.4009932552598907, "grad_norm": 1.084637999534607, "learning_rate": 2.173039043800206e-05, "loss": 0.4945, "step": 20876 }, { "epoch": 1.4011274789436596, "grad_norm": 1.0683140754699707, "learning_rate": 2.172142668544745e-05, "loss": 0.4652, "step": 20878 }, { "epoch": 1.4012617026274286, "grad_norm": 1.0307785272598267, "learning_rate": 2.1712464268999826e-05, "loss": 0.4868, "step": 20880 }, { "epoch": 1.4013959263111977, "grad_norm": 1.1262508630752563, "learning_rate": 2.1703503189082634e-05, "loss": 0.5312, "step": 20882 }, { "epoch": 1.4015301499949666, "grad_norm": 0.9825021028518677, "learning_rate": 2.1694543446119265e-05, "loss": 0.4495, "step": 20884 }, { "epoch": 1.4016643736787355, "grad_norm": 0.9609806537628174, "learning_rate": 2.1685585040533075e-05, "loss": 0.5075, "step": 20886 }, { "epoch": 1.4017985973625047, "grad_norm": 1.0035909414291382, "learning_rate": 2.167662797274729e-05, "loss": 0.5096, "step": 20888 }, { "epoch": 1.4019328210462736, "grad_norm": 0.925609290599823, "learning_rate": 2.1667672243185178e-05, "loss": 0.4324, "step": 20890 }, { "epoch": 1.4020670447300425, "grad_norm": 0.98987877368927, "learning_rate": 2.16587178522698e-05, "loss": 0.4612, "step": 20892 }, { "epoch": 1.4022012684138117, "grad_norm": 0.9895521998405457, "learning_rate": 2.1649764800424294e-05, "loss": 0.5452, "step": 20894 }, { "epoch": 1.4023354920975806, "grad_norm": 1.0596760511398315, "learning_rate": 2.1640813088071633e-05, "loss": 0.5384, "step": 20896 }, { "epoch": 1.4024697157813497, "grad_norm": 0.9556659460067749, "learning_rate": 2.163186271563481e-05, "loss": 0.4398, "step": 20898 }, { "epoch": 1.4026039394651186, "grad_norm": 1.061688780784607, "learning_rate": 2.1622913683536684e-05, "loss": 0.4534, "step": 20900 }, { "epoch": 1.4027381631488876, "grad_norm": 1.1543545722961426, "learning_rate": 2.1613965992200087e-05, "loss": 0.4923, "step": 20902 }, { "epoch": 1.4028723868326567, "grad_norm": 1.1429158449172974, "learning_rate": 2.1605019642047765e-05, "loss": 0.5098, "step": 20904 }, { "epoch": 1.4030066105164256, "grad_norm": 1.0910015106201172, "learning_rate": 2.159607463350244e-05, "loss": 0.4345, "step": 20906 }, { "epoch": 1.4031408342001948, "grad_norm": 1.1227449178695679, "learning_rate": 2.1587130966986735e-05, "loss": 0.4461, "step": 20908 }, { "epoch": 1.4032750578839637, "grad_norm": 1.3753069639205933, "learning_rate": 2.1578188642923225e-05, "loss": 0.5575, "step": 20910 }, { "epoch": 1.4034092815677326, "grad_norm": 1.1705611944198608, "learning_rate": 2.1569247661734392e-05, "loss": 0.5633, "step": 20912 }, { "epoch": 1.4035435052515015, "grad_norm": 0.9577829241752625, "learning_rate": 2.1560308023842717e-05, "loss": 0.4806, "step": 20914 }, { "epoch": 1.4036777289352707, "grad_norm": 1.0731220245361328, "learning_rate": 2.1551369729670546e-05, "loss": 0.4483, "step": 20916 }, { "epoch": 1.4038119526190396, "grad_norm": 0.9627668857574463, "learning_rate": 2.1542432779640255e-05, "loss": 0.4701, "step": 20918 }, { "epoch": 1.4039461763028087, "grad_norm": 1.0893503427505493, "learning_rate": 2.153349717417401e-05, "loss": 0.4905, "step": 20920 }, { "epoch": 1.4040803999865776, "grad_norm": 0.9726977348327637, "learning_rate": 2.1524562913694074e-05, "loss": 0.4862, "step": 20922 }, { "epoch": 1.4042146236703466, "grad_norm": 0.8646559119224548, "learning_rate": 2.151562999862252e-05, "loss": 0.4984, "step": 20924 }, { "epoch": 1.4043488473541157, "grad_norm": 1.0155696868896484, "learning_rate": 2.150669842938146e-05, "loss": 0.5444, "step": 20926 }, { "epoch": 1.4044830710378846, "grad_norm": 1.0304315090179443, "learning_rate": 2.1497768206392864e-05, "loss": 0.5069, "step": 20928 }, { "epoch": 1.4046172947216538, "grad_norm": 0.9800500869750977, "learning_rate": 2.148883933007868e-05, "loss": 0.4949, "step": 20930 }, { "epoch": 1.4047515184054227, "grad_norm": 0.8693739771842957, "learning_rate": 2.1479911800860752e-05, "loss": 0.4786, "step": 20932 }, { "epoch": 1.4048857420891916, "grad_norm": 1.0007768869400024, "learning_rate": 2.1470985619160926e-05, "loss": 0.444, "step": 20934 }, { "epoch": 1.4050199657729605, "grad_norm": 0.8693901300430298, "learning_rate": 2.146206078540093e-05, "loss": 0.4498, "step": 20936 }, { "epoch": 1.4051541894567297, "grad_norm": 0.9130423665046692, "learning_rate": 2.1453137300002445e-05, "loss": 0.4634, "step": 20938 }, { "epoch": 1.4052884131404986, "grad_norm": 1.1779518127441406, "learning_rate": 2.1444215163387067e-05, "loss": 0.5016, "step": 20940 }, { "epoch": 1.4054226368242677, "grad_norm": 1.0261186361312866, "learning_rate": 2.143529437597639e-05, "loss": 0.524, "step": 20942 }, { "epoch": 1.4055568605080366, "grad_norm": 1.036157488822937, "learning_rate": 2.1426374938191884e-05, "loss": 0.5194, "step": 20944 }, { "epoch": 1.4056910841918056, "grad_norm": 1.007396936416626, "learning_rate": 2.1417456850454976e-05, "loss": 0.4604, "step": 20946 }, { "epoch": 1.4058253078755747, "grad_norm": 0.9180713295936584, "learning_rate": 2.1408540113187027e-05, "loss": 0.4729, "step": 20948 }, { "epoch": 1.4059595315593436, "grad_norm": 1.113799810409546, "learning_rate": 2.139962472680931e-05, "loss": 0.4736, "step": 20950 }, { "epoch": 1.4060937552431128, "grad_norm": 1.0503449440002441, "learning_rate": 2.139071069174311e-05, "loss": 0.4933, "step": 20952 }, { "epoch": 1.4062279789268817, "grad_norm": 1.029120922088623, "learning_rate": 2.1381798008409547e-05, "loss": 0.4722, "step": 20954 }, { "epoch": 1.4063622026106506, "grad_norm": 1.1079305410385132, "learning_rate": 2.1372886677229775e-05, "loss": 0.4751, "step": 20956 }, { "epoch": 1.4064964262944197, "grad_norm": 1.1420440673828125, "learning_rate": 2.1363976698624815e-05, "loss": 0.5038, "step": 20958 }, { "epoch": 1.4066306499781887, "grad_norm": 1.000579595565796, "learning_rate": 2.135506807301565e-05, "loss": 0.4933, "step": 20960 }, { "epoch": 1.4067648736619576, "grad_norm": 0.9605050683021545, "learning_rate": 2.1346160800823166e-05, "loss": 0.4728, "step": 20962 }, { "epoch": 1.4068990973457267, "grad_norm": 1.0144368410110474, "learning_rate": 2.133725488246826e-05, "loss": 0.4661, "step": 20964 }, { "epoch": 1.4070333210294956, "grad_norm": 0.853814423084259, "learning_rate": 2.13283503183717e-05, "loss": 0.4718, "step": 20966 }, { "epoch": 1.4071675447132646, "grad_norm": 0.8243944644927979, "learning_rate": 2.13194471089542e-05, "loss": 0.451, "step": 20968 }, { "epoch": 1.4073017683970337, "grad_norm": 0.9485942125320435, "learning_rate": 2.1310545254636412e-05, "loss": 0.4761, "step": 20970 }, { "epoch": 1.4074359920808026, "grad_norm": 0.95245361328125, "learning_rate": 2.130164475583896e-05, "loss": 0.5054, "step": 20972 }, { "epoch": 1.4075702157645718, "grad_norm": 1.0146323442459106, "learning_rate": 2.129274561298237e-05, "loss": 0.5032, "step": 20974 }, { "epoch": 1.4077044394483407, "grad_norm": 1.0180444717407227, "learning_rate": 2.1283847826487092e-05, "loss": 0.5308, "step": 20976 }, { "epoch": 1.4078386631321096, "grad_norm": 0.9613975882530212, "learning_rate": 2.1274951396773517e-05, "loss": 0.4744, "step": 20978 }, { "epoch": 1.4079728868158787, "grad_norm": 1.1436967849731445, "learning_rate": 2.1266056324262023e-05, "loss": 0.5142, "step": 20980 }, { "epoch": 1.4081071104996477, "grad_norm": 0.9791481494903564, "learning_rate": 2.125716260937285e-05, "loss": 0.4777, "step": 20982 }, { "epoch": 1.4082413341834168, "grad_norm": 1.1079564094543457, "learning_rate": 2.1248270252526237e-05, "loss": 0.4608, "step": 20984 }, { "epoch": 1.4083755578671857, "grad_norm": 1.0518410205841064, "learning_rate": 2.1239379254142323e-05, "loss": 0.4951, "step": 20986 }, { "epoch": 1.4085097815509546, "grad_norm": 0.8152596950531006, "learning_rate": 2.1230489614641174e-05, "loss": 0.4868, "step": 20988 }, { "epoch": 1.4086440052347236, "grad_norm": 1.0190281867980957, "learning_rate": 2.122160133444281e-05, "loss": 0.5007, "step": 20990 }, { "epoch": 1.4087782289184927, "grad_norm": 1.1111348867416382, "learning_rate": 2.1212714413967206e-05, "loss": 0.5638, "step": 20992 }, { "epoch": 1.4089124526022616, "grad_norm": 1.0240122079849243, "learning_rate": 2.120382885363424e-05, "loss": 0.4973, "step": 20994 }, { "epoch": 1.4090466762860308, "grad_norm": 1.0007127523422241, "learning_rate": 2.119494465386373e-05, "loss": 0.4804, "step": 20996 }, { "epoch": 1.4091808999697997, "grad_norm": 0.9426189064979553, "learning_rate": 2.1186061815075424e-05, "loss": 0.5261, "step": 20998 }, { "epoch": 1.4093151236535686, "grad_norm": 0.9891473054885864, "learning_rate": 2.117718033768906e-05, "loss": 0.5182, "step": 21000 }, { "epoch": 1.4094493473373377, "grad_norm": 1.1265980005264282, "learning_rate": 2.1168300222124248e-05, "loss": 0.5239, "step": 21002 }, { "epoch": 1.4095835710211067, "grad_norm": 0.9827826023101807, "learning_rate": 2.1159421468800556e-05, "loss": 0.4719, "step": 21004 }, { "epoch": 1.4097177947048758, "grad_norm": 1.1680190563201904, "learning_rate": 2.115054407813747e-05, "loss": 0.5362, "step": 21006 }, { "epoch": 1.4098520183886447, "grad_norm": 1.0358234643936157, "learning_rate": 2.1141668050554457e-05, "loss": 0.5439, "step": 21008 }, { "epoch": 1.4099862420724136, "grad_norm": 1.0975145101547241, "learning_rate": 2.1132793386470867e-05, "loss": 0.4816, "step": 21010 }, { "epoch": 1.4101204657561826, "grad_norm": 0.9598863124847412, "learning_rate": 2.112392008630604e-05, "loss": 0.5637, "step": 21012 }, { "epoch": 1.4102546894399517, "grad_norm": 1.0278030633926392, "learning_rate": 2.1115048150479233e-05, "loss": 0.4563, "step": 21014 }, { "epoch": 1.4103889131237206, "grad_norm": 1.017699956893921, "learning_rate": 2.110617757940956e-05, "loss": 0.5587, "step": 21016 }, { "epoch": 1.4105231368074898, "grad_norm": 0.8422526121139526, "learning_rate": 2.1097308373516194e-05, "loss": 0.4379, "step": 21018 }, { "epoch": 1.4106573604912587, "grad_norm": 0.9482269287109375, "learning_rate": 2.1088440533218156e-05, "loss": 0.419, "step": 21020 }, { "epoch": 1.4107915841750276, "grad_norm": 1.0598201751708984, "learning_rate": 2.107957405893447e-05, "loss": 0.4747, "step": 21022 }, { "epoch": 1.4109258078587967, "grad_norm": 1.0010422468185425, "learning_rate": 2.107070895108404e-05, "loss": 0.4989, "step": 21024 }, { "epoch": 1.4110600315425657, "grad_norm": 0.8625784516334534, "learning_rate": 2.1061845210085728e-05, "loss": 0.4603, "step": 21026 }, { "epoch": 1.4111942552263348, "grad_norm": 1.1818901300430298, "learning_rate": 2.1052982836358303e-05, "loss": 0.4763, "step": 21028 }, { "epoch": 1.4113284789101037, "grad_norm": 1.0573731660842896, "learning_rate": 2.1044121830320535e-05, "loss": 0.4394, "step": 21030 }, { "epoch": 1.4114627025938726, "grad_norm": 1.0304356813430786, "learning_rate": 2.1035262192391077e-05, "loss": 0.4655, "step": 21032 }, { "epoch": 1.4115969262776418, "grad_norm": 0.9857569336891174, "learning_rate": 2.102640392298852e-05, "loss": 0.4512, "step": 21034 }, { "epoch": 1.4117311499614107, "grad_norm": 1.1267962455749512, "learning_rate": 2.101754702253138e-05, "loss": 0.5103, "step": 21036 }, { "epoch": 1.4118653736451796, "grad_norm": 0.9692539572715759, "learning_rate": 2.1008691491438175e-05, "loss": 0.5152, "step": 21038 }, { "epoch": 1.4119995973289488, "grad_norm": 1.0739535093307495, "learning_rate": 2.099983733012727e-05, "loss": 0.4613, "step": 21040 }, { "epoch": 1.4121338210127177, "grad_norm": 1.01617431640625, "learning_rate": 2.0990984539017062e-05, "loss": 0.5276, "step": 21042 }, { "epoch": 1.4122680446964866, "grad_norm": 1.0768928527832031, "learning_rate": 2.0982133118525755e-05, "loss": 0.4781, "step": 21044 }, { "epoch": 1.4124022683802557, "grad_norm": 1.089753270149231, "learning_rate": 2.0973283069071608e-05, "loss": 0.5187, "step": 21046 }, { "epoch": 1.4125364920640247, "grad_norm": 1.1406461000442505, "learning_rate": 2.0964434391072745e-05, "loss": 0.4936, "step": 21048 }, { "epoch": 1.4126707157477938, "grad_norm": 1.0846813917160034, "learning_rate": 2.0955587084947282e-05, "loss": 0.4411, "step": 21050 }, { "epoch": 1.4128049394315627, "grad_norm": 0.9188988208770752, "learning_rate": 2.0946741151113215e-05, "loss": 0.4442, "step": 21052 }, { "epoch": 1.4129391631153316, "grad_norm": 1.0002262592315674, "learning_rate": 2.0937896589988494e-05, "loss": 0.4859, "step": 21054 }, { "epoch": 1.4130733867991008, "grad_norm": 0.9710934162139893, "learning_rate": 2.0929053401990995e-05, "loss": 0.5224, "step": 21056 }, { "epoch": 1.4132076104828697, "grad_norm": 0.9674138426780701, "learning_rate": 2.0920211587538573e-05, "loss": 0.5951, "step": 21058 }, { "epoch": 1.4133418341666388, "grad_norm": 1.0755289793014526, "learning_rate": 2.091137114704897e-05, "loss": 0.4637, "step": 21060 }, { "epoch": 1.4134760578504078, "grad_norm": 0.8497282266616821, "learning_rate": 2.0902532080939886e-05, "loss": 0.4851, "step": 21062 }, { "epoch": 1.4136102815341767, "grad_norm": 0.9734933972358704, "learning_rate": 2.089369438962892e-05, "loss": 0.4962, "step": 21064 }, { "epoch": 1.4137445052179456, "grad_norm": 0.9598963856697083, "learning_rate": 2.0884858073533674e-05, "loss": 0.5254, "step": 21066 }, { "epoch": 1.4138787289017147, "grad_norm": 1.0404390096664429, "learning_rate": 2.087602313307162e-05, "loss": 0.5203, "step": 21068 }, { "epoch": 1.4140129525854837, "grad_norm": 0.8939399719238281, "learning_rate": 2.086718956866024e-05, "loss": 0.5101, "step": 21070 }, { "epoch": 1.4141471762692528, "grad_norm": 0.8685775399208069, "learning_rate": 2.0858357380716826e-05, "loss": 0.4528, "step": 21072 }, { "epoch": 1.4142813999530217, "grad_norm": 1.0026849508285522, "learning_rate": 2.084952656965874e-05, "loss": 0.5312, "step": 21074 }, { "epoch": 1.4144156236367906, "grad_norm": 1.0558052062988281, "learning_rate": 2.084069713590318e-05, "loss": 0.4603, "step": 21076 }, { "epoch": 1.4145498473205598, "grad_norm": 1.0284819602966309, "learning_rate": 2.0831869079867368e-05, "loss": 0.4986, "step": 21078 }, { "epoch": 1.4146840710043287, "grad_norm": 1.1609539985656738, "learning_rate": 2.0823042401968374e-05, "loss": 0.504, "step": 21080 }, { "epoch": 1.4148182946880978, "grad_norm": 1.082035779953003, "learning_rate": 2.0814217102623264e-05, "loss": 0.5068, "step": 21082 }, { "epoch": 1.4149525183718668, "grad_norm": 1.0110067129135132, "learning_rate": 2.0805393182248995e-05, "loss": 0.4733, "step": 21084 }, { "epoch": 1.4150867420556357, "grad_norm": 1.0480561256408691, "learning_rate": 2.0796570641262476e-05, "loss": 0.4388, "step": 21086 }, { "epoch": 1.4152209657394046, "grad_norm": 0.9707655310630798, "learning_rate": 2.0787749480080587e-05, "loss": 0.4946, "step": 21088 }, { "epoch": 1.4153551894231737, "grad_norm": 0.9844449162483215, "learning_rate": 2.0778929699120088e-05, "loss": 0.5224, "step": 21090 }, { "epoch": 1.4154894131069427, "grad_norm": 2.456658124923706, "learning_rate": 2.0770111298797702e-05, "loss": 0.5376, "step": 21092 }, { "epoch": 1.4156236367907118, "grad_norm": 1.0447827577590942, "learning_rate": 2.076129427953006e-05, "loss": 0.5156, "step": 21094 }, { "epoch": 1.4157578604744807, "grad_norm": 1.0012128353118896, "learning_rate": 2.0752478641733785e-05, "loss": 0.4915, "step": 21096 }, { "epoch": 1.4158920841582496, "grad_norm": 1.5245203971862793, "learning_rate": 2.0743664385825385e-05, "loss": 0.6781, "step": 21098 }, { "epoch": 1.4160263078420188, "grad_norm": 1.121191382408142, "learning_rate": 2.073485151222131e-05, "loss": 0.5262, "step": 21100 }, { "epoch": 1.4161605315257877, "grad_norm": 0.9988177418708801, "learning_rate": 2.072604002133794e-05, "loss": 0.4631, "step": 21102 }, { "epoch": 1.4162947552095568, "grad_norm": 1.0113762617111206, "learning_rate": 2.071722991359163e-05, "loss": 0.4545, "step": 21104 }, { "epoch": 1.4164289788933258, "grad_norm": 0.9512695670127869, "learning_rate": 2.0708421189398604e-05, "loss": 0.4821, "step": 21106 }, { "epoch": 1.4165632025770947, "grad_norm": 1.1240195035934448, "learning_rate": 2.0699613849175114e-05, "loss": 0.4688, "step": 21108 }, { "epoch": 1.4166974262608638, "grad_norm": 0.9319685101509094, "learning_rate": 2.0690807893337217e-05, "loss": 0.4531, "step": 21110 }, { "epoch": 1.4168316499446327, "grad_norm": 0.9338365197181702, "learning_rate": 2.068200332230103e-05, "loss": 0.4689, "step": 21112 }, { "epoch": 1.4169658736284017, "grad_norm": 1.0448791980743408, "learning_rate": 2.0673200136482518e-05, "loss": 0.461, "step": 21114 }, { "epoch": 1.4171000973121708, "grad_norm": 1.1320430040359497, "learning_rate": 2.0664398336297646e-05, "loss": 0.5149, "step": 21116 }, { "epoch": 1.4172343209959397, "grad_norm": 0.934168815612793, "learning_rate": 2.0655597922162274e-05, "loss": 0.4847, "step": 21118 }, { "epoch": 1.4173685446797086, "grad_norm": 1.030619740486145, "learning_rate": 2.064679889449219e-05, "loss": 0.4648, "step": 21120 }, { "epoch": 1.4175027683634778, "grad_norm": 1.2354577779769897, "learning_rate": 2.063800125370312e-05, "loss": 0.4735, "step": 21122 }, { "epoch": 1.4176369920472467, "grad_norm": 0.9623546600341797, "learning_rate": 2.062920500021078e-05, "loss": 0.5488, "step": 21124 }, { "epoch": 1.4177712157310158, "grad_norm": 0.9811288118362427, "learning_rate": 2.062041013443074e-05, "loss": 0.509, "step": 21126 }, { "epoch": 1.4179054394147848, "grad_norm": 1.1046497821807861, "learning_rate": 2.0611616656778547e-05, "loss": 0.4268, "step": 21128 }, { "epoch": 1.4180396630985537, "grad_norm": 1.0107942819595337, "learning_rate": 2.0602824567669664e-05, "loss": 0.5002, "step": 21130 }, { "epoch": 1.4181738867823228, "grad_norm": 1.008588194847107, "learning_rate": 2.059403386751953e-05, "loss": 0.4647, "step": 21132 }, { "epoch": 1.4183081104660917, "grad_norm": 0.8859527111053467, "learning_rate": 2.0585244556743448e-05, "loss": 0.4628, "step": 21134 }, { "epoch": 1.4184423341498609, "grad_norm": 1.128159523010254, "learning_rate": 2.0576456635756762e-05, "loss": 0.5023, "step": 21136 }, { "epoch": 1.4185765578336298, "grad_norm": 0.9821992516517639, "learning_rate": 2.0567670104974607e-05, "loss": 0.5079, "step": 21138 }, { "epoch": 1.4187107815173987, "grad_norm": 1.0721951723098755, "learning_rate": 2.0558884964812174e-05, "loss": 0.5065, "step": 21140 }, { "epoch": 1.4188450052011676, "grad_norm": 0.9656804203987122, "learning_rate": 2.055010121568452e-05, "loss": 0.4979, "step": 21142 }, { "epoch": 1.4189792288849368, "grad_norm": 0.9861791133880615, "learning_rate": 2.0541318858006685e-05, "loss": 0.4696, "step": 21144 }, { "epoch": 1.4191134525687057, "grad_norm": 0.8539104461669922, "learning_rate": 2.0532537892193615e-05, "loss": 0.4893, "step": 21146 }, { "epoch": 1.4192476762524748, "grad_norm": 1.1190543174743652, "learning_rate": 2.052375831866018e-05, "loss": 0.5314, "step": 21148 }, { "epoch": 1.4193818999362438, "grad_norm": 1.093247652053833, "learning_rate": 2.05149801378212e-05, "loss": 0.493, "step": 21150 }, { "epoch": 1.4195161236200127, "grad_norm": 0.9135884642601013, "learning_rate": 2.0506203350091414e-05, "loss": 0.4278, "step": 21152 }, { "epoch": 1.4196503473037818, "grad_norm": 1.005814790725708, "learning_rate": 2.0497427955885533e-05, "loss": 0.5174, "step": 21154 }, { "epoch": 1.4197845709875507, "grad_norm": 1.111291527748108, "learning_rate": 2.0488653955618174e-05, "loss": 0.4829, "step": 21156 }, { "epoch": 1.4199187946713199, "grad_norm": 0.9510667324066162, "learning_rate": 2.0479881349703883e-05, "loss": 0.4606, "step": 21158 }, { "epoch": 1.4200530183550888, "grad_norm": 1.0444616079330444, "learning_rate": 2.047111013855713e-05, "loss": 0.4972, "step": 21160 }, { "epoch": 1.4201872420388577, "grad_norm": 1.0435190200805664, "learning_rate": 2.0462340322592378e-05, "loss": 0.4813, "step": 21162 }, { "epoch": 1.4203214657226266, "grad_norm": 0.8452038764953613, "learning_rate": 2.0453571902223938e-05, "loss": 0.4287, "step": 21164 }, { "epoch": 1.4204556894063958, "grad_norm": 0.8310977816581726, "learning_rate": 2.044480487786617e-05, "loss": 0.3898, "step": 21166 }, { "epoch": 1.4205899130901647, "grad_norm": 1.0622636079788208, "learning_rate": 2.0436039249933208e-05, "loss": 0.4817, "step": 21168 }, { "epoch": 1.4207241367739338, "grad_norm": 1.041193962097168, "learning_rate": 2.042727501883928e-05, "loss": 0.5178, "step": 21170 }, { "epoch": 1.4208583604577028, "grad_norm": 1.0471394062042236, "learning_rate": 2.041851218499844e-05, "loss": 0.5063, "step": 21172 }, { "epoch": 1.4209925841414717, "grad_norm": 1.1910574436187744, "learning_rate": 2.040975074882474e-05, "loss": 0.5231, "step": 21174 }, { "epoch": 1.4211268078252408, "grad_norm": 1.1655224561691284, "learning_rate": 2.040099071073214e-05, "loss": 0.5138, "step": 21176 }, { "epoch": 1.4212610315090097, "grad_norm": 1.1665756702423096, "learning_rate": 2.0392232071134517e-05, "loss": 0.4695, "step": 21178 }, { "epoch": 1.4213952551927789, "grad_norm": 1.0353599786758423, "learning_rate": 2.0383474830445692e-05, "loss": 0.481, "step": 21180 }, { "epoch": 1.4215294788765478, "grad_norm": 1.0987224578857422, "learning_rate": 2.037471898907946e-05, "loss": 0.4692, "step": 21182 }, { "epoch": 1.4216637025603167, "grad_norm": 1.1034603118896484, "learning_rate": 2.0365964547449502e-05, "loss": 0.4786, "step": 21184 }, { "epoch": 1.4217979262440856, "grad_norm": 0.8870950937271118, "learning_rate": 2.035721150596944e-05, "loss": 0.4515, "step": 21186 }, { "epoch": 1.4219321499278548, "grad_norm": 0.970938503742218, "learning_rate": 2.0348459865052832e-05, "loss": 0.5585, "step": 21188 }, { "epoch": 1.4220663736116237, "grad_norm": 1.1109589338302612, "learning_rate": 2.0339709625113208e-05, "loss": 0.4898, "step": 21190 }, { "epoch": 1.4222005972953928, "grad_norm": 1.042851448059082, "learning_rate": 2.0330960786563963e-05, "loss": 0.4403, "step": 21192 }, { "epoch": 1.4223348209791618, "grad_norm": 1.2549662590026855, "learning_rate": 2.0322213349818507e-05, "loss": 0.4719, "step": 21194 }, { "epoch": 1.4224690446629307, "grad_norm": 0.9628381729125977, "learning_rate": 2.031346731529008e-05, "loss": 0.4539, "step": 21196 }, { "epoch": 1.4226032683466998, "grad_norm": 1.042144775390625, "learning_rate": 2.0304722683391964e-05, "loss": 0.4781, "step": 21198 }, { "epoch": 1.4227374920304687, "grad_norm": 1.130577564239502, "learning_rate": 2.029597945453729e-05, "loss": 0.5006, "step": 21200 }, { "epoch": 1.4228717157142379, "grad_norm": 1.0895018577575684, "learning_rate": 2.0287237629139192e-05, "loss": 0.493, "step": 21202 }, { "epoch": 1.4230059393980068, "grad_norm": 0.9593971967697144, "learning_rate": 2.0278497207610697e-05, "loss": 0.4971, "step": 21204 }, { "epoch": 1.4231401630817757, "grad_norm": 1.0314475297927856, "learning_rate": 2.0269758190364758e-05, "loss": 0.4966, "step": 21206 }, { "epoch": 1.4232743867655449, "grad_norm": 1.0766124725341797, "learning_rate": 2.026102057781426e-05, "loss": 0.4696, "step": 21208 }, { "epoch": 1.4234086104493138, "grad_norm": 0.9447959065437317, "learning_rate": 2.0252284370372087e-05, "loss": 0.4751, "step": 21210 }, { "epoch": 1.423542834133083, "grad_norm": 0.9836453199386597, "learning_rate": 2.0243549568450974e-05, "loss": 0.5005, "step": 21212 }, { "epoch": 1.4236770578168518, "grad_norm": 1.0962803363800049, "learning_rate": 2.023481617246363e-05, "loss": 0.4753, "step": 21214 }, { "epoch": 1.4238112815006208, "grad_norm": 1.0244379043579102, "learning_rate": 2.022608418282268e-05, "loss": 0.4903, "step": 21216 }, { "epoch": 1.4239455051843897, "grad_norm": 0.9502069354057312, "learning_rate": 2.021735359994072e-05, "loss": 0.4726, "step": 21218 }, { "epoch": 1.4240797288681588, "grad_norm": 1.0028274059295654, "learning_rate": 2.020862442423024e-05, "loss": 0.4824, "step": 21220 }, { "epoch": 1.4242139525519277, "grad_norm": 1.1517637968063354, "learning_rate": 2.0199896656103666e-05, "loss": 0.498, "step": 21222 }, { "epoch": 1.4243481762356969, "grad_norm": 0.9527151584625244, "learning_rate": 2.0191170295973382e-05, "loss": 0.4545, "step": 21224 }, { "epoch": 1.4244823999194658, "grad_norm": 1.1030168533325195, "learning_rate": 2.0182445344251667e-05, "loss": 0.5273, "step": 21226 }, { "epoch": 1.4246166236032347, "grad_norm": 0.9871190786361694, "learning_rate": 2.0173721801350787e-05, "loss": 0.4946, "step": 21228 }, { "epoch": 1.4247508472870039, "grad_norm": 1.0557881593704224, "learning_rate": 2.0164999667682887e-05, "loss": 0.5172, "step": 21230 }, { "epoch": 1.4248850709707728, "grad_norm": 1.0631259679794312, "learning_rate": 2.0156278943660123e-05, "loss": 0.4653, "step": 21232 }, { "epoch": 1.425019294654542, "grad_norm": 0.9954543113708496, "learning_rate": 2.0147559629694455e-05, "loss": 0.4941, "step": 21234 }, { "epoch": 1.4251535183383108, "grad_norm": 0.9826318621635437, "learning_rate": 2.013884172619791e-05, "loss": 0.4995, "step": 21236 }, { "epoch": 1.4252877420220798, "grad_norm": 1.0855181217193604, "learning_rate": 2.013012523358236e-05, "loss": 0.4978, "step": 21238 }, { "epoch": 1.4254219657058487, "grad_norm": 0.9283624291419983, "learning_rate": 2.012141015225967e-05, "loss": 0.4722, "step": 21240 }, { "epoch": 1.4255561893896178, "grad_norm": 1.0445226430892944, "learning_rate": 2.0112696482641596e-05, "loss": 0.5212, "step": 21242 }, { "epoch": 1.4256904130733867, "grad_norm": 1.0525439977645874, "learning_rate": 2.0103984225139845e-05, "loss": 0.4703, "step": 21244 }, { "epoch": 1.4258246367571559, "grad_norm": 0.9168682098388672, "learning_rate": 2.0095273380166036e-05, "loss": 0.4566, "step": 21246 }, { "epoch": 1.4259588604409248, "grad_norm": 1.0187290906906128, "learning_rate": 2.0086563948131765e-05, "loss": 0.493, "step": 21248 }, { "epoch": 1.4260930841246937, "grad_norm": 1.0676566362380981, "learning_rate": 2.007785592944853e-05, "loss": 0.4371, "step": 21250 }, { "epoch": 1.4262273078084629, "grad_norm": 1.172776699066162, "learning_rate": 2.0069149324527763e-05, "loss": 0.4856, "step": 21252 }, { "epoch": 1.4263615314922318, "grad_norm": 0.997073233127594, "learning_rate": 2.0060444133780816e-05, "loss": 0.4938, "step": 21254 }, { "epoch": 1.426495755176001, "grad_norm": 1.026848316192627, "learning_rate": 2.0051740357619024e-05, "loss": 0.4413, "step": 21256 }, { "epoch": 1.4266299788597698, "grad_norm": 1.0656684637069702, "learning_rate": 2.0043037996453597e-05, "loss": 0.5141, "step": 21258 }, { "epoch": 1.4267642025435388, "grad_norm": 0.9959386587142944, "learning_rate": 2.0034337050695752e-05, "loss": 0.4947, "step": 21260 }, { "epoch": 1.4268984262273077, "grad_norm": 0.9422990083694458, "learning_rate": 2.0025637520756518e-05, "loss": 0.4286, "step": 21262 }, { "epoch": 1.4270326499110768, "grad_norm": 1.0149545669555664, "learning_rate": 2.0016939407046987e-05, "loss": 0.555, "step": 21264 }, { "epoch": 1.4271668735948457, "grad_norm": 1.128385066986084, "learning_rate": 2.000824270997809e-05, "loss": 0.4297, "step": 21266 }, { "epoch": 1.4273010972786149, "grad_norm": 1.0133975744247437, "learning_rate": 1.9999547429960774e-05, "loss": 0.4741, "step": 21268 }, { "epoch": 1.4274353209623838, "grad_norm": 1.1769696474075317, "learning_rate": 1.9990853567405843e-05, "loss": 0.5217, "step": 21270 }, { "epoch": 1.4275695446461527, "grad_norm": 1.0363843441009521, "learning_rate": 1.998216112272407e-05, "loss": 0.477, "step": 21272 }, { "epoch": 1.4277037683299219, "grad_norm": 0.99687260389328, "learning_rate": 1.997347009632614e-05, "loss": 0.5187, "step": 21274 }, { "epoch": 1.4278379920136908, "grad_norm": 1.007660150527954, "learning_rate": 1.996478048862272e-05, "loss": 0.4331, "step": 21276 }, { "epoch": 1.42797221569746, "grad_norm": 0.9654832482337952, "learning_rate": 1.9956092300024364e-05, "loss": 0.4636, "step": 21278 }, { "epoch": 1.4281064393812288, "grad_norm": 1.1670583486557007, "learning_rate": 1.9947405530941565e-05, "loss": 0.514, "step": 21280 }, { "epoch": 1.4282406630649978, "grad_norm": 0.8719428777694702, "learning_rate": 1.9938720181784743e-05, "loss": 0.4627, "step": 21282 }, { "epoch": 1.428374886748767, "grad_norm": 1.3694096803665161, "learning_rate": 1.9930036252964295e-05, "loss": 0.4919, "step": 21284 }, { "epoch": 1.4285091104325358, "grad_norm": 1.0464633703231812, "learning_rate": 1.9921353744890488e-05, "loss": 0.4696, "step": 21286 }, { "epoch": 1.428643334116305, "grad_norm": 0.984787106513977, "learning_rate": 1.99126726579736e-05, "loss": 0.5129, "step": 21288 }, { "epoch": 1.4287775578000739, "grad_norm": 1.0427882671356201, "learning_rate": 1.9903992992623755e-05, "loss": 0.4904, "step": 21290 }, { "epoch": 1.4289117814838428, "grad_norm": 1.1129236221313477, "learning_rate": 1.9895314749251038e-05, "loss": 0.4712, "step": 21292 }, { "epoch": 1.4290460051676117, "grad_norm": 1.205096960067749, "learning_rate": 1.9886637928265516e-05, "loss": 0.5254, "step": 21294 }, { "epoch": 1.4291802288513809, "grad_norm": 1.0272270441055298, "learning_rate": 1.9877962530077123e-05, "loss": 0.4748, "step": 21296 }, { "epoch": 1.4293144525351498, "grad_norm": 1.138121247291565, "learning_rate": 1.986928855509578e-05, "loss": 0.4687, "step": 21298 }, { "epoch": 1.429448676218919, "grad_norm": 1.0026321411132812, "learning_rate": 1.9860616003731307e-05, "loss": 0.4459, "step": 21300 }, { "epoch": 1.4295828999026878, "grad_norm": 0.9387633204460144, "learning_rate": 1.9851944876393463e-05, "loss": 0.4546, "step": 21302 }, { "epoch": 1.4297171235864568, "grad_norm": 1.1133606433868408, "learning_rate": 1.984327517349192e-05, "loss": 0.4649, "step": 21304 }, { "epoch": 1.429851347270226, "grad_norm": 0.9046503901481628, "learning_rate": 1.983460689543634e-05, "loss": 0.537, "step": 21306 }, { "epoch": 1.4299855709539948, "grad_norm": 0.9311491847038269, "learning_rate": 1.982594004263627e-05, "loss": 0.4314, "step": 21308 }, { "epoch": 1.430119794637764, "grad_norm": 1.051015853881836, "learning_rate": 1.9817274615501204e-05, "loss": 0.5037, "step": 21310 }, { "epoch": 1.4302540183215329, "grad_norm": 1.1479281187057495, "learning_rate": 1.9808610614440532e-05, "loss": 0.4923, "step": 21312 }, { "epoch": 1.4303882420053018, "grad_norm": 1.0123236179351807, "learning_rate": 1.9799948039863666e-05, "loss": 0.4856, "step": 21314 }, { "epoch": 1.4305224656890707, "grad_norm": 0.9054355621337891, "learning_rate": 1.979128689217986e-05, "loss": 0.4404, "step": 21316 }, { "epoch": 1.4306566893728399, "grad_norm": 1.3398513793945312, "learning_rate": 1.9782627171798353e-05, "loss": 0.4903, "step": 21318 }, { "epoch": 1.4307909130566088, "grad_norm": 0.9859731793403625, "learning_rate": 1.9773968879128272e-05, "loss": 0.4297, "step": 21320 }, { "epoch": 1.430925136740378, "grad_norm": 0.9070358276367188, "learning_rate": 1.976531201457874e-05, "loss": 0.4997, "step": 21322 }, { "epoch": 1.4310593604241468, "grad_norm": 1.1439933776855469, "learning_rate": 1.9756656578558746e-05, "loss": 0.4739, "step": 21324 }, { "epoch": 1.4311935841079158, "grad_norm": 0.9748851656913757, "learning_rate": 1.9748002571477277e-05, "loss": 0.4601, "step": 21326 }, { "epoch": 1.431327807791685, "grad_norm": 1.045695424079895, "learning_rate": 1.9739349993743194e-05, "loss": 0.4841, "step": 21328 }, { "epoch": 1.4314620314754538, "grad_norm": 0.9456130266189575, "learning_rate": 1.973069884576532e-05, "loss": 0.4816, "step": 21330 }, { "epoch": 1.431596255159223, "grad_norm": 1.1043306589126587, "learning_rate": 1.972204912795239e-05, "loss": 0.4705, "step": 21332 }, { "epoch": 1.4317304788429919, "grad_norm": 0.9588850140571594, "learning_rate": 1.9713400840713116e-05, "loss": 0.4906, "step": 21334 }, { "epoch": 1.4318647025267608, "grad_norm": 1.0731778144836426, "learning_rate": 1.9704753984456094e-05, "loss": 0.5061, "step": 21336 }, { "epoch": 1.4319989262105297, "grad_norm": 1.1636130809783936, "learning_rate": 1.9696108559589875e-05, "loss": 0.4515, "step": 21338 }, { "epoch": 1.4321331498942989, "grad_norm": 0.8831275701522827, "learning_rate": 1.968746456652292e-05, "loss": 0.4468, "step": 21340 }, { "epoch": 1.4322673735780678, "grad_norm": 1.1297000646591187, "learning_rate": 1.967882200566367e-05, "loss": 0.5227, "step": 21342 }, { "epoch": 1.432401597261837, "grad_norm": 0.9723655581474304, "learning_rate": 1.967018087742044e-05, "loss": 0.4799, "step": 21344 }, { "epoch": 1.4325358209456058, "grad_norm": 0.8861226439476013, "learning_rate": 1.966154118220156e-05, "loss": 0.4335, "step": 21346 }, { "epoch": 1.4326700446293748, "grad_norm": 1.075437307357788, "learning_rate": 1.9652902920415167e-05, "loss": 0.4959, "step": 21348 }, { "epoch": 1.432804268313144, "grad_norm": 1.0862118005752563, "learning_rate": 1.964426609246945e-05, "loss": 0.5556, "step": 21350 }, { "epoch": 1.4329384919969128, "grad_norm": 1.2012962102890015, "learning_rate": 1.9635630698772455e-05, "loss": 0.5405, "step": 21352 }, { "epoch": 1.433072715680682, "grad_norm": 1.0934321880340576, "learning_rate": 1.9626996739732212e-05, "loss": 0.4795, "step": 21354 }, { "epoch": 1.4332069393644509, "grad_norm": 0.8701134920120239, "learning_rate": 1.961836421575667e-05, "loss": 0.445, "step": 21356 }, { "epoch": 1.4333411630482198, "grad_norm": 1.0055067539215088, "learning_rate": 1.960973312725364e-05, "loss": 0.5013, "step": 21358 }, { "epoch": 1.433475386731989, "grad_norm": 0.9744080305099487, "learning_rate": 1.960110347463097e-05, "loss": 0.4911, "step": 21360 }, { "epoch": 1.4336096104157579, "grad_norm": 1.0852227210998535, "learning_rate": 1.959247525829638e-05, "loss": 0.5211, "step": 21362 }, { "epoch": 1.433743834099527, "grad_norm": 0.905018150806427, "learning_rate": 1.9583848478657544e-05, "loss": 0.5219, "step": 21364 }, { "epoch": 1.433878057783296, "grad_norm": 0.9459779262542725, "learning_rate": 1.957522313612207e-05, "loss": 0.5036, "step": 21366 }, { "epoch": 1.4340122814670648, "grad_norm": 1.0793790817260742, "learning_rate": 1.9566599231097467e-05, "loss": 0.4279, "step": 21368 }, { "epoch": 1.4341465051508338, "grad_norm": 0.9977946281433105, "learning_rate": 1.9557976763991188e-05, "loss": 0.4438, "step": 21370 }, { "epoch": 1.434280728834603, "grad_norm": 1.025618314743042, "learning_rate": 1.9549355735210663e-05, "loss": 0.5266, "step": 21372 }, { "epoch": 1.4344149525183718, "grad_norm": 1.0904392004013062, "learning_rate": 1.95407361451632e-05, "loss": 0.5137, "step": 21374 }, { "epoch": 1.434549176202141, "grad_norm": 1.053537130355835, "learning_rate": 1.953211799425606e-05, "loss": 0.4902, "step": 21376 }, { "epoch": 1.4346833998859099, "grad_norm": 1.0914679765701294, "learning_rate": 1.9523501282896417e-05, "loss": 0.5094, "step": 21378 }, { "epoch": 1.4348176235696788, "grad_norm": 0.898318350315094, "learning_rate": 1.9514886011491422e-05, "loss": 0.4391, "step": 21380 }, { "epoch": 1.434951847253448, "grad_norm": 0.9205227494239807, "learning_rate": 1.9506272180448104e-05, "loss": 0.4686, "step": 21382 }, { "epoch": 1.4350860709372169, "grad_norm": 1.480831503868103, "learning_rate": 1.9497659790173495e-05, "loss": 0.489, "step": 21384 }, { "epoch": 1.435220294620986, "grad_norm": 0.9175587892532349, "learning_rate": 1.9489048841074444e-05, "loss": 0.4238, "step": 21386 }, { "epoch": 1.435354518304755, "grad_norm": 1.1250370740890503, "learning_rate": 1.9480439333557855e-05, "loss": 0.4379, "step": 21388 }, { "epoch": 1.4354887419885238, "grad_norm": 1.0786669254302979, "learning_rate": 1.9471831268030478e-05, "loss": 0.5135, "step": 21390 }, { "epoch": 1.4356229656722928, "grad_norm": 0.9290297031402588, "learning_rate": 1.9463224644899057e-05, "loss": 0.3903, "step": 21392 }, { "epoch": 1.435757189356062, "grad_norm": 0.8532508015632629, "learning_rate": 1.945461946457023e-05, "loss": 0.3723, "step": 21394 }, { "epoch": 1.4358914130398308, "grad_norm": 1.1052356958389282, "learning_rate": 1.944601572745056e-05, "loss": 0.5119, "step": 21396 }, { "epoch": 1.4360256367236, "grad_norm": 1.059980034828186, "learning_rate": 1.9437413433946556e-05, "loss": 0.538, "step": 21398 }, { "epoch": 1.4361598604073689, "grad_norm": 0.925700843334198, "learning_rate": 1.942881258446468e-05, "loss": 0.4563, "step": 21400 }, { "epoch": 1.4362940840911378, "grad_norm": 1.0991933345794678, "learning_rate": 1.9420213179411295e-05, "loss": 0.5543, "step": 21402 }, { "epoch": 1.436428307774907, "grad_norm": 1.028889536857605, "learning_rate": 1.941161521919271e-05, "loss": 0.5328, "step": 21404 }, { "epoch": 1.4365625314586759, "grad_norm": 1.015349268913269, "learning_rate": 1.940301870421513e-05, "loss": 0.4829, "step": 21406 }, { "epoch": 1.436696755142445, "grad_norm": 1.159861445426941, "learning_rate": 1.9394423634884772e-05, "loss": 0.4627, "step": 21408 }, { "epoch": 1.436830978826214, "grad_norm": 1.0292245149612427, "learning_rate": 1.9385830011607698e-05, "loss": 0.4361, "step": 21410 }, { "epoch": 1.4369652025099828, "grad_norm": 1.166775107383728, "learning_rate": 1.9377237834789987e-05, "loss": 0.4988, "step": 21412 }, { "epoch": 1.4370994261937517, "grad_norm": 1.0735512971878052, "learning_rate": 1.936864710483754e-05, "loss": 0.5168, "step": 21414 }, { "epoch": 1.437233649877521, "grad_norm": 1.0026241540908813, "learning_rate": 1.9360057822156298e-05, "loss": 0.4931, "step": 21416 }, { "epoch": 1.4373678735612898, "grad_norm": 0.9932929277420044, "learning_rate": 1.9351469987152056e-05, "loss": 0.5066, "step": 21418 }, { "epoch": 1.437502097245059, "grad_norm": 1.0896600484848022, "learning_rate": 1.934288360023061e-05, "loss": 0.4603, "step": 21420 }, { "epoch": 1.4376363209288279, "grad_norm": 1.0472451448440552, "learning_rate": 1.9334298661797623e-05, "loss": 0.4696, "step": 21422 }, { "epoch": 1.4377705446125968, "grad_norm": 1.1072280406951904, "learning_rate": 1.932571517225873e-05, "loss": 0.51, "step": 21424 }, { "epoch": 1.437904768296366, "grad_norm": 1.0572043657302856, "learning_rate": 1.9317133132019472e-05, "loss": 0.4363, "step": 21426 }, { "epoch": 1.4380389919801349, "grad_norm": 0.983066201210022, "learning_rate": 1.930855254148532e-05, "loss": 0.4479, "step": 21428 }, { "epoch": 1.438173215663904, "grad_norm": 0.9452990293502808, "learning_rate": 1.9299973401061726e-05, "loss": 0.4991, "step": 21430 }, { "epoch": 1.438307439347673, "grad_norm": 1.0300242900848389, "learning_rate": 1.9291395711154025e-05, "loss": 0.5181, "step": 21432 }, { "epoch": 1.4384416630314418, "grad_norm": 1.1217944622039795, "learning_rate": 1.9282819472167484e-05, "loss": 0.5238, "step": 21434 }, { "epoch": 1.438575886715211, "grad_norm": 0.9125789403915405, "learning_rate": 1.9274244684507307e-05, "loss": 0.4463, "step": 21436 }, { "epoch": 1.43871011039898, "grad_norm": 1.03207266330719, "learning_rate": 1.9265671348578667e-05, "loss": 0.5244, "step": 21438 }, { "epoch": 1.438844334082749, "grad_norm": 1.0886263847351074, "learning_rate": 1.9257099464786615e-05, "loss": 0.4924, "step": 21440 }, { "epoch": 1.438978557766518, "grad_norm": 0.8789401650428772, "learning_rate": 1.9248529033536162e-05, "loss": 0.5126, "step": 21442 }, { "epoch": 1.4391127814502869, "grad_norm": 1.0428028106689453, "learning_rate": 1.9239960055232226e-05, "loss": 0.4548, "step": 21444 }, { "epoch": 1.4392470051340558, "grad_norm": 1.0682729482650757, "learning_rate": 1.9231392530279706e-05, "loss": 0.4665, "step": 21446 }, { "epoch": 1.439381228817825, "grad_norm": 0.9879252314567566, "learning_rate": 1.922282645908337e-05, "loss": 0.5348, "step": 21448 }, { "epoch": 1.4395154525015939, "grad_norm": 1.1306565999984741, "learning_rate": 1.9214261842047982e-05, "loss": 0.4947, "step": 21450 }, { "epoch": 1.439649676185363, "grad_norm": 1.0321546792984009, "learning_rate": 1.920569867957818e-05, "loss": 0.482, "step": 21452 }, { "epoch": 1.439783899869132, "grad_norm": 1.0249627828598022, "learning_rate": 1.9197136972078563e-05, "loss": 0.4649, "step": 21454 }, { "epoch": 1.4399181235529008, "grad_norm": 1.0898289680480957, "learning_rate": 1.9188576719953633e-05, "loss": 0.5052, "step": 21456 }, { "epoch": 1.44005234723667, "grad_norm": 0.9657506942749023, "learning_rate": 1.9180017923607886e-05, "loss": 0.428, "step": 21458 }, { "epoch": 1.440186570920439, "grad_norm": 0.9635042548179626, "learning_rate": 1.9171460583445684e-05, "loss": 0.3944, "step": 21460 }, { "epoch": 1.440320794604208, "grad_norm": 0.8400821089744568, "learning_rate": 1.9162904699871347e-05, "loss": 0.4102, "step": 21462 }, { "epoch": 1.440455018287977, "grad_norm": 1.024552583694458, "learning_rate": 1.9154350273289113e-05, "loss": 0.5273, "step": 21464 }, { "epoch": 1.4405892419717459, "grad_norm": 0.8710123300552368, "learning_rate": 1.9145797304103186e-05, "loss": 0.5207, "step": 21466 }, { "epoch": 1.4407234656555148, "grad_norm": 1.0175509452819824, "learning_rate": 1.9137245792717668e-05, "loss": 0.4739, "step": 21468 }, { "epoch": 1.440857689339284, "grad_norm": 1.0981627702713013, "learning_rate": 1.912869573953659e-05, "loss": 0.4928, "step": 21470 }, { "epoch": 1.4409919130230529, "grad_norm": 1.0668026208877563, "learning_rate": 1.9120147144963918e-05, "loss": 0.5096, "step": 21472 }, { "epoch": 1.441126136706822, "grad_norm": 0.9977552890777588, "learning_rate": 1.9111600009403592e-05, "loss": 0.4646, "step": 21474 }, { "epoch": 1.441260360390591, "grad_norm": 1.1157846450805664, "learning_rate": 1.910305433325941e-05, "loss": 0.5493, "step": 21476 }, { "epoch": 1.4413945840743598, "grad_norm": 1.1162174940109253, "learning_rate": 1.9094510116935167e-05, "loss": 0.4907, "step": 21478 }, { "epoch": 1.441528807758129, "grad_norm": 1.1077864170074463, "learning_rate": 1.9085967360834544e-05, "loss": 0.5888, "step": 21480 }, { "epoch": 1.441663031441898, "grad_norm": 0.912291944026947, "learning_rate": 1.907742606536118e-05, "loss": 0.421, "step": 21482 }, { "epoch": 1.441797255125667, "grad_norm": 1.0846011638641357, "learning_rate": 1.9068886230918608e-05, "loss": 0.4901, "step": 21484 }, { "epoch": 1.441931478809436, "grad_norm": 1.0536340475082397, "learning_rate": 1.906034785791036e-05, "loss": 0.5485, "step": 21486 }, { "epoch": 1.4420657024932049, "grad_norm": 1.0287697315216064, "learning_rate": 1.9051810946739828e-05, "loss": 0.5602, "step": 21488 }, { "epoch": 1.4421999261769738, "grad_norm": 0.951287567615509, "learning_rate": 1.9043275497810377e-05, "loss": 0.4924, "step": 21490 }, { "epoch": 1.442334149860743, "grad_norm": 0.9405043721199036, "learning_rate": 1.903474151152527e-05, "loss": 0.4611, "step": 21492 }, { "epoch": 1.4424683735445119, "grad_norm": 0.859656035900116, "learning_rate": 1.9026208988287746e-05, "loss": 0.4543, "step": 21494 }, { "epoch": 1.442602597228281, "grad_norm": 0.9674857258796692, "learning_rate": 1.901767792850094e-05, "loss": 0.5117, "step": 21496 }, { "epoch": 1.44273682091205, "grad_norm": 1.019100546836853, "learning_rate": 1.9009148332567932e-05, "loss": 0.4471, "step": 21498 }, { "epoch": 1.4428710445958188, "grad_norm": 0.9753833413124084, "learning_rate": 1.900062020089173e-05, "loss": 0.5191, "step": 21500 }, { "epoch": 1.443005268279588, "grad_norm": 0.9421480894088745, "learning_rate": 1.899209353387524e-05, "loss": 0.419, "step": 21502 }, { "epoch": 1.443139491963357, "grad_norm": 0.9760271310806274, "learning_rate": 1.8983568331921375e-05, "loss": 0.4939, "step": 21504 }, { "epoch": 1.443273715647126, "grad_norm": 0.9397803544998169, "learning_rate": 1.89750445954329e-05, "loss": 0.4635, "step": 21506 }, { "epoch": 1.443407939330895, "grad_norm": 1.3463327884674072, "learning_rate": 1.896652232481259e-05, "loss": 0.5534, "step": 21508 }, { "epoch": 1.4435421630146639, "grad_norm": 1.047634243965149, "learning_rate": 1.8958001520463043e-05, "loss": 0.5032, "step": 21510 }, { "epoch": 1.443676386698433, "grad_norm": 1.0969831943511963, "learning_rate": 1.8949482182786904e-05, "loss": 0.4102, "step": 21512 }, { "epoch": 1.443810610382202, "grad_norm": 0.9968197345733643, "learning_rate": 1.8940964312186653e-05, "loss": 0.4793, "step": 21514 }, { "epoch": 1.443944834065971, "grad_norm": 1.0413786172866821, "learning_rate": 1.8932447909064775e-05, "loss": 0.4534, "step": 21516 }, { "epoch": 1.44407905774974, "grad_norm": 0.8649364709854126, "learning_rate": 1.892393297382365e-05, "loss": 0.4605, "step": 21518 }, { "epoch": 1.444213281433509, "grad_norm": 1.0816700458526611, "learning_rate": 1.8915419506865583e-05, "loss": 0.4633, "step": 21520 }, { "epoch": 1.4443475051172778, "grad_norm": 1.1534231901168823, "learning_rate": 1.8906907508592798e-05, "loss": 0.5159, "step": 21522 }, { "epoch": 1.444481728801047, "grad_norm": 0.9998898506164551, "learning_rate": 1.889839697940751e-05, "loss": 0.4935, "step": 21524 }, { "epoch": 1.444615952484816, "grad_norm": 1.0361120700836182, "learning_rate": 1.8889887919711808e-05, "loss": 0.5168, "step": 21526 }, { "epoch": 1.444750176168585, "grad_norm": 0.9519708156585693, "learning_rate": 1.8881380329907722e-05, "loss": 0.4998, "step": 21528 }, { "epoch": 1.444884399852354, "grad_norm": 1.0586285591125488, "learning_rate": 1.8872874210397213e-05, "loss": 0.4256, "step": 21530 }, { "epoch": 1.4450186235361229, "grad_norm": 1.076627254486084, "learning_rate": 1.88643695615822e-05, "loss": 0.5524, "step": 21532 }, { "epoch": 1.445152847219892, "grad_norm": 0.9264193177223206, "learning_rate": 1.8855866383864483e-05, "loss": 0.4912, "step": 21534 }, { "epoch": 1.445287070903661, "grad_norm": 1.0108731985092163, "learning_rate": 1.8847364677645873e-05, "loss": 0.5871, "step": 21536 }, { "epoch": 1.44542129458743, "grad_norm": 1.1152682304382324, "learning_rate": 1.8838864443327986e-05, "loss": 0.4531, "step": 21538 }, { "epoch": 1.445555518271199, "grad_norm": 1.0338091850280762, "learning_rate": 1.8830365681312502e-05, "loss": 0.5172, "step": 21540 }, { "epoch": 1.445689741954968, "grad_norm": 1.0477180480957031, "learning_rate": 1.8821868392000925e-05, "loss": 0.4847, "step": 21542 }, { "epoch": 1.4458239656387368, "grad_norm": 1.0267977714538574, "learning_rate": 1.8813372575794768e-05, "loss": 0.5348, "step": 21544 }, { "epoch": 1.445958189322506, "grad_norm": 0.9160050749778748, "learning_rate": 1.880487823309544e-05, "loss": 0.5006, "step": 21546 }, { "epoch": 1.446092413006275, "grad_norm": 0.9413403868675232, "learning_rate": 1.879638536430427e-05, "loss": 0.4925, "step": 21548 }, { "epoch": 1.446226636690044, "grad_norm": 1.0063854455947876, "learning_rate": 1.8787893969822517e-05, "loss": 0.468, "step": 21550 }, { "epoch": 1.446360860373813, "grad_norm": 0.9902949929237366, "learning_rate": 1.8779404050051413e-05, "loss": 0.4644, "step": 21552 }, { "epoch": 1.4464950840575819, "grad_norm": 0.7921978235244751, "learning_rate": 1.8770915605392077e-05, "loss": 0.4609, "step": 21554 }, { "epoch": 1.446629307741351, "grad_norm": 1.0576163530349731, "learning_rate": 1.8762428636245564e-05, "loss": 0.4959, "step": 21556 }, { "epoch": 1.44676353142512, "grad_norm": 0.8750025033950806, "learning_rate": 1.8753943143012864e-05, "loss": 0.4538, "step": 21558 }, { "epoch": 1.446897755108889, "grad_norm": 0.9878489971160889, "learning_rate": 1.8745459126094915e-05, "loss": 0.5063, "step": 21560 }, { "epoch": 1.447031978792658, "grad_norm": 1.0368151664733887, "learning_rate": 1.8736976585892572e-05, "loss": 0.4613, "step": 21562 }, { "epoch": 1.447166202476427, "grad_norm": 1.1094083786010742, "learning_rate": 1.8728495522806606e-05, "loss": 0.4851, "step": 21564 }, { "epoch": 1.4473004261601958, "grad_norm": 0.9020206928253174, "learning_rate": 1.8720015937237733e-05, "loss": 0.4275, "step": 21566 }, { "epoch": 1.447434649843965, "grad_norm": 1.0404009819030762, "learning_rate": 1.871153782958658e-05, "loss": 0.4766, "step": 21568 }, { "epoch": 1.447568873527734, "grad_norm": 1.043221354484558, "learning_rate": 1.870306120025375e-05, "loss": 0.4849, "step": 21570 }, { "epoch": 1.447703097211503, "grad_norm": 0.9704181551933289, "learning_rate": 1.869458604963973e-05, "loss": 0.446, "step": 21572 }, { "epoch": 1.447837320895272, "grad_norm": 0.889318585395813, "learning_rate": 1.8686112378144988e-05, "loss": 0.4764, "step": 21574 }, { "epoch": 1.4479715445790409, "grad_norm": 0.9584651589393616, "learning_rate": 1.867764018616982e-05, "loss": 0.5097, "step": 21576 }, { "epoch": 1.44810576826281, "grad_norm": 1.079387903213501, "learning_rate": 1.8669169474114582e-05, "loss": 0.5277, "step": 21578 }, { "epoch": 1.448239991946579, "grad_norm": 0.998833954334259, "learning_rate": 1.8660700242379463e-05, "loss": 0.4426, "step": 21580 }, { "epoch": 1.448374215630348, "grad_norm": 0.927009642124176, "learning_rate": 1.8652232491364648e-05, "loss": 0.5055, "step": 21582 }, { "epoch": 1.448508439314117, "grad_norm": 1.0209903717041016, "learning_rate": 1.8643766221470204e-05, "loss": 0.5451, "step": 21584 }, { "epoch": 1.448642662997886, "grad_norm": 1.0737221240997314, "learning_rate": 1.863530143309615e-05, "loss": 0.488, "step": 21586 }, { "epoch": 1.448776886681655, "grad_norm": 1.0311989784240723, "learning_rate": 1.8626838126642403e-05, "loss": 0.4725, "step": 21588 }, { "epoch": 1.448911110365424, "grad_norm": 1.0966482162475586, "learning_rate": 1.861837630250888e-05, "loss": 0.5068, "step": 21590 }, { "epoch": 1.4490453340491931, "grad_norm": 1.0530413389205933, "learning_rate": 1.860991596109537e-05, "loss": 0.5359, "step": 21592 }, { "epoch": 1.449179557732962, "grad_norm": 0.965501070022583, "learning_rate": 1.8601457102801608e-05, "loss": 0.4393, "step": 21594 }, { "epoch": 1.449313781416731, "grad_norm": 1.1146641969680786, "learning_rate": 1.8592999728027234e-05, "loss": 0.5026, "step": 21596 }, { "epoch": 1.4494480051004999, "grad_norm": 1.087504267692566, "learning_rate": 1.858454383717188e-05, "loss": 0.5362, "step": 21598 }, { "epoch": 1.449582228784269, "grad_norm": 0.9007117748260498, "learning_rate": 1.857608943063504e-05, "loss": 0.4667, "step": 21600 }, { "epoch": 1.449716452468038, "grad_norm": 1.2094337940216064, "learning_rate": 1.8567636508816212e-05, "loss": 0.587, "step": 21602 }, { "epoch": 1.449850676151807, "grad_norm": 1.0666933059692383, "learning_rate": 1.855918507211472e-05, "loss": 0.4908, "step": 21604 }, { "epoch": 1.449984899835576, "grad_norm": 0.9486268758773804, "learning_rate": 1.8550735120929926e-05, "loss": 0.4853, "step": 21606 }, { "epoch": 1.450119123519345, "grad_norm": 1.0211158990859985, "learning_rate": 1.8542286655661027e-05, "loss": 0.5098, "step": 21608 }, { "epoch": 1.450253347203114, "grad_norm": 1.0262904167175293, "learning_rate": 1.8533839676707255e-05, "loss": 0.5266, "step": 21610 }, { "epoch": 1.450387570886883, "grad_norm": 1.088341474533081, "learning_rate": 1.8525394184467676e-05, "loss": 0.4646, "step": 21612 }, { "epoch": 1.4505217945706521, "grad_norm": 1.045545220375061, "learning_rate": 1.851695017934133e-05, "loss": 0.5073, "step": 21614 }, { "epoch": 1.450656018254421, "grad_norm": 1.011045217514038, "learning_rate": 1.8508507661727163e-05, "loss": 0.5643, "step": 21616 }, { "epoch": 1.45079024193819, "grad_norm": 0.9788286685943604, "learning_rate": 1.85000666320241e-05, "loss": 0.5108, "step": 21618 }, { "epoch": 1.4509244656219589, "grad_norm": 1.0276010036468506, "learning_rate": 1.8491627090630948e-05, "loss": 0.4665, "step": 21620 }, { "epoch": 1.451058689305728, "grad_norm": 0.9841398000717163, "learning_rate": 1.848318903794646e-05, "loss": 0.4942, "step": 21622 }, { "epoch": 1.451192912989497, "grad_norm": 1.0520919561386108, "learning_rate": 1.8474752474369296e-05, "loss": 0.4844, "step": 21624 }, { "epoch": 1.451327136673266, "grad_norm": 0.9467720985412598, "learning_rate": 1.84663174002981e-05, "loss": 0.4829, "step": 21626 }, { "epoch": 1.451461360357035, "grad_norm": 0.9637964963912964, "learning_rate": 1.845788381613138e-05, "loss": 0.4828, "step": 21628 }, { "epoch": 1.451595584040804, "grad_norm": 1.0604698657989502, "learning_rate": 1.8449451722267646e-05, "loss": 0.492, "step": 21630 }, { "epoch": 1.451729807724573, "grad_norm": 1.139906406402588, "learning_rate": 1.844102111910529e-05, "loss": 0.532, "step": 21632 }, { "epoch": 1.451864031408342, "grad_norm": 1.1570615768432617, "learning_rate": 1.84325920070426e-05, "loss": 0.4849, "step": 21634 }, { "epoch": 1.4519982550921111, "grad_norm": 1.039461374282837, "learning_rate": 1.8424164386477878e-05, "loss": 0.4266, "step": 21636 }, { "epoch": 1.45213247877588, "grad_norm": 1.10236394405365, "learning_rate": 1.8415738257809275e-05, "loss": 0.5226, "step": 21638 }, { "epoch": 1.452266702459649, "grad_norm": 1.047585129737854, "learning_rate": 1.8407313621434952e-05, "loss": 0.4525, "step": 21640 }, { "epoch": 1.4524009261434179, "grad_norm": 1.1584211587905884, "learning_rate": 1.839889047775294e-05, "loss": 0.5172, "step": 21642 }, { "epoch": 1.452535149827187, "grad_norm": 1.0993553400039673, "learning_rate": 1.8390468827161207e-05, "loss": 0.5157, "step": 21644 }, { "epoch": 1.452669373510956, "grad_norm": 1.1074891090393066, "learning_rate": 1.838204867005765e-05, "loss": 0.5238, "step": 21646 }, { "epoch": 1.452803597194725, "grad_norm": 0.9702417850494385, "learning_rate": 1.837363000684013e-05, "loss": 0.4804, "step": 21648 }, { "epoch": 1.452937820878494, "grad_norm": 1.046984076499939, "learning_rate": 1.836521283790641e-05, "loss": 0.4525, "step": 21650 }, { "epoch": 1.453072044562263, "grad_norm": 1.0746935606002808, "learning_rate": 1.835679716365417e-05, "loss": 0.4884, "step": 21652 }, { "epoch": 1.453206268246032, "grad_norm": 1.0536320209503174, "learning_rate": 1.8348382984481034e-05, "loss": 0.4848, "step": 21654 }, { "epoch": 1.453340491929801, "grad_norm": 1.8641531467437744, "learning_rate": 1.833997030078457e-05, "loss": 0.5027, "step": 21656 }, { "epoch": 1.4534747156135701, "grad_norm": 1.090524435043335, "learning_rate": 1.8331559112962232e-05, "loss": 0.4615, "step": 21658 }, { "epoch": 1.453608939297339, "grad_norm": 1.1034295558929443, "learning_rate": 1.8323149421411497e-05, "loss": 0.5082, "step": 21660 }, { "epoch": 1.453743162981108, "grad_norm": 1.0330030918121338, "learning_rate": 1.831474122652962e-05, "loss": 0.4812, "step": 21662 }, { "epoch": 1.453877386664877, "grad_norm": 0.9497284293174744, "learning_rate": 1.830633452871393e-05, "loss": 0.4553, "step": 21664 }, { "epoch": 1.454011610348646, "grad_norm": 0.9797527194023132, "learning_rate": 1.8297929328361594e-05, "loss": 0.4597, "step": 21666 }, { "epoch": 1.4541458340324152, "grad_norm": 0.9925074577331543, "learning_rate": 1.828952562586978e-05, "loss": 0.4644, "step": 21668 }, { "epoch": 1.454280057716184, "grad_norm": 0.9066833853721619, "learning_rate": 1.828112342163552e-05, "loss": 0.4259, "step": 21670 }, { "epoch": 1.454414281399953, "grad_norm": 1.0383461713790894, "learning_rate": 1.827272271605581e-05, "loss": 0.4895, "step": 21672 }, { "epoch": 1.454548505083722, "grad_norm": 1.1868698596954346, "learning_rate": 1.8264323509527543e-05, "loss": 0.4826, "step": 21674 }, { "epoch": 1.454682728767491, "grad_norm": 0.9397610425949097, "learning_rate": 1.8255925802447604e-05, "loss": 0.4659, "step": 21676 }, { "epoch": 1.45481695245126, "grad_norm": 0.8566227555274963, "learning_rate": 1.8247529595212743e-05, "loss": 0.4663, "step": 21678 }, { "epoch": 1.4549511761350291, "grad_norm": 1.125922441482544, "learning_rate": 1.8239134888219676e-05, "loss": 0.5208, "step": 21680 }, { "epoch": 1.455085399818798, "grad_norm": 1.0006755590438843, "learning_rate": 1.823074168186501e-05, "loss": 0.5075, "step": 21682 }, { "epoch": 1.455219623502567, "grad_norm": 1.1265826225280762, "learning_rate": 1.8222349976545343e-05, "loss": 0.5357, "step": 21684 }, { "epoch": 1.455353847186336, "grad_norm": 1.0235995054244995, "learning_rate": 1.8213959772657142e-05, "loss": 0.4938, "step": 21686 }, { "epoch": 1.455488070870105, "grad_norm": 0.8493221998214722, "learning_rate": 1.8205571070596867e-05, "loss": 0.4864, "step": 21688 }, { "epoch": 1.4556222945538742, "grad_norm": 1.1292208433151245, "learning_rate": 1.8197183870760803e-05, "loss": 0.4735, "step": 21690 }, { "epoch": 1.455756518237643, "grad_norm": 1.0905543565750122, "learning_rate": 1.818879817354528e-05, "loss": 0.4877, "step": 21692 }, { "epoch": 1.455890741921412, "grad_norm": 0.9483680725097656, "learning_rate": 1.818041397934647e-05, "loss": 0.4189, "step": 21694 }, { "epoch": 1.456024965605181, "grad_norm": 0.9809787273406982, "learning_rate": 1.8172031288560554e-05, "loss": 0.4545, "step": 21696 }, { "epoch": 1.45615918928895, "grad_norm": 0.9923301935195923, "learning_rate": 1.8163650101583567e-05, "loss": 0.5628, "step": 21698 }, { "epoch": 1.456293412972719, "grad_norm": 1.0795013904571533, "learning_rate": 1.815527041881151e-05, "loss": 0.4643, "step": 21700 }, { "epoch": 1.4564276366564881, "grad_norm": 1.05745530128479, "learning_rate": 1.8146892240640307e-05, "loss": 0.4644, "step": 21702 }, { "epoch": 1.456561860340257, "grad_norm": 1.1342967748641968, "learning_rate": 1.8138515567465793e-05, "loss": 0.5214, "step": 21704 }, { "epoch": 1.456696084024026, "grad_norm": 1.1231117248535156, "learning_rate": 1.8130140399683776e-05, "loss": 0.429, "step": 21706 }, { "epoch": 1.456830307707795, "grad_norm": 0.9478837847709656, "learning_rate": 1.812176673768996e-05, "loss": 0.5035, "step": 21708 }, { "epoch": 1.456964531391564, "grad_norm": 1.055141806602478, "learning_rate": 1.8113394581879985e-05, "loss": 0.4912, "step": 21710 }, { "epoch": 1.4570987550753332, "grad_norm": 0.8722517490386963, "learning_rate": 1.810502393264939e-05, "loss": 0.425, "step": 21712 }, { "epoch": 1.457232978759102, "grad_norm": 1.0529042482376099, "learning_rate": 1.8096654790393714e-05, "loss": 0.4594, "step": 21714 }, { "epoch": 1.457367202442871, "grad_norm": 1.0669209957122803, "learning_rate": 1.808828715550836e-05, "loss": 0.5019, "step": 21716 }, { "epoch": 1.45750142612664, "grad_norm": 0.9976727366447449, "learning_rate": 1.8079921028388692e-05, "loss": 0.4755, "step": 21718 }, { "epoch": 1.457635649810409, "grad_norm": 1.0423046350479126, "learning_rate": 1.8071556409429968e-05, "loss": 0.4495, "step": 21720 }, { "epoch": 1.457769873494178, "grad_norm": 1.0592374801635742, "learning_rate": 1.8063193299027436e-05, "loss": 0.5846, "step": 21722 }, { "epoch": 1.4579040971779471, "grad_norm": 1.0062223672866821, "learning_rate": 1.8054831697576203e-05, "loss": 0.4799, "step": 21724 }, { "epoch": 1.458038320861716, "grad_norm": 0.9556512832641602, "learning_rate": 1.8046471605471393e-05, "loss": 0.4769, "step": 21726 }, { "epoch": 1.458172544545485, "grad_norm": 0.8796365261077881, "learning_rate": 1.8038113023107928e-05, "loss": 0.4946, "step": 21728 }, { "epoch": 1.458306768229254, "grad_norm": 1.0229606628417969, "learning_rate": 1.8029755950880787e-05, "loss": 0.4577, "step": 21730 }, { "epoch": 1.458440991913023, "grad_norm": 1.0130605697631836, "learning_rate": 1.8021400389184795e-05, "loss": 0.5172, "step": 21732 }, { "epoch": 1.4585752155967922, "grad_norm": 2.721203088760376, "learning_rate": 1.801304633841477e-05, "loss": 0.5908, "step": 21734 }, { "epoch": 1.458709439280561, "grad_norm": 1.1897770166397095, "learning_rate": 1.8004693798965404e-05, "loss": 0.5511, "step": 21736 }, { "epoch": 1.45884366296433, "grad_norm": 1.0730276107788086, "learning_rate": 1.799634277123134e-05, "loss": 0.474, "step": 21738 }, { "epoch": 1.4589778866480991, "grad_norm": 1.1049566268920898, "learning_rate": 1.7987993255607132e-05, "loss": 0.5343, "step": 21740 }, { "epoch": 1.459112110331868, "grad_norm": 0.9912712574005127, "learning_rate": 1.797964525248731e-05, "loss": 0.4348, "step": 21742 }, { "epoch": 1.4592463340156372, "grad_norm": 1.0087511539459229, "learning_rate": 1.7971298762266287e-05, "loss": 0.5086, "step": 21744 }, { "epoch": 1.4593805576994061, "grad_norm": 1.1275442838668823, "learning_rate": 1.796295378533841e-05, "loss": 0.4361, "step": 21746 }, { "epoch": 1.459514781383175, "grad_norm": 0.9750168323516846, "learning_rate": 1.7954610322097953e-05, "loss": 0.4603, "step": 21748 }, { "epoch": 1.459649005066944, "grad_norm": 1.0766751766204834, "learning_rate": 1.794626837293916e-05, "loss": 0.503, "step": 21750 }, { "epoch": 1.459783228750713, "grad_norm": 0.9707712531089783, "learning_rate": 1.7937927938256143e-05, "loss": 0.4528, "step": 21752 }, { "epoch": 1.459917452434482, "grad_norm": 1.1949578523635864, "learning_rate": 1.7929589018443016e-05, "loss": 0.4502, "step": 21754 }, { "epoch": 1.4600516761182512, "grad_norm": 1.1604313850402832, "learning_rate": 1.7921251613893715e-05, "loss": 0.5451, "step": 21756 }, { "epoch": 1.46018589980202, "grad_norm": 1.018720269203186, "learning_rate": 1.7912915725002205e-05, "loss": 0.4937, "step": 21758 }, { "epoch": 1.460320123485789, "grad_norm": 0.9990526437759399, "learning_rate": 1.7904581352162312e-05, "loss": 0.529, "step": 21760 }, { "epoch": 1.4604543471695581, "grad_norm": 0.8207942247390747, "learning_rate": 1.789624849576786e-05, "loss": 0.4089, "step": 21762 }, { "epoch": 1.460588570853327, "grad_norm": 1.1114898920059204, "learning_rate": 1.7887917156212532e-05, "loss": 0.5099, "step": 21764 }, { "epoch": 1.4607227945370962, "grad_norm": 1.045689582824707, "learning_rate": 1.7879587333889975e-05, "loss": 0.4647, "step": 21766 }, { "epoch": 1.4608570182208651, "grad_norm": 1.1530375480651855, "learning_rate": 1.7871259029193754e-05, "loss": 0.4636, "step": 21768 }, { "epoch": 1.460991241904634, "grad_norm": 1.0641634464263916, "learning_rate": 1.786293224251735e-05, "loss": 0.456, "step": 21770 }, { "epoch": 1.461125465588403, "grad_norm": 0.8483332991600037, "learning_rate": 1.785460697425422e-05, "loss": 0.4027, "step": 21772 }, { "epoch": 1.461259689272172, "grad_norm": 1.3350237607955933, "learning_rate": 1.7846283224797698e-05, "loss": 0.4661, "step": 21774 }, { "epoch": 1.461393912955941, "grad_norm": 1.0425379276275635, "learning_rate": 1.7837960994541063e-05, "loss": 0.4859, "step": 21776 }, { "epoch": 1.4615281366397102, "grad_norm": 1.0305285453796387, "learning_rate": 1.7829640283877514e-05, "loss": 0.4955, "step": 21778 }, { "epoch": 1.461662360323479, "grad_norm": 0.9458545446395874, "learning_rate": 1.7821321093200217e-05, "loss": 0.4683, "step": 21780 }, { "epoch": 1.461796584007248, "grad_norm": 0.9453576803207397, "learning_rate": 1.7813003422902224e-05, "loss": 0.4282, "step": 21782 }, { "epoch": 1.4619308076910171, "grad_norm": 0.9296939969062805, "learning_rate": 1.7804687273376526e-05, "loss": 0.425, "step": 21784 }, { "epoch": 1.462065031374786, "grad_norm": 1.0925058126449585, "learning_rate": 1.7796372645016024e-05, "loss": 0.5591, "step": 21786 }, { "epoch": 1.4621992550585552, "grad_norm": 1.0471255779266357, "learning_rate": 1.778805953821361e-05, "loss": 0.4467, "step": 21788 }, { "epoch": 1.462333478742324, "grad_norm": 0.9129785299301147, "learning_rate": 1.777974795336202e-05, "loss": 0.5041, "step": 21790 }, { "epoch": 1.462467702426093, "grad_norm": 0.9907031655311584, "learning_rate": 1.7771437890854e-05, "loss": 0.4398, "step": 21792 }, { "epoch": 1.462601926109862, "grad_norm": 1.1017745733261108, "learning_rate": 1.7763129351082165e-05, "loss": 0.4732, "step": 21794 }, { "epoch": 1.462736149793631, "grad_norm": 1.0758988857269287, "learning_rate": 1.7754822334439075e-05, "loss": 0.4765, "step": 21796 }, { "epoch": 1.4628703734774, "grad_norm": 0.9974637627601624, "learning_rate": 1.7746516841317207e-05, "loss": 0.4644, "step": 21798 }, { "epoch": 1.4630045971611692, "grad_norm": 0.8979473114013672, "learning_rate": 1.773821287210901e-05, "loss": 0.4921, "step": 21800 }, { "epoch": 1.463138820844938, "grad_norm": 1.1205793619155884, "learning_rate": 1.772991042720682e-05, "loss": 0.5132, "step": 21802 }, { "epoch": 1.463273044528707, "grad_norm": 0.9702668190002441, "learning_rate": 1.77216095070029e-05, "loss": 0.4651, "step": 21804 }, { "epoch": 1.4634072682124761, "grad_norm": 1.0302200317382812, "learning_rate": 1.7713310111889443e-05, "loss": 0.5258, "step": 21806 }, { "epoch": 1.463541491896245, "grad_norm": 0.9386501312255859, "learning_rate": 1.7705012242258613e-05, "loss": 0.5057, "step": 21808 }, { "epoch": 1.4636757155800142, "grad_norm": 0.9929801225662231, "learning_rate": 1.769671589850243e-05, "loss": 0.4907, "step": 21810 }, { "epoch": 1.463809939263783, "grad_norm": 0.9299373626708984, "learning_rate": 1.768842108101293e-05, "loss": 0.4233, "step": 21812 }, { "epoch": 1.463944162947552, "grad_norm": 0.9028799533843994, "learning_rate": 1.7680127790181967e-05, "loss": 0.451, "step": 21814 }, { "epoch": 1.4640783866313212, "grad_norm": 1.065677523612976, "learning_rate": 1.767183602640143e-05, "loss": 0.5507, "step": 21816 }, { "epoch": 1.46421261031509, "grad_norm": 0.7841746211051941, "learning_rate": 1.7663545790063047e-05, "loss": 0.4868, "step": 21818 }, { "epoch": 1.4643468339988592, "grad_norm": 0.9826595783233643, "learning_rate": 1.765525708155856e-05, "loss": 0.4955, "step": 21820 }, { "epoch": 1.4644810576826282, "grad_norm": 0.8774657249450684, "learning_rate": 1.764696990127957e-05, "loss": 0.4306, "step": 21822 }, { "epoch": 1.464615281366397, "grad_norm": 0.9303463697433472, "learning_rate": 1.7638684249617632e-05, "loss": 0.4163, "step": 21824 }, { "epoch": 1.464749505050166, "grad_norm": 1.0600500106811523, "learning_rate": 1.7630400126964214e-05, "loss": 0.4468, "step": 21826 }, { "epoch": 1.4648837287339351, "grad_norm": 1.0781666040420532, "learning_rate": 1.7622117533710752e-05, "loss": 0.4558, "step": 21828 }, { "epoch": 1.465017952417704, "grad_norm": 0.9212883710861206, "learning_rate": 1.7613836470248575e-05, "loss": 0.4128, "step": 21830 }, { "epoch": 1.4651521761014732, "grad_norm": 1.0324337482452393, "learning_rate": 1.760555693696893e-05, "loss": 0.4796, "step": 21832 }, { "epoch": 1.465286399785242, "grad_norm": 0.9544461369514465, "learning_rate": 1.7597278934263007e-05, "loss": 0.5213, "step": 21834 }, { "epoch": 1.465420623469011, "grad_norm": 1.010330080986023, "learning_rate": 1.7589002462521954e-05, "loss": 0.4643, "step": 21836 }, { "epoch": 1.4655548471527802, "grad_norm": 0.9494339227676392, "learning_rate": 1.7580727522136804e-05, "loss": 0.4516, "step": 21838 }, { "epoch": 1.465689070836549, "grad_norm": 1.0267274379730225, "learning_rate": 1.757245411349852e-05, "loss": 0.5265, "step": 21840 }, { "epoch": 1.4658232945203182, "grad_norm": 1.125027060508728, "learning_rate": 1.7564182236998024e-05, "loss": 0.5227, "step": 21842 }, { "epoch": 1.4659575182040872, "grad_norm": 1.034874439239502, "learning_rate": 1.755591189302611e-05, "loss": 0.5408, "step": 21844 }, { "epoch": 1.466091741887856, "grad_norm": 1.0111058950424194, "learning_rate": 1.754764308197358e-05, "loss": 0.5175, "step": 21846 }, { "epoch": 1.466225965571625, "grad_norm": 0.921784520149231, "learning_rate": 1.7539375804231083e-05, "loss": 0.4944, "step": 21848 }, { "epoch": 1.4663601892553941, "grad_norm": 1.0426305532455444, "learning_rate": 1.7531110060189283e-05, "loss": 0.5336, "step": 21850 }, { "epoch": 1.466494412939163, "grad_norm": 1.0855768918991089, "learning_rate": 1.752284585023865e-05, "loss": 0.4675, "step": 21852 }, { "epoch": 1.4666286366229322, "grad_norm": 1.0991650819778442, "learning_rate": 1.75145831747697e-05, "loss": 0.4888, "step": 21854 }, { "epoch": 1.466762860306701, "grad_norm": 1.0111898183822632, "learning_rate": 1.7506322034172808e-05, "loss": 0.537, "step": 21856 }, { "epoch": 1.46689708399047, "grad_norm": 0.8977757096290588, "learning_rate": 1.7498062428838314e-05, "loss": 0.4458, "step": 21858 }, { "epoch": 1.4670313076742392, "grad_norm": 0.9779512882232666, "learning_rate": 1.7489804359156458e-05, "loss": 0.5115, "step": 21860 }, { "epoch": 1.467165531358008, "grad_norm": 1.0374683141708374, "learning_rate": 1.748154782551742e-05, "loss": 0.5079, "step": 21862 }, { "epoch": 1.4672997550417772, "grad_norm": 1.0324270725250244, "learning_rate": 1.7473292828311282e-05, "loss": 0.5083, "step": 21864 }, { "epoch": 1.4674339787255462, "grad_norm": 1.0627511739730835, "learning_rate": 1.7465039367928117e-05, "loss": 0.476, "step": 21866 }, { "epoch": 1.467568202409315, "grad_norm": 1.0572152137756348, "learning_rate": 1.745678744475786e-05, "loss": 0.4595, "step": 21868 }, { "epoch": 1.467702426093084, "grad_norm": 1.024880051612854, "learning_rate": 1.7448537059190407e-05, "loss": 0.4777, "step": 21870 }, { "epoch": 1.4678366497768531, "grad_norm": 1.0588107109069824, "learning_rate": 1.7440288211615553e-05, "loss": 0.5856, "step": 21872 }, { "epoch": 1.467970873460622, "grad_norm": 1.052733302116394, "learning_rate": 1.743204090242307e-05, "loss": 0.4873, "step": 21874 }, { "epoch": 1.4681050971443912, "grad_norm": 1.1226145029067993, "learning_rate": 1.742379513200259e-05, "loss": 0.536, "step": 21876 }, { "epoch": 1.46823932082816, "grad_norm": 0.8805328011512756, "learning_rate": 1.741555090074377e-05, "loss": 0.4685, "step": 21878 }, { "epoch": 1.468373544511929, "grad_norm": 0.8982160687446594, "learning_rate": 1.7407308209036066e-05, "loss": 0.4502, "step": 21880 }, { "epoch": 1.4685077681956982, "grad_norm": 0.996573269367218, "learning_rate": 1.739906705726897e-05, "loss": 0.492, "step": 21882 }, { "epoch": 1.468641991879467, "grad_norm": 1.0391132831573486, "learning_rate": 1.739082744583183e-05, "loss": 0.4687, "step": 21884 }, { "epoch": 1.4687762155632362, "grad_norm": 0.8928671479225159, "learning_rate": 1.738258937511399e-05, "loss": 0.4398, "step": 21886 }, { "epoch": 1.4689104392470052, "grad_norm": 1.0958343744277954, "learning_rate": 1.737435284550466e-05, "loss": 0.4913, "step": 21888 }, { "epoch": 1.469044662930774, "grad_norm": 1.0855900049209595, "learning_rate": 1.7366117857393007e-05, "loss": 0.4774, "step": 21890 }, { "epoch": 1.4691788866145432, "grad_norm": 0.9509583115577698, "learning_rate": 1.7357884411168097e-05, "loss": 0.4433, "step": 21892 }, { "epoch": 1.4693131102983121, "grad_norm": 1.227495551109314, "learning_rate": 1.734965250721897e-05, "loss": 0.514, "step": 21894 }, { "epoch": 1.4694473339820813, "grad_norm": 0.9190131425857544, "learning_rate": 1.7341422145934568e-05, "loss": 0.4864, "step": 21896 }, { "epoch": 1.4695815576658502, "grad_norm": 1.0067321062088013, "learning_rate": 1.7333193327703738e-05, "loss": 0.4545, "step": 21898 }, { "epoch": 1.469715781349619, "grad_norm": 0.9455028772354126, "learning_rate": 1.7324966052915274e-05, "loss": 0.4393, "step": 21900 }, { "epoch": 1.469850005033388, "grad_norm": 1.0392711162567139, "learning_rate": 1.7316740321957937e-05, "loss": 0.5109, "step": 21902 }, { "epoch": 1.4699842287171572, "grad_norm": 1.030715823173523, "learning_rate": 1.7308516135220325e-05, "loss": 0.5246, "step": 21904 }, { "epoch": 1.470118452400926, "grad_norm": 0.9867245554924011, "learning_rate": 1.7300293493091074e-05, "loss": 0.4647, "step": 21906 }, { "epoch": 1.4702526760846952, "grad_norm": 0.9754754900932312, "learning_rate": 1.7292072395958643e-05, "loss": 0.4182, "step": 21908 }, { "epoch": 1.4703868997684642, "grad_norm": 1.0586143732070923, "learning_rate": 1.728385284421145e-05, "loss": 0.5041, "step": 21910 }, { "epoch": 1.470521123452233, "grad_norm": 1.1397920846939087, "learning_rate": 1.7275634838237897e-05, "loss": 0.5264, "step": 21912 }, { "epoch": 1.4706553471360022, "grad_norm": 1.073852777481079, "learning_rate": 1.7267418378426232e-05, "loss": 0.4827, "step": 21914 }, { "epoch": 1.4707895708197711, "grad_norm": 1.0655097961425781, "learning_rate": 1.72592034651647e-05, "loss": 0.5417, "step": 21916 }, { "epoch": 1.4709237945035403, "grad_norm": 0.9804315567016602, "learning_rate": 1.7250990098841417e-05, "loss": 0.4795, "step": 21918 }, { "epoch": 1.4710580181873092, "grad_norm": 0.9913179874420166, "learning_rate": 1.7242778279844463e-05, "loss": 0.509, "step": 21920 }, { "epoch": 1.471192241871078, "grad_norm": 0.9221985936164856, "learning_rate": 1.7234568008561797e-05, "loss": 0.4097, "step": 21922 }, { "epoch": 1.471326465554847, "grad_norm": 0.860504150390625, "learning_rate": 1.722635928538138e-05, "loss": 0.4928, "step": 21924 }, { "epoch": 1.4714606892386162, "grad_norm": 1.0933024883270264, "learning_rate": 1.7218152110691044e-05, "loss": 0.4646, "step": 21926 }, { "epoch": 1.471594912922385, "grad_norm": 0.9904744625091553, "learning_rate": 1.7209946484878554e-05, "loss": 0.4596, "step": 21928 }, { "epoch": 1.4717291366061542, "grad_norm": 1.0993026494979858, "learning_rate": 1.72017424083316e-05, "loss": 0.4669, "step": 21930 }, { "epoch": 1.4718633602899232, "grad_norm": 0.9404697418212891, "learning_rate": 1.7193539881437837e-05, "loss": 0.4174, "step": 21932 }, { "epoch": 1.471997583973692, "grad_norm": 1.056062936782837, "learning_rate": 1.71853389045848e-05, "loss": 0.4875, "step": 21934 }, { "epoch": 1.4721318076574612, "grad_norm": 0.9709733128547668, "learning_rate": 1.7177139478159977e-05, "loss": 0.4813, "step": 21936 }, { "epoch": 1.4722660313412301, "grad_norm": 0.9787045121192932, "learning_rate": 1.7168941602550754e-05, "loss": 0.5001, "step": 21938 }, { "epoch": 1.4724002550249993, "grad_norm": 1.0178070068359375, "learning_rate": 1.7160745278144498e-05, "loss": 0.4499, "step": 21940 }, { "epoch": 1.4725344787087682, "grad_norm": 1.0335322618484497, "learning_rate": 1.715255050532843e-05, "loss": 0.4557, "step": 21942 }, { "epoch": 1.472668702392537, "grad_norm": 1.0430026054382324, "learning_rate": 1.7144357284489782e-05, "loss": 0.4808, "step": 21944 }, { "epoch": 1.472802926076306, "grad_norm": 1.0066791772842407, "learning_rate": 1.7136165616015636e-05, "loss": 0.516, "step": 21946 }, { "epoch": 1.4729371497600752, "grad_norm": 1.0189493894577026, "learning_rate": 1.712797550029305e-05, "loss": 0.4766, "step": 21948 }, { "epoch": 1.473071373443844, "grad_norm": 1.0610727071762085, "learning_rate": 1.711978693770896e-05, "loss": 0.5111, "step": 21950 }, { "epoch": 1.4732055971276132, "grad_norm": 0.9555049538612366, "learning_rate": 1.7111599928650302e-05, "loss": 0.4599, "step": 21952 }, { "epoch": 1.4733398208113822, "grad_norm": 1.0025465488433838, "learning_rate": 1.710341447350387e-05, "loss": 0.5389, "step": 21954 }, { "epoch": 1.473474044495151, "grad_norm": 1.009305477142334, "learning_rate": 1.7095230572656418e-05, "loss": 0.515, "step": 21956 }, { "epoch": 1.4736082681789202, "grad_norm": 0.9866924285888672, "learning_rate": 1.70870482264946e-05, "loss": 0.4518, "step": 21958 }, { "epoch": 1.4737424918626891, "grad_norm": 1.1503498554229736, "learning_rate": 1.707886743540505e-05, "loss": 0.5341, "step": 21960 }, { "epoch": 1.4738767155464583, "grad_norm": 1.3890480995178223, "learning_rate": 1.7070688199774275e-05, "loss": 0.4966, "step": 21962 }, { "epoch": 1.4740109392302272, "grad_norm": 1.063452959060669, "learning_rate": 1.7062510519988727e-05, "loss": 0.4917, "step": 21964 }, { "epoch": 1.474145162913996, "grad_norm": 1.2469446659088135, "learning_rate": 1.705433439643478e-05, "loss": 0.4613, "step": 21966 }, { "epoch": 1.4742793865977653, "grad_norm": 1.0755219459533691, "learning_rate": 1.704615982949876e-05, "loss": 0.4906, "step": 21968 }, { "epoch": 1.4744136102815342, "grad_norm": 1.1139458417892456, "learning_rate": 1.703798681956687e-05, "loss": 0.4987, "step": 21970 }, { "epoch": 1.4745478339653033, "grad_norm": 0.9543841481208801, "learning_rate": 1.7029815367025304e-05, "loss": 0.4754, "step": 21972 }, { "epoch": 1.4746820576490722, "grad_norm": 1.0988030433654785, "learning_rate": 1.7021645472260146e-05, "loss": 0.5489, "step": 21974 }, { "epoch": 1.4748162813328412, "grad_norm": 0.9184737801551819, "learning_rate": 1.701347713565735e-05, "loss": 0.5104, "step": 21976 }, { "epoch": 1.47495050501661, "grad_norm": 1.095395565032959, "learning_rate": 1.7005310357602916e-05, "loss": 0.4972, "step": 21978 }, { "epoch": 1.4750847287003792, "grad_norm": 1.0471221208572388, "learning_rate": 1.6997145138482674e-05, "loss": 0.479, "step": 21980 }, { "epoch": 1.4752189523841481, "grad_norm": 1.0793814659118652, "learning_rate": 1.6988981478682436e-05, "loss": 0.5019, "step": 21982 }, { "epoch": 1.4753531760679173, "grad_norm": 0.9716320633888245, "learning_rate": 1.6980819378587915e-05, "loss": 0.4508, "step": 21984 }, { "epoch": 1.4754873997516862, "grad_norm": 1.0645676851272583, "learning_rate": 1.697265883858475e-05, "loss": 0.5126, "step": 21986 }, { "epoch": 1.475621623435455, "grad_norm": 1.1257740259170532, "learning_rate": 1.6964499859058485e-05, "loss": 0.525, "step": 21988 }, { "epoch": 1.4757558471192243, "grad_norm": 1.1400830745697021, "learning_rate": 1.6956342440394663e-05, "loss": 0.4899, "step": 21990 }, { "epoch": 1.4758900708029932, "grad_norm": 1.0845667123794556, "learning_rate": 1.6948186582978683e-05, "loss": 0.5192, "step": 21992 }, { "epoch": 1.4760242944867623, "grad_norm": 0.9828360080718994, "learning_rate": 1.694003228719589e-05, "loss": 0.4691, "step": 21994 }, { "epoch": 1.4761585181705312, "grad_norm": 1.1349446773529053, "learning_rate": 1.693187955343155e-05, "loss": 0.5609, "step": 21996 }, { "epoch": 1.4762927418543002, "grad_norm": 0.8819553852081299, "learning_rate": 1.6923728382070886e-05, "loss": 0.4321, "step": 21998 }, { "epoch": 1.476426965538069, "grad_norm": 0.9215646386146545, "learning_rate": 1.6915578773499003e-05, "loss": 0.5229, "step": 22000 }, { "epoch": 1.4765611892218382, "grad_norm": 1.0390326976776123, "learning_rate": 1.6907430728101e-05, "loss": 0.4688, "step": 22002 }, { "epoch": 1.4766954129056071, "grad_norm": 0.9488692879676819, "learning_rate": 1.689928424626178e-05, "loss": 0.5094, "step": 22004 }, { "epoch": 1.4768296365893763, "grad_norm": 1.4182480573654175, "learning_rate": 1.6891139328366313e-05, "loss": 0.5009, "step": 22006 }, { "epoch": 1.4769638602731452, "grad_norm": 1.041573405265808, "learning_rate": 1.6882995974799387e-05, "loss": 0.474, "step": 22008 }, { "epoch": 1.477098083956914, "grad_norm": 1.0602591037750244, "learning_rate": 1.68748541859458e-05, "loss": 0.496, "step": 22010 }, { "epoch": 1.4772323076406833, "grad_norm": 0.9753900170326233, "learning_rate": 1.686671396219021e-05, "loss": 0.5225, "step": 22012 }, { "epoch": 1.4773665313244522, "grad_norm": 1.1660841703414917, "learning_rate": 1.6858575303917234e-05, "loss": 0.4788, "step": 22014 }, { "epoch": 1.4775007550082213, "grad_norm": 0.9335696697235107, "learning_rate": 1.685043821151139e-05, "loss": 0.4973, "step": 22016 }, { "epoch": 1.4776349786919902, "grad_norm": 1.10598886013031, "learning_rate": 1.6842302685357165e-05, "loss": 0.4726, "step": 22018 }, { "epoch": 1.4777692023757591, "grad_norm": 0.9315410256385803, "learning_rate": 1.683416872583894e-05, "loss": 0.4898, "step": 22020 }, { "epoch": 1.477903426059528, "grad_norm": 1.0776035785675049, "learning_rate": 1.6826036333341027e-05, "loss": 0.5003, "step": 22022 }, { "epoch": 1.4780376497432972, "grad_norm": 0.8986199498176575, "learning_rate": 1.6817905508247643e-05, "loss": 0.5129, "step": 22024 }, { "epoch": 1.4781718734270661, "grad_norm": 1.0102131366729736, "learning_rate": 1.6809776250942995e-05, "loss": 0.4638, "step": 22026 }, { "epoch": 1.4783060971108353, "grad_norm": 1.0412551164627075, "learning_rate": 1.6801648561811133e-05, "loss": 0.4676, "step": 22028 }, { "epoch": 1.4784403207946042, "grad_norm": 0.9449609518051147, "learning_rate": 1.679352244123613e-05, "loss": 0.5366, "step": 22030 }, { "epoch": 1.478574544478373, "grad_norm": 1.1150097846984863, "learning_rate": 1.678539788960186e-05, "loss": 0.4236, "step": 22032 }, { "epoch": 1.4787087681621423, "grad_norm": 1.0405479669570923, "learning_rate": 1.6777274907292245e-05, "loss": 0.5033, "step": 22034 }, { "epoch": 1.4788429918459112, "grad_norm": 1.066287875175476, "learning_rate": 1.6769153494691043e-05, "loss": 0.5122, "step": 22036 }, { "epoch": 1.4789772155296803, "grad_norm": 1.1990296840667725, "learning_rate": 1.6761033652182008e-05, "loss": 0.5075, "step": 22038 }, { "epoch": 1.4791114392134492, "grad_norm": 1.2186036109924316, "learning_rate": 1.675291538014877e-05, "loss": 0.4539, "step": 22040 }, { "epoch": 1.4792456628972181, "grad_norm": 1.1169241666793823, "learning_rate": 1.6744798678974906e-05, "loss": 0.4672, "step": 22042 }, { "epoch": 1.4793798865809873, "grad_norm": 1.2180031538009644, "learning_rate": 1.673668354904391e-05, "loss": 0.4853, "step": 22044 }, { "epoch": 1.4795141102647562, "grad_norm": 1.022362232208252, "learning_rate": 1.672856999073919e-05, "loss": 0.4604, "step": 22046 }, { "epoch": 1.4796483339485254, "grad_norm": 2.732006311416626, "learning_rate": 1.672045800444413e-05, "loss": 0.4882, "step": 22048 }, { "epoch": 1.4797825576322943, "grad_norm": 1.036034345626831, "learning_rate": 1.6712347590541983e-05, "loss": 0.5039, "step": 22050 }, { "epoch": 1.4799167813160632, "grad_norm": 0.9300113916397095, "learning_rate": 1.6704238749415957e-05, "loss": 0.4571, "step": 22052 }, { "epoch": 1.480051004999832, "grad_norm": 1.0871950387954712, "learning_rate": 1.6696131481449162e-05, "loss": 0.5488, "step": 22054 }, { "epoch": 1.4801852286836013, "grad_norm": 1.0557066202163696, "learning_rate": 1.668802578702468e-05, "loss": 0.477, "step": 22056 }, { "epoch": 1.4803194523673702, "grad_norm": 1.0013145208358765, "learning_rate": 1.6679921666525476e-05, "loss": 0.4437, "step": 22058 }, { "epoch": 1.4804536760511393, "grad_norm": 0.9436190128326416, "learning_rate": 1.6671819120334454e-05, "loss": 0.4432, "step": 22060 }, { "epoch": 1.4805878997349082, "grad_norm": 1.2498823404312134, "learning_rate": 1.6663718148834423e-05, "loss": 0.5072, "step": 22062 }, { "epoch": 1.4807221234186771, "grad_norm": 1.0579122304916382, "learning_rate": 1.6655618752408176e-05, "loss": 0.408, "step": 22064 }, { "epoch": 1.4808563471024463, "grad_norm": 1.0340778827667236, "learning_rate": 1.6647520931438355e-05, "loss": 0.5089, "step": 22066 }, { "epoch": 1.4809905707862152, "grad_norm": 1.081233263015747, "learning_rate": 1.6639424686307613e-05, "loss": 0.5374, "step": 22068 }, { "epoch": 1.4811247944699844, "grad_norm": 0.8616448640823364, "learning_rate": 1.663133001739843e-05, "loss": 0.4137, "step": 22070 }, { "epoch": 1.4812590181537533, "grad_norm": 0.9921708106994629, "learning_rate": 1.6623236925093293e-05, "loss": 0.5115, "step": 22072 }, { "epoch": 1.4813932418375222, "grad_norm": 0.9645135402679443, "learning_rate": 1.661514540977457e-05, "loss": 0.4393, "step": 22074 }, { "epoch": 1.481527465521291, "grad_norm": 1.0653139352798462, "learning_rate": 1.6607055471824595e-05, "loss": 0.4785, "step": 22076 }, { "epoch": 1.4816616892050603, "grad_norm": 0.9894845485687256, "learning_rate": 1.659896711162558e-05, "loss": 0.4881, "step": 22078 }, { "epoch": 1.4817959128888292, "grad_norm": 0.9394354820251465, "learning_rate": 1.6590880329559692e-05, "loss": 0.531, "step": 22080 }, { "epoch": 1.4819301365725983, "grad_norm": 0.9779031872749329, "learning_rate": 1.658279512600899e-05, "loss": 0.4629, "step": 22082 }, { "epoch": 1.4820643602563672, "grad_norm": 1.0840893983840942, "learning_rate": 1.6574711501355534e-05, "loss": 0.4881, "step": 22084 }, { "epoch": 1.4821985839401361, "grad_norm": 1.404558777809143, "learning_rate": 1.6566629455981218e-05, "loss": 0.4181, "step": 22086 }, { "epoch": 1.4823328076239053, "grad_norm": 0.9741535186767578, "learning_rate": 1.6558548990267925e-05, "loss": 0.4326, "step": 22088 }, { "epoch": 1.4824670313076742, "grad_norm": 0.9967256784439087, "learning_rate": 1.6550470104597417e-05, "loss": 0.4607, "step": 22090 }, { "epoch": 1.4826012549914434, "grad_norm": 1.3023165464401245, "learning_rate": 1.654239279935143e-05, "loss": 0.4302, "step": 22092 }, { "epoch": 1.4827354786752123, "grad_norm": 0.9432088732719421, "learning_rate": 1.6534317074911582e-05, "loss": 0.4792, "step": 22094 }, { "epoch": 1.4828697023589812, "grad_norm": 1.0441328287124634, "learning_rate": 1.652624293165946e-05, "loss": 0.4332, "step": 22096 }, { "epoch": 1.48300392604275, "grad_norm": 1.0635572671890259, "learning_rate": 1.6518170369976532e-05, "loss": 0.466, "step": 22098 }, { "epoch": 1.4831381497265193, "grad_norm": 0.9838083386421204, "learning_rate": 1.651009939024422e-05, "loss": 0.4276, "step": 22100 }, { "epoch": 1.4832723734102882, "grad_norm": 1.1172091960906982, "learning_rate": 1.6502029992843836e-05, "loss": 0.4367, "step": 22102 }, { "epoch": 1.4834065970940573, "grad_norm": 1.242810606956482, "learning_rate": 1.6493962178156685e-05, "loss": 0.4783, "step": 22104 }, { "epoch": 1.4835408207778262, "grad_norm": 1.1038495302200317, "learning_rate": 1.6485895946563927e-05, "loss": 0.4885, "step": 22106 }, { "epoch": 1.4836750444615951, "grad_norm": 0.980767548084259, "learning_rate": 1.647783129844669e-05, "loss": 0.5184, "step": 22108 }, { "epoch": 1.4838092681453643, "grad_norm": 0.9729624390602112, "learning_rate": 1.646976823418599e-05, "loss": 0.5311, "step": 22110 }, { "epoch": 1.4839434918291332, "grad_norm": 1.0298727750778198, "learning_rate": 1.646170675416282e-05, "loss": 0.4504, "step": 22112 }, { "epoch": 1.4840777155129024, "grad_norm": 1.0202456712722778, "learning_rate": 1.6453646858758055e-05, "loss": 0.5316, "step": 22114 }, { "epoch": 1.4842119391966713, "grad_norm": 1.0979762077331543, "learning_rate": 1.644558854835251e-05, "loss": 0.5239, "step": 22116 }, { "epoch": 1.4843461628804402, "grad_norm": 0.9957572221755981, "learning_rate": 1.6437531823326922e-05, "loss": 0.4982, "step": 22118 }, { "epoch": 1.4844803865642093, "grad_norm": 0.9726836085319519, "learning_rate": 1.6429476684061945e-05, "loss": 0.4833, "step": 22120 }, { "epoch": 1.4846146102479783, "grad_norm": 1.643353819847107, "learning_rate": 1.6421423130938196e-05, "loss": 0.4703, "step": 22122 }, { "epoch": 1.4847488339317474, "grad_norm": 0.9165441989898682, "learning_rate": 1.6413371164336154e-05, "loss": 0.4287, "step": 22124 }, { "epoch": 1.4848830576155163, "grad_norm": 0.9640105366706848, "learning_rate": 1.6405320784636318e-05, "loss": 0.4866, "step": 22126 }, { "epoch": 1.4850172812992852, "grad_norm": 0.8589420318603516, "learning_rate": 1.6397271992218982e-05, "loss": 0.4146, "step": 22128 }, { "epoch": 1.4851515049830541, "grad_norm": 1.0952810049057007, "learning_rate": 1.638922478746448e-05, "loss": 0.5149, "step": 22130 }, { "epoch": 1.4852857286668233, "grad_norm": 0.9134923815727234, "learning_rate": 1.6381179170752997e-05, "loss": 0.5017, "step": 22132 }, { "epoch": 1.4854199523505922, "grad_norm": 0.9884053468704224, "learning_rate": 1.6373135142464707e-05, "loss": 0.4738, "step": 22134 }, { "epoch": 1.4855541760343614, "grad_norm": 1.1389747858047485, "learning_rate": 1.636509270297966e-05, "loss": 0.4923, "step": 22136 }, { "epoch": 1.4856883997181303, "grad_norm": 1.044872522354126, "learning_rate": 1.635705185267784e-05, "loss": 0.4902, "step": 22138 }, { "epoch": 1.4858226234018992, "grad_norm": 1.0619035959243774, "learning_rate": 1.634901259193915e-05, "loss": 0.5227, "step": 22140 }, { "epoch": 1.4859568470856683, "grad_norm": 0.9813964366912842, "learning_rate": 1.6340974921143455e-05, "loss": 0.4091, "step": 22142 }, { "epoch": 1.4860910707694373, "grad_norm": 1.0402342081069946, "learning_rate": 1.6332938840670508e-05, "loss": 0.4614, "step": 22144 }, { "epoch": 1.4862252944532064, "grad_norm": 0.9421043395996094, "learning_rate": 1.63249043509e-05, "loss": 0.5051, "step": 22146 }, { "epoch": 1.4863595181369753, "grad_norm": 1.063668131828308, "learning_rate": 1.6316871452211524e-05, "loss": 0.4944, "step": 22148 }, { "epoch": 1.4864937418207442, "grad_norm": 1.1434531211853027, "learning_rate": 1.6308840144984656e-05, "loss": 0.5422, "step": 22150 }, { "epoch": 1.4866279655045131, "grad_norm": 0.9865716099739075, "learning_rate": 1.6300810429598823e-05, "loss": 0.497, "step": 22152 }, { "epoch": 1.4867621891882823, "grad_norm": 1.030135989189148, "learning_rate": 1.6292782306433462e-05, "loss": 0.4521, "step": 22154 }, { "epoch": 1.4868964128720512, "grad_norm": 1.0379871129989624, "learning_rate": 1.628475577586782e-05, "loss": 0.4689, "step": 22156 }, { "epoch": 1.4870306365558204, "grad_norm": 1.1710013151168823, "learning_rate": 1.627673083828119e-05, "loss": 0.499, "step": 22158 }, { "epoch": 1.4871648602395893, "grad_norm": 0.8846861124038696, "learning_rate": 1.62687074940527e-05, "loss": 0.4483, "step": 22160 }, { "epoch": 1.4872990839233582, "grad_norm": 0.9259506464004517, "learning_rate": 1.626068574356146e-05, "loss": 0.4439, "step": 22162 }, { "epoch": 1.4874333076071273, "grad_norm": 1.6477583646774292, "learning_rate": 1.6252665587186477e-05, "loss": 0.4561, "step": 22164 }, { "epoch": 1.4875675312908962, "grad_norm": 0.9675223231315613, "learning_rate": 1.624464702530668e-05, "loss": 0.4391, "step": 22166 }, { "epoch": 1.4877017549746654, "grad_norm": 0.9893090724945068, "learning_rate": 1.6236630058300924e-05, "loss": 0.4569, "step": 22168 }, { "epoch": 1.4878359786584343, "grad_norm": 0.9624961614608765, "learning_rate": 1.6228614686548018e-05, "loss": 0.474, "step": 22170 }, { "epoch": 1.4879702023422032, "grad_norm": 0.9343721270561218, "learning_rate": 1.622060091042666e-05, "loss": 0.481, "step": 22172 }, { "epoch": 1.4881044260259721, "grad_norm": 1.0352492332458496, "learning_rate": 1.621258873031548e-05, "loss": 0.5404, "step": 22174 }, { "epoch": 1.4882386497097413, "grad_norm": 1.141648769378662, "learning_rate": 1.620457814659303e-05, "loss": 0.4337, "step": 22176 }, { "epoch": 1.4883728733935102, "grad_norm": 0.9643620848655701, "learning_rate": 1.6196569159637825e-05, "loss": 0.529, "step": 22178 }, { "epoch": 1.4885070970772794, "grad_norm": 1.030743956565857, "learning_rate": 1.6188561769828252e-05, "loss": 0.422, "step": 22180 }, { "epoch": 1.4886413207610483, "grad_norm": 1.0663295984268188, "learning_rate": 1.6180555977542655e-05, "loss": 0.5016, "step": 22182 }, { "epoch": 1.4887755444448172, "grad_norm": 0.8622217774391174, "learning_rate": 1.6172551783159278e-05, "loss": 0.4585, "step": 22184 }, { "epoch": 1.4889097681285863, "grad_norm": 1.099914312362671, "learning_rate": 1.6164549187056294e-05, "loss": 0.5106, "step": 22186 }, { "epoch": 1.4890439918123552, "grad_norm": 0.9957973957061768, "learning_rate": 1.6156548189611847e-05, "loss": 0.4241, "step": 22188 }, { "epoch": 1.4891782154961244, "grad_norm": 0.8730512857437134, "learning_rate": 1.6148548791203926e-05, "loss": 0.4678, "step": 22190 }, { "epoch": 1.4893124391798933, "grad_norm": 1.048096776008606, "learning_rate": 1.6140550992210545e-05, "loss": 0.5086, "step": 22192 }, { "epoch": 1.4894466628636622, "grad_norm": 0.9279083609580994, "learning_rate": 1.6132554793009514e-05, "loss": 0.4457, "step": 22194 }, { "epoch": 1.4895808865474314, "grad_norm": 1.15582275390625, "learning_rate": 1.612456019397869e-05, "loss": 0.5107, "step": 22196 }, { "epoch": 1.4897151102312003, "grad_norm": 1.0184177160263062, "learning_rate": 1.6116567195495767e-05, "loss": 0.5176, "step": 22198 }, { "epoch": 1.4898493339149694, "grad_norm": 1.0488437414169312, "learning_rate": 1.610857579793843e-05, "loss": 0.4874, "step": 22200 }, { "epoch": 1.4899835575987384, "grad_norm": 1.0049000978469849, "learning_rate": 1.610058600168424e-05, "loss": 0.4562, "step": 22202 }, { "epoch": 1.4901177812825073, "grad_norm": 0.9642466306686401, "learning_rate": 1.6092597807110703e-05, "loss": 0.4392, "step": 22204 }, { "epoch": 1.4902520049662762, "grad_norm": 1.002695918083191, "learning_rate": 1.608461121459523e-05, "loss": 0.4841, "step": 22206 }, { "epoch": 1.4903862286500453, "grad_norm": 1.1459479331970215, "learning_rate": 1.6076626224515196e-05, "loss": 0.4775, "step": 22208 }, { "epoch": 1.4905204523338142, "grad_norm": 1.008082628250122, "learning_rate": 1.6068642837247872e-05, "loss": 0.456, "step": 22210 }, { "epoch": 1.4906546760175834, "grad_norm": 0.9226283431053162, "learning_rate": 1.6060661053170444e-05, "loss": 0.4293, "step": 22212 }, { "epoch": 1.4907888997013523, "grad_norm": 1.091553807258606, "learning_rate": 1.6052680872660032e-05, "loss": 0.4298, "step": 22214 }, { "epoch": 1.4909231233851212, "grad_norm": 1.1263148784637451, "learning_rate": 1.6044702296093707e-05, "loss": 0.5747, "step": 22216 }, { "epoch": 1.4910573470688904, "grad_norm": 0.9613867998123169, "learning_rate": 1.603672532384841e-05, "loss": 0.4229, "step": 22218 }, { "epoch": 1.4911915707526593, "grad_norm": 1.1696033477783203, "learning_rate": 1.6028749956301094e-05, "loss": 0.4472, "step": 22220 }, { "epoch": 1.4913257944364284, "grad_norm": 0.8974775671958923, "learning_rate": 1.6020776193828506e-05, "loss": 0.4678, "step": 22222 }, { "epoch": 1.4914600181201974, "grad_norm": 1.1368629932403564, "learning_rate": 1.601280403680744e-05, "loss": 0.5267, "step": 22224 }, { "epoch": 1.4915942418039663, "grad_norm": 0.9700947999954224, "learning_rate": 1.6004833485614536e-05, "loss": 0.4524, "step": 22226 }, { "epoch": 1.4917284654877352, "grad_norm": 1.0958696603775024, "learning_rate": 1.5996864540626416e-05, "loss": 0.5448, "step": 22228 }, { "epoch": 1.4918626891715043, "grad_norm": 0.923024594783783, "learning_rate": 1.5988897202219582e-05, "loss": 0.4223, "step": 22230 }, { "epoch": 1.4919969128552732, "grad_norm": 0.9782229065895081, "learning_rate": 1.5980931470770476e-05, "loss": 0.4957, "step": 22232 }, { "epoch": 1.4921311365390424, "grad_norm": 1.0762877464294434, "learning_rate": 1.5972967346655448e-05, "loss": 0.483, "step": 22234 }, { "epoch": 1.4922653602228113, "grad_norm": 1.0898056030273438, "learning_rate": 1.5965004830250814e-05, "loss": 0.5673, "step": 22236 }, { "epoch": 1.4923995839065802, "grad_norm": 0.9333029985427856, "learning_rate": 1.595704392193278e-05, "loss": 0.5022, "step": 22238 }, { "epoch": 1.4925338075903494, "grad_norm": 1.0515496730804443, "learning_rate": 1.5949084622077472e-05, "loss": 0.5338, "step": 22240 }, { "epoch": 1.4926680312741183, "grad_norm": 0.9734745025634766, "learning_rate": 1.5941126931060947e-05, "loss": 0.482, "step": 22242 }, { "epoch": 1.4928022549578874, "grad_norm": 0.9645838737487793, "learning_rate": 1.593317084925921e-05, "loss": 0.4808, "step": 22244 }, { "epoch": 1.4929364786416564, "grad_norm": 1.1264073848724365, "learning_rate": 1.5925216377048153e-05, "loss": 0.4864, "step": 22246 }, { "epoch": 1.4930707023254253, "grad_norm": 0.9437454342842102, "learning_rate": 1.5917263514803643e-05, "loss": 0.4782, "step": 22248 }, { "epoch": 1.4932049260091942, "grad_norm": 0.9792948365211487, "learning_rate": 1.5909312262901395e-05, "loss": 0.469, "step": 22250 }, { "epoch": 1.4933391496929633, "grad_norm": 1.0361741781234741, "learning_rate": 1.590136262171709e-05, "loss": 0.5108, "step": 22252 }, { "epoch": 1.4934733733767322, "grad_norm": 1.7859095335006714, "learning_rate": 1.5893414591626372e-05, "loss": 0.4976, "step": 22254 }, { "epoch": 1.4936075970605014, "grad_norm": 1.0242053270339966, "learning_rate": 1.5885468173004726e-05, "loss": 0.4742, "step": 22256 }, { "epoch": 1.4937418207442703, "grad_norm": 1.0312488079071045, "learning_rate": 1.5877523366227635e-05, "loss": 0.5077, "step": 22258 }, { "epoch": 1.4938760444280392, "grad_norm": 1.1025859117507935, "learning_rate": 1.586958017167047e-05, "loss": 0.4627, "step": 22260 }, { "epoch": 1.4940102681118084, "grad_norm": 1.0530234575271606, "learning_rate": 1.5861638589708534e-05, "loss": 0.4536, "step": 22262 }, { "epoch": 1.4941444917955773, "grad_norm": 1.0849812030792236, "learning_rate": 1.585369862071702e-05, "loss": 0.501, "step": 22264 }, { "epoch": 1.4942787154793464, "grad_norm": 0.8678374886512756, "learning_rate": 1.5845760265071125e-05, "loss": 0.4688, "step": 22266 }, { "epoch": 1.4944129391631154, "grad_norm": 0.9843668937683105, "learning_rate": 1.583782352314589e-05, "loss": 0.4795, "step": 22268 }, { "epoch": 1.4945471628468843, "grad_norm": 1.0516777038574219, "learning_rate": 1.5829888395316317e-05, "loss": 0.4421, "step": 22270 }, { "epoch": 1.4946813865306534, "grad_norm": 1.0073456764221191, "learning_rate": 1.582195488195731e-05, "loss": 0.4981, "step": 22272 }, { "epoch": 1.4948156102144223, "grad_norm": 0.9583938121795654, "learning_rate": 1.5814022983443743e-05, "loss": 0.4775, "step": 22274 }, { "epoch": 1.4949498338981915, "grad_norm": 1.054783582687378, "learning_rate": 1.580609270015035e-05, "loss": 0.5537, "step": 22276 }, { "epoch": 1.4950840575819604, "grad_norm": 0.8921435475349426, "learning_rate": 1.5798164032451867e-05, "loss": 0.4924, "step": 22278 }, { "epoch": 1.4952182812657293, "grad_norm": 1.042488694190979, "learning_rate": 1.579023698072285e-05, "loss": 0.4765, "step": 22280 }, { "epoch": 1.4953525049494982, "grad_norm": 1.041841983795166, "learning_rate": 1.578231154533788e-05, "loss": 0.4815, "step": 22282 }, { "epoch": 1.4954867286332674, "grad_norm": 1.0122069120407104, "learning_rate": 1.5774387726671385e-05, "loss": 0.4836, "step": 22284 }, { "epoch": 1.4956209523170363, "grad_norm": 1.0777119398117065, "learning_rate": 1.5766465525097784e-05, "loss": 0.4443, "step": 22286 }, { "epoch": 1.4957551760008054, "grad_norm": 0.9867185354232788, "learning_rate": 1.575854494099137e-05, "loss": 0.4634, "step": 22288 }, { "epoch": 1.4958893996845744, "grad_norm": 1.0630228519439697, "learning_rate": 1.575062597472638e-05, "loss": 0.5764, "step": 22290 }, { "epoch": 1.4960236233683433, "grad_norm": 1.1396502256393433, "learning_rate": 1.5742708626676943e-05, "loss": 0.4397, "step": 22292 }, { "epoch": 1.4961578470521124, "grad_norm": 0.9992977976799011, "learning_rate": 1.5734792897217178e-05, "loss": 0.5039, "step": 22294 }, { "epoch": 1.4962920707358813, "grad_norm": 1.0234202146530151, "learning_rate": 1.5726878786721067e-05, "loss": 0.4732, "step": 22296 }, { "epoch": 1.4964262944196505, "grad_norm": 1.0589426755905151, "learning_rate": 1.5718966295562538e-05, "loss": 0.4807, "step": 22298 }, { "epoch": 1.4965605181034194, "grad_norm": 0.9448699951171875, "learning_rate": 1.5711055424115424e-05, "loss": 0.4663, "step": 22300 }, { "epoch": 1.4966947417871883, "grad_norm": 1.0552382469177246, "learning_rate": 1.570314617275353e-05, "loss": 0.4461, "step": 22302 }, { "epoch": 1.4968289654709572, "grad_norm": 1.1160061359405518, "learning_rate": 1.5695238541850526e-05, "loss": 0.5051, "step": 22304 }, { "epoch": 1.4969631891547264, "grad_norm": 0.9328937530517578, "learning_rate": 1.568733253178007e-05, "loss": 0.4966, "step": 22306 }, { "epoch": 1.4970974128384953, "grad_norm": 0.9663674831390381, "learning_rate": 1.5679428142915652e-05, "loss": 0.4867, "step": 22308 }, { "epoch": 1.4972316365222644, "grad_norm": 1.0076817274093628, "learning_rate": 1.567152537563078e-05, "loss": 0.527, "step": 22310 }, { "epoch": 1.4973658602060334, "grad_norm": 0.8996386528015137, "learning_rate": 1.566362423029881e-05, "loss": 0.4739, "step": 22312 }, { "epoch": 1.4975000838898023, "grad_norm": 1.0612578392028809, "learning_rate": 1.5655724707293096e-05, "loss": 0.4517, "step": 22314 }, { "epoch": 1.4976343075735714, "grad_norm": 0.9685842990875244, "learning_rate": 1.5647826806986853e-05, "loss": 0.4856, "step": 22316 }, { "epoch": 1.4977685312573403, "grad_norm": 1.0554558038711548, "learning_rate": 1.5639930529753244e-05, "loss": 0.4535, "step": 22318 }, { "epoch": 1.4979027549411095, "grad_norm": 1.0142149925231934, "learning_rate": 1.563203587596535e-05, "loss": 0.4464, "step": 22320 }, { "epoch": 1.4980369786248784, "grad_norm": 1.0849847793579102, "learning_rate": 1.5624142845996165e-05, "loss": 0.497, "step": 22322 }, { "epoch": 1.4981712023086473, "grad_norm": 1.0530681610107422, "learning_rate": 1.561625144021865e-05, "loss": 0.5238, "step": 22324 }, { "epoch": 1.4983054259924162, "grad_norm": 1.2183722257614136, "learning_rate": 1.5608361659005637e-05, "loss": 0.4877, "step": 22326 }, { "epoch": 1.4984396496761854, "grad_norm": 1.033240795135498, "learning_rate": 1.560047350272991e-05, "loss": 0.511, "step": 22328 }, { "epoch": 1.4985738733599543, "grad_norm": 0.9958361983299255, "learning_rate": 1.559258697176415e-05, "loss": 0.4966, "step": 22330 }, { "epoch": 1.4987080970437234, "grad_norm": 0.8929667472839355, "learning_rate": 1.5584702066481016e-05, "loss": 0.3835, "step": 22332 }, { "epoch": 1.4988423207274923, "grad_norm": 1.1883955001831055, "learning_rate": 1.5576818787253027e-05, "loss": 0.5171, "step": 22334 }, { "epoch": 1.4989765444112613, "grad_norm": 1.052208423614502, "learning_rate": 1.5568937134452664e-05, "loss": 0.4719, "step": 22336 }, { "epoch": 1.4991107680950304, "grad_norm": 0.8502485752105713, "learning_rate": 1.5561057108452304e-05, "loss": 0.4843, "step": 22338 }, { "epoch": 1.4992449917787993, "grad_norm": 0.9813560247421265, "learning_rate": 1.5553178709624283e-05, "loss": 0.4513, "step": 22340 }, { "epoch": 1.4993792154625685, "grad_norm": 1.0747052431106567, "learning_rate": 1.5545301938340817e-05, "loss": 0.5191, "step": 22342 }, { "epoch": 1.4995134391463374, "grad_norm": 1.043516993522644, "learning_rate": 1.553742679497412e-05, "loss": 0.542, "step": 22344 }, { "epoch": 1.4996476628301063, "grad_norm": 1.3962687253952026, "learning_rate": 1.5529553279896202e-05, "loss": 0.4693, "step": 22346 }, { "epoch": 1.4997818865138755, "grad_norm": 1.1393555402755737, "learning_rate": 1.5521681393479126e-05, "loss": 0.5109, "step": 22348 }, { "epoch": 1.4999161101976444, "grad_norm": 1.1023056507110596, "learning_rate": 1.5513811136094787e-05, "loss": 0.4991, "step": 22350 }, { "epoch": 1.5000503338814135, "grad_norm": 1.081140160560608, "learning_rate": 1.5505942508115073e-05, "loss": 0.4659, "step": 22352 }, { "epoch": 1.5001845575651824, "grad_norm": 1.0859403610229492, "learning_rate": 1.5498075509911745e-05, "loss": 0.4666, "step": 22354 }, { "epoch": 1.5003187812489513, "grad_norm": 1.193873405456543, "learning_rate": 1.5490210141856508e-05, "loss": 0.5178, "step": 22356 }, { "epoch": 1.5004530049327203, "grad_norm": 1.0530229806900024, "learning_rate": 1.5482346404320967e-05, "loss": 0.4997, "step": 22358 }, { "epoch": 1.5005872286164894, "grad_norm": 1.0302584171295166, "learning_rate": 1.5474484297676694e-05, "loss": 0.4758, "step": 22360 }, { "epoch": 1.5007214523002583, "grad_norm": 1.1433876752853394, "learning_rate": 1.546662382229515e-05, "loss": 0.5066, "step": 22362 }, { "epoch": 1.5008556759840275, "grad_norm": 1.156192421913147, "learning_rate": 1.5458764978547718e-05, "loss": 0.5427, "step": 22364 }, { "epoch": 1.5009898996677964, "grad_norm": 1.0290182828903198, "learning_rate": 1.545090776680571e-05, "loss": 0.4451, "step": 22366 }, { "epoch": 1.5011241233515653, "grad_norm": 0.9857996702194214, "learning_rate": 1.544305218744038e-05, "loss": 0.4972, "step": 22368 }, { "epoch": 1.5012583470353342, "grad_norm": 1.1398378610610962, "learning_rate": 1.5435198240822873e-05, "loss": 0.4968, "step": 22370 }, { "epoch": 1.5013925707191034, "grad_norm": 0.9778451323509216, "learning_rate": 1.5427345927324305e-05, "loss": 0.4606, "step": 22372 }, { "epoch": 1.5015267944028725, "grad_norm": 1.0325839519500732, "learning_rate": 1.541949524731563e-05, "loss": 0.5229, "step": 22374 }, { "epoch": 1.5016610180866414, "grad_norm": 1.1207289695739746, "learning_rate": 1.5411646201167817e-05, "loss": 0.5016, "step": 22376 }, { "epoch": 1.5017952417704103, "grad_norm": 1.1190552711486816, "learning_rate": 1.5403798789251695e-05, "loss": 0.5034, "step": 22378 }, { "epoch": 1.5019294654541793, "grad_norm": 0.903392493724823, "learning_rate": 1.5395953011938063e-05, "loss": 0.4447, "step": 22380 }, { "epoch": 1.5020636891379484, "grad_norm": 0.9386855959892273, "learning_rate": 1.5388108869597605e-05, "loss": 0.4992, "step": 22382 }, { "epoch": 1.5021979128217176, "grad_norm": 0.9510099291801453, "learning_rate": 1.5380266362600943e-05, "loss": 0.4438, "step": 22384 }, { "epoch": 1.5023321365054865, "grad_norm": 1.1823641061782837, "learning_rate": 1.5372425491318615e-05, "loss": 0.5579, "step": 22386 }, { "epoch": 1.5024663601892554, "grad_norm": 1.0836889743804932, "learning_rate": 1.5364586256121076e-05, "loss": 0.5182, "step": 22388 }, { "epoch": 1.5026005838730243, "grad_norm": 0.9870573878288269, "learning_rate": 1.535674865737875e-05, "loss": 0.4395, "step": 22390 }, { "epoch": 1.5027348075567935, "grad_norm": 1.028084397315979, "learning_rate": 1.534891269546192e-05, "loss": 0.519, "step": 22392 }, { "epoch": 1.5028690312405624, "grad_norm": 0.9285738468170166, "learning_rate": 1.534107837074083e-05, "loss": 0.4851, "step": 22394 }, { "epoch": 1.5030032549243315, "grad_norm": 0.9531024098396301, "learning_rate": 1.5333245683585618e-05, "loss": 0.4493, "step": 22396 }, { "epoch": 1.5031374786081004, "grad_norm": 0.9427856206893921, "learning_rate": 1.5325414634366392e-05, "loss": 0.4815, "step": 22398 }, { "epoch": 1.5032717022918693, "grad_norm": 1.107208251953125, "learning_rate": 1.531758522345314e-05, "loss": 0.496, "step": 22400 }, { "epoch": 1.5034059259756383, "grad_norm": 0.9453824162483215, "learning_rate": 1.530975745121579e-05, "loss": 0.5032, "step": 22402 }, { "epoch": 1.5035401496594074, "grad_norm": 1.037557601928711, "learning_rate": 1.5301931318024166e-05, "loss": 0.5256, "step": 22404 }, { "epoch": 1.5036743733431766, "grad_norm": 0.9742701649665833, "learning_rate": 1.5294106824248065e-05, "loss": 0.4578, "step": 22406 }, { "epoch": 1.5038085970269455, "grad_norm": 1.0854389667510986, "learning_rate": 1.5286283970257166e-05, "loss": 0.4644, "step": 22408 }, { "epoch": 1.5039428207107144, "grad_norm": 1.0835214853286743, "learning_rate": 1.52784627564211e-05, "loss": 0.4463, "step": 22410 }, { "epoch": 1.5040770443944833, "grad_norm": 1.093213677406311, "learning_rate": 1.527064318310939e-05, "loss": 0.5273, "step": 22412 }, { "epoch": 1.5042112680782525, "grad_norm": 1.007411003112793, "learning_rate": 1.5262825250691497e-05, "loss": 0.4554, "step": 22414 }, { "epoch": 1.5043454917620214, "grad_norm": 1.095778465270996, "learning_rate": 1.5255008959536787e-05, "loss": 0.4905, "step": 22416 }, { "epoch": 1.5044797154457905, "grad_norm": 1.0032175779342651, "learning_rate": 1.5247194310014601e-05, "loss": 0.4516, "step": 22418 }, { "epoch": 1.5046139391295594, "grad_norm": 1.103121042251587, "learning_rate": 1.5239381302494144e-05, "loss": 0.4593, "step": 22420 }, { "epoch": 1.5047481628133283, "grad_norm": 1.0171631574630737, "learning_rate": 1.5231569937344564e-05, "loss": 0.5937, "step": 22422 }, { "epoch": 1.5048823864970973, "grad_norm": 1.1355721950531006, "learning_rate": 1.5223760214934917e-05, "loss": 0.5546, "step": 22424 }, { "epoch": 1.5050166101808664, "grad_norm": 0.8735546469688416, "learning_rate": 1.5215952135634237e-05, "loss": 0.436, "step": 22426 }, { "epoch": 1.5051508338646356, "grad_norm": 1.1134432554244995, "learning_rate": 1.5208145699811415e-05, "loss": 0.4899, "step": 22428 }, { "epoch": 1.5052850575484045, "grad_norm": 1.0774012804031372, "learning_rate": 1.5200340907835298e-05, "loss": 0.4783, "step": 22430 }, { "epoch": 1.5054192812321734, "grad_norm": 1.0947202444076538, "learning_rate": 1.519253776007462e-05, "loss": 0.462, "step": 22432 }, { "epoch": 1.5055535049159423, "grad_norm": 0.9434705972671509, "learning_rate": 1.5184736256898107e-05, "loss": 0.474, "step": 22434 }, { "epoch": 1.5056877285997115, "grad_norm": 1.147038221359253, "learning_rate": 1.517693639867433e-05, "loss": 0.5078, "step": 22436 }, { "epoch": 1.5058219522834804, "grad_norm": 1.1056617498397827, "learning_rate": 1.5169138185771841e-05, "loss": 0.4686, "step": 22438 }, { "epoch": 1.5059561759672495, "grad_norm": 0.9053165316581726, "learning_rate": 1.5161341618559088e-05, "loss": 0.4727, "step": 22440 }, { "epoch": 1.5060903996510184, "grad_norm": 1.0582969188690186, "learning_rate": 1.515354669740443e-05, "loss": 0.4678, "step": 22442 }, { "epoch": 1.5062246233347873, "grad_norm": 0.9727494120597839, "learning_rate": 1.5145753422676157e-05, "loss": 0.4288, "step": 22444 }, { "epoch": 1.5063588470185563, "grad_norm": 0.9582500457763672, "learning_rate": 1.5137961794742511e-05, "loss": 0.4534, "step": 22446 }, { "epoch": 1.5064930707023254, "grad_norm": 1.2127413749694824, "learning_rate": 1.5130171813971617e-05, "loss": 0.4995, "step": 22448 }, { "epoch": 1.5066272943860946, "grad_norm": 0.9376811385154724, "learning_rate": 1.5122383480731533e-05, "loss": 0.459, "step": 22450 }, { "epoch": 1.5067615180698635, "grad_norm": 1.0548505783081055, "learning_rate": 1.5114596795390234e-05, "loss": 0.4784, "step": 22452 }, { "epoch": 1.5068957417536324, "grad_norm": 1.0130548477172852, "learning_rate": 1.5106811758315654e-05, "loss": 0.5134, "step": 22454 }, { "epoch": 1.5070299654374013, "grad_norm": 1.0775259733200073, "learning_rate": 1.50990283698756e-05, "loss": 0.5642, "step": 22456 }, { "epoch": 1.5071641891211705, "grad_norm": 1.0384955406188965, "learning_rate": 1.5091246630437827e-05, "loss": 0.5506, "step": 22458 }, { "epoch": 1.5072984128049396, "grad_norm": 1.0623197555541992, "learning_rate": 1.5083466540370006e-05, "loss": 0.4947, "step": 22460 }, { "epoch": 1.5074326364887085, "grad_norm": 1.0416051149368286, "learning_rate": 1.5075688100039715e-05, "loss": 0.4979, "step": 22462 }, { "epoch": 1.5075668601724774, "grad_norm": 1.0377678871154785, "learning_rate": 1.5067911309814503e-05, "loss": 0.4405, "step": 22464 }, { "epoch": 1.5077010838562463, "grad_norm": 0.9336119890213013, "learning_rate": 1.5060136170061773e-05, "loss": 0.4473, "step": 22466 }, { "epoch": 1.5078353075400155, "grad_norm": 0.8464655876159668, "learning_rate": 1.5052362681148935e-05, "loss": 0.4769, "step": 22468 }, { "epoch": 1.5079695312237844, "grad_norm": 1.09187650680542, "learning_rate": 1.5044590843443207e-05, "loss": 0.4561, "step": 22470 }, { "epoch": 1.5081037549075536, "grad_norm": 0.9741097092628479, "learning_rate": 1.5036820657311839e-05, "loss": 0.4661, "step": 22472 }, { "epoch": 1.5082379785913225, "grad_norm": 0.9991913437843323, "learning_rate": 1.5029052123121928e-05, "loss": 0.5184, "step": 22474 }, { "epoch": 1.5083722022750914, "grad_norm": 0.9883608222007751, "learning_rate": 1.5021285241240552e-05, "loss": 0.4325, "step": 22476 }, { "epoch": 1.5085064259588603, "grad_norm": 0.9305347204208374, "learning_rate": 1.5013520012034665e-05, "loss": 0.508, "step": 22478 }, { "epoch": 1.5086406496426294, "grad_norm": 1.0459718704223633, "learning_rate": 1.5005756435871165e-05, "loss": 0.4844, "step": 22480 }, { "epoch": 1.5087748733263986, "grad_norm": 1.0773015022277832, "learning_rate": 1.4997994513116836e-05, "loss": 0.6178, "step": 22482 }, { "epoch": 1.5089090970101675, "grad_norm": 1.004028558731079, "learning_rate": 1.4990234244138457e-05, "loss": 0.4982, "step": 22484 }, { "epoch": 1.5090433206939364, "grad_norm": 0.8696219325065613, "learning_rate": 1.4982475629302667e-05, "loss": 0.4384, "step": 22486 }, { "epoch": 1.5091775443777053, "grad_norm": 0.9148178696632385, "learning_rate": 1.4974718668976046e-05, "loss": 0.464, "step": 22488 }, { "epoch": 1.5093117680614745, "grad_norm": 1.0425653457641602, "learning_rate": 1.4966963363525077e-05, "loss": 0.428, "step": 22490 }, { "epoch": 1.5094459917452434, "grad_norm": 0.8826553821563721, "learning_rate": 1.4959209713316213e-05, "loss": 0.4278, "step": 22492 }, { "epoch": 1.5095802154290126, "grad_norm": 1.0622694492340088, "learning_rate": 1.4951457718715772e-05, "loss": 0.477, "step": 22494 }, { "epoch": 1.5097144391127815, "grad_norm": 1.084007740020752, "learning_rate": 1.4943707380090067e-05, "loss": 0.5435, "step": 22496 }, { "epoch": 1.5098486627965504, "grad_norm": 1.0633143186569214, "learning_rate": 1.493595869780522e-05, "loss": 0.56, "step": 22498 }, { "epoch": 1.5099828864803193, "grad_norm": 1.3784126043319702, "learning_rate": 1.4928211672227387e-05, "loss": 0.4376, "step": 22500 }, { "epoch": 1.5101171101640884, "grad_norm": 1.055816888809204, "learning_rate": 1.4920466303722568e-05, "loss": 0.4615, "step": 22502 }, { "epoch": 1.5102513338478576, "grad_norm": 1.0887091159820557, "learning_rate": 1.4912722592656758e-05, "loss": 0.511, "step": 22504 }, { "epoch": 1.5103855575316265, "grad_norm": 1.1009773015975952, "learning_rate": 1.4904980539395808e-05, "loss": 0.4833, "step": 22506 }, { "epoch": 1.5105197812153954, "grad_norm": 1.0618146657943726, "learning_rate": 1.4897240144305513e-05, "loss": 0.4793, "step": 22508 }, { "epoch": 1.5106540048991643, "grad_norm": 1.0707080364227295, "learning_rate": 1.4889501407751583e-05, "loss": 0.5317, "step": 22510 }, { "epoch": 1.5107882285829335, "grad_norm": 0.9482539296150208, "learning_rate": 1.4881764330099685e-05, "loss": 0.515, "step": 22512 }, { "epoch": 1.5109224522667024, "grad_norm": 0.8839782476425171, "learning_rate": 1.487402891171536e-05, "loss": 0.4213, "step": 22514 }, { "epoch": 1.5110566759504716, "grad_norm": 0.9936320781707764, "learning_rate": 1.4866295152964104e-05, "loss": 0.4851, "step": 22516 }, { "epoch": 1.5111908996342405, "grad_norm": 0.9398537278175354, "learning_rate": 1.4858563054211294e-05, "loss": 0.4829, "step": 22518 }, { "epoch": 1.5113251233180094, "grad_norm": 1.0250434875488281, "learning_rate": 1.4850832615822297e-05, "loss": 0.475, "step": 22520 }, { "epoch": 1.5114593470017783, "grad_norm": 1.359682559967041, "learning_rate": 1.484310383816232e-05, "loss": 0.5049, "step": 22522 }, { "epoch": 1.5115935706855474, "grad_norm": 1.0129345655441284, "learning_rate": 1.4835376721596588e-05, "loss": 0.4854, "step": 22524 }, { "epoch": 1.5117277943693166, "grad_norm": 1.093867540359497, "learning_rate": 1.482765126649014e-05, "loss": 0.5006, "step": 22526 }, { "epoch": 1.5118620180530855, "grad_norm": 0.9952700138092041, "learning_rate": 1.481992747320799e-05, "loss": 0.5021, "step": 22528 }, { "epoch": 1.5119962417368544, "grad_norm": 0.9017931222915649, "learning_rate": 1.4812205342115104e-05, "loss": 0.4701, "step": 22530 }, { "epoch": 1.5121304654206233, "grad_norm": 1.0444830656051636, "learning_rate": 1.4804484873576314e-05, "loss": 0.4944, "step": 22532 }, { "epoch": 1.5122646891043925, "grad_norm": 1.1205003261566162, "learning_rate": 1.4796766067956424e-05, "loss": 0.5356, "step": 22534 }, { "epoch": 1.5123989127881616, "grad_norm": 1.121140480041504, "learning_rate": 1.4789048925620091e-05, "loss": 0.4984, "step": 22536 }, { "epoch": 1.5125331364719306, "grad_norm": 0.7770562171936035, "learning_rate": 1.4781333446931966e-05, "loss": 0.3848, "step": 22538 }, { "epoch": 1.5126673601556995, "grad_norm": 0.9135148525238037, "learning_rate": 1.4773619632256568e-05, "loss": 0.4568, "step": 22540 }, { "epoch": 1.5128015838394684, "grad_norm": 0.8369489908218384, "learning_rate": 1.476590748195839e-05, "loss": 0.4315, "step": 22542 }, { "epoch": 1.5129358075232375, "grad_norm": 1.1326509714126587, "learning_rate": 1.4758196996401796e-05, "loss": 0.429, "step": 22544 }, { "epoch": 1.5130700312070064, "grad_norm": 1.1862406730651855, "learning_rate": 1.4750488175951099e-05, "loss": 0.4852, "step": 22546 }, { "epoch": 1.5132042548907756, "grad_norm": 1.0056285858154297, "learning_rate": 1.4742781020970498e-05, "loss": 0.4809, "step": 22548 }, { "epoch": 1.5133384785745445, "grad_norm": 0.997708261013031, "learning_rate": 1.4735075531824183e-05, "loss": 0.4832, "step": 22550 }, { "epoch": 1.5134727022583134, "grad_norm": 1.0812458992004395, "learning_rate": 1.4727371708876203e-05, "loss": 0.5126, "step": 22552 }, { "epoch": 1.5136069259420823, "grad_norm": 1.117914080619812, "learning_rate": 1.4719669552490551e-05, "loss": 0.4548, "step": 22554 }, { "epoch": 1.5137411496258515, "grad_norm": 1.194248914718628, "learning_rate": 1.4711969063031117e-05, "loss": 0.5727, "step": 22556 }, { "epoch": 1.5138753733096206, "grad_norm": 0.9323875308036804, "learning_rate": 1.4704270240861773e-05, "loss": 0.4637, "step": 22558 }, { "epoch": 1.5140095969933896, "grad_norm": 1.0517685413360596, "learning_rate": 1.4696573086346237e-05, "loss": 0.4656, "step": 22560 }, { "epoch": 1.5141438206771585, "grad_norm": 0.9463817477226257, "learning_rate": 1.4688877599848216e-05, "loss": 0.4301, "step": 22562 }, { "epoch": 1.5142780443609274, "grad_norm": 1.0993897914886475, "learning_rate": 1.4681183781731294e-05, "loss": 0.4534, "step": 22564 }, { "epoch": 1.5144122680446965, "grad_norm": 1.0667643547058105, "learning_rate": 1.4673491632358987e-05, "loss": 0.4823, "step": 22566 }, { "epoch": 1.5145464917284654, "grad_norm": 1.004535436630249, "learning_rate": 1.4665801152094715e-05, "loss": 0.4562, "step": 22568 }, { "epoch": 1.5146807154122346, "grad_norm": 1.5485291481018066, "learning_rate": 1.4658112341301878e-05, "loss": 0.4401, "step": 22570 }, { "epoch": 1.5148149390960035, "grad_norm": 1.1256248950958252, "learning_rate": 1.4650425200343732e-05, "loss": 0.4754, "step": 22572 }, { "epoch": 1.5149491627797724, "grad_norm": 1.0484271049499512, "learning_rate": 1.4642739729583482e-05, "loss": 0.5324, "step": 22574 }, { "epoch": 1.5150833864635413, "grad_norm": 0.9325432777404785, "learning_rate": 1.463505592938424e-05, "loss": 0.4541, "step": 22576 }, { "epoch": 1.5152176101473105, "grad_norm": 0.9707252383232117, "learning_rate": 1.4627373800109078e-05, "loss": 0.4435, "step": 22578 }, { "epoch": 1.5153518338310796, "grad_norm": 1.595065951347351, "learning_rate": 1.4619693342120943e-05, "loss": 0.4888, "step": 22580 }, { "epoch": 1.5154860575148486, "grad_norm": 1.0771530866622925, "learning_rate": 1.4612014555782733e-05, "loss": 0.4859, "step": 22582 }, { "epoch": 1.5156202811986175, "grad_norm": 0.9721912741661072, "learning_rate": 1.4604337441457227e-05, "loss": 0.4599, "step": 22584 }, { "epoch": 1.5157545048823864, "grad_norm": 0.9945108294487, "learning_rate": 1.459666199950719e-05, "loss": 0.4889, "step": 22586 }, { "epoch": 1.5158887285661555, "grad_norm": 0.9683254957199097, "learning_rate": 1.4588988230295242e-05, "loss": 0.4873, "step": 22588 }, { "epoch": 1.5160229522499244, "grad_norm": 1.6426539421081543, "learning_rate": 1.458131613418398e-05, "loss": 0.4694, "step": 22590 }, { "epoch": 1.5161571759336936, "grad_norm": 1.1235136985778809, "learning_rate": 1.4573645711535899e-05, "loss": 0.4858, "step": 22592 }, { "epoch": 1.5162913996174625, "grad_norm": 1.0458322763442993, "learning_rate": 1.4565976962713369e-05, "loss": 0.5101, "step": 22594 }, { "epoch": 1.5164256233012314, "grad_norm": 0.9468684196472168, "learning_rate": 1.4558309888078758e-05, "loss": 0.4137, "step": 22596 }, { "epoch": 1.5165598469850003, "grad_norm": 1.0720895528793335, "learning_rate": 1.4550644487994303e-05, "loss": 0.469, "step": 22598 }, { "epoch": 1.5166940706687695, "grad_norm": 0.8745453357696533, "learning_rate": 1.45429807628222e-05, "loss": 0.4279, "step": 22600 }, { "epoch": 1.5168282943525386, "grad_norm": 1.0765306949615479, "learning_rate": 1.4535318712924529e-05, "loss": 0.4812, "step": 22602 }, { "epoch": 1.5169625180363076, "grad_norm": 0.9445536136627197, "learning_rate": 1.4527658338663313e-05, "loss": 0.4573, "step": 22604 }, { "epoch": 1.5170967417200765, "grad_norm": 1.193177342414856, "learning_rate": 1.4519999640400478e-05, "loss": 0.5268, "step": 22606 }, { "epoch": 1.5172309654038454, "grad_norm": 1.078736424446106, "learning_rate": 1.4512342618497898e-05, "loss": 0.5245, "step": 22608 }, { "epoch": 1.5173651890876145, "grad_norm": 1.1003497838974, "learning_rate": 1.4504687273317352e-05, "loss": 0.4627, "step": 22610 }, { "epoch": 1.5174994127713837, "grad_norm": 1.018603801727295, "learning_rate": 1.4497033605220529e-05, "loss": 0.4807, "step": 22612 }, { "epoch": 1.5176336364551526, "grad_norm": 0.9425628781318665, "learning_rate": 1.4489381614569043e-05, "loss": 0.436, "step": 22614 }, { "epoch": 1.5177678601389215, "grad_norm": 1.0539040565490723, "learning_rate": 1.4481731301724461e-05, "loss": 0.4773, "step": 22616 }, { "epoch": 1.5179020838226904, "grad_norm": 1.076253890991211, "learning_rate": 1.4474082667048217e-05, "loss": 0.4542, "step": 22618 }, { "epoch": 1.5180363075064596, "grad_norm": 1.0536024570465088, "learning_rate": 1.4466435710901737e-05, "loss": 0.5279, "step": 22620 }, { "epoch": 1.5181705311902285, "grad_norm": 1.0149723291397095, "learning_rate": 1.4458790433646263e-05, "loss": 0.5016, "step": 22622 }, { "epoch": 1.5183047548739976, "grad_norm": 4.249631881713867, "learning_rate": 1.445114683564307e-05, "loss": 0.5644, "step": 22624 }, { "epoch": 1.5184389785577665, "grad_norm": 0.9433045983314514, "learning_rate": 1.4443504917253275e-05, "loss": 0.4131, "step": 22626 }, { "epoch": 1.5185732022415355, "grad_norm": 1.0704541206359863, "learning_rate": 1.4435864678837962e-05, "loss": 0.4985, "step": 22628 }, { "epoch": 1.5187074259253044, "grad_norm": 1.1655170917510986, "learning_rate": 1.4428226120758114e-05, "loss": 0.5102, "step": 22630 }, { "epoch": 1.5188416496090735, "grad_norm": 0.979566216468811, "learning_rate": 1.4420589243374626e-05, "loss": 0.4862, "step": 22632 }, { "epoch": 1.5189758732928427, "grad_norm": 0.9434431791305542, "learning_rate": 1.4412954047048326e-05, "loss": 0.4373, "step": 22634 }, { "epoch": 1.5191100969766116, "grad_norm": 1.0647971630096436, "learning_rate": 1.4405320532139976e-05, "loss": 0.4929, "step": 22636 }, { "epoch": 1.5192443206603805, "grad_norm": 0.9849785566329956, "learning_rate": 1.4397688699010241e-05, "loss": 0.4674, "step": 22638 }, { "epoch": 1.5193785443441494, "grad_norm": 1.1018036603927612, "learning_rate": 1.4390058548019708e-05, "loss": 0.5187, "step": 22640 }, { "epoch": 1.5195127680279186, "grad_norm": 1.1140117645263672, "learning_rate": 1.4382430079528864e-05, "loss": 0.4234, "step": 22642 }, { "epoch": 1.5196469917116875, "grad_norm": 1.0272324085235596, "learning_rate": 1.4374803293898181e-05, "loss": 0.4543, "step": 22644 }, { "epoch": 1.5197812153954566, "grad_norm": 1.0630650520324707, "learning_rate": 1.4367178191487968e-05, "loss": 0.5103, "step": 22646 }, { "epoch": 1.5199154390792255, "grad_norm": 0.9447331428527832, "learning_rate": 1.4359554772658552e-05, "loss": 0.482, "step": 22648 }, { "epoch": 1.5200496627629945, "grad_norm": 1.052130103111267, "learning_rate": 1.4351933037770049e-05, "loss": 0.5365, "step": 22650 }, { "epoch": 1.5201838864467634, "grad_norm": 1.0532441139221191, "learning_rate": 1.4344312987182635e-05, "loss": 0.521, "step": 22652 }, { "epoch": 1.5203181101305325, "grad_norm": 1.086188554763794, "learning_rate": 1.4336694621256303e-05, "loss": 0.5064, "step": 22654 }, { "epoch": 1.5204523338143017, "grad_norm": 1.1076205968856812, "learning_rate": 1.4329077940351031e-05, "loss": 0.4888, "step": 22656 }, { "epoch": 1.5205865574980706, "grad_norm": 1.1020424365997314, "learning_rate": 1.4321462944826685e-05, "loss": 0.4586, "step": 22658 }, { "epoch": 1.5207207811818395, "grad_norm": 0.9768147468566895, "learning_rate": 1.4313849635043058e-05, "loss": 0.4714, "step": 22660 }, { "epoch": 1.5208550048656084, "grad_norm": 0.9518616795539856, "learning_rate": 1.4306238011359863e-05, "loss": 0.4481, "step": 22662 }, { "epoch": 1.5209892285493776, "grad_norm": 0.9238994717597961, "learning_rate": 1.4298628074136717e-05, "loss": 0.4517, "step": 22664 }, { "epoch": 1.5211234522331465, "grad_norm": 0.9104543328285217, "learning_rate": 1.4291019823733203e-05, "loss": 0.4736, "step": 22666 }, { "epoch": 1.5212576759169156, "grad_norm": 1.152108907699585, "learning_rate": 1.4283413260508788e-05, "loss": 0.5382, "step": 22668 }, { "epoch": 1.5213918996006845, "grad_norm": 1.0099257230758667, "learning_rate": 1.4275808384822858e-05, "loss": 0.4768, "step": 22670 }, { "epoch": 1.5215261232844535, "grad_norm": 1.1115846633911133, "learning_rate": 1.4268205197034717e-05, "loss": 0.4871, "step": 22672 }, { "epoch": 1.5216603469682224, "grad_norm": 0.9350446462631226, "learning_rate": 1.4260603697503638e-05, "loss": 0.4816, "step": 22674 }, { "epoch": 1.5217945706519915, "grad_norm": 1.0540435314178467, "learning_rate": 1.4253003886588751e-05, "loss": 0.4726, "step": 22676 }, { "epoch": 1.5219287943357607, "grad_norm": 0.9183579683303833, "learning_rate": 1.4245405764649138e-05, "loss": 0.4748, "step": 22678 }, { "epoch": 1.5220630180195296, "grad_norm": 1.0179717540740967, "learning_rate": 1.4237809332043784e-05, "loss": 0.4622, "step": 22680 }, { "epoch": 1.5221972417032985, "grad_norm": 1.0048412084579468, "learning_rate": 1.4230214589131625e-05, "loss": 0.4691, "step": 22682 }, { "epoch": 1.5223314653870674, "grad_norm": 1.0105338096618652, "learning_rate": 1.4222621536271474e-05, "loss": 0.4468, "step": 22684 }, { "epoch": 1.5224656890708366, "grad_norm": 0.9902668595314026, "learning_rate": 1.4215030173822136e-05, "loss": 0.4491, "step": 22686 }, { "epoch": 1.5225999127546057, "grad_norm": 1.0982518196105957, "learning_rate": 1.4207440502142222e-05, "loss": 0.4566, "step": 22688 }, { "epoch": 1.5227341364383746, "grad_norm": 1.0153194665908813, "learning_rate": 1.4199852521590373e-05, "loss": 0.4912, "step": 22690 }, { "epoch": 1.5228683601221435, "grad_norm": 0.9769749045372009, "learning_rate": 1.419226623252508e-05, "loss": 0.4997, "step": 22692 }, { "epoch": 1.5230025838059125, "grad_norm": 0.9678530097007751, "learning_rate": 1.4184681635304814e-05, "loss": 0.4807, "step": 22694 }, { "epoch": 1.5231368074896814, "grad_norm": 0.9360699653625488, "learning_rate": 1.4177098730287914e-05, "loss": 0.4441, "step": 22696 }, { "epoch": 1.5232710311734505, "grad_norm": 1.1069527864456177, "learning_rate": 1.4169517517832654e-05, "loss": 0.4684, "step": 22698 }, { "epoch": 1.5234052548572197, "grad_norm": 1.3689544200897217, "learning_rate": 1.4161937998297214e-05, "loss": 0.5736, "step": 22700 }, { "epoch": 1.5235394785409886, "grad_norm": 0.9946498274803162, "learning_rate": 1.4154360172039749e-05, "loss": 0.5132, "step": 22702 }, { "epoch": 1.5236737022247575, "grad_norm": 0.970460832118988, "learning_rate": 1.4146784039418282e-05, "loss": 0.4773, "step": 22704 }, { "epoch": 1.5238079259085264, "grad_norm": 0.9970105886459351, "learning_rate": 1.4139209600790766e-05, "loss": 0.4294, "step": 22706 }, { "epoch": 1.5239421495922956, "grad_norm": 1.1605664491653442, "learning_rate": 1.4131636856515056e-05, "loss": 0.455, "step": 22708 }, { "epoch": 1.5240763732760647, "grad_norm": 1.0581333637237549, "learning_rate": 1.4124065806948994e-05, "loss": 0.4962, "step": 22710 }, { "epoch": 1.5242105969598336, "grad_norm": 0.9241947531700134, "learning_rate": 1.4116496452450262e-05, "loss": 0.3972, "step": 22712 }, { "epoch": 1.5243448206436025, "grad_norm": 1.2357085943222046, "learning_rate": 1.4108928793376542e-05, "loss": 0.441, "step": 22714 }, { "epoch": 1.5244790443273715, "grad_norm": 1.1028046607971191, "learning_rate": 1.4101362830085324e-05, "loss": 0.4704, "step": 22716 }, { "epoch": 1.5246132680111406, "grad_norm": 0.9638956189155579, "learning_rate": 1.4093798562934141e-05, "loss": 0.4772, "step": 22718 }, { "epoch": 1.5247474916949095, "grad_norm": 1.06745445728302, "learning_rate": 1.408623599228035e-05, "loss": 0.4541, "step": 22720 }, { "epoch": 1.5248817153786787, "grad_norm": 1.1657145023345947, "learning_rate": 1.4078675118481305e-05, "loss": 0.5244, "step": 22722 }, { "epoch": 1.5250159390624476, "grad_norm": 1.0456702709197998, "learning_rate": 1.407111594189422e-05, "loss": 0.4596, "step": 22724 }, { "epoch": 1.5251501627462165, "grad_norm": 1.0020138025283813, "learning_rate": 1.4063558462876259e-05, "loss": 0.5044, "step": 22726 }, { "epoch": 1.5252843864299854, "grad_norm": 0.9499043822288513, "learning_rate": 1.4056002681784496e-05, "loss": 0.4474, "step": 22728 }, { "epoch": 1.5254186101137546, "grad_norm": 0.9909380674362183, "learning_rate": 1.4048448598975905e-05, "loss": 0.4658, "step": 22730 }, { "epoch": 1.5255528337975237, "grad_norm": 1.068764090538025, "learning_rate": 1.4040896214807442e-05, "loss": 0.4693, "step": 22732 }, { "epoch": 1.5256870574812926, "grad_norm": 0.9431993365287781, "learning_rate": 1.4033345529635922e-05, "loss": 0.4253, "step": 22734 }, { "epoch": 1.5258212811650615, "grad_norm": 1.0659258365631104, "learning_rate": 1.4025796543818098e-05, "loss": 0.5087, "step": 22736 }, { "epoch": 1.5259555048488305, "grad_norm": 1.0343708992004395, "learning_rate": 1.4018249257710636e-05, "loss": 0.4301, "step": 22738 }, { "epoch": 1.5260897285325996, "grad_norm": 1.1548924446105957, "learning_rate": 1.4010703671670155e-05, "loss": 0.4605, "step": 22740 }, { "epoch": 1.5262239522163685, "grad_norm": 1.0230505466461182, "learning_rate": 1.4003159786053139e-05, "loss": 0.4107, "step": 22742 }, { "epoch": 1.5263581759001377, "grad_norm": 1.0957962274551392, "learning_rate": 1.3995617601216066e-05, "loss": 0.488, "step": 22744 }, { "epoch": 1.5264923995839066, "grad_norm": 0.9980576038360596, "learning_rate": 1.3988077117515236e-05, "loss": 0.4498, "step": 22746 }, { "epoch": 1.5266266232676755, "grad_norm": 0.8710897564888, "learning_rate": 1.3980538335306964e-05, "loss": 0.5065, "step": 22748 }, { "epoch": 1.5267608469514444, "grad_norm": 1.0267164707183838, "learning_rate": 1.3973001254947404e-05, "loss": 0.4917, "step": 22750 }, { "epoch": 1.5268950706352136, "grad_norm": 1.0731956958770752, "learning_rate": 1.3965465876792716e-05, "loss": 0.5146, "step": 22752 }, { "epoch": 1.5270292943189827, "grad_norm": 0.9476215839385986, "learning_rate": 1.3957932201198897e-05, "loss": 0.4683, "step": 22754 }, { "epoch": 1.5271635180027516, "grad_norm": 0.9859458208084106, "learning_rate": 1.3950400228521914e-05, "loss": 0.4408, "step": 22756 }, { "epoch": 1.5272977416865205, "grad_norm": 0.9938817620277405, "learning_rate": 1.3942869959117616e-05, "loss": 0.4758, "step": 22758 }, { "epoch": 1.5274319653702895, "grad_norm": 1.087532877922058, "learning_rate": 1.3935341393341822e-05, "loss": 0.5726, "step": 22760 }, { "epoch": 1.5275661890540586, "grad_norm": 1.059373378753662, "learning_rate": 1.3927814531550226e-05, "loss": 0.4676, "step": 22762 }, { "epoch": 1.5277004127378278, "grad_norm": 1.0563775300979614, "learning_rate": 1.3920289374098467e-05, "loss": 0.4405, "step": 22764 }, { "epoch": 1.5278346364215967, "grad_norm": 0.9323393702507019, "learning_rate": 1.391276592134207e-05, "loss": 0.4398, "step": 22766 }, { "epoch": 1.5279688601053656, "grad_norm": 0.9752359390258789, "learning_rate": 1.3905244173636534e-05, "loss": 0.4928, "step": 22768 }, { "epoch": 1.5281030837891345, "grad_norm": 0.9924960732460022, "learning_rate": 1.389772413133722e-05, "loss": 0.5054, "step": 22770 }, { "epoch": 1.5282373074729034, "grad_norm": 1.0688010454177856, "learning_rate": 1.3890205794799476e-05, "loss": 0.517, "step": 22772 }, { "epoch": 1.5283715311566726, "grad_norm": 1.0437699556350708, "learning_rate": 1.3882689164378477e-05, "loss": 0.482, "step": 22774 }, { "epoch": 1.5285057548404417, "grad_norm": 1.0509988069534302, "learning_rate": 1.3875174240429405e-05, "loss": 0.4936, "step": 22776 }, { "epoch": 1.5286399785242106, "grad_norm": 1.0045628547668457, "learning_rate": 1.3867661023307298e-05, "loss": 0.474, "step": 22778 }, { "epoch": 1.5287742022079795, "grad_norm": 1.0354044437408447, "learning_rate": 1.3860149513367171e-05, "loss": 0.507, "step": 22780 }, { "epoch": 1.5289084258917485, "grad_norm": 1.0339772701263428, "learning_rate": 1.3852639710963916e-05, "loss": 0.4641, "step": 22782 }, { "epoch": 1.5290426495755176, "grad_norm": 1.1421676874160767, "learning_rate": 1.384513161645235e-05, "loss": 0.5663, "step": 22784 }, { "epoch": 1.5291768732592868, "grad_norm": 1.1045042276382446, "learning_rate": 1.3837625230187212e-05, "loss": 0.4658, "step": 22786 }, { "epoch": 1.5293110969430557, "grad_norm": 1.1218268871307373, "learning_rate": 1.383012055252318e-05, "loss": 0.4573, "step": 22788 }, { "epoch": 1.5294453206268246, "grad_norm": 0.9117696285247803, "learning_rate": 1.3822617583814828e-05, "loss": 0.4597, "step": 22790 }, { "epoch": 1.5295795443105935, "grad_norm": 0.9316728711128235, "learning_rate": 1.3815116324416654e-05, "loss": 0.4361, "step": 22792 }, { "epoch": 1.5297137679943626, "grad_norm": 0.9336127042770386, "learning_rate": 1.380761677468307e-05, "loss": 0.455, "step": 22794 }, { "epoch": 1.5298479916781316, "grad_norm": 0.9192490577697754, "learning_rate": 1.3800118934968436e-05, "loss": 0.4215, "step": 22796 }, { "epoch": 1.5299822153619007, "grad_norm": 1.175935983657837, "learning_rate": 1.3792622805626998e-05, "loss": 0.4492, "step": 22798 }, { "epoch": 1.5301164390456696, "grad_norm": 1.0037862062454224, "learning_rate": 1.378512838701293e-05, "loss": 0.4946, "step": 22800 }, { "epoch": 1.5302506627294385, "grad_norm": 1.108101725578308, "learning_rate": 1.377763567948034e-05, "loss": 0.4989, "step": 22802 }, { "epoch": 1.5303848864132075, "grad_norm": 1.0684829950332642, "learning_rate": 1.3770144683383213e-05, "loss": 0.506, "step": 22804 }, { "epoch": 1.5305191100969766, "grad_norm": 0.9487861394882202, "learning_rate": 1.3762655399075524e-05, "loss": 0.4489, "step": 22806 }, { "epoch": 1.5306533337807458, "grad_norm": 1.1091814041137695, "learning_rate": 1.3755167826911097e-05, "loss": 0.4866, "step": 22808 }, { "epoch": 1.5307875574645147, "grad_norm": 1.2332231998443604, "learning_rate": 1.3747681967243748e-05, "loss": 0.4827, "step": 22810 }, { "epoch": 1.5309217811482836, "grad_norm": 1.0854482650756836, "learning_rate": 1.3740197820427108e-05, "loss": 0.4717, "step": 22812 }, { "epoch": 1.5310560048320525, "grad_norm": 1.0635013580322266, "learning_rate": 1.3732715386814832e-05, "loss": 0.4622, "step": 22814 }, { "epoch": 1.5311902285158216, "grad_norm": 1.0756031274795532, "learning_rate": 1.3725234666760428e-05, "loss": 0.4947, "step": 22816 }, { "epoch": 1.5313244521995906, "grad_norm": 1.1334630250930786, "learning_rate": 1.3717755660617365e-05, "loss": 0.4707, "step": 22818 }, { "epoch": 1.5314586758833597, "grad_norm": 0.9022170305252075, "learning_rate": 1.3710278368738998e-05, "loss": 0.4441, "step": 22820 }, { "epoch": 1.5315928995671286, "grad_norm": 0.9382498264312744, "learning_rate": 1.3702802791478625e-05, "loss": 0.4669, "step": 22822 }, { "epoch": 1.5317271232508975, "grad_norm": 1.1221257448196411, "learning_rate": 1.3695328929189426e-05, "loss": 0.5197, "step": 22824 }, { "epoch": 1.5318613469346665, "grad_norm": 0.9527440071105957, "learning_rate": 1.3687856782224562e-05, "loss": 0.4608, "step": 22826 }, { "epoch": 1.5319955706184356, "grad_norm": 0.9356170892715454, "learning_rate": 1.3680386350937058e-05, "loss": 0.5009, "step": 22828 }, { "epoch": 1.5321297943022048, "grad_norm": 1.005391001701355, "learning_rate": 1.3672917635679877e-05, "loss": 0.4767, "step": 22830 }, { "epoch": 1.5322640179859737, "grad_norm": 0.804820716381073, "learning_rate": 1.3665450636805893e-05, "loss": 0.4131, "step": 22832 }, { "epoch": 1.5323982416697426, "grad_norm": 0.9496227502822876, "learning_rate": 1.3657985354667929e-05, "loss": 0.5042, "step": 22834 }, { "epoch": 1.5325324653535115, "grad_norm": 1.1346501111984253, "learning_rate": 1.365052178961868e-05, "loss": 0.5019, "step": 22836 }, { "epoch": 1.5326666890372806, "grad_norm": 1.0846158266067505, "learning_rate": 1.3643059942010832e-05, "loss": 0.4948, "step": 22838 }, { "epoch": 1.5328009127210498, "grad_norm": 1.0905262231826782, "learning_rate": 1.3635599812196875e-05, "loss": 0.4883, "step": 22840 }, { "epoch": 1.5329351364048187, "grad_norm": 0.9363589882850647, "learning_rate": 1.3628141400529337e-05, "loss": 0.494, "step": 22842 }, { "epoch": 1.5330693600885876, "grad_norm": 1.1014615297317505, "learning_rate": 1.3620684707360576e-05, "loss": 0.4778, "step": 22844 }, { "epoch": 1.5332035837723565, "grad_norm": 1.0319340229034424, "learning_rate": 1.3613229733042943e-05, "loss": 0.4386, "step": 22846 }, { "epoch": 1.5333378074561255, "grad_norm": 1.0542309284210205, "learning_rate": 1.3605776477928656e-05, "loss": 0.4262, "step": 22848 }, { "epoch": 1.5334720311398946, "grad_norm": 1.1086591482162476, "learning_rate": 1.359832494236986e-05, "loss": 0.4776, "step": 22850 }, { "epoch": 1.5336062548236638, "grad_norm": 0.8893815875053406, "learning_rate": 1.3590875126718621e-05, "loss": 0.4307, "step": 22852 }, { "epoch": 1.5337404785074327, "grad_norm": 1.0791600942611694, "learning_rate": 1.3583427031326946e-05, "loss": 0.4935, "step": 22854 }, { "epoch": 1.5338747021912016, "grad_norm": 0.8816147446632385, "learning_rate": 1.3575980656546732e-05, "loss": 0.444, "step": 22856 }, { "epoch": 1.5340089258749705, "grad_norm": 1.0307036638259888, "learning_rate": 1.3568536002729815e-05, "loss": 0.4971, "step": 22858 }, { "epoch": 1.5341431495587396, "grad_norm": 1.0213000774383545, "learning_rate": 1.3561093070227909e-05, "loss": 0.4832, "step": 22860 }, { "epoch": 1.5342773732425088, "grad_norm": 1.0029380321502686, "learning_rate": 1.3553651859392725e-05, "loss": 0.4448, "step": 22862 }, { "epoch": 1.5344115969262777, "grad_norm": 0.9618352651596069, "learning_rate": 1.35462123705758e-05, "loss": 0.4694, "step": 22864 }, { "epoch": 1.5345458206100466, "grad_norm": 1.014799952507019, "learning_rate": 1.353877460412869e-05, "loss": 0.4678, "step": 22866 }, { "epoch": 1.5346800442938155, "grad_norm": 1.0170156955718994, "learning_rate": 1.3531338560402767e-05, "loss": 0.4646, "step": 22868 }, { "epoch": 1.5348142679775847, "grad_norm": 0.9131841659545898, "learning_rate": 1.352390423974937e-05, "loss": 0.4263, "step": 22870 }, { "epoch": 1.5349484916613536, "grad_norm": 1.0565253496170044, "learning_rate": 1.3516471642519784e-05, "loss": 0.4633, "step": 22872 }, { "epoch": 1.5350827153451228, "grad_norm": 1.152448058128357, "learning_rate": 1.3509040769065157e-05, "loss": 0.5157, "step": 22874 }, { "epoch": 1.5352169390288917, "grad_norm": 0.9575986862182617, "learning_rate": 1.3501611619736609e-05, "loss": 0.4592, "step": 22876 }, { "epoch": 1.5353511627126606, "grad_norm": 1.0137993097305298, "learning_rate": 1.3494184194885146e-05, "loss": 0.4733, "step": 22878 }, { "epoch": 1.5354853863964295, "grad_norm": 1.088549256324768, "learning_rate": 1.3486758494861696e-05, "loss": 0.4896, "step": 22880 }, { "epoch": 1.5356196100801986, "grad_norm": 0.902563214302063, "learning_rate": 1.3479334520017084e-05, "loss": 0.4272, "step": 22882 }, { "epoch": 1.5357538337639678, "grad_norm": 0.9075942039489746, "learning_rate": 1.347191227070212e-05, "loss": 0.4132, "step": 22884 }, { "epoch": 1.5358880574477367, "grad_norm": 1.0854324102401733, "learning_rate": 1.3464491747267472e-05, "loss": 0.4918, "step": 22886 }, { "epoch": 1.5360222811315056, "grad_norm": 0.8842144012451172, "learning_rate": 1.3457072950063748e-05, "loss": 0.5058, "step": 22888 }, { "epoch": 1.5361565048152745, "grad_norm": 1.0799119472503662, "learning_rate": 1.3449655879441447e-05, "loss": 0.5155, "step": 22890 }, { "epoch": 1.5362907284990437, "grad_norm": 1.0394150018692017, "learning_rate": 1.3442240535751049e-05, "loss": 0.4633, "step": 22892 }, { "epoch": 1.5364249521828126, "grad_norm": 0.9008004069328308, "learning_rate": 1.3434826919342901e-05, "loss": 0.439, "step": 22894 }, { "epoch": 1.5365591758665818, "grad_norm": 1.079035997390747, "learning_rate": 1.342741503056728e-05, "loss": 0.4822, "step": 22896 }, { "epoch": 1.5366933995503507, "grad_norm": 1.0192890167236328, "learning_rate": 1.3420004869774367e-05, "loss": 0.4212, "step": 22898 }, { "epoch": 1.5368276232341196, "grad_norm": 0.8894795775413513, "learning_rate": 1.3412596437314302e-05, "loss": 0.499, "step": 22900 }, { "epoch": 1.5369618469178885, "grad_norm": 1.1050899028778076, "learning_rate": 1.3405189733537098e-05, "loss": 0.419, "step": 22902 }, { "epoch": 1.5370960706016576, "grad_norm": 0.9837499260902405, "learning_rate": 1.3397784758792731e-05, "loss": 0.4764, "step": 22904 }, { "epoch": 1.5372302942854268, "grad_norm": 1.1928001642227173, "learning_rate": 1.339038151343106e-05, "loss": 0.5221, "step": 22906 }, { "epoch": 1.5373645179691957, "grad_norm": 1.0429415702819824, "learning_rate": 1.3382979997801876e-05, "loss": 0.4655, "step": 22908 }, { "epoch": 1.5374987416529646, "grad_norm": 1.0762752294540405, "learning_rate": 1.3375580212254862e-05, "loss": 0.4728, "step": 22910 }, { "epoch": 1.5376329653367335, "grad_norm": 1.103614091873169, "learning_rate": 1.3368182157139686e-05, "loss": 0.4927, "step": 22912 }, { "epoch": 1.5377671890205027, "grad_norm": 1.0256223678588867, "learning_rate": 1.3360785832805867e-05, "loss": 0.5132, "step": 22914 }, { "epoch": 1.5379014127042718, "grad_norm": 1.0534138679504395, "learning_rate": 1.335339123960287e-05, "loss": 0.4351, "step": 22916 }, { "epoch": 1.5380356363880407, "grad_norm": 1.0414409637451172, "learning_rate": 1.3345998377880053e-05, "loss": 0.5019, "step": 22918 }, { "epoch": 1.5381698600718097, "grad_norm": 0.9938732981681824, "learning_rate": 1.3338607247986756e-05, "loss": 0.5434, "step": 22920 }, { "epoch": 1.5383040837555786, "grad_norm": 1.0894412994384766, "learning_rate": 1.3331217850272176e-05, "loss": 0.4831, "step": 22922 }, { "epoch": 1.5384383074393475, "grad_norm": 0.8352552652359009, "learning_rate": 1.3323830185085445e-05, "loss": 0.405, "step": 22924 }, { "epoch": 1.5385725311231166, "grad_norm": 1.027392864227295, "learning_rate": 1.3316444252775601e-05, "loss": 0.4724, "step": 22926 }, { "epoch": 1.5387067548068858, "grad_norm": 1.2071126699447632, "learning_rate": 1.3309060053691646e-05, "loss": 0.4669, "step": 22928 }, { "epoch": 1.5388409784906547, "grad_norm": 1.0582983493804932, "learning_rate": 1.3301677588182442e-05, "loss": 0.4437, "step": 22930 }, { "epoch": 1.5389752021744236, "grad_norm": 1.1398028135299683, "learning_rate": 1.3294296856596821e-05, "loss": 0.5096, "step": 22932 }, { "epoch": 1.5391094258581925, "grad_norm": 1.09684419631958, "learning_rate": 1.328691785928351e-05, "loss": 0.5131, "step": 22934 }, { "epoch": 1.5392436495419617, "grad_norm": 1.0865932703018188, "learning_rate": 1.3279540596591105e-05, "loss": 0.4987, "step": 22936 }, { "epoch": 1.5393778732257308, "grad_norm": 1.120383381843567, "learning_rate": 1.327216506886822e-05, "loss": 0.4713, "step": 22938 }, { "epoch": 1.5395120969094997, "grad_norm": 0.9693374037742615, "learning_rate": 1.326479127646329e-05, "loss": 0.4503, "step": 22940 }, { "epoch": 1.5396463205932687, "grad_norm": 0.9845932722091675, "learning_rate": 1.325741921972476e-05, "loss": 0.5201, "step": 22942 }, { "epoch": 1.5397805442770376, "grad_norm": 1.0494850873947144, "learning_rate": 1.325004889900091e-05, "loss": 0.4806, "step": 22944 }, { "epoch": 1.5399147679608067, "grad_norm": 1.0713417530059814, "learning_rate": 1.3242680314639993e-05, "loss": 0.5529, "step": 22946 }, { "epoch": 1.5400489916445756, "grad_norm": 0.932652473449707, "learning_rate": 1.3235313466990129e-05, "loss": 0.5251, "step": 22948 }, { "epoch": 1.5401832153283448, "grad_norm": 1.5986257791519165, "learning_rate": 1.3227948356399428e-05, "loss": 0.5229, "step": 22950 }, { "epoch": 1.5403174390121137, "grad_norm": 1.0638889074325562, "learning_rate": 1.3220584983215855e-05, "loss": 0.4649, "step": 22952 }, { "epoch": 1.5404516626958826, "grad_norm": 1.0206716060638428, "learning_rate": 1.3213223347787324e-05, "loss": 0.446, "step": 22954 }, { "epoch": 1.5405858863796515, "grad_norm": 1.1154367923736572, "learning_rate": 1.320586345046163e-05, "loss": 0.485, "step": 22956 }, { "epoch": 1.5407201100634207, "grad_norm": 0.9781999588012695, "learning_rate": 1.3198505291586555e-05, "loss": 0.4917, "step": 22958 }, { "epoch": 1.5408543337471898, "grad_norm": 0.9744774699211121, "learning_rate": 1.3191148871509724e-05, "loss": 0.5047, "step": 22960 }, { "epoch": 1.5409885574309587, "grad_norm": 1.0790060758590698, "learning_rate": 1.3183794190578752e-05, "loss": 0.4558, "step": 22962 }, { "epoch": 1.5411227811147277, "grad_norm": 1.945683479309082, "learning_rate": 1.317644124914108e-05, "loss": 0.5305, "step": 22964 }, { "epoch": 1.5412570047984966, "grad_norm": 1.2510796785354614, "learning_rate": 1.3169090047544164e-05, "loss": 0.4542, "step": 22966 }, { "epoch": 1.5413912284822657, "grad_norm": 1.0230785608291626, "learning_rate": 1.3161740586135308e-05, "loss": 0.4038, "step": 22968 }, { "epoch": 1.5415254521660346, "grad_norm": 1.131100058555603, "learning_rate": 1.3154392865261778e-05, "loss": 0.4862, "step": 22970 }, { "epoch": 1.5416596758498038, "grad_norm": 1.1616319417953491, "learning_rate": 1.3147046885270736e-05, "loss": 0.4893, "step": 22972 }, { "epoch": 1.5417938995335727, "grad_norm": 0.9876658320426941, "learning_rate": 1.3139702646509255e-05, "loss": 0.437, "step": 22974 }, { "epoch": 1.5419281232173416, "grad_norm": 0.9533802270889282, "learning_rate": 1.313236014932433e-05, "loss": 0.4538, "step": 22976 }, { "epoch": 1.5420623469011105, "grad_norm": 0.9141824841499329, "learning_rate": 1.312501939406291e-05, "loss": 0.5134, "step": 22978 }, { "epoch": 1.5421965705848797, "grad_norm": 0.8759487271308899, "learning_rate": 1.3117680381071807e-05, "loss": 0.4456, "step": 22980 }, { "epoch": 1.5423307942686488, "grad_norm": 1.0206835269927979, "learning_rate": 1.3110343110697782e-05, "loss": 0.4491, "step": 22982 }, { "epoch": 1.5424650179524177, "grad_norm": 1.1610170602798462, "learning_rate": 1.3103007583287486e-05, "loss": 0.476, "step": 22984 }, { "epoch": 1.5425992416361867, "grad_norm": 1.086210012435913, "learning_rate": 1.3095673799187546e-05, "loss": 0.4752, "step": 22986 }, { "epoch": 1.5427334653199556, "grad_norm": 1.1827083826065063, "learning_rate": 1.3088341758744432e-05, "loss": 0.4967, "step": 22988 }, { "epoch": 1.5428676890037247, "grad_norm": 1.1646783351898193, "learning_rate": 1.3081011462304615e-05, "loss": 0.4897, "step": 22990 }, { "epoch": 1.5430019126874939, "grad_norm": 0.93353670835495, "learning_rate": 1.3073682910214386e-05, "loss": 0.4634, "step": 22992 }, { "epoch": 1.5431361363712628, "grad_norm": 1.0688941478729248, "learning_rate": 1.3066356102820032e-05, "loss": 0.5364, "step": 22994 }, { "epoch": 1.5432703600550317, "grad_norm": 0.9025923609733582, "learning_rate": 1.3059031040467717e-05, "loss": 0.4392, "step": 22996 }, { "epoch": 1.5434045837388006, "grad_norm": 1.08487069606781, "learning_rate": 1.305170772350356e-05, "loss": 0.457, "step": 22998 }, { "epoch": 1.5435388074225695, "grad_norm": 0.9283562302589417, "learning_rate": 1.3044386152273552e-05, "loss": 0.4543, "step": 23000 }, { "epoch": 1.5436730311063387, "grad_norm": 1.0412951707839966, "learning_rate": 1.303706632712363e-05, "loss": 0.5077, "step": 23002 }, { "epoch": 1.5438072547901078, "grad_norm": 1.063987374305725, "learning_rate": 1.3029748248399642e-05, "loss": 0.4588, "step": 23004 }, { "epoch": 1.5439414784738767, "grad_norm": 1.0530637502670288, "learning_rate": 1.302243191644733e-05, "loss": 0.4607, "step": 23006 }, { "epoch": 1.5440757021576457, "grad_norm": 0.9383824467658997, "learning_rate": 1.3015117331612415e-05, "loss": 0.4724, "step": 23008 }, { "epoch": 1.5442099258414146, "grad_norm": 1.0184880495071411, "learning_rate": 1.3007804494240478e-05, "loss": 0.5048, "step": 23010 }, { "epoch": 1.5443441495251837, "grad_norm": 0.978066623210907, "learning_rate": 1.3000493404677039e-05, "loss": 0.4338, "step": 23012 }, { "epoch": 1.5444783732089529, "grad_norm": 1.1344259977340698, "learning_rate": 1.2993184063267516e-05, "loss": 0.4571, "step": 23014 }, { "epoch": 1.5446125968927218, "grad_norm": 1.0928936004638672, "learning_rate": 1.2985876470357284e-05, "loss": 0.5347, "step": 23016 }, { "epoch": 1.5447468205764907, "grad_norm": 0.9233540296554565, "learning_rate": 1.2978570626291609e-05, "loss": 0.4887, "step": 23018 }, { "epoch": 1.5448810442602596, "grad_norm": 0.9370667338371277, "learning_rate": 1.2971266531415672e-05, "loss": 0.4634, "step": 23020 }, { "epoch": 1.5450152679440288, "grad_norm": 1.0324769020080566, "learning_rate": 1.2963964186074563e-05, "loss": 0.5203, "step": 23022 }, { "epoch": 1.5451494916277977, "grad_norm": 1.148537039756775, "learning_rate": 1.2956663590613327e-05, "loss": 0.5763, "step": 23024 }, { "epoch": 1.5452837153115668, "grad_norm": 1.030452013015747, "learning_rate": 1.2949364745376884e-05, "loss": 0.4983, "step": 23026 }, { "epoch": 1.5454179389953357, "grad_norm": 0.9592177271842957, "learning_rate": 1.2942067650710116e-05, "loss": 0.4368, "step": 23028 }, { "epoch": 1.5455521626791047, "grad_norm": 0.7843078970909119, "learning_rate": 1.2934772306957776e-05, "loss": 0.4538, "step": 23030 }, { "epoch": 1.5456863863628736, "grad_norm": 1.0391765832901, "learning_rate": 1.2927478714464559e-05, "loss": 0.4873, "step": 23032 }, { "epoch": 1.5458206100466427, "grad_norm": 1.0403895378112793, "learning_rate": 1.2920186873575064e-05, "loss": 0.4878, "step": 23034 }, { "epoch": 1.5459548337304119, "grad_norm": 1.0897223949432373, "learning_rate": 1.2912896784633833e-05, "loss": 0.5198, "step": 23036 }, { "epoch": 1.5460890574141808, "grad_norm": 1.030831217765808, "learning_rate": 1.2905608447985301e-05, "loss": 0.4858, "step": 23038 }, { "epoch": 1.5462232810979497, "grad_norm": 1.1313307285308838, "learning_rate": 1.2898321863973829e-05, "loss": 0.5108, "step": 23040 }, { "epoch": 1.5463575047817186, "grad_norm": 0.9930167198181152, "learning_rate": 1.2891037032943676e-05, "loss": 0.51, "step": 23042 }, { "epoch": 1.5464917284654878, "grad_norm": 0.9952108860015869, "learning_rate": 1.288375395523907e-05, "loss": 0.4509, "step": 23044 }, { "epoch": 1.5466259521492567, "grad_norm": 1.1315922737121582, "learning_rate": 1.2876472631204095e-05, "loss": 0.5316, "step": 23046 }, { "epoch": 1.5467601758330258, "grad_norm": 1.0469512939453125, "learning_rate": 1.2869193061182794e-05, "loss": 0.492, "step": 23048 }, { "epoch": 1.5468943995167947, "grad_norm": 1.2232224941253662, "learning_rate": 1.286191524551909e-05, "loss": 0.4547, "step": 23050 }, { "epoch": 1.5470286232005637, "grad_norm": 1.02517831325531, "learning_rate": 1.2854639184556877e-05, "loss": 0.4744, "step": 23052 }, { "epoch": 1.5471628468843326, "grad_norm": 1.056570053100586, "learning_rate": 1.2847364878639905e-05, "loss": 0.5329, "step": 23054 }, { "epoch": 1.5472970705681017, "grad_norm": 0.9619842171669006, "learning_rate": 1.2840092328111903e-05, "loss": 0.5317, "step": 23056 }, { "epoch": 1.5474312942518709, "grad_norm": 1.1084604263305664, "learning_rate": 1.2832821533316464e-05, "loss": 0.4555, "step": 23058 }, { "epoch": 1.5475655179356398, "grad_norm": 0.9658335447311401, "learning_rate": 1.282555249459712e-05, "loss": 0.5221, "step": 23060 }, { "epoch": 1.5476997416194087, "grad_norm": 0.9604290127754211, "learning_rate": 1.2818285212297303e-05, "loss": 0.4113, "step": 23062 }, { "epoch": 1.5478339653031776, "grad_norm": 1.0296306610107422, "learning_rate": 1.2811019686760412e-05, "loss": 0.4564, "step": 23064 }, { "epoch": 1.5479681889869468, "grad_norm": 1.0469577312469482, "learning_rate": 1.2803755918329712e-05, "loss": 0.4442, "step": 23066 }, { "epoch": 1.548102412670716, "grad_norm": 1.0704536437988281, "learning_rate": 1.2796493907348406e-05, "loss": 0.5484, "step": 23068 }, { "epoch": 1.5482366363544848, "grad_norm": 0.9309592247009277, "learning_rate": 1.2789233654159582e-05, "loss": 0.4486, "step": 23070 }, { "epoch": 1.5483708600382537, "grad_norm": 1.0256258249282837, "learning_rate": 1.2781975159106319e-05, "loss": 0.476, "step": 23072 }, { "epoch": 1.5485050837220227, "grad_norm": 0.9599533081054688, "learning_rate": 1.277471842253154e-05, "loss": 0.4906, "step": 23074 }, { "epoch": 1.5486393074057916, "grad_norm": 0.8366543650627136, "learning_rate": 1.2767463444778117e-05, "loss": 0.4661, "step": 23076 }, { "epoch": 1.5487735310895607, "grad_norm": 0.9993776082992554, "learning_rate": 1.276021022618883e-05, "loss": 0.4571, "step": 23078 }, { "epoch": 1.5489077547733299, "grad_norm": 1.1004295349121094, "learning_rate": 1.2752958767106366e-05, "loss": 0.482, "step": 23080 }, { "epoch": 1.5490419784570988, "grad_norm": 0.9002885222434998, "learning_rate": 1.2745709067873369e-05, "loss": 0.4416, "step": 23082 }, { "epoch": 1.5491762021408677, "grad_norm": 1.0409138202667236, "learning_rate": 1.2738461128832353e-05, "loss": 0.468, "step": 23084 }, { "epoch": 1.5493104258246366, "grad_norm": 1.0249680280685425, "learning_rate": 1.2731214950325799e-05, "loss": 0.4723, "step": 23086 }, { "epoch": 1.5494446495084058, "grad_norm": 1.1449722051620483, "learning_rate": 1.2723970532696027e-05, "loss": 0.4917, "step": 23088 }, { "epoch": 1.549578873192175, "grad_norm": 1.0396345853805542, "learning_rate": 1.271672787628536e-05, "loss": 0.4533, "step": 23090 }, { "epoch": 1.5497130968759438, "grad_norm": 0.9743292927742004, "learning_rate": 1.2709486981435976e-05, "loss": 0.457, "step": 23092 }, { "epoch": 1.5498473205597127, "grad_norm": 0.9149103760719299, "learning_rate": 1.2702247848490012e-05, "loss": 0.4378, "step": 23094 }, { "epoch": 1.5499815442434817, "grad_norm": 1.0200872421264648, "learning_rate": 1.2695010477789498e-05, "loss": 0.4957, "step": 23096 }, { "epoch": 1.5501157679272508, "grad_norm": 0.9662615060806274, "learning_rate": 1.2687774869676384e-05, "loss": 0.4914, "step": 23098 }, { "epoch": 1.5502499916110197, "grad_norm": 0.8781123161315918, "learning_rate": 1.268054102449252e-05, "loss": 0.3764, "step": 23100 }, { "epoch": 1.5503842152947889, "grad_norm": 0.9832098484039307, "learning_rate": 1.267330894257972e-05, "loss": 0.4445, "step": 23102 }, { "epoch": 1.5505184389785578, "grad_norm": 1.0590614080429077, "learning_rate": 1.2666078624279676e-05, "loss": 0.4758, "step": 23104 }, { "epoch": 1.5506526626623267, "grad_norm": 0.986400842666626, "learning_rate": 1.2658850069933998e-05, "loss": 0.4434, "step": 23106 }, { "epoch": 1.5507868863460956, "grad_norm": 1.0914992094039917, "learning_rate": 1.2651623279884211e-05, "loss": 0.5054, "step": 23108 }, { "epoch": 1.5509211100298648, "grad_norm": 0.9587206840515137, "learning_rate": 1.2644398254471801e-05, "loss": 0.4642, "step": 23110 }, { "epoch": 1.551055333713634, "grad_norm": 0.964069664478302, "learning_rate": 1.2637174994038104e-05, "loss": 0.4808, "step": 23112 }, { "epoch": 1.5511895573974028, "grad_norm": 0.9514328837394714, "learning_rate": 1.262995349892444e-05, "loss": 0.4808, "step": 23114 }, { "epoch": 1.5513237810811717, "grad_norm": 0.950463056564331, "learning_rate": 1.2622733769471962e-05, "loss": 0.4498, "step": 23116 }, { "epoch": 1.5514580047649407, "grad_norm": 1.109168291091919, "learning_rate": 1.2615515806021827e-05, "loss": 0.4549, "step": 23118 }, { "epoch": 1.5515922284487098, "grad_norm": 0.9191727638244629, "learning_rate": 1.2608299608915047e-05, "loss": 0.492, "step": 23120 }, { "epoch": 1.5517264521324787, "grad_norm": 0.8966755270957947, "learning_rate": 1.2601085178492594e-05, "loss": 0.4732, "step": 23122 }, { "epoch": 1.5518606758162479, "grad_norm": 1.1773089170455933, "learning_rate": 1.2593872515095323e-05, "loss": 0.4923, "step": 23124 }, { "epoch": 1.5519948995000168, "grad_norm": 1.02364182472229, "learning_rate": 1.2586661619064022e-05, "loss": 0.5059, "step": 23126 }, { "epoch": 1.5521291231837857, "grad_norm": 1.1185792684555054, "learning_rate": 1.2579452490739374e-05, "loss": 0.528, "step": 23128 }, { "epoch": 1.5522633468675546, "grad_norm": 0.9567335844039917, "learning_rate": 1.257224513046203e-05, "loss": 0.4232, "step": 23130 }, { "epoch": 1.5523975705513238, "grad_norm": 1.0627226829528809, "learning_rate": 1.2565039538572499e-05, "loss": 0.5006, "step": 23132 }, { "epoch": 1.552531794235093, "grad_norm": 0.9607114791870117, "learning_rate": 1.2557835715411237e-05, "loss": 0.5347, "step": 23134 }, { "epoch": 1.5526660179188618, "grad_norm": 1.0322315692901611, "learning_rate": 1.25506336613186e-05, "loss": 0.5135, "step": 23136 }, { "epoch": 1.5528002416026307, "grad_norm": 1.1543500423431396, "learning_rate": 1.2543433376634894e-05, "loss": 0.5045, "step": 23138 }, { "epoch": 1.5529344652863997, "grad_norm": 1.1164662837982178, "learning_rate": 1.2536234861700307e-05, "loss": 0.5522, "step": 23140 }, { "epoch": 1.5530686889701688, "grad_norm": 1.0598642826080322, "learning_rate": 1.2529038116854951e-05, "loss": 0.4391, "step": 23142 }, { "epoch": 1.553202912653938, "grad_norm": 0.9816542267799377, "learning_rate": 1.2521843142438867e-05, "loss": 0.4365, "step": 23144 }, { "epoch": 1.5533371363377069, "grad_norm": 0.9278079867362976, "learning_rate": 1.2514649938791978e-05, "loss": 0.4136, "step": 23146 }, { "epoch": 1.5534713600214758, "grad_norm": 0.929638147354126, "learning_rate": 1.2507458506254178e-05, "loss": 0.4902, "step": 23148 }, { "epoch": 1.5536055837052447, "grad_norm": 0.9098523259162903, "learning_rate": 1.250026884516523e-05, "loss": 0.4483, "step": 23150 }, { "epoch": 1.5537398073890136, "grad_norm": 1.2403209209442139, "learning_rate": 1.249308095586486e-05, "loss": 0.5198, "step": 23152 }, { "epoch": 1.5538740310727828, "grad_norm": 0.8855780363082886, "learning_rate": 1.2485894838692641e-05, "loss": 0.4567, "step": 23154 }, { "epoch": 1.554008254756552, "grad_norm": 1.0350875854492188, "learning_rate": 1.2478710493988133e-05, "loss": 0.4493, "step": 23156 }, { "epoch": 1.5541424784403208, "grad_norm": 0.9332207441329956, "learning_rate": 1.2471527922090754e-05, "loss": 0.5363, "step": 23158 }, { "epoch": 1.5542767021240897, "grad_norm": 0.9449737668037415, "learning_rate": 1.2464347123339904e-05, "loss": 0.5492, "step": 23160 }, { "epoch": 1.5544109258078587, "grad_norm": 1.0071675777435303, "learning_rate": 1.2457168098074839e-05, "loss": 0.5195, "step": 23162 }, { "epoch": 1.5545451494916278, "grad_norm": 1.085398554801941, "learning_rate": 1.2449990846634757e-05, "loss": 0.4877, "step": 23164 }, { "epoch": 1.554679373175397, "grad_norm": 1.099508285522461, "learning_rate": 1.244281536935875e-05, "loss": 0.5213, "step": 23166 }, { "epoch": 1.5548135968591659, "grad_norm": 1.03630793094635, "learning_rate": 1.2435641666585884e-05, "loss": 0.4711, "step": 23168 }, { "epoch": 1.5549478205429348, "grad_norm": 0.9316409826278687, "learning_rate": 1.242846973865508e-05, "loss": 0.4284, "step": 23170 }, { "epoch": 1.5550820442267037, "grad_norm": 0.9820182919502258, "learning_rate": 1.24212995859052e-05, "loss": 0.504, "step": 23172 }, { "epoch": 1.5552162679104728, "grad_norm": 1.2011853456497192, "learning_rate": 1.2414131208675006e-05, "loss": 0.5168, "step": 23174 }, { "epoch": 1.5553504915942418, "grad_norm": 0.9982286691665649, "learning_rate": 1.2406964607303212e-05, "loss": 0.4828, "step": 23176 }, { "epoch": 1.555484715278011, "grad_norm": 1.0185546875, "learning_rate": 1.2399799782128407e-05, "loss": 0.4556, "step": 23178 }, { "epoch": 1.5556189389617798, "grad_norm": 0.9433539509773254, "learning_rate": 1.2392636733489155e-05, "loss": 0.5052, "step": 23180 }, { "epoch": 1.5557531626455487, "grad_norm": 1.020642876625061, "learning_rate": 1.2385475461723833e-05, "loss": 0.5012, "step": 23182 }, { "epoch": 1.5558873863293177, "grad_norm": 0.9557784795761108, "learning_rate": 1.2378315967170845e-05, "loss": 0.4644, "step": 23184 }, { "epoch": 1.5560216100130868, "grad_norm": 1.135974645614624, "learning_rate": 1.2371158250168436e-05, "loss": 0.5054, "step": 23186 }, { "epoch": 1.556155833696856, "grad_norm": 0.9932438731193542, "learning_rate": 1.2364002311054818e-05, "loss": 0.4858, "step": 23188 }, { "epoch": 1.5562900573806249, "grad_norm": 1.077278971672058, "learning_rate": 1.2356848150168082e-05, "loss": 0.4898, "step": 23190 }, { "epoch": 1.5564242810643938, "grad_norm": 1.1726289987564087, "learning_rate": 1.2349695767846247e-05, "loss": 0.491, "step": 23192 }, { "epoch": 1.5565585047481627, "grad_norm": 1.0936367511749268, "learning_rate": 1.2342545164427244e-05, "loss": 0.4557, "step": 23194 }, { "epoch": 1.5566927284319318, "grad_norm": 1.0036332607269287, "learning_rate": 1.233539634024894e-05, "loss": 0.4511, "step": 23196 }, { "epoch": 1.5568269521157008, "grad_norm": 1.285355806350708, "learning_rate": 1.2328249295649097e-05, "loss": 0.4976, "step": 23198 }, { "epoch": 1.55696117579947, "grad_norm": 1.0038576126098633, "learning_rate": 1.232110403096539e-05, "loss": 0.4419, "step": 23200 }, { "epoch": 1.5570953994832388, "grad_norm": 0.9209941029548645, "learning_rate": 1.2313960546535419e-05, "loss": 0.4389, "step": 23202 }, { "epoch": 1.5572296231670077, "grad_norm": 1.1058903932571411, "learning_rate": 1.2306818842696716e-05, "loss": 0.51, "step": 23204 }, { "epoch": 1.5573638468507767, "grad_norm": 1.073699712753296, "learning_rate": 1.2299678919786689e-05, "loss": 0.5003, "step": 23206 }, { "epoch": 1.5574980705345458, "grad_norm": 1.0794081687927246, "learning_rate": 1.229254077814273e-05, "loss": 0.5043, "step": 23208 }, { "epoch": 1.557632294218315, "grad_norm": 0.9014645218849182, "learning_rate": 1.2285404418102053e-05, "loss": 0.4542, "step": 23210 }, { "epoch": 1.5577665179020839, "grad_norm": 0.9933221936225891, "learning_rate": 1.2278269840001838e-05, "loss": 0.4572, "step": 23212 }, { "epoch": 1.5579007415858528, "grad_norm": 1.0882283449172974, "learning_rate": 1.2271137044179215e-05, "loss": 0.4931, "step": 23214 }, { "epoch": 1.5580349652696217, "grad_norm": 1.047158122062683, "learning_rate": 1.226400603097116e-05, "loss": 0.4966, "step": 23216 }, { "epoch": 1.5581691889533908, "grad_norm": 1.0260157585144043, "learning_rate": 1.2256876800714633e-05, "loss": 0.4813, "step": 23218 }, { "epoch": 1.55830341263716, "grad_norm": 0.9270790219306946, "learning_rate": 1.2249749353746454e-05, "loss": 0.4752, "step": 23220 }, { "epoch": 1.558437636320929, "grad_norm": 0.8672318458557129, "learning_rate": 1.2242623690403387e-05, "loss": 0.4314, "step": 23222 }, { "epoch": 1.5585718600046978, "grad_norm": 0.9131848216056824, "learning_rate": 1.223549981102209e-05, "loss": 0.4286, "step": 23224 }, { "epoch": 1.5587060836884667, "grad_norm": 1.1069892644882202, "learning_rate": 1.2228377715939183e-05, "loss": 0.4977, "step": 23226 }, { "epoch": 1.5588403073722357, "grad_norm": 1.1032196283340454, "learning_rate": 1.2221257405491148e-05, "loss": 0.4791, "step": 23228 }, { "epoch": 1.5589745310560048, "grad_norm": 1.0574241876602173, "learning_rate": 1.2214138880014414e-05, "loss": 0.5097, "step": 23230 }, { "epoch": 1.559108754739774, "grad_norm": 0.9770507216453552, "learning_rate": 1.2207022139845303e-05, "loss": 0.4435, "step": 23232 }, { "epoch": 1.5592429784235429, "grad_norm": 1.1765968799591064, "learning_rate": 1.2199907185320091e-05, "loss": 0.4514, "step": 23234 }, { "epoch": 1.5593772021073118, "grad_norm": 1.059708595275879, "learning_rate": 1.2192794016774922e-05, "loss": 0.3983, "step": 23236 }, { "epoch": 1.5595114257910807, "grad_norm": 1.7546675205230713, "learning_rate": 1.218568263454592e-05, "loss": 0.4223, "step": 23238 }, { "epoch": 1.5596456494748498, "grad_norm": 1.036199927330017, "learning_rate": 1.2178573038969026e-05, "loss": 0.4571, "step": 23240 }, { "epoch": 1.559779873158619, "grad_norm": 1.1719951629638672, "learning_rate": 1.2171465230380196e-05, "loss": 0.4786, "step": 23242 }, { "epoch": 1.559914096842388, "grad_norm": 0.9336151480674744, "learning_rate": 1.2164359209115234e-05, "loss": 0.45, "step": 23244 }, { "epoch": 1.5600483205261568, "grad_norm": 0.9916703104972839, "learning_rate": 1.2157254975509912e-05, "loss": 0.4858, "step": 23246 }, { "epoch": 1.5601825442099257, "grad_norm": 0.9358370304107666, "learning_rate": 1.2150152529899878e-05, "loss": 0.4647, "step": 23248 }, { "epoch": 1.560316767893695, "grad_norm": 1.0746142864227295, "learning_rate": 1.2143051872620708e-05, "loss": 0.4617, "step": 23250 }, { "epoch": 1.5604509915774638, "grad_norm": 1.036489486694336, "learning_rate": 1.2135953004007882e-05, "loss": 0.486, "step": 23252 }, { "epoch": 1.560585215261233, "grad_norm": 1.023247241973877, "learning_rate": 1.2128855924396837e-05, "loss": 0.5029, "step": 23254 }, { "epoch": 1.5607194389450019, "grad_norm": 0.9809913039207458, "learning_rate": 1.2121760634122875e-05, "loss": 0.4899, "step": 23256 }, { "epoch": 1.5608536626287708, "grad_norm": 0.9748601317405701, "learning_rate": 1.211466713352124e-05, "loss": 0.4277, "step": 23258 }, { "epoch": 1.5609878863125397, "grad_norm": 1.083095908164978, "learning_rate": 1.210757542292707e-05, "loss": 0.5058, "step": 23260 }, { "epoch": 1.5611221099963088, "grad_norm": 1.1028939485549927, "learning_rate": 1.2100485502675463e-05, "loss": 0.5041, "step": 23262 }, { "epoch": 1.561256333680078, "grad_norm": 1.063094139099121, "learning_rate": 1.2093397373101373e-05, "loss": 0.5128, "step": 23264 }, { "epoch": 1.561390557363847, "grad_norm": 1.0166436433792114, "learning_rate": 1.2086311034539749e-05, "loss": 0.5278, "step": 23266 }, { "epoch": 1.5615247810476158, "grad_norm": 1.1538267135620117, "learning_rate": 1.2079226487325347e-05, "loss": 0.4639, "step": 23268 }, { "epoch": 1.5616590047313847, "grad_norm": 1.1826285123825073, "learning_rate": 1.2072143731792934e-05, "loss": 0.4692, "step": 23270 }, { "epoch": 1.561793228415154, "grad_norm": 0.9804865717887878, "learning_rate": 1.2065062768277135e-05, "loss": 0.4773, "step": 23272 }, { "epoch": 1.5619274520989228, "grad_norm": 1.1251349449157715, "learning_rate": 1.2057983597112543e-05, "loss": 0.4574, "step": 23274 }, { "epoch": 1.562061675782692, "grad_norm": 1.0487046241760254, "learning_rate": 1.2050906218633617e-05, "loss": 0.5116, "step": 23276 }, { "epoch": 1.5621958994664609, "grad_norm": 1.1605690717697144, "learning_rate": 1.2043830633174747e-05, "loss": 0.4846, "step": 23278 }, { "epoch": 1.5623301231502298, "grad_norm": 0.9828898310661316, "learning_rate": 1.203675684107024e-05, "loss": 0.4914, "step": 23280 }, { "epoch": 1.5624643468339987, "grad_norm": 1.118901252746582, "learning_rate": 1.2029684842654304e-05, "loss": 0.4927, "step": 23282 }, { "epoch": 1.5625985705177678, "grad_norm": 1.0152097940444946, "learning_rate": 1.2022614638261109e-05, "loss": 0.5329, "step": 23284 }, { "epoch": 1.562732794201537, "grad_norm": 1.0980987548828125, "learning_rate": 1.20155462282247e-05, "loss": 0.4771, "step": 23286 }, { "epoch": 1.562867017885306, "grad_norm": 1.060423731803894, "learning_rate": 1.2008479612879032e-05, "loss": 0.5249, "step": 23288 }, { "epoch": 1.5630012415690748, "grad_norm": 1.042399287223816, "learning_rate": 1.200141479255798e-05, "loss": 0.5566, "step": 23290 }, { "epoch": 1.5631354652528437, "grad_norm": 1.145865797996521, "learning_rate": 1.1994351767595374e-05, "loss": 0.5369, "step": 23292 }, { "epoch": 1.5632696889366129, "grad_norm": 1.0151365995407104, "learning_rate": 1.1987290538324914e-05, "loss": 0.4685, "step": 23294 }, { "epoch": 1.563403912620382, "grad_norm": 1.0187491178512573, "learning_rate": 1.1980231105080225e-05, "loss": 0.5318, "step": 23296 }, { "epoch": 1.563538136304151, "grad_norm": 0.933469295501709, "learning_rate": 1.1973173468194843e-05, "loss": 0.4409, "step": 23298 }, { "epoch": 1.5636723599879199, "grad_norm": 1.0887510776519775, "learning_rate": 1.1966117628002249e-05, "loss": 0.521, "step": 23300 }, { "epoch": 1.5638065836716888, "grad_norm": 0.9163106083869934, "learning_rate": 1.1959063584835795e-05, "loss": 0.4308, "step": 23302 }, { "epoch": 1.5639408073554577, "grad_norm": 0.9544594287872314, "learning_rate": 1.1952011339028807e-05, "loss": 0.48, "step": 23304 }, { "epoch": 1.5640750310392268, "grad_norm": 1.0309133529663086, "learning_rate": 1.1944960890914442e-05, "loss": 0.4892, "step": 23306 }, { "epoch": 1.564209254722996, "grad_norm": 1.001133918762207, "learning_rate": 1.1937912240825849e-05, "loss": 0.4814, "step": 23308 }, { "epoch": 1.564343478406765, "grad_norm": 0.9937286972999573, "learning_rate": 1.1930865389096052e-05, "loss": 0.5181, "step": 23310 }, { "epoch": 1.5644777020905338, "grad_norm": 0.9627927541732788, "learning_rate": 1.1923820336058012e-05, "loss": 0.4482, "step": 23312 }, { "epoch": 1.5646119257743027, "grad_norm": 1.146790623664856, "learning_rate": 1.1916777082044588e-05, "loss": 0.4795, "step": 23314 }, { "epoch": 1.5647461494580719, "grad_norm": 1.0631041526794434, "learning_rate": 1.1909735627388562e-05, "loss": 0.4174, "step": 23316 }, { "epoch": 1.564880373141841, "grad_norm": 1.1096659898757935, "learning_rate": 1.1902695972422618e-05, "loss": 0.4753, "step": 23318 }, { "epoch": 1.56501459682561, "grad_norm": 1.0855131149291992, "learning_rate": 1.189565811747938e-05, "loss": 0.5567, "step": 23320 }, { "epoch": 1.5651488205093789, "grad_norm": 1.2113926410675049, "learning_rate": 1.188862206289137e-05, "loss": 0.4981, "step": 23322 }, { "epoch": 1.5652830441931478, "grad_norm": 0.9407163858413696, "learning_rate": 1.1881587808991023e-05, "loss": 0.5004, "step": 23324 }, { "epoch": 1.565417267876917, "grad_norm": 1.029340386390686, "learning_rate": 1.1874555356110684e-05, "loss": 0.5376, "step": 23326 }, { "epoch": 1.5655514915606858, "grad_norm": 1.0241527557373047, "learning_rate": 1.1867524704582644e-05, "loss": 0.4501, "step": 23328 }, { "epoch": 1.565685715244455, "grad_norm": 1.0108983516693115, "learning_rate": 1.1860495854739067e-05, "loss": 0.4492, "step": 23330 }, { "epoch": 1.565819938928224, "grad_norm": 1.048672080039978, "learning_rate": 1.1853468806912088e-05, "loss": 0.4283, "step": 23332 }, { "epoch": 1.5659541626119928, "grad_norm": 1.066028356552124, "learning_rate": 1.1846443561433673e-05, "loss": 0.427, "step": 23334 }, { "epoch": 1.5660883862957617, "grad_norm": 0.9435437917709351, "learning_rate": 1.1839420118635785e-05, "loss": 0.4447, "step": 23336 }, { "epoch": 1.5662226099795309, "grad_norm": 0.912183940410614, "learning_rate": 1.1832398478850243e-05, "loss": 0.4356, "step": 23338 }, { "epoch": 1.5663568336633, "grad_norm": 0.9043480157852173, "learning_rate": 1.182537864240883e-05, "loss": 0.4056, "step": 23340 }, { "epoch": 1.566491057347069, "grad_norm": 0.9847351312637329, "learning_rate": 1.1818360609643215e-05, "loss": 0.4958, "step": 23342 }, { "epoch": 1.5666252810308379, "grad_norm": 1.0925720930099487, "learning_rate": 1.1811344380884975e-05, "loss": 0.5238, "step": 23344 }, { "epoch": 1.5667595047146068, "grad_norm": 1.146865963935852, "learning_rate": 1.180432995646562e-05, "loss": 0.4531, "step": 23346 }, { "epoch": 1.566893728398376, "grad_norm": 0.9299564361572266, "learning_rate": 1.1797317336716552e-05, "loss": 0.5076, "step": 23348 }, { "epoch": 1.5670279520821448, "grad_norm": 1.033992886543274, "learning_rate": 1.179030652196913e-05, "loss": 0.4787, "step": 23350 }, { "epoch": 1.567162175765914, "grad_norm": 1.011711835861206, "learning_rate": 1.1783297512554591e-05, "loss": 0.5177, "step": 23352 }, { "epoch": 1.567296399449683, "grad_norm": 0.9572144746780396, "learning_rate": 1.1776290308804095e-05, "loss": 0.4684, "step": 23354 }, { "epoch": 1.5674306231334518, "grad_norm": 1.0200315713882446, "learning_rate": 1.1769284911048706e-05, "loss": 0.4596, "step": 23356 }, { "epoch": 1.5675648468172207, "grad_norm": 0.9919617772102356, "learning_rate": 1.176228131961944e-05, "loss": 0.451, "step": 23358 }, { "epoch": 1.5676990705009899, "grad_norm": 1.0359539985656738, "learning_rate": 1.1755279534847186e-05, "loss": 0.4625, "step": 23360 }, { "epoch": 1.567833294184759, "grad_norm": 1.0746967792510986, "learning_rate": 1.1748279557062775e-05, "loss": 0.4682, "step": 23362 }, { "epoch": 1.567967517868528, "grad_norm": 1.1900519132614136, "learning_rate": 1.1741281386596913e-05, "loss": 0.4418, "step": 23364 }, { "epoch": 1.5681017415522969, "grad_norm": 0.9637075066566467, "learning_rate": 1.1734285023780295e-05, "loss": 0.4454, "step": 23366 }, { "epoch": 1.5682359652360658, "grad_norm": 1.0022313594818115, "learning_rate": 1.1727290468943447e-05, "loss": 0.4795, "step": 23368 }, { "epoch": 1.568370188919835, "grad_norm": 1.1569160223007202, "learning_rate": 1.172029772241688e-05, "loss": 0.5055, "step": 23370 }, { "epoch": 1.568504412603604, "grad_norm": 0.9874303936958313, "learning_rate": 1.171330678453097e-05, "loss": 0.4996, "step": 23372 }, { "epoch": 1.568638636287373, "grad_norm": 1.0189257860183716, "learning_rate": 1.1706317655616029e-05, "loss": 0.427, "step": 23374 }, { "epoch": 1.568772859971142, "grad_norm": 0.8920593857765198, "learning_rate": 1.1699330336002261e-05, "loss": 0.4697, "step": 23376 }, { "epoch": 1.5689070836549108, "grad_norm": 1.0679326057434082, "learning_rate": 1.1692344826019835e-05, "loss": 0.4625, "step": 23378 }, { "epoch": 1.5690413073386797, "grad_norm": 1.0527321100234985, "learning_rate": 1.168536112599879e-05, "loss": 0.5308, "step": 23380 }, { "epoch": 1.5691755310224489, "grad_norm": 1.0946030616760254, "learning_rate": 1.1678379236269082e-05, "loss": 0.4846, "step": 23382 }, { "epoch": 1.569309754706218, "grad_norm": 1.0754079818725586, "learning_rate": 1.1671399157160589e-05, "loss": 0.417, "step": 23384 }, { "epoch": 1.569443978389987, "grad_norm": 1.1126391887664795, "learning_rate": 1.1664420889003131e-05, "loss": 0.4788, "step": 23386 }, { "epoch": 1.5695782020737559, "grad_norm": 0.9987939596176147, "learning_rate": 1.1657444432126403e-05, "loss": 0.4408, "step": 23388 }, { "epoch": 1.5697124257575248, "grad_norm": 1.1336610317230225, "learning_rate": 1.1650469786860025e-05, "loss": 0.5384, "step": 23390 }, { "epoch": 1.569846649441294, "grad_norm": 0.9724118709564209, "learning_rate": 1.1643496953533522e-05, "loss": 0.459, "step": 23392 }, { "epoch": 1.569980873125063, "grad_norm": 0.9910349249839783, "learning_rate": 1.1636525932476382e-05, "loss": 0.4318, "step": 23394 }, { "epoch": 1.570115096808832, "grad_norm": 0.9914292693138123, "learning_rate": 1.1629556724017936e-05, "loss": 0.4409, "step": 23396 }, { "epoch": 1.570249320492601, "grad_norm": 0.9634431600570679, "learning_rate": 1.1622589328487504e-05, "loss": 0.5045, "step": 23398 }, { "epoch": 1.5703835441763698, "grad_norm": 1.0638256072998047, "learning_rate": 1.1615623746214255e-05, "loss": 0.4842, "step": 23400 }, { "epoch": 1.570517767860139, "grad_norm": 0.990250825881958, "learning_rate": 1.1608659977527303e-05, "loss": 0.4729, "step": 23402 }, { "epoch": 1.5706519915439079, "grad_norm": 0.9492442607879639, "learning_rate": 1.1601698022755658e-05, "loss": 0.4816, "step": 23404 }, { "epoch": 1.570786215227677, "grad_norm": 1.1385966539382935, "learning_rate": 1.159473788222829e-05, "loss": 0.4708, "step": 23406 }, { "epoch": 1.570920438911446, "grad_norm": 1.0014199018478394, "learning_rate": 1.1587779556274042e-05, "loss": 0.4839, "step": 23408 }, { "epoch": 1.5710546625952149, "grad_norm": 0.9885188341140747, "learning_rate": 1.1580823045221673e-05, "loss": 0.4584, "step": 23410 }, { "epoch": 1.5711888862789838, "grad_norm": 0.9447250366210938, "learning_rate": 1.1573868349399848e-05, "loss": 0.4484, "step": 23412 }, { "epoch": 1.571323109962753, "grad_norm": 0.9757140278816223, "learning_rate": 1.1566915469137201e-05, "loss": 0.4863, "step": 23414 }, { "epoch": 1.571457333646522, "grad_norm": 1.0615960359573364, "learning_rate": 1.155996440476222e-05, "loss": 0.4747, "step": 23416 }, { "epoch": 1.571591557330291, "grad_norm": 0.9987040162086487, "learning_rate": 1.1553015156603331e-05, "loss": 0.4632, "step": 23418 }, { "epoch": 1.57172578101406, "grad_norm": 1.0166984796524048, "learning_rate": 1.1546067724988873e-05, "loss": 0.4359, "step": 23420 }, { "epoch": 1.5718600046978288, "grad_norm": 0.9812948107719421, "learning_rate": 1.1539122110247086e-05, "loss": 0.4624, "step": 23422 }, { "epoch": 1.571994228381598, "grad_norm": 1.099785327911377, "learning_rate": 1.1532178312706166e-05, "loss": 0.4843, "step": 23424 }, { "epoch": 1.5721284520653669, "grad_norm": 1.042169213294983, "learning_rate": 1.1525236332694162e-05, "loss": 0.5418, "step": 23426 }, { "epoch": 1.572262675749136, "grad_norm": 1.0316282510757446, "learning_rate": 1.1518296170539105e-05, "loss": 0.4221, "step": 23428 }, { "epoch": 1.572396899432905, "grad_norm": 1.0206512212753296, "learning_rate": 1.1511357826568864e-05, "loss": 0.4744, "step": 23430 }, { "epoch": 1.5725311231166739, "grad_norm": 1.0751756429672241, "learning_rate": 1.150442130111129e-05, "loss": 0.532, "step": 23432 }, { "epoch": 1.5726653468004428, "grad_norm": 1.0585800409317017, "learning_rate": 1.14974865944941e-05, "loss": 0.4456, "step": 23434 }, { "epoch": 1.572799570484212, "grad_norm": 1.1338976621627808, "learning_rate": 1.1490553707044965e-05, "loss": 0.5337, "step": 23436 }, { "epoch": 1.572933794167981, "grad_norm": 0.9720898270606995, "learning_rate": 1.1483622639091446e-05, "loss": 0.5049, "step": 23438 }, { "epoch": 1.57306801785175, "grad_norm": 1.065352439880371, "learning_rate": 1.1476693390961019e-05, "loss": 0.4294, "step": 23440 }, { "epoch": 1.573202241535519, "grad_norm": 0.9880411028862, "learning_rate": 1.146976596298106e-05, "loss": 0.4533, "step": 23442 }, { "epoch": 1.5733364652192878, "grad_norm": 1.2746909856796265, "learning_rate": 1.146284035547891e-05, "loss": 0.4535, "step": 23444 }, { "epoch": 1.573470688903057, "grad_norm": 0.8341091275215149, "learning_rate": 1.1455916568781772e-05, "loss": 0.4299, "step": 23446 }, { "epoch": 1.573604912586826, "grad_norm": 1.156506061553955, "learning_rate": 1.1448994603216779e-05, "loss": 0.4808, "step": 23448 }, { "epoch": 1.573739136270595, "grad_norm": 1.1684461832046509, "learning_rate": 1.1442074459110974e-05, "loss": 0.4736, "step": 23450 }, { "epoch": 1.573873359954364, "grad_norm": 1.2712520360946655, "learning_rate": 1.1435156136791342e-05, "loss": 0.5259, "step": 23452 }, { "epoch": 1.5740075836381329, "grad_norm": 1.3483922481536865, "learning_rate": 1.1428239636584732e-05, "loss": 0.509, "step": 23454 }, { "epoch": 1.5741418073219018, "grad_norm": 1.0553919076919556, "learning_rate": 1.142132495881798e-05, "loss": 0.4452, "step": 23456 }, { "epoch": 1.574276031005671, "grad_norm": 1.154910683631897, "learning_rate": 1.141441210381774e-05, "loss": 0.5168, "step": 23458 }, { "epoch": 1.57441025468944, "grad_norm": 0.8882455229759216, "learning_rate": 1.140750107191066e-05, "loss": 0.4366, "step": 23460 }, { "epoch": 1.574544478373209, "grad_norm": 1.1063082218170166, "learning_rate": 1.1400591863423254e-05, "loss": 0.5155, "step": 23462 }, { "epoch": 1.574678702056978, "grad_norm": 1.195164442062378, "learning_rate": 1.1393684478681993e-05, "loss": 0.4859, "step": 23464 }, { "epoch": 1.5748129257407468, "grad_norm": 1.1390354633331299, "learning_rate": 1.1386778918013225e-05, "loss": 0.5171, "step": 23466 }, { "epoch": 1.574947149424516, "grad_norm": 1.1175415515899658, "learning_rate": 1.1379875181743222e-05, "loss": 0.477, "step": 23468 }, { "epoch": 1.575081373108285, "grad_norm": 1.0612244606018066, "learning_rate": 1.1372973270198163e-05, "loss": 0.5073, "step": 23470 }, { "epoch": 1.575215596792054, "grad_norm": 1.0456383228302002, "learning_rate": 1.1366073183704167e-05, "loss": 0.4893, "step": 23472 }, { "epoch": 1.575349820475823, "grad_norm": 1.1216797828674316, "learning_rate": 1.1359174922587251e-05, "loss": 0.5365, "step": 23474 }, { "epoch": 1.5754840441595919, "grad_norm": 0.8650084137916565, "learning_rate": 1.1352278487173328e-05, "loss": 0.4372, "step": 23476 }, { "epoch": 1.575618267843361, "grad_norm": 1.2271318435668945, "learning_rate": 1.1345383877788241e-05, "loss": 0.4567, "step": 23478 }, { "epoch": 1.57575249152713, "grad_norm": 1.0574833154678345, "learning_rate": 1.133849109475777e-05, "loss": 0.5015, "step": 23480 }, { "epoch": 1.575886715210899, "grad_norm": 1.1210253238677979, "learning_rate": 1.133160013840755e-05, "loss": 0.4641, "step": 23482 }, { "epoch": 1.576020938894668, "grad_norm": 0.9176300168037415, "learning_rate": 1.132471100906322e-05, "loss": 0.4495, "step": 23484 }, { "epoch": 1.576155162578437, "grad_norm": 1.1437222957611084, "learning_rate": 1.1317823707050224e-05, "loss": 0.5005, "step": 23486 }, { "epoch": 1.5762893862622058, "grad_norm": 0.9332422018051147, "learning_rate": 1.131093823269398e-05, "loss": 0.4629, "step": 23488 }, { "epoch": 1.576423609945975, "grad_norm": 1.0111188888549805, "learning_rate": 1.130405458631984e-05, "loss": 0.4381, "step": 23490 }, { "epoch": 1.576557833629744, "grad_norm": 1.0104448795318604, "learning_rate": 1.1297172768253018e-05, "loss": 0.4717, "step": 23492 }, { "epoch": 1.576692057313513, "grad_norm": 1.1603244543075562, "learning_rate": 1.1290292778818684e-05, "loss": 0.4849, "step": 23494 }, { "epoch": 1.576826280997282, "grad_norm": 1.1314606666564941, "learning_rate": 1.1283414618341903e-05, "loss": 0.5394, "step": 23496 }, { "epoch": 1.5769605046810509, "grad_norm": 1.1823599338531494, "learning_rate": 1.1276538287147642e-05, "loss": 0.4741, "step": 23498 }, { "epoch": 1.57709472836482, "grad_norm": 1.140844702720642, "learning_rate": 1.1269663785560792e-05, "loss": 0.5258, "step": 23500 }, { "epoch": 1.577228952048589, "grad_norm": 1.1149667501449585, "learning_rate": 1.1262791113906179e-05, "loss": 0.5388, "step": 23502 }, { "epoch": 1.577363175732358, "grad_norm": 0.9421783089637756, "learning_rate": 1.1255920272508519e-05, "loss": 0.4208, "step": 23504 }, { "epoch": 1.577497399416127, "grad_norm": 1.1093504428863525, "learning_rate": 1.1249051261692433e-05, "loss": 0.511, "step": 23506 }, { "epoch": 1.577631623099896, "grad_norm": 1.1066049337387085, "learning_rate": 1.124218408178246e-05, "loss": 0.4689, "step": 23508 }, { "epoch": 1.5777658467836648, "grad_norm": 1.0475133657455444, "learning_rate": 1.1235318733103089e-05, "loss": 0.5488, "step": 23510 }, { "epoch": 1.577900070467434, "grad_norm": 1.338498830795288, "learning_rate": 1.1228455215978683e-05, "loss": 0.448, "step": 23512 }, { "epoch": 1.578034294151203, "grad_norm": 0.8669447302818298, "learning_rate": 1.1221593530733527e-05, "loss": 0.4772, "step": 23514 }, { "epoch": 1.578168517834972, "grad_norm": 1.027613878250122, "learning_rate": 1.1214733677691808e-05, "loss": 0.5029, "step": 23516 }, { "epoch": 1.578302741518741, "grad_norm": 0.9990667700767517, "learning_rate": 1.1207875657177663e-05, "loss": 0.4554, "step": 23518 }, { "epoch": 1.5784369652025099, "grad_norm": 1.090484380722046, "learning_rate": 1.1201019469515106e-05, "loss": 0.4944, "step": 23520 }, { "epoch": 1.578571188886279, "grad_norm": 0.8236868381500244, "learning_rate": 1.1194165115028094e-05, "loss": 0.453, "step": 23522 }, { "epoch": 1.5787054125700481, "grad_norm": 0.9789796471595764, "learning_rate": 1.1187312594040472e-05, "loss": 0.5029, "step": 23524 }, { "epoch": 1.578839636253817, "grad_norm": 0.8634899258613586, "learning_rate": 1.118046190687601e-05, "loss": 0.4639, "step": 23526 }, { "epoch": 1.578973859937586, "grad_norm": 1.1111963987350464, "learning_rate": 1.1173613053858373e-05, "loss": 0.5827, "step": 23528 }, { "epoch": 1.579108083621355, "grad_norm": 0.9517291784286499, "learning_rate": 1.1166766035311182e-05, "loss": 0.4568, "step": 23530 }, { "epoch": 1.5792423073051238, "grad_norm": 1.070487380027771, "learning_rate": 1.1159920851557937e-05, "loss": 0.4755, "step": 23532 }, { "epoch": 1.579376530988893, "grad_norm": 1.080098271369934, "learning_rate": 1.1153077502922055e-05, "loss": 0.4911, "step": 23534 }, { "epoch": 1.579510754672662, "grad_norm": 1.1324330568313599, "learning_rate": 1.1146235989726856e-05, "loss": 0.5065, "step": 23536 }, { "epoch": 1.579644978356431, "grad_norm": 1.0070408582687378, "learning_rate": 1.113939631229562e-05, "loss": 0.4729, "step": 23538 }, { "epoch": 1.5797792020402, "grad_norm": 1.0733522176742554, "learning_rate": 1.1132558470951487e-05, "loss": 0.5337, "step": 23540 }, { "epoch": 1.5799134257239689, "grad_norm": 1.0693459510803223, "learning_rate": 1.1125722466017547e-05, "loss": 0.4989, "step": 23542 }, { "epoch": 1.580047649407738, "grad_norm": 1.1518325805664062, "learning_rate": 1.1118888297816754e-05, "loss": 0.4878, "step": 23544 }, { "epoch": 1.5801818730915071, "grad_norm": 1.1432942152023315, "learning_rate": 1.1112055966672053e-05, "loss": 0.461, "step": 23546 }, { "epoch": 1.580316096775276, "grad_norm": 0.9708120822906494, "learning_rate": 1.1105225472906216e-05, "loss": 0.4498, "step": 23548 }, { "epoch": 1.580450320459045, "grad_norm": 0.8708906173706055, "learning_rate": 1.109839681684201e-05, "loss": 0.4934, "step": 23550 }, { "epoch": 1.580584544142814, "grad_norm": 0.9430227279663086, "learning_rate": 1.1091569998802075e-05, "loss": 0.4438, "step": 23552 }, { "epoch": 1.580718767826583, "grad_norm": 0.990928590297699, "learning_rate": 1.1084745019108916e-05, "loss": 0.4344, "step": 23554 }, { "epoch": 1.580852991510352, "grad_norm": 1.0264774560928345, "learning_rate": 1.1077921878085047e-05, "loss": 0.4769, "step": 23556 }, { "epoch": 1.580987215194121, "grad_norm": 1.0623157024383545, "learning_rate": 1.1071100576052818e-05, "loss": 0.5127, "step": 23558 }, { "epoch": 1.58112143887789, "grad_norm": 1.053694486618042, "learning_rate": 1.1064281113334546e-05, "loss": 0.4593, "step": 23560 }, { "epoch": 1.581255662561659, "grad_norm": 1.0246875286102295, "learning_rate": 1.1057463490252434e-05, "loss": 0.5409, "step": 23562 }, { "epoch": 1.5813898862454279, "grad_norm": 1.1175333261489868, "learning_rate": 1.1050647707128592e-05, "loss": 0.5502, "step": 23564 }, { "epoch": 1.581524109929197, "grad_norm": 0.8779823184013367, "learning_rate": 1.104383376428504e-05, "loss": 0.4051, "step": 23566 }, { "epoch": 1.5816583336129661, "grad_norm": 1.0429855585098267, "learning_rate": 1.1037021662043756e-05, "loss": 0.497, "step": 23568 }, { "epoch": 1.581792557296735, "grad_norm": 1.1017266511917114, "learning_rate": 1.1030211400726576e-05, "loss": 0.4846, "step": 23570 }, { "epoch": 1.581926780980504, "grad_norm": 1.1748697757720947, "learning_rate": 1.1023402980655279e-05, "loss": 0.4775, "step": 23572 }, { "epoch": 1.582061004664273, "grad_norm": 1.116491675376892, "learning_rate": 1.1016596402151536e-05, "loss": 0.5414, "step": 23574 }, { "epoch": 1.582195228348042, "grad_norm": 1.145172357559204, "learning_rate": 1.1009791665536968e-05, "loss": 0.5369, "step": 23576 }, { "epoch": 1.582329452031811, "grad_norm": 1.0083330869674683, "learning_rate": 1.1002988771133054e-05, "loss": 0.4583, "step": 23578 }, { "epoch": 1.58246367571558, "grad_norm": 0.9352924823760986, "learning_rate": 1.0996187719261269e-05, "loss": 0.4951, "step": 23580 }, { "epoch": 1.582597899399349, "grad_norm": 1.0740264654159546, "learning_rate": 1.0989388510242887e-05, "loss": 0.4206, "step": 23582 }, { "epoch": 1.582732123083118, "grad_norm": 1.1314347982406616, "learning_rate": 1.0982591144399201e-05, "loss": 0.451, "step": 23584 }, { "epoch": 1.5828663467668869, "grad_norm": 1.1169883012771606, "learning_rate": 1.097579562205135e-05, "loss": 0.5466, "step": 23586 }, { "epoch": 1.583000570450656, "grad_norm": 1.0791853666305542, "learning_rate": 1.0969001943520424e-05, "loss": 0.4515, "step": 23588 }, { "epoch": 1.5831347941344251, "grad_norm": 1.1795690059661865, "learning_rate": 1.0962210109127407e-05, "loss": 0.4691, "step": 23590 }, { "epoch": 1.583269017818194, "grad_norm": 1.0334551334381104, "learning_rate": 1.0955420119193199e-05, "loss": 0.4492, "step": 23592 }, { "epoch": 1.583403241501963, "grad_norm": 0.8433769941329956, "learning_rate": 1.0948631974038604e-05, "loss": 0.3866, "step": 23594 }, { "epoch": 1.583537465185732, "grad_norm": 1.1193410158157349, "learning_rate": 1.0941845673984364e-05, "loss": 0.4858, "step": 23596 }, { "epoch": 1.583671688869501, "grad_norm": 0.9980464577674866, "learning_rate": 1.0935061219351111e-05, "loss": 0.4543, "step": 23598 }, { "epoch": 1.5838059125532702, "grad_norm": 0.8986864686012268, "learning_rate": 1.0928278610459398e-05, "loss": 0.4081, "step": 23600 }, { "epoch": 1.583940136237039, "grad_norm": 1.094754695892334, "learning_rate": 1.092149784762968e-05, "loss": 0.4822, "step": 23602 }, { "epoch": 1.584074359920808, "grad_norm": 0.8928935527801514, "learning_rate": 1.0914718931182355e-05, "loss": 0.4363, "step": 23604 }, { "epoch": 1.584208583604577, "grad_norm": 1.0659420490264893, "learning_rate": 1.0907941861437688e-05, "loss": 0.503, "step": 23606 }, { "epoch": 1.5843428072883459, "grad_norm": 1.1274199485778809, "learning_rate": 1.0901166638715926e-05, "loss": 0.5279, "step": 23608 }, { "epoch": 1.584477030972115, "grad_norm": 1.0495855808258057, "learning_rate": 1.0894393263337127e-05, "loss": 0.5288, "step": 23610 }, { "epoch": 1.5846112546558841, "grad_norm": 1.0033987760543823, "learning_rate": 1.0887621735621362e-05, "loss": 0.4518, "step": 23612 }, { "epoch": 1.584745478339653, "grad_norm": 0.9234064221382141, "learning_rate": 1.0880852055888547e-05, "loss": 0.4399, "step": 23614 }, { "epoch": 1.584879702023422, "grad_norm": 0.9354954957962036, "learning_rate": 1.0874084224458558e-05, "loss": 0.4442, "step": 23616 }, { "epoch": 1.585013925707191, "grad_norm": 0.978236973285675, "learning_rate": 1.0867318241651154e-05, "loss": 0.4884, "step": 23618 }, { "epoch": 1.58514814939096, "grad_norm": 0.9640846848487854, "learning_rate": 1.0860554107786015e-05, "loss": 0.52, "step": 23620 }, { "epoch": 1.5852823730747292, "grad_norm": 1.090498447418213, "learning_rate": 1.0853791823182723e-05, "loss": 0.4359, "step": 23622 }, { "epoch": 1.585416596758498, "grad_norm": 1.0842543840408325, "learning_rate": 1.0847031388160777e-05, "loss": 0.4861, "step": 23624 }, { "epoch": 1.585550820442267, "grad_norm": 0.9848273396492004, "learning_rate": 1.0840272803039625e-05, "loss": 0.4802, "step": 23626 }, { "epoch": 1.585685044126036, "grad_norm": 0.8913401365280151, "learning_rate": 1.0833516068138577e-05, "loss": 0.3745, "step": 23628 }, { "epoch": 1.585819267809805, "grad_norm": 0.9378396272659302, "learning_rate": 1.0826761183776879e-05, "loss": 0.4543, "step": 23630 }, { "epoch": 1.585953491493574, "grad_norm": 1.1678546667099, "learning_rate": 1.0820008150273669e-05, "loss": 0.4911, "step": 23632 }, { "epoch": 1.5860877151773431, "grad_norm": 1.0119303464889526, "learning_rate": 1.0813256967948044e-05, "loss": 0.4856, "step": 23634 }, { "epoch": 1.586221938861112, "grad_norm": 1.1532948017120361, "learning_rate": 1.0806507637118967e-05, "loss": 0.482, "step": 23636 }, { "epoch": 1.586356162544881, "grad_norm": 1.0109304189682007, "learning_rate": 1.0799760158105337e-05, "loss": 0.5013, "step": 23638 }, { "epoch": 1.58649038622865, "grad_norm": 1.0066771507263184, "learning_rate": 1.0793014531225937e-05, "loss": 0.4438, "step": 23640 }, { "epoch": 1.586624609912419, "grad_norm": 1.0120000839233398, "learning_rate": 1.0786270756799522e-05, "loss": 0.4615, "step": 23642 }, { "epoch": 1.5867588335961882, "grad_norm": 1.0950095653533936, "learning_rate": 1.0779528835144686e-05, "loss": 0.4697, "step": 23644 }, { "epoch": 1.586893057279957, "grad_norm": 0.864580512046814, "learning_rate": 1.0772788766580022e-05, "loss": 0.4016, "step": 23646 }, { "epoch": 1.587027280963726, "grad_norm": 0.970075249671936, "learning_rate": 1.0766050551423917e-05, "loss": 0.5004, "step": 23648 }, { "epoch": 1.587161504647495, "grad_norm": 1.0387160778045654, "learning_rate": 1.075931418999479e-05, "loss": 0.4854, "step": 23650 }, { "epoch": 1.587295728331264, "grad_norm": 0.913539469242096, "learning_rate": 1.0752579682610891e-05, "loss": 0.4409, "step": 23652 }, { "epoch": 1.587429952015033, "grad_norm": 1.9731390476226807, "learning_rate": 1.0745847029590439e-05, "loss": 0.5259, "step": 23654 }, { "epoch": 1.5875641756988021, "grad_norm": 0.968388557434082, "learning_rate": 1.0739116231251523e-05, "loss": 0.4799, "step": 23656 }, { "epoch": 1.587698399382571, "grad_norm": 0.9847589135169983, "learning_rate": 1.0732387287912166e-05, "loss": 0.4816, "step": 23658 }, { "epoch": 1.58783262306634, "grad_norm": 1.0640642642974854, "learning_rate": 1.0725660199890275e-05, "loss": 0.4835, "step": 23660 }, { "epoch": 1.587966846750109, "grad_norm": 0.9867693781852722, "learning_rate": 1.0718934967503724e-05, "loss": 0.4932, "step": 23662 }, { "epoch": 1.588101070433878, "grad_norm": 1.11772882938385, "learning_rate": 1.0712211591070254e-05, "loss": 0.4645, "step": 23664 }, { "epoch": 1.5882352941176472, "grad_norm": 1.0750466585159302, "learning_rate": 1.070549007090753e-05, "loss": 0.4911, "step": 23666 }, { "epoch": 1.588369517801416, "grad_norm": 1.048567771911621, "learning_rate": 1.069877040733312e-05, "loss": 0.4705, "step": 23668 }, { "epoch": 1.588503741485185, "grad_norm": 1.1166553497314453, "learning_rate": 1.0692052600664537e-05, "loss": 0.4695, "step": 23670 }, { "epoch": 1.588637965168954, "grad_norm": 1.2060980796813965, "learning_rate": 1.0685336651219158e-05, "loss": 0.4708, "step": 23672 }, { "epoch": 1.588772188852723, "grad_norm": 1.026139259338379, "learning_rate": 1.0678622559314344e-05, "loss": 0.4734, "step": 23674 }, { "epoch": 1.5889064125364922, "grad_norm": 1.078305721282959, "learning_rate": 1.067191032526726e-05, "loss": 0.4849, "step": 23676 }, { "epoch": 1.5890406362202611, "grad_norm": 0.9637233018875122, "learning_rate": 1.0665199949395093e-05, "loss": 0.4589, "step": 23678 }, { "epoch": 1.58917485990403, "grad_norm": 1.070265531539917, "learning_rate": 1.0658491432014867e-05, "loss": 0.5019, "step": 23680 }, { "epoch": 1.589309083587799, "grad_norm": 1.0106004476547241, "learning_rate": 1.0651784773443573e-05, "loss": 0.4511, "step": 23682 }, { "epoch": 1.589443307271568, "grad_norm": 0.9935443997383118, "learning_rate": 1.0645079973998073e-05, "loss": 0.4821, "step": 23684 }, { "epoch": 1.589577530955337, "grad_norm": 1.0848814249038696, "learning_rate": 1.0638377033995156e-05, "loss": 0.4612, "step": 23686 }, { "epoch": 1.5897117546391062, "grad_norm": 1.0245535373687744, "learning_rate": 1.0631675953751508e-05, "loss": 0.5094, "step": 23688 }, { "epoch": 1.589845978322875, "grad_norm": 1.1181143522262573, "learning_rate": 1.0624976733583775e-05, "loss": 0.4962, "step": 23690 }, { "epoch": 1.589980202006644, "grad_norm": 1.008202314376831, "learning_rate": 1.0618279373808459e-05, "loss": 0.4608, "step": 23692 }, { "epoch": 1.590114425690413, "grad_norm": 1.0486910343170166, "learning_rate": 1.0611583874742004e-05, "loss": 0.5284, "step": 23694 }, { "epoch": 1.590248649374182, "grad_norm": 1.1446926593780518, "learning_rate": 1.0604890236700753e-05, "loss": 0.4721, "step": 23696 }, { "epoch": 1.5903828730579512, "grad_norm": 1.0472633838653564, "learning_rate": 1.0598198460000963e-05, "loss": 0.4813, "step": 23698 }, { "epoch": 1.5905170967417201, "grad_norm": 1.083215355873108, "learning_rate": 1.0591508544958823e-05, "loss": 0.4383, "step": 23700 }, { "epoch": 1.590651320425489, "grad_norm": 1.0399837493896484, "learning_rate": 1.0584820491890402e-05, "loss": 0.4771, "step": 23702 }, { "epoch": 1.590785544109258, "grad_norm": 1.2904225587844849, "learning_rate": 1.0578134301111731e-05, "loss": 0.472, "step": 23704 }, { "epoch": 1.5909197677930271, "grad_norm": 1.0123722553253174, "learning_rate": 1.0571449972938668e-05, "loss": 0.4639, "step": 23706 }, { "epoch": 1.591053991476796, "grad_norm": 1.1034024953842163, "learning_rate": 1.0564767507687078e-05, "loss": 0.5012, "step": 23708 }, { "epoch": 1.5911882151605652, "grad_norm": 0.9454026222229004, "learning_rate": 1.055808690567266e-05, "loss": 0.4836, "step": 23710 }, { "epoch": 1.591322438844334, "grad_norm": 0.9937888979911804, "learning_rate": 1.055140816721109e-05, "loss": 0.432, "step": 23712 }, { "epoch": 1.591456662528103, "grad_norm": 1.0560818910598755, "learning_rate": 1.054473129261791e-05, "loss": 0.4543, "step": 23714 }, { "epoch": 1.591590886211872, "grad_norm": 1.067166805267334, "learning_rate": 1.0538056282208598e-05, "loss": 0.4555, "step": 23716 }, { "epoch": 1.591725109895641, "grad_norm": 1.1397252082824707, "learning_rate": 1.0531383136298507e-05, "loss": 0.5167, "step": 23718 }, { "epoch": 1.5918593335794102, "grad_norm": 0.9467281699180603, "learning_rate": 1.0524711855202967e-05, "loss": 0.532, "step": 23720 }, { "epoch": 1.5919935572631791, "grad_norm": 1.033605933189392, "learning_rate": 1.051804243923717e-05, "loss": 0.4741, "step": 23722 }, { "epoch": 1.592127780946948, "grad_norm": 0.9673511385917664, "learning_rate": 1.0511374888716224e-05, "loss": 0.5159, "step": 23724 }, { "epoch": 1.592262004630717, "grad_norm": 0.8267554044723511, "learning_rate": 1.0504709203955155e-05, "loss": 0.4169, "step": 23726 }, { "epoch": 1.5923962283144861, "grad_norm": 1.0662051439285278, "learning_rate": 1.0498045385268923e-05, "loss": 0.4861, "step": 23728 }, { "epoch": 1.592530451998255, "grad_norm": 1.031984806060791, "learning_rate": 1.0491383432972357e-05, "loss": 0.4379, "step": 23730 }, { "epoch": 1.5926646756820242, "grad_norm": 1.0029823780059814, "learning_rate": 1.048472334738026e-05, "loss": 0.4919, "step": 23732 }, { "epoch": 1.592798899365793, "grad_norm": 1.0263690948486328, "learning_rate": 1.0478065128807251e-05, "loss": 0.5601, "step": 23734 }, { "epoch": 1.592933123049562, "grad_norm": 0.8970464468002319, "learning_rate": 1.0471408777567965e-05, "loss": 0.4435, "step": 23736 }, { "epoch": 1.593067346733331, "grad_norm": 1.0362054109573364, "learning_rate": 1.0464754293976875e-05, "loss": 0.4712, "step": 23738 }, { "epoch": 1.5932015704171, "grad_norm": 1.1117849349975586, "learning_rate": 1.045810167834841e-05, "loss": 0.5359, "step": 23740 }, { "epoch": 1.5933357941008692, "grad_norm": 1.1014670133590698, "learning_rate": 1.0451450930996887e-05, "loss": 0.4716, "step": 23742 }, { "epoch": 1.5934700177846381, "grad_norm": 0.8141437768936157, "learning_rate": 1.0444802052236535e-05, "loss": 0.4373, "step": 23744 }, { "epoch": 1.593604241468407, "grad_norm": 0.9930576682090759, "learning_rate": 1.0438155042381491e-05, "loss": 0.4522, "step": 23746 }, { "epoch": 1.593738465152176, "grad_norm": 0.8991205096244812, "learning_rate": 1.0431509901745846e-05, "loss": 0.461, "step": 23748 }, { "epoch": 1.5938726888359451, "grad_norm": 1.0563803911209106, "learning_rate": 1.0424866630643542e-05, "loss": 0.4278, "step": 23750 }, { "epoch": 1.5940069125197143, "grad_norm": 1.1721192598342896, "learning_rate": 1.0418225229388479e-05, "loss": 0.452, "step": 23752 }, { "epoch": 1.5941411362034832, "grad_norm": 1.0518708229064941, "learning_rate": 1.0411585698294418e-05, "loss": 0.4441, "step": 23754 }, { "epoch": 1.594275359887252, "grad_norm": 1.1242610216140747, "learning_rate": 1.04049480376751e-05, "loss": 0.4785, "step": 23756 }, { "epoch": 1.594409583571021, "grad_norm": 1.1258399486541748, "learning_rate": 1.0398312247844127e-05, "loss": 0.5362, "step": 23758 }, { "epoch": 1.59454380725479, "grad_norm": 1.0580230951309204, "learning_rate": 1.0391678329115028e-05, "loss": 0.4615, "step": 23760 }, { "epoch": 1.594678030938559, "grad_norm": 1.0128093957901, "learning_rate": 1.0385046281801242e-05, "loss": 0.494, "step": 23762 }, { "epoch": 1.5948122546223282, "grad_norm": 1.00709068775177, "learning_rate": 1.0378416106216105e-05, "loss": 0.5047, "step": 23764 }, { "epoch": 1.5949464783060971, "grad_norm": 1.025793194770813, "learning_rate": 1.0371787802672906e-05, "loss": 0.4645, "step": 23766 }, { "epoch": 1.595080701989866, "grad_norm": 1.0294287204742432, "learning_rate": 1.03651613714848e-05, "loss": 0.5148, "step": 23768 }, { "epoch": 1.595214925673635, "grad_norm": 1.018163800239563, "learning_rate": 1.0358536812964904e-05, "loss": 0.457, "step": 23770 }, { "epoch": 1.5953491493574041, "grad_norm": 1.067983627319336, "learning_rate": 1.0351914127426166e-05, "loss": 0.4174, "step": 23772 }, { "epoch": 1.5954833730411733, "grad_norm": 1.0045945644378662, "learning_rate": 1.034529331518153e-05, "loss": 0.4896, "step": 23774 }, { "epoch": 1.5956175967249422, "grad_norm": 1.0037025213241577, "learning_rate": 1.03386743765438e-05, "loss": 0.4742, "step": 23776 }, { "epoch": 1.595751820408711, "grad_norm": 1.0527470111846924, "learning_rate": 1.0332057311825726e-05, "loss": 0.4628, "step": 23778 }, { "epoch": 1.59588604409248, "grad_norm": 1.3437060117721558, "learning_rate": 1.032544212133994e-05, "loss": 0.4746, "step": 23780 }, { "epoch": 1.5960202677762492, "grad_norm": 1.1082535982131958, "learning_rate": 1.0318828805398994e-05, "loss": 0.5561, "step": 23782 }, { "epoch": 1.596154491460018, "grad_norm": 1.046541690826416, "learning_rate": 1.0312217364315351e-05, "loss": 0.5194, "step": 23784 }, { "epoch": 1.5962887151437872, "grad_norm": 1.1011675596237183, "learning_rate": 1.03056077984014e-05, "loss": 0.5635, "step": 23786 }, { "epoch": 1.5964229388275561, "grad_norm": 1.070259928703308, "learning_rate": 1.029900010796943e-05, "loss": 0.4599, "step": 23788 }, { "epoch": 1.596557162511325, "grad_norm": 1.0323693752288818, "learning_rate": 1.0292394293331637e-05, "loss": 0.4672, "step": 23790 }, { "epoch": 1.596691386195094, "grad_norm": 0.905518651008606, "learning_rate": 1.0285790354800118e-05, "loss": 0.4294, "step": 23792 }, { "epoch": 1.5968256098788631, "grad_norm": 1.046562671661377, "learning_rate": 1.027918829268692e-05, "loss": 0.4776, "step": 23794 }, { "epoch": 1.5969598335626323, "grad_norm": 0.9571558833122253, "learning_rate": 1.027258810730396e-05, "loss": 0.4368, "step": 23796 }, { "epoch": 1.5970940572464012, "grad_norm": 1.0320895910263062, "learning_rate": 1.0265989798963116e-05, "loss": 0.4882, "step": 23798 }, { "epoch": 1.59722828093017, "grad_norm": 0.9254117012023926, "learning_rate": 1.025939336797609e-05, "loss": 0.4507, "step": 23800 }, { "epoch": 1.597362504613939, "grad_norm": 1.1301331520080566, "learning_rate": 1.0252798814654597e-05, "loss": 0.4758, "step": 23802 }, { "epoch": 1.5974967282977082, "grad_norm": 1.0007113218307495, "learning_rate": 1.0246206139310187e-05, "loss": 0.4802, "step": 23804 }, { "epoch": 1.597630951981477, "grad_norm": 1.1954693794250488, "learning_rate": 1.0239615342254378e-05, "loss": 0.4899, "step": 23806 }, { "epoch": 1.5977651756652462, "grad_norm": 0.9859517812728882, "learning_rate": 1.0233026423798559e-05, "loss": 0.5094, "step": 23808 }, { "epoch": 1.5978993993490151, "grad_norm": 1.156038522720337, "learning_rate": 1.0226439384254044e-05, "loss": 0.5374, "step": 23810 }, { "epoch": 1.598033623032784, "grad_norm": 1.1276850700378418, "learning_rate": 1.0219854223932041e-05, "loss": 0.5342, "step": 23812 }, { "epoch": 1.598167846716553, "grad_norm": 1.047963261604309, "learning_rate": 1.0213270943143716e-05, "loss": 0.4327, "step": 23814 }, { "epoch": 1.5983020704003221, "grad_norm": 1.077088475227356, "learning_rate": 1.02066895422001e-05, "loss": 0.453, "step": 23816 }, { "epoch": 1.5984362940840913, "grad_norm": 1.0303544998168945, "learning_rate": 1.020011002141215e-05, "loss": 0.5176, "step": 23818 }, { "epoch": 1.5985705177678602, "grad_norm": 0.9186301827430725, "learning_rate": 1.0193532381090732e-05, "loss": 0.4509, "step": 23820 }, { "epoch": 1.598704741451629, "grad_norm": 1.0409269332885742, "learning_rate": 1.0186956621546645e-05, "loss": 0.5433, "step": 23822 }, { "epoch": 1.598838965135398, "grad_norm": 1.086869239807129, "learning_rate": 1.0180382743090555e-05, "loss": 0.4621, "step": 23824 }, { "epoch": 1.5989731888191672, "grad_norm": 0.9349266886711121, "learning_rate": 1.0173810746033103e-05, "loss": 0.4674, "step": 23826 }, { "epoch": 1.5991074125029363, "grad_norm": 0.9937050938606262, "learning_rate": 1.0167240630684765e-05, "loss": 0.4908, "step": 23828 }, { "epoch": 1.5992416361867052, "grad_norm": 0.9550259709358215, "learning_rate": 1.0160672397355969e-05, "loss": 0.4843, "step": 23830 }, { "epoch": 1.5993758598704741, "grad_norm": 1.048781156539917, "learning_rate": 1.0154106046357071e-05, "loss": 0.5059, "step": 23832 }, { "epoch": 1.599510083554243, "grad_norm": 0.9172594547271729, "learning_rate": 1.0147541577998298e-05, "loss": 0.4637, "step": 23834 }, { "epoch": 1.599644307238012, "grad_norm": 1.0190455913543701, "learning_rate": 1.0140978992589833e-05, "loss": 0.4182, "step": 23836 }, { "epoch": 1.5997785309217811, "grad_norm": 0.9304594993591309, "learning_rate": 1.0134418290441728e-05, "loss": 0.4749, "step": 23838 }, { "epoch": 1.5999127546055503, "grad_norm": 0.9567098021507263, "learning_rate": 1.012785947186397e-05, "loss": 0.4793, "step": 23840 }, { "epoch": 1.6000469782893192, "grad_norm": 1.0866206884384155, "learning_rate": 1.0121302537166433e-05, "loss": 0.4727, "step": 23842 }, { "epoch": 1.600181201973088, "grad_norm": 1.1256355047225952, "learning_rate": 1.0114747486658943e-05, "loss": 0.5181, "step": 23844 }, { "epoch": 1.600315425656857, "grad_norm": 1.1933351755142212, "learning_rate": 1.0108194320651205e-05, "loss": 0.5037, "step": 23846 }, { "epoch": 1.6004496493406262, "grad_norm": 1.0349985361099243, "learning_rate": 1.010164303945284e-05, "loss": 0.5087, "step": 23848 }, { "epoch": 1.6005838730243953, "grad_norm": 0.9755839109420776, "learning_rate": 1.0095093643373377e-05, "loss": 0.5202, "step": 23850 }, { "epoch": 1.6007180967081642, "grad_norm": 1.032321810722351, "learning_rate": 1.0088546132722276e-05, "loss": 0.5722, "step": 23852 }, { "epoch": 1.6008523203919331, "grad_norm": 0.9793102145195007, "learning_rate": 1.0082000507808892e-05, "loss": 0.5009, "step": 23854 }, { "epoch": 1.600986544075702, "grad_norm": 1.093875765800476, "learning_rate": 1.0075456768942488e-05, "loss": 0.5367, "step": 23856 }, { "epoch": 1.6011207677594712, "grad_norm": 1.1801968812942505, "learning_rate": 1.0068914916432231e-05, "loss": 0.5578, "step": 23858 }, { "epoch": 1.6012549914432401, "grad_norm": 0.9556602835655212, "learning_rate": 1.0062374950587234e-05, "loss": 0.4843, "step": 23860 }, { "epoch": 1.6013892151270093, "grad_norm": 1.0572758913040161, "learning_rate": 1.0055836871716473e-05, "loss": 0.4997, "step": 23862 }, { "epoch": 1.6015234388107782, "grad_norm": 1.0739139318466187, "learning_rate": 1.0049300680128887e-05, "loss": 0.5168, "step": 23864 }, { "epoch": 1.601657662494547, "grad_norm": 1.0001834630966187, "learning_rate": 1.004276637613329e-05, "loss": 0.4611, "step": 23866 }, { "epoch": 1.601791886178316, "grad_norm": 0.9812644720077515, "learning_rate": 1.0036233960038399e-05, "loss": 0.5198, "step": 23868 }, { "epoch": 1.6019261098620852, "grad_norm": 1.0067847967147827, "learning_rate": 1.0029703432152859e-05, "loss": 0.4752, "step": 23870 }, { "epoch": 1.6020603335458543, "grad_norm": 1.0018914937973022, "learning_rate": 1.002317479278525e-05, "loss": 0.4786, "step": 23872 }, { "epoch": 1.6021945572296232, "grad_norm": 1.049195408821106, "learning_rate": 1.001664804224402e-05, "loss": 0.4891, "step": 23874 }, { "epoch": 1.6023287809133921, "grad_norm": 1.0486708879470825, "learning_rate": 1.0010123180837544e-05, "loss": 0.4587, "step": 23876 }, { "epoch": 1.602463004597161, "grad_norm": 0.9982337355613708, "learning_rate": 1.0003600208874098e-05, "loss": 0.466, "step": 23878 }, { "epoch": 1.6025972282809302, "grad_norm": 1.1180343627929688, "learning_rate": 9.997079126661907e-06, "loss": 0.4489, "step": 23880 }, { "epoch": 1.6027314519646991, "grad_norm": 1.0404959917068481, "learning_rate": 9.990559934509053e-06, "loss": 0.4944, "step": 23882 }, { "epoch": 1.6028656756484683, "grad_norm": 1.0978857278823853, "learning_rate": 9.984042632723589e-06, "loss": 0.532, "step": 23884 }, { "epoch": 1.6029998993322372, "grad_norm": 1.035715103149414, "learning_rate": 9.977527221613397e-06, "loss": 0.5087, "step": 23886 }, { "epoch": 1.603134123016006, "grad_norm": 1.1274858713150024, "learning_rate": 9.97101370148636e-06, "loss": 0.5453, "step": 23888 }, { "epoch": 1.603268346699775, "grad_norm": 1.0599273443222046, "learning_rate": 9.964502072650195e-06, "loss": 0.4714, "step": 23890 }, { "epoch": 1.6034025703835442, "grad_norm": 1.046848177909851, "learning_rate": 9.957992335412597e-06, "loss": 0.4718, "step": 23892 }, { "epoch": 1.6035367940673133, "grad_norm": 1.0222798585891724, "learning_rate": 9.95148449008112e-06, "loss": 0.4411, "step": 23894 }, { "epoch": 1.6036710177510822, "grad_norm": 1.1524306535720825, "learning_rate": 9.944978536963246e-06, "loss": 0.4696, "step": 23896 }, { "epoch": 1.6038052414348511, "grad_norm": 0.9545031189918518, "learning_rate": 9.938474476366378e-06, "loss": 0.4327, "step": 23898 }, { "epoch": 1.60393946511862, "grad_norm": 0.9605699777603149, "learning_rate": 9.931972308597792e-06, "loss": 0.4458, "step": 23900 }, { "epoch": 1.6040736888023892, "grad_norm": 0.9968399405479431, "learning_rate": 9.925472033964744e-06, "loss": 0.4516, "step": 23902 }, { "epoch": 1.6042079124861581, "grad_norm": 1.0261261463165283, "learning_rate": 9.918973652774339e-06, "loss": 0.5041, "step": 23904 }, { "epoch": 1.6043421361699273, "grad_norm": 1.03523850440979, "learning_rate": 9.912477165333612e-06, "loss": 0.4782, "step": 23906 }, { "epoch": 1.6044763598536962, "grad_norm": 1.0323772430419922, "learning_rate": 9.905982571949491e-06, "loss": 0.4811, "step": 23908 }, { "epoch": 1.604610583537465, "grad_norm": 0.9295010566711426, "learning_rate": 9.899489872928874e-06, "loss": 0.4352, "step": 23910 }, { "epoch": 1.604744807221234, "grad_norm": 1.0068014860153198, "learning_rate": 9.892999068578502e-06, "loss": 0.4196, "step": 23912 }, { "epoch": 1.6048790309050032, "grad_norm": 0.9357097744941711, "learning_rate": 9.886510159205053e-06, "loss": 0.4026, "step": 23914 }, { "epoch": 1.6050132545887723, "grad_norm": 0.9535422921180725, "learning_rate": 9.880023145115114e-06, "loss": 0.4632, "step": 23916 }, { "epoch": 1.6051474782725412, "grad_norm": 1.1852741241455078, "learning_rate": 9.873538026615198e-06, "loss": 0.4818, "step": 23918 }, { "epoch": 1.6052817019563101, "grad_norm": 1.123728632926941, "learning_rate": 9.867054804011693e-06, "loss": 0.4643, "step": 23920 }, { "epoch": 1.605415925640079, "grad_norm": 0.9284471869468689, "learning_rate": 9.860573477610952e-06, "loss": 0.484, "step": 23922 }, { "epoch": 1.6055501493238482, "grad_norm": 1.1171112060546875, "learning_rate": 9.854094047719164e-06, "loss": 0.457, "step": 23924 }, { "epoch": 1.6056843730076173, "grad_norm": 1.0885698795318604, "learning_rate": 9.8476165146425e-06, "loss": 0.4984, "step": 23926 }, { "epoch": 1.6058185966913863, "grad_norm": 1.0228629112243652, "learning_rate": 9.841140878686983e-06, "loss": 0.4652, "step": 23928 }, { "epoch": 1.6059528203751552, "grad_norm": 0.9912495017051697, "learning_rate": 9.834667140158605e-06, "loss": 0.5383, "step": 23930 }, { "epoch": 1.606087044058924, "grad_norm": 1.0985013246536255, "learning_rate": 9.828195299363224e-06, "loss": 0.4345, "step": 23932 }, { "epoch": 1.6062212677426932, "grad_norm": 1.0537720918655396, "learning_rate": 9.82172535660662e-06, "loss": 0.4558, "step": 23934 }, { "epoch": 1.6063554914264622, "grad_norm": 0.8829530477523804, "learning_rate": 9.81525731219447e-06, "loss": 0.4826, "step": 23936 }, { "epoch": 1.6064897151102313, "grad_norm": 1.0897839069366455, "learning_rate": 9.808791166432413e-06, "loss": 0.4997, "step": 23938 }, { "epoch": 1.6066239387940002, "grad_norm": 1.4052565097808838, "learning_rate": 9.802326919625931e-06, "loss": 0.4729, "step": 23940 }, { "epoch": 1.6067581624777691, "grad_norm": 1.0854122638702393, "learning_rate": 9.795864572080466e-06, "loss": 0.4899, "step": 23942 }, { "epoch": 1.606892386161538, "grad_norm": 0.8918828368186951, "learning_rate": 9.78940412410132e-06, "loss": 0.4255, "step": 23944 }, { "epoch": 1.6070266098453072, "grad_norm": 1.6517807245254517, "learning_rate": 9.782945575993779e-06, "loss": 0.421, "step": 23946 }, { "epoch": 1.6071608335290763, "grad_norm": 0.9856716990470886, "learning_rate": 9.77648892806296e-06, "loss": 0.4596, "step": 23948 }, { "epoch": 1.6072950572128453, "grad_norm": 1.058132290840149, "learning_rate": 9.770034180613968e-06, "loss": 0.514, "step": 23950 }, { "epoch": 1.6074292808966142, "grad_norm": 0.9935703277587891, "learning_rate": 9.763581333951727e-06, "loss": 0.4523, "step": 23952 }, { "epoch": 1.607563504580383, "grad_norm": 0.9994638562202454, "learning_rate": 9.757130388381158e-06, "loss": 0.4745, "step": 23954 }, { "epoch": 1.6076977282641522, "grad_norm": 1.0030862092971802, "learning_rate": 9.750681344207035e-06, "loss": 0.4799, "step": 23956 }, { "epoch": 1.6078319519479212, "grad_norm": 1.0018229484558105, "learning_rate": 9.744234201734082e-06, "loss": 0.4709, "step": 23958 }, { "epoch": 1.6079661756316903, "grad_norm": 0.9601824879646301, "learning_rate": 9.737788961266903e-06, "loss": 0.4588, "step": 23960 }, { "epoch": 1.6081003993154592, "grad_norm": 1.0627379417419434, "learning_rate": 9.731345623110022e-06, "loss": 0.4816, "step": 23962 }, { "epoch": 1.6082346229992281, "grad_norm": 1.1042258739471436, "learning_rate": 9.724904187567879e-06, "loss": 0.533, "step": 23964 }, { "epoch": 1.608368846682997, "grad_norm": 1.1249784231185913, "learning_rate": 9.718464654944798e-06, "loss": 0.4703, "step": 23966 }, { "epoch": 1.6085030703667662, "grad_norm": 0.8959927558898926, "learning_rate": 9.712027025545067e-06, "loss": 0.4851, "step": 23968 }, { "epoch": 1.6086372940505353, "grad_norm": 1.1174625158309937, "learning_rate": 9.70559129967284e-06, "loss": 0.4609, "step": 23970 }, { "epoch": 1.6087715177343043, "grad_norm": 1.1044148206710815, "learning_rate": 9.699157477632181e-06, "loss": 0.455, "step": 23972 }, { "epoch": 1.6089057414180732, "grad_norm": 1.0346393585205078, "learning_rate": 9.692725559727072e-06, "loss": 0.5299, "step": 23974 }, { "epoch": 1.609039965101842, "grad_norm": 1.0218905210494995, "learning_rate": 9.686295546261436e-06, "loss": 0.4798, "step": 23976 }, { "epoch": 1.6091741887856112, "grad_norm": 0.9859256148338318, "learning_rate": 9.679867437539063e-06, "loss": 0.4614, "step": 23978 }, { "epoch": 1.6093084124693802, "grad_norm": 1.0233776569366455, "learning_rate": 9.673441233863662e-06, "loss": 0.4599, "step": 23980 }, { "epoch": 1.6094426361531493, "grad_norm": 1.170893907546997, "learning_rate": 9.667016935538859e-06, "loss": 0.4715, "step": 23982 }, { "epoch": 1.6095768598369182, "grad_norm": 1.0487552881240845, "learning_rate": 9.6605945428682e-06, "loss": 0.4594, "step": 23984 }, { "epoch": 1.6097110835206871, "grad_norm": 1.0074599981307983, "learning_rate": 9.654174056155113e-06, "loss": 0.4731, "step": 23986 }, { "epoch": 1.609845307204456, "grad_norm": 0.9991166591644287, "learning_rate": 9.64775547570298e-06, "loss": 0.4205, "step": 23988 }, { "epoch": 1.6099795308882252, "grad_norm": 0.9638586640357971, "learning_rate": 9.641338801815048e-06, "loss": 0.463, "step": 23990 }, { "epoch": 1.6101137545719943, "grad_norm": 1.2994322776794434, "learning_rate": 9.634924034794501e-06, "loss": 0.4747, "step": 23992 }, { "epoch": 1.6102479782557633, "grad_norm": 1.1774928569793701, "learning_rate": 9.628511174944404e-06, "loss": 0.5106, "step": 23994 }, { "epoch": 1.6103822019395322, "grad_norm": 1.0639322996139526, "learning_rate": 9.622100222567775e-06, "loss": 0.4535, "step": 23996 }, { "epoch": 1.610516425623301, "grad_norm": 1.0801434516906738, "learning_rate": 9.615691177967518e-06, "loss": 0.4782, "step": 23998 }, { "epoch": 1.6106506493070702, "grad_norm": 0.9594919085502625, "learning_rate": 9.609284041446438e-06, "loss": 0.5241, "step": 24000 }, { "epoch": 1.6107848729908394, "grad_norm": 2.1199254989624023, "learning_rate": 9.602878813307249e-06, "loss": 0.4488, "step": 24002 }, { "epoch": 1.6109190966746083, "grad_norm": 0.9908333420753479, "learning_rate": 9.596475493852608e-06, "loss": 0.481, "step": 24004 }, { "epoch": 1.6110533203583772, "grad_norm": 1.0409706830978394, "learning_rate": 9.590074083385053e-06, "loss": 0.457, "step": 24006 }, { "epoch": 1.6111875440421461, "grad_norm": 1.0337355136871338, "learning_rate": 9.583674582207036e-06, "loss": 0.4516, "step": 24008 }, { "epoch": 1.6113217677259153, "grad_norm": 0.9755725860595703, "learning_rate": 9.577276990620903e-06, "loss": 0.4258, "step": 24010 }, { "epoch": 1.6114559914096842, "grad_norm": 1.0953158140182495, "learning_rate": 9.570881308928958e-06, "loss": 0.504, "step": 24012 }, { "epoch": 1.6115902150934533, "grad_norm": 1.0453033447265625, "learning_rate": 9.564487537433365e-06, "loss": 0.4813, "step": 24014 }, { "epoch": 1.6117244387772223, "grad_norm": 0.9101317524909973, "learning_rate": 9.55809567643623e-06, "loss": 0.4578, "step": 24016 }, { "epoch": 1.6118586624609912, "grad_norm": 1.0483723878860474, "learning_rate": 9.551705726239546e-06, "loss": 0.4429, "step": 24018 }, { "epoch": 1.61199288614476, "grad_norm": 1.000253438949585, "learning_rate": 9.545317687145232e-06, "loss": 0.4399, "step": 24020 }, { "epoch": 1.6121271098285292, "grad_norm": 0.8763706088066101, "learning_rate": 9.538931559455095e-06, "loss": 0.4658, "step": 24022 }, { "epoch": 1.6122613335122984, "grad_norm": 0.9557889103889465, "learning_rate": 9.532547343470889e-06, "loss": 0.4444, "step": 24024 }, { "epoch": 1.6123955571960673, "grad_norm": 1.0888750553131104, "learning_rate": 9.526165039494244e-06, "loss": 0.4544, "step": 24026 }, { "epoch": 1.6125297808798362, "grad_norm": 1.0589221715927124, "learning_rate": 9.519784647826713e-06, "loss": 0.4658, "step": 24028 }, { "epoch": 1.6126640045636051, "grad_norm": 1.1489931344985962, "learning_rate": 9.513406168769745e-06, "loss": 0.4601, "step": 24030 }, { "epoch": 1.6127982282473743, "grad_norm": 0.9511253833770752, "learning_rate": 9.507029602624734e-06, "loss": 0.4467, "step": 24032 }, { "epoch": 1.6129324519311432, "grad_norm": 1.063330888748169, "learning_rate": 9.50065494969295e-06, "loss": 0.4993, "step": 24034 }, { "epoch": 1.6130666756149123, "grad_norm": 0.9531103372573853, "learning_rate": 9.494282210275579e-06, "loss": 0.5202, "step": 24036 }, { "epoch": 1.6132008992986813, "grad_norm": 1.212205410003662, "learning_rate": 9.48791138467372e-06, "loss": 0.5128, "step": 24038 }, { "epoch": 1.6133351229824502, "grad_norm": 1.049813985824585, "learning_rate": 9.481542473188377e-06, "loss": 0.5109, "step": 24040 }, { "epoch": 1.613469346666219, "grad_norm": 1.0941503047943115, "learning_rate": 9.475175476120484e-06, "loss": 0.4704, "step": 24042 }, { "epoch": 1.6136035703499882, "grad_norm": 0.9977181553840637, "learning_rate": 9.468810393770856e-06, "loss": 0.4586, "step": 24044 }, { "epoch": 1.6137377940337574, "grad_norm": 1.051984190940857, "learning_rate": 9.462447226440252e-06, "loss": 0.4489, "step": 24046 }, { "epoch": 1.6138720177175263, "grad_norm": 1.109163761138916, "learning_rate": 9.456085974429286e-06, "loss": 0.5115, "step": 24048 }, { "epoch": 1.6140062414012952, "grad_norm": 0.9957579374313354, "learning_rate": 9.449726638038536e-06, "loss": 0.4676, "step": 24050 }, { "epoch": 1.6141404650850641, "grad_norm": 1.0215259790420532, "learning_rate": 9.443369217568455e-06, "loss": 0.4786, "step": 24052 }, { "epoch": 1.6142746887688333, "grad_norm": 0.995062530040741, "learning_rate": 9.437013713319437e-06, "loss": 0.4429, "step": 24054 }, { "epoch": 1.6144089124526022, "grad_norm": 0.9971685409545898, "learning_rate": 9.430660125591762e-06, "loss": 0.4857, "step": 24056 }, { "epoch": 1.6145431361363713, "grad_norm": 0.99956876039505, "learning_rate": 9.424308454685622e-06, "loss": 0.4322, "step": 24058 }, { "epoch": 1.6146773598201403, "grad_norm": 1.0475598573684692, "learning_rate": 9.417958700901097e-06, "loss": 0.4886, "step": 24060 }, { "epoch": 1.6148115835039092, "grad_norm": 1.055567979812622, "learning_rate": 9.411610864538239e-06, "loss": 0.4814, "step": 24062 }, { "epoch": 1.614945807187678, "grad_norm": 1.0338261127471924, "learning_rate": 9.405264945896959e-06, "loss": 0.4787, "step": 24064 }, { "epoch": 1.6150800308714472, "grad_norm": 1.0155824422836304, "learning_rate": 9.398920945277079e-06, "loss": 0.4907, "step": 24066 }, { "epoch": 1.6152142545552164, "grad_norm": 1.0634373426437378, "learning_rate": 9.392578862978334e-06, "loss": 0.4434, "step": 24068 }, { "epoch": 1.6153484782389853, "grad_norm": 1.0663130283355713, "learning_rate": 9.386238699300398e-06, "loss": 0.4752, "step": 24070 }, { "epoch": 1.6154827019227542, "grad_norm": 1.0315016508102417, "learning_rate": 9.37990045454281e-06, "loss": 0.4798, "step": 24072 }, { "epoch": 1.6156169256065231, "grad_norm": 0.9357467293739319, "learning_rate": 9.373564129005074e-06, "loss": 0.5043, "step": 24074 }, { "epoch": 1.6157511492902923, "grad_norm": 0.9481315016746521, "learning_rate": 9.36722972298652e-06, "loss": 0.526, "step": 24076 }, { "epoch": 1.6158853729740614, "grad_norm": 0.9957374930381775, "learning_rate": 9.36089723678647e-06, "loss": 0.4725, "step": 24078 }, { "epoch": 1.6160195966578303, "grad_norm": 1.7620972394943237, "learning_rate": 9.354566670704102e-06, "loss": 0.4567, "step": 24080 }, { "epoch": 1.6161538203415993, "grad_norm": 1.176615834236145, "learning_rate": 9.348238025038547e-06, "loss": 0.4465, "step": 24082 }, { "epoch": 1.6162880440253682, "grad_norm": 0.9767720699310303, "learning_rate": 9.341911300088807e-06, "loss": 0.3975, "step": 24084 }, { "epoch": 1.6164222677091373, "grad_norm": 0.9831427335739136, "learning_rate": 9.335586496153804e-06, "loss": 0.4522, "step": 24086 }, { "epoch": 1.6165564913929062, "grad_norm": 1.091477394104004, "learning_rate": 9.329263613532363e-06, "loss": 0.4849, "step": 24088 }, { "epoch": 1.6166907150766754, "grad_norm": 1.0562587976455688, "learning_rate": 9.322942652523259e-06, "loss": 0.4562, "step": 24090 }, { "epoch": 1.6168249387604443, "grad_norm": 0.8952846527099609, "learning_rate": 9.316623613425119e-06, "loss": 0.4132, "step": 24092 }, { "epoch": 1.6169591624442132, "grad_norm": 1.0176690816879272, "learning_rate": 9.310306496536519e-06, "loss": 0.4767, "step": 24094 }, { "epoch": 1.6170933861279821, "grad_norm": 1.0066777467727661, "learning_rate": 9.303991302155907e-06, "loss": 0.4267, "step": 24096 }, { "epoch": 1.6172276098117513, "grad_norm": 0.9026097655296326, "learning_rate": 9.297678030581697e-06, "loss": 0.4117, "step": 24098 }, { "epoch": 1.6173618334955204, "grad_norm": 0.9847307801246643, "learning_rate": 9.291366682112152e-06, "loss": 0.4353, "step": 24100 }, { "epoch": 1.6174960571792893, "grad_norm": 0.9909016489982605, "learning_rate": 9.285057257045498e-06, "loss": 0.4191, "step": 24102 }, { "epoch": 1.6176302808630583, "grad_norm": 1.0340805053710938, "learning_rate": 9.278749755679823e-06, "loss": 0.4293, "step": 24104 }, { "epoch": 1.6177645045468272, "grad_norm": 1.0646992921829224, "learning_rate": 9.272444178313127e-06, "loss": 0.4774, "step": 24106 }, { "epoch": 1.6178987282305963, "grad_norm": 0.9061216115951538, "learning_rate": 9.266140525243373e-06, "loss": 0.3749, "step": 24108 }, { "epoch": 1.6180329519143652, "grad_norm": 1.1270420551300049, "learning_rate": 9.259838796768367e-06, "loss": 0.4538, "step": 24110 }, { "epoch": 1.6181671755981344, "grad_norm": 0.9613937735557556, "learning_rate": 9.25353899318589e-06, "loss": 0.4073, "step": 24112 }, { "epoch": 1.6183013992819033, "grad_norm": 1.1744015216827393, "learning_rate": 9.247241114793543e-06, "loss": 0.523, "step": 24114 }, { "epoch": 1.6184356229656722, "grad_norm": 0.8820990324020386, "learning_rate": 9.240945161888931e-06, "loss": 0.4776, "step": 24116 }, { "epoch": 1.6185698466494411, "grad_norm": 0.978901743888855, "learning_rate": 9.2346511347695e-06, "loss": 0.4847, "step": 24118 }, { "epoch": 1.6187040703332103, "grad_norm": 1.159218192100525, "learning_rate": 9.228359033732653e-06, "loss": 0.4507, "step": 24120 }, { "epoch": 1.6188382940169794, "grad_norm": 1.0147674083709717, "learning_rate": 9.222068859075667e-06, "loss": 0.4269, "step": 24122 }, { "epoch": 1.6189725177007483, "grad_norm": 1.0214875936508179, "learning_rate": 9.215780611095743e-06, "loss": 0.4861, "step": 24124 }, { "epoch": 1.6191067413845173, "grad_norm": 8.175945281982422, "learning_rate": 9.209494290089971e-06, "loss": 0.5368, "step": 24126 }, { "epoch": 1.6192409650682862, "grad_norm": 1.0697429180145264, "learning_rate": 9.203209896355398e-06, "loss": 0.5083, "step": 24128 }, { "epoch": 1.6193751887520553, "grad_norm": 1.0672308206558228, "learning_rate": 9.196927430188929e-06, "loss": 0.5229, "step": 24130 }, { "epoch": 1.6195094124358242, "grad_norm": 0.8495045304298401, "learning_rate": 9.190646891887405e-06, "loss": 0.4541, "step": 24132 }, { "epoch": 1.6196436361195934, "grad_norm": 1.102070689201355, "learning_rate": 9.184368281747557e-06, "loss": 0.5542, "step": 24134 }, { "epoch": 1.6197778598033623, "grad_norm": 1.296924114227295, "learning_rate": 9.178091600066063e-06, "loss": 0.5097, "step": 24136 }, { "epoch": 1.6199120834871312, "grad_norm": 1.095659852027893, "learning_rate": 9.171816847139448e-06, "loss": 0.4479, "step": 24138 }, { "epoch": 1.6200463071709001, "grad_norm": 1.1159011125564575, "learning_rate": 9.165544023264233e-06, "loss": 0.4538, "step": 24140 }, { "epoch": 1.6201805308546693, "grad_norm": 0.9968121647834778, "learning_rate": 9.159273128736734e-06, "loss": 0.4992, "step": 24142 }, { "epoch": 1.6203147545384384, "grad_norm": 1.04972243309021, "learning_rate": 9.15300416385329e-06, "loss": 0.5159, "step": 24144 }, { "epoch": 1.6204489782222073, "grad_norm": 1.0225540399551392, "learning_rate": 9.14673712891006e-06, "loss": 0.4906, "step": 24146 }, { "epoch": 1.6205832019059763, "grad_norm": 1.0914386510849, "learning_rate": 9.140472024203179e-06, "loss": 0.4838, "step": 24148 }, { "epoch": 1.6207174255897452, "grad_norm": 0.9849649667739868, "learning_rate": 9.134208850028647e-06, "loss": 0.4953, "step": 24150 }, { "epoch": 1.6208516492735143, "grad_norm": 1.0372711420059204, "learning_rate": 9.127947606682391e-06, "loss": 0.4938, "step": 24152 }, { "epoch": 1.6209858729572835, "grad_norm": 1.1415741443634033, "learning_rate": 9.121688294460223e-06, "loss": 0.4554, "step": 24154 }, { "epoch": 1.6211200966410524, "grad_norm": 1.046114206314087, "learning_rate": 9.115430913657912e-06, "loss": 0.5092, "step": 24156 }, { "epoch": 1.6212543203248213, "grad_norm": 1.081341028213501, "learning_rate": 9.109175464571096e-06, "loss": 0.469, "step": 24158 }, { "epoch": 1.6213885440085902, "grad_norm": 0.9061887264251709, "learning_rate": 9.10292194749533e-06, "loss": 0.426, "step": 24160 }, { "epoch": 1.6215227676923594, "grad_norm": 1.0150114297866821, "learning_rate": 9.096670362726073e-06, "loss": 0.4992, "step": 24162 }, { "epoch": 1.6216569913761283, "grad_norm": 1.1075929403305054, "learning_rate": 9.090420710558718e-06, "loss": 0.4694, "step": 24164 }, { "epoch": 1.6217912150598974, "grad_norm": 1.0864084959030151, "learning_rate": 9.084172991288525e-06, "loss": 0.5109, "step": 24166 }, { "epoch": 1.6219254387436663, "grad_norm": 1.1024622917175293, "learning_rate": 9.077927205210712e-06, "loss": 0.4718, "step": 24168 }, { "epoch": 1.6220596624274353, "grad_norm": 1.1553581953048706, "learning_rate": 9.071683352620385e-06, "loss": 0.4834, "step": 24170 }, { "epoch": 1.6221938861112042, "grad_norm": 1.1230138540267944, "learning_rate": 9.06544143381251e-06, "loss": 0.4815, "step": 24172 }, { "epoch": 1.6223281097949733, "grad_norm": 0.9074229001998901, "learning_rate": 9.059201449082045e-06, "loss": 0.4646, "step": 24174 }, { "epoch": 1.6224623334787425, "grad_norm": 0.9188847541809082, "learning_rate": 9.052963398723796e-06, "loss": 0.4414, "step": 24176 }, { "epoch": 1.6225965571625114, "grad_norm": 1.0981385707855225, "learning_rate": 9.046727283032519e-06, "loss": 0.4777, "step": 24178 }, { "epoch": 1.6227307808462803, "grad_norm": 1.112444519996643, "learning_rate": 9.040493102302844e-06, "loss": 0.4312, "step": 24180 }, { "epoch": 1.6228650045300492, "grad_norm": 0.9772467613220215, "learning_rate": 9.03426085682933e-06, "loss": 0.4385, "step": 24182 }, { "epoch": 1.6229992282138184, "grad_norm": 0.9029442071914673, "learning_rate": 9.028030546906419e-06, "loss": 0.4033, "step": 24184 }, { "epoch": 1.6231334518975873, "grad_norm": 1.0854231119155884, "learning_rate": 9.021802172828509e-06, "loss": 0.4883, "step": 24186 }, { "epoch": 1.6232676755813564, "grad_norm": 0.9764794707298279, "learning_rate": 9.01557573488987e-06, "loss": 0.474, "step": 24188 }, { "epoch": 1.6234018992651253, "grad_norm": 1.0908037424087524, "learning_rate": 9.009351233384684e-06, "loss": 0.4767, "step": 24190 }, { "epoch": 1.6235361229488943, "grad_norm": 1.0757347345352173, "learning_rate": 9.003128668607031e-06, "loss": 0.5152, "step": 24192 }, { "epoch": 1.6236703466326632, "grad_norm": 1.0436749458312988, "learning_rate": 8.99690804085095e-06, "loss": 0.4605, "step": 24194 }, { "epoch": 1.6238045703164323, "grad_norm": 1.0184334516525269, "learning_rate": 8.990689350410314e-06, "loss": 0.4616, "step": 24196 }, { "epoch": 1.6239387940002015, "grad_norm": 1.022641658782959, "learning_rate": 8.984472597578997e-06, "loss": 0.4667, "step": 24198 }, { "epoch": 1.6240730176839704, "grad_norm": 0.9187766313552856, "learning_rate": 8.978257782650668e-06, "loss": 0.4686, "step": 24200 }, { "epoch": 1.6242072413677393, "grad_norm": 0.908483624458313, "learning_rate": 8.972044905919008e-06, "loss": 0.4349, "step": 24202 }, { "epoch": 1.6243414650515082, "grad_norm": 1.0387719869613647, "learning_rate": 8.965833967677534e-06, "loss": 0.4219, "step": 24204 }, { "epoch": 1.6244756887352774, "grad_norm": 0.919877827167511, "learning_rate": 8.959624968219732e-06, "loss": 0.4656, "step": 24206 }, { "epoch": 1.6246099124190463, "grad_norm": 1.0006935596466064, "learning_rate": 8.95341790783894e-06, "loss": 0.4791, "step": 24208 }, { "epoch": 1.6247441361028154, "grad_norm": 1.0531569719314575, "learning_rate": 8.94721278682844e-06, "loss": 0.4467, "step": 24210 }, { "epoch": 1.6248783597865843, "grad_norm": 0.9406517744064331, "learning_rate": 8.941009605481398e-06, "loss": 0.4366, "step": 24212 }, { "epoch": 1.6250125834703533, "grad_norm": 0.979747474193573, "learning_rate": 8.934808364090924e-06, "loss": 0.4619, "step": 24214 }, { "epoch": 1.6251468071541222, "grad_norm": 0.9257476925849915, "learning_rate": 8.928609062950005e-06, "loss": 0.5277, "step": 24216 }, { "epoch": 1.6252810308378913, "grad_norm": 1.0857957601547241, "learning_rate": 8.922411702351546e-06, "loss": 0.4364, "step": 24218 }, { "epoch": 1.6254152545216605, "grad_norm": 0.9188510775566101, "learning_rate": 8.916216282588341e-06, "loss": 0.4383, "step": 24220 }, { "epoch": 1.6255494782054294, "grad_norm": 1.0596593618392944, "learning_rate": 8.910022803953144e-06, "loss": 0.4746, "step": 24222 }, { "epoch": 1.6256837018891983, "grad_norm": 1.073279857635498, "learning_rate": 8.90383126673856e-06, "loss": 0.4814, "step": 24224 }, { "epoch": 1.6258179255729672, "grad_norm": 1.12808096408844, "learning_rate": 8.897641671237156e-06, "loss": 0.4481, "step": 24226 }, { "epoch": 1.6259521492567364, "grad_norm": 0.8533748984336853, "learning_rate": 8.891454017741341e-06, "loss": 0.4077, "step": 24228 }, { "epoch": 1.6260863729405055, "grad_norm": 1.05867338180542, "learning_rate": 8.885268306543492e-06, "loss": 0.5289, "step": 24230 }, { "epoch": 1.6262205966242744, "grad_norm": 0.8957057595252991, "learning_rate": 8.879084537935855e-06, "loss": 0.4112, "step": 24232 }, { "epoch": 1.6263548203080433, "grad_norm": 1.0482287406921387, "learning_rate": 8.872902712210628e-06, "loss": 0.4893, "step": 24234 }, { "epoch": 1.6264890439918123, "grad_norm": 1.076281189918518, "learning_rate": 8.86672282965988e-06, "loss": 0.434, "step": 24236 }, { "epoch": 1.6266232676755814, "grad_norm": 1.1521581411361694, "learning_rate": 8.860544890575584e-06, "loss": 0.4854, "step": 24238 }, { "epoch": 1.6267574913593503, "grad_norm": 0.9764477014541626, "learning_rate": 8.85436889524965e-06, "loss": 0.4432, "step": 24240 }, { "epoch": 1.6268917150431195, "grad_norm": 1.2391307353973389, "learning_rate": 8.848194843973862e-06, "loss": 0.5098, "step": 24242 }, { "epoch": 1.6270259387268884, "grad_norm": 0.8947152495384216, "learning_rate": 8.842022737039957e-06, "loss": 0.4335, "step": 24244 }, { "epoch": 1.6271601624106573, "grad_norm": 1.1010911464691162, "learning_rate": 8.835852574739544e-06, "loss": 0.5048, "step": 24246 }, { "epoch": 1.6272943860944262, "grad_norm": 1.032392978668213, "learning_rate": 8.829684357364153e-06, "loss": 0.4737, "step": 24248 }, { "epoch": 1.6274286097781954, "grad_norm": 1.0642104148864746, "learning_rate": 8.823518085205206e-06, "loss": 0.519, "step": 24250 }, { "epoch": 1.6275628334619645, "grad_norm": 0.9958031177520752, "learning_rate": 8.817353758554075e-06, "loss": 0.4677, "step": 24252 }, { "epoch": 1.6276970571457334, "grad_norm": 0.9674162268638611, "learning_rate": 8.811191377701995e-06, "loss": 0.4031, "step": 24254 }, { "epoch": 1.6278312808295023, "grad_norm": 1.0566951036453247, "learning_rate": 8.805030942940123e-06, "loss": 0.4425, "step": 24256 }, { "epoch": 1.6279655045132713, "grad_norm": 1.0692909955978394, "learning_rate": 8.79887245455952e-06, "loss": 0.4803, "step": 24258 }, { "epoch": 1.6280997281970404, "grad_norm": 0.8982757925987244, "learning_rate": 8.792715912851196e-06, "loss": 0.4266, "step": 24260 }, { "epoch": 1.6282339518808093, "grad_norm": 1.057623028755188, "learning_rate": 8.786561318105996e-06, "loss": 0.4831, "step": 24262 }, { "epoch": 1.6283681755645785, "grad_norm": 1.1221034526824951, "learning_rate": 8.780408670614753e-06, "loss": 0.5228, "step": 24264 }, { "epoch": 1.6285023992483474, "grad_norm": 0.9700652360916138, "learning_rate": 8.774257970668127e-06, "loss": 0.4641, "step": 24266 }, { "epoch": 1.6286366229321163, "grad_norm": 1.0552066564559937, "learning_rate": 8.768109218556753e-06, "loss": 0.485, "step": 24268 }, { "epoch": 1.6287708466158852, "grad_norm": 1.1140886545181274, "learning_rate": 8.76196241457113e-06, "loss": 0.4931, "step": 24270 }, { "epoch": 1.6289050702996544, "grad_norm": 0.9525741338729858, "learning_rate": 8.7558175590017e-06, "loss": 0.4955, "step": 24272 }, { "epoch": 1.6290392939834235, "grad_norm": 0.972881555557251, "learning_rate": 8.74967465213879e-06, "loss": 0.4408, "step": 24274 }, { "epoch": 1.6291735176671924, "grad_norm": 1.0415549278259277, "learning_rate": 8.743533694272638e-06, "loss": 0.4838, "step": 24276 }, { "epoch": 1.6293077413509613, "grad_norm": 1.0316137075424194, "learning_rate": 8.737394685693378e-06, "loss": 0.4813, "step": 24278 }, { "epoch": 1.6294419650347303, "grad_norm": 1.128528356552124, "learning_rate": 8.731257626691092e-06, "loss": 0.4627, "step": 24280 }, { "epoch": 1.6295761887184994, "grad_norm": 0.9621889591217041, "learning_rate": 8.725122517555734e-06, "loss": 0.443, "step": 24282 }, { "epoch": 1.6297104124022683, "grad_norm": 1.0137451887130737, "learning_rate": 8.718989358577167e-06, "loss": 0.503, "step": 24284 }, { "epoch": 1.6298446360860375, "grad_norm": 1.077551245689392, "learning_rate": 8.712858150045172e-06, "loss": 0.4421, "step": 24286 }, { "epoch": 1.6299788597698064, "grad_norm": 1.0536868572235107, "learning_rate": 8.706728892249449e-06, "loss": 0.4551, "step": 24288 }, { "epoch": 1.6301130834535753, "grad_norm": 1.1465702056884766, "learning_rate": 8.700601585479579e-06, "loss": 0.4962, "step": 24290 }, { "epoch": 1.6302473071373442, "grad_norm": 0.9732354283332825, "learning_rate": 8.694476230025094e-06, "loss": 0.4252, "step": 24292 }, { "epoch": 1.6303815308211134, "grad_norm": 0.9033833146095276, "learning_rate": 8.68835282617536e-06, "loss": 0.4583, "step": 24294 }, { "epoch": 1.6305157545048825, "grad_norm": 1.121772050857544, "learning_rate": 8.68223137421973e-06, "loss": 0.4867, "step": 24296 }, { "epoch": 1.6306499781886514, "grad_norm": 1.051953911781311, "learning_rate": 8.676111874447407e-06, "loss": 0.4999, "step": 24298 }, { "epoch": 1.6307842018724203, "grad_norm": 1.007792592048645, "learning_rate": 8.669994327147552e-06, "loss": 0.541, "step": 24300 }, { "epoch": 1.6309184255561893, "grad_norm": 0.9095066785812378, "learning_rate": 8.663878732609187e-06, "loss": 0.4635, "step": 24302 }, { "epoch": 1.6310526492399584, "grad_norm": 0.9789410829544067, "learning_rate": 8.657765091121273e-06, "loss": 0.4391, "step": 24304 }, { "epoch": 1.6311868729237275, "grad_norm": 1.049037218093872, "learning_rate": 8.651653402972654e-06, "loss": 0.4506, "step": 24306 }, { "epoch": 1.6313210966074965, "grad_norm": 0.935494065284729, "learning_rate": 8.645543668452112e-06, "loss": 0.4854, "step": 24308 }, { "epoch": 1.6314553202912654, "grad_norm": 1.0862587690353394, "learning_rate": 8.639435887848307e-06, "loss": 0.4681, "step": 24310 }, { "epoch": 1.6315895439750343, "grad_norm": 0.9944185614585876, "learning_rate": 8.63333006144983e-06, "loss": 0.436, "step": 24312 }, { "epoch": 1.6317237676588034, "grad_norm": 1.0487542152404785, "learning_rate": 8.627226189545162e-06, "loss": 0.4708, "step": 24314 }, { "epoch": 1.6318579913425724, "grad_norm": 1.168813705444336, "learning_rate": 8.621124272422688e-06, "loss": 0.4989, "step": 24316 }, { "epoch": 1.6319922150263415, "grad_norm": 1.0781807899475098, "learning_rate": 8.61502431037074e-06, "loss": 0.5246, "step": 24318 }, { "epoch": 1.6321264387101104, "grad_norm": 1.1735255718231201, "learning_rate": 8.608926303677506e-06, "loss": 0.5324, "step": 24320 }, { "epoch": 1.6322606623938793, "grad_norm": 1.054585337638855, "learning_rate": 8.602830252631116e-06, "loss": 0.4316, "step": 24322 }, { "epoch": 1.6323948860776483, "grad_norm": 1.0092709064483643, "learning_rate": 8.59673615751958e-06, "loss": 0.4439, "step": 24324 }, { "epoch": 1.6325291097614174, "grad_norm": 0.9695367813110352, "learning_rate": 8.590644018630861e-06, "loss": 0.4566, "step": 24326 }, { "epoch": 1.6326633334451865, "grad_norm": 1.1469132900238037, "learning_rate": 8.584553836252768e-06, "loss": 0.503, "step": 24328 }, { "epoch": 1.6327975571289555, "grad_norm": 0.9535269737243652, "learning_rate": 8.578465610673076e-06, "loss": 0.4698, "step": 24330 }, { "epoch": 1.6329317808127244, "grad_norm": 1.1265140771865845, "learning_rate": 8.572379342179437e-06, "loss": 0.4746, "step": 24332 }, { "epoch": 1.6330660044964933, "grad_norm": 0.9789575934410095, "learning_rate": 8.566295031059407e-06, "loss": 0.4916, "step": 24334 }, { "epoch": 1.6332002281802624, "grad_norm": 1.0417834520339966, "learning_rate": 8.560212677600448e-06, "loss": 0.4428, "step": 24336 }, { "epoch": 1.6333344518640314, "grad_norm": 1.0458998680114746, "learning_rate": 8.554132282089967e-06, "loss": 0.4488, "step": 24338 }, { "epoch": 1.6334686755478005, "grad_norm": 1.9805901050567627, "learning_rate": 8.548053844815236e-06, "loss": 0.4567, "step": 24340 }, { "epoch": 1.6336028992315694, "grad_norm": 1.0965425968170166, "learning_rate": 8.541977366063448e-06, "loss": 0.4532, "step": 24342 }, { "epoch": 1.6337371229153383, "grad_norm": 0.953021764755249, "learning_rate": 8.53590284612169e-06, "loss": 0.497, "step": 24344 }, { "epoch": 1.6338713465991073, "grad_norm": 1.0408341884613037, "learning_rate": 8.529830285277001e-06, "loss": 0.4347, "step": 24346 }, { "epoch": 1.6340055702828764, "grad_norm": 1.0203832387924194, "learning_rate": 8.523759683816274e-06, "loss": 0.491, "step": 24348 }, { "epoch": 1.6341397939666455, "grad_norm": 0.9675636887550354, "learning_rate": 8.517691042026365e-06, "loss": 0.5657, "step": 24350 }, { "epoch": 1.6342740176504145, "grad_norm": 0.9288319945335388, "learning_rate": 8.51162436019396e-06, "loss": 0.4978, "step": 24352 }, { "epoch": 1.6344082413341834, "grad_norm": 1.0077799558639526, "learning_rate": 8.505559638605732e-06, "loss": 0.4438, "step": 24354 }, { "epoch": 1.6345424650179523, "grad_norm": 0.9253379106521606, "learning_rate": 8.499496877548202e-06, "loss": 0.4669, "step": 24356 }, { "epoch": 1.6346766887017214, "grad_norm": 1.0141042470932007, "learning_rate": 8.493436077307848e-06, "loss": 0.4449, "step": 24358 }, { "epoch": 1.6348109123854904, "grad_norm": 0.9649704098701477, "learning_rate": 8.487377238171024e-06, "loss": 0.4532, "step": 24360 }, { "epoch": 1.6349451360692595, "grad_norm": 1.0539889335632324, "learning_rate": 8.481320360423994e-06, "loss": 0.4502, "step": 24362 }, { "epoch": 1.6350793597530284, "grad_norm": 1.0667953491210938, "learning_rate": 8.47526544435292e-06, "loss": 0.4614, "step": 24364 }, { "epoch": 1.6352135834367973, "grad_norm": 1.1892738342285156, "learning_rate": 8.469212490243911e-06, "loss": 0.4995, "step": 24366 }, { "epoch": 1.6353478071205663, "grad_norm": 0.9447067379951477, "learning_rate": 8.463161498382948e-06, "loss": 0.4993, "step": 24368 }, { "epoch": 1.6354820308043354, "grad_norm": 0.9807162880897522, "learning_rate": 8.457112469055923e-06, "loss": 0.5024, "step": 24370 }, { "epoch": 1.6356162544881045, "grad_norm": 1.0578080415725708, "learning_rate": 8.45106540254863e-06, "loss": 0.4119, "step": 24372 }, { "epoch": 1.6357504781718735, "grad_norm": 0.9289562702178955, "learning_rate": 8.445020299146812e-06, "loss": 0.4189, "step": 24374 }, { "epoch": 1.6358847018556424, "grad_norm": 0.8946725726127625, "learning_rate": 8.438977159136063e-06, "loss": 0.443, "step": 24376 }, { "epoch": 1.6360189255394113, "grad_norm": 1.1086225509643555, "learning_rate": 8.432935982801921e-06, "loss": 0.477, "step": 24378 }, { "epoch": 1.6361531492231804, "grad_norm": 1.088677167892456, "learning_rate": 8.426896770429815e-06, "loss": 0.4793, "step": 24380 }, { "epoch": 1.6362873729069496, "grad_norm": 1.0079419612884521, "learning_rate": 8.420859522305075e-06, "loss": 0.4715, "step": 24382 }, { "epoch": 1.6364215965907185, "grad_norm": 0.9994024634361267, "learning_rate": 8.41482423871297e-06, "loss": 0.4382, "step": 24384 }, { "epoch": 1.6365558202744874, "grad_norm": 1.1206235885620117, "learning_rate": 8.408790919938636e-06, "loss": 0.4492, "step": 24386 }, { "epoch": 1.6366900439582563, "grad_norm": 1.0920964479446411, "learning_rate": 8.402759566267171e-06, "loss": 0.4001, "step": 24388 }, { "epoch": 1.6368242676420255, "grad_norm": 1.0045682191848755, "learning_rate": 8.396730177983497e-06, "loss": 0.543, "step": 24390 }, { "epoch": 1.6369584913257944, "grad_norm": 1.0271632671356201, "learning_rate": 8.39070275537252e-06, "loss": 0.4419, "step": 24392 }, { "epoch": 1.6370927150095635, "grad_norm": 0.9291853904724121, "learning_rate": 8.384677298719006e-06, "loss": 0.5088, "step": 24394 }, { "epoch": 1.6372269386933325, "grad_norm": 1.037255883216858, "learning_rate": 8.378653808307668e-06, "loss": 0.4374, "step": 24396 }, { "epoch": 1.6373611623771014, "grad_norm": 1.1152249574661255, "learning_rate": 8.372632284423094e-06, "loss": 0.4607, "step": 24398 }, { "epoch": 1.6374953860608703, "grad_norm": 0.9363716244697571, "learning_rate": 8.366612727349787e-06, "loss": 0.503, "step": 24400 }, { "epoch": 1.6376296097446394, "grad_norm": 1.0684161186218262, "learning_rate": 8.360595137372151e-06, "loss": 0.5139, "step": 24402 }, { "epoch": 1.6377638334284086, "grad_norm": 0.9547727704048157, "learning_rate": 8.354579514774525e-06, "loss": 0.4621, "step": 24404 }, { "epoch": 1.6378980571121775, "grad_norm": 1.076237440109253, "learning_rate": 8.348565859841123e-06, "loss": 0.5022, "step": 24406 }, { "epoch": 1.6380322807959464, "grad_norm": 1.1423180103302002, "learning_rate": 8.342554172856076e-06, "loss": 0.5077, "step": 24408 }, { "epoch": 1.6381665044797153, "grad_norm": 0.9638442397117615, "learning_rate": 8.336544454103424e-06, "loss": 0.4551, "step": 24410 }, { "epoch": 1.6383007281634845, "grad_norm": 0.9665816426277161, "learning_rate": 8.330536703867126e-06, "loss": 0.4097, "step": 24412 }, { "epoch": 1.6384349518472534, "grad_norm": 0.9710249900817871, "learning_rate": 8.324530922431018e-06, "loss": 0.454, "step": 24414 }, { "epoch": 1.6385691755310225, "grad_norm": 0.9155290126800537, "learning_rate": 8.3185271100789e-06, "loss": 0.4525, "step": 24416 }, { "epoch": 1.6387033992147915, "grad_norm": 1.1133970022201538, "learning_rate": 8.312525267094385e-06, "loss": 0.4567, "step": 24418 }, { "epoch": 1.6388376228985604, "grad_norm": 1.2855124473571777, "learning_rate": 8.306525393761095e-06, "loss": 0.4322, "step": 24420 }, { "epoch": 1.6389718465823293, "grad_norm": 1.0277609825134277, "learning_rate": 8.300527490362476e-06, "loss": 0.4718, "step": 24422 }, { "epoch": 1.6391060702660984, "grad_norm": 1.107844591140747, "learning_rate": 8.294531557181945e-06, "loss": 0.518, "step": 24424 }, { "epoch": 1.6392402939498676, "grad_norm": 1.0599548816680908, "learning_rate": 8.288537594502788e-06, "loss": 0.4354, "step": 24426 }, { "epoch": 1.6393745176336365, "grad_norm": 1.0279606580734253, "learning_rate": 8.282545602608211e-06, "loss": 0.4515, "step": 24428 }, { "epoch": 1.6395087413174054, "grad_norm": 1.03935968875885, "learning_rate": 8.276555581781303e-06, "loss": 0.4835, "step": 24430 }, { "epoch": 1.6396429650011743, "grad_norm": 1.0039790868759155, "learning_rate": 8.270567532305118e-06, "loss": 0.461, "step": 24432 }, { "epoch": 1.6397771886849435, "grad_norm": 0.9484822154045105, "learning_rate": 8.264581454462555e-06, "loss": 0.4623, "step": 24434 }, { "epoch": 1.6399114123687124, "grad_norm": 0.9461115002632141, "learning_rate": 8.25859734853645e-06, "loss": 0.4257, "step": 24436 }, { "epoch": 1.6400456360524815, "grad_norm": 1.037699818611145, "learning_rate": 8.252615214809528e-06, "loss": 0.511, "step": 24438 }, { "epoch": 1.6401798597362505, "grad_norm": 1.0340139865875244, "learning_rate": 8.246635053564455e-06, "loss": 0.5021, "step": 24440 }, { "epoch": 1.6403140834200194, "grad_norm": 1.0507210493087769, "learning_rate": 8.240656865083757e-06, "loss": 0.4525, "step": 24442 }, { "epoch": 1.6404483071037883, "grad_norm": 0.9403718113899231, "learning_rate": 8.234680649649935e-06, "loss": 0.4646, "step": 24444 }, { "epoch": 1.6405825307875574, "grad_norm": 0.8724382519721985, "learning_rate": 8.228706407545306e-06, "loss": 0.467, "step": 24446 }, { "epoch": 1.6407167544713266, "grad_norm": 1.0470285415649414, "learning_rate": 8.222734139052152e-06, "loss": 0.5314, "step": 24448 }, { "epoch": 1.6408509781550955, "grad_norm": 0.9047724604606628, "learning_rate": 8.21676384445267e-06, "loss": 0.4392, "step": 24450 }, { "epoch": 1.6409852018388644, "grad_norm": 0.9392143487930298, "learning_rate": 8.210795524028924e-06, "loss": 0.5069, "step": 24452 }, { "epoch": 1.6411194255226333, "grad_norm": 1.0471280813217163, "learning_rate": 8.204829178062923e-06, "loss": 0.4784, "step": 24454 }, { "epoch": 1.6412536492064025, "grad_norm": 1.150148868560791, "learning_rate": 8.198864806836553e-06, "loss": 0.4711, "step": 24456 }, { "epoch": 1.6413878728901716, "grad_norm": 1.0669646263122559, "learning_rate": 8.192902410631632e-06, "loss": 0.4197, "step": 24458 }, { "epoch": 1.6415220965739405, "grad_norm": 1.0547282695770264, "learning_rate": 8.18694198972984e-06, "loss": 0.4555, "step": 24460 }, { "epoch": 1.6416563202577095, "grad_norm": 1.076779842376709, "learning_rate": 8.180983544412834e-06, "loss": 0.4827, "step": 24462 }, { "epoch": 1.6417905439414784, "grad_norm": 0.961208701133728, "learning_rate": 8.175027074962122e-06, "loss": 0.4254, "step": 24464 }, { "epoch": 1.6419247676252475, "grad_norm": 1.114872932434082, "learning_rate": 8.169072581659137e-06, "loss": 0.4758, "step": 24466 }, { "epoch": 1.6420589913090164, "grad_norm": 1.0683403015136719, "learning_rate": 8.163120064785201e-06, "loss": 0.4474, "step": 24468 }, { "epoch": 1.6421932149927856, "grad_norm": 0.8533875942230225, "learning_rate": 8.157169524621583e-06, "loss": 0.501, "step": 24470 }, { "epoch": 1.6423274386765545, "grad_norm": 1.0208030939102173, "learning_rate": 8.15122096144943e-06, "loss": 0.4891, "step": 24472 }, { "epoch": 1.6424616623603234, "grad_norm": 0.895160973072052, "learning_rate": 8.145274375549794e-06, "loss": 0.4192, "step": 24474 }, { "epoch": 1.6425958860440923, "grad_norm": 0.9571309089660645, "learning_rate": 8.139329767203624e-06, "loss": 0.4343, "step": 24476 }, { "epoch": 1.6427301097278615, "grad_norm": 1.2458082437515259, "learning_rate": 8.133387136691822e-06, "loss": 0.4861, "step": 24478 }, { "epoch": 1.6428643334116306, "grad_norm": 1.117855191230774, "learning_rate": 8.127446484295137e-06, "loss": 0.4511, "step": 24480 }, { "epoch": 1.6429985570953995, "grad_norm": 0.957265317440033, "learning_rate": 8.121507810294276e-06, "loss": 0.4072, "step": 24482 }, { "epoch": 1.6431327807791685, "grad_norm": 0.9388834238052368, "learning_rate": 8.115571114969828e-06, "loss": 0.4677, "step": 24484 }, { "epoch": 1.6432670044629374, "grad_norm": 1.0905842781066895, "learning_rate": 8.109636398602276e-06, "loss": 0.5119, "step": 24486 }, { "epoch": 1.6434012281467065, "grad_norm": 1.0503604412078857, "learning_rate": 8.103703661472023e-06, "loss": 0.4732, "step": 24488 }, { "epoch": 1.6435354518304754, "grad_norm": 1.1269357204437256, "learning_rate": 8.09777290385939e-06, "loss": 0.5235, "step": 24490 }, { "epoch": 1.6436696755142446, "grad_norm": 0.9764650464057922, "learning_rate": 8.091844126044601e-06, "loss": 0.5161, "step": 24492 }, { "epoch": 1.6438038991980135, "grad_norm": 1.036859393119812, "learning_rate": 8.085917328307763e-06, "loss": 0.5158, "step": 24494 }, { "epoch": 1.6439381228817824, "grad_norm": 1.1827311515808105, "learning_rate": 8.079992510928897e-06, "loss": 0.4658, "step": 24496 }, { "epoch": 1.6440723465655513, "grad_norm": 1.084754228591919, "learning_rate": 8.074069674187968e-06, "loss": 0.4841, "step": 24498 }, { "epoch": 1.6442065702493205, "grad_norm": 0.7590323686599731, "learning_rate": 8.068148818364802e-06, "loss": 0.3998, "step": 24500 }, { "epoch": 1.6443407939330896, "grad_norm": 0.8609249591827393, "learning_rate": 8.062229943739148e-06, "loss": 0.4364, "step": 24502 }, { "epoch": 1.6444750176168585, "grad_norm": 1.0397735834121704, "learning_rate": 8.056313050590652e-06, "loss": 0.4556, "step": 24504 }, { "epoch": 1.6446092413006275, "grad_norm": 0.9933218359947205, "learning_rate": 8.050398139198895e-06, "loss": 0.4548, "step": 24506 }, { "epoch": 1.6447434649843964, "grad_norm": 1.1229078769683838, "learning_rate": 8.044485209843327e-06, "loss": 0.4544, "step": 24508 }, { "epoch": 1.6448776886681655, "grad_norm": 1.0707441568374634, "learning_rate": 8.03857426280334e-06, "loss": 0.5081, "step": 24510 }, { "epoch": 1.6450119123519344, "grad_norm": 1.0590848922729492, "learning_rate": 8.032665298358205e-06, "loss": 0.4747, "step": 24512 }, { "epoch": 1.6451461360357036, "grad_norm": 0.9965388774871826, "learning_rate": 8.026758316787108e-06, "loss": 0.4613, "step": 24514 }, { "epoch": 1.6452803597194725, "grad_norm": 1.0541249513626099, "learning_rate": 8.020853318369149e-06, "loss": 0.4614, "step": 24516 }, { "epoch": 1.6454145834032414, "grad_norm": 1.0630964040756226, "learning_rate": 8.014950303383305e-06, "loss": 0.4383, "step": 24518 }, { "epoch": 1.6455488070870103, "grad_norm": 1.1048094034194946, "learning_rate": 8.009049272108516e-06, "loss": 0.4591, "step": 24520 }, { "epoch": 1.6456830307707795, "grad_norm": 1.175954818725586, "learning_rate": 8.003150224823574e-06, "loss": 0.4995, "step": 24522 }, { "epoch": 1.6458172544545486, "grad_norm": 0.9277005195617676, "learning_rate": 7.997253161807205e-06, "loss": 0.414, "step": 24524 }, { "epoch": 1.6459514781383175, "grad_norm": 0.9522114396095276, "learning_rate": 7.991358083338013e-06, "loss": 0.4443, "step": 24526 }, { "epoch": 1.6460857018220865, "grad_norm": 0.9481935501098633, "learning_rate": 7.985464989694558e-06, "loss": 0.4982, "step": 24528 }, { "epoch": 1.6462199255058554, "grad_norm": 1.0402849912643433, "learning_rate": 7.979573881155261e-06, "loss": 0.4713, "step": 24530 }, { "epoch": 1.6463541491896245, "grad_norm": 1.096053957939148, "learning_rate": 7.973684757998472e-06, "loss": 0.5413, "step": 24532 }, { "epoch": 1.6464883728733937, "grad_norm": 1.1752251386642456, "learning_rate": 7.967797620502426e-06, "loss": 0.4956, "step": 24534 }, { "epoch": 1.6466225965571626, "grad_norm": 1.226892113685608, "learning_rate": 7.961912468945304e-06, "loss": 0.4958, "step": 24536 }, { "epoch": 1.6467568202409315, "grad_norm": 1.0735069513320923, "learning_rate": 7.956029303605134e-06, "loss": 0.4744, "step": 24538 }, { "epoch": 1.6468910439247004, "grad_norm": 0.8980748057365417, "learning_rate": 7.950148124759932e-06, "loss": 0.4323, "step": 24540 }, { "epoch": 1.6470252676084696, "grad_norm": 1.162540316581726, "learning_rate": 7.944268932687521e-06, "loss": 0.5048, "step": 24542 }, { "epoch": 1.6471594912922385, "grad_norm": 0.966513454914093, "learning_rate": 7.938391727665712e-06, "loss": 0.4728, "step": 24544 }, { "epoch": 1.6472937149760076, "grad_norm": 1.059916615486145, "learning_rate": 7.93251650997217e-06, "loss": 0.4523, "step": 24546 }, { "epoch": 1.6474279386597765, "grad_norm": 1.071812391281128, "learning_rate": 7.926643279884521e-06, "loss": 0.533, "step": 24548 }, { "epoch": 1.6475621623435455, "grad_norm": 0.8966371417045593, "learning_rate": 7.920772037680236e-06, "loss": 0.4953, "step": 24550 }, { "epoch": 1.6476963860273144, "grad_norm": 0.9865714907646179, "learning_rate": 7.914902783636729e-06, "loss": 0.4738, "step": 24552 }, { "epoch": 1.6478306097110835, "grad_norm": 0.9470037221908569, "learning_rate": 7.909035518031299e-06, "loss": 0.4447, "step": 24554 }, { "epoch": 1.6479648333948527, "grad_norm": 1.0400514602661133, "learning_rate": 7.903170241141178e-06, "loss": 0.4488, "step": 24556 }, { "epoch": 1.6480990570786216, "grad_norm": 0.9073011875152588, "learning_rate": 7.897306953243489e-06, "loss": 0.4574, "step": 24558 }, { "epoch": 1.6482332807623905, "grad_norm": 0.8621808886528015, "learning_rate": 7.891445654615254e-06, "loss": 0.4469, "step": 24560 }, { "epoch": 1.6483675044461594, "grad_norm": 0.9729363918304443, "learning_rate": 7.885586345533397e-06, "loss": 0.4661, "step": 24562 }, { "epoch": 1.6485017281299286, "grad_norm": 1.0364904403686523, "learning_rate": 7.879729026274779e-06, "loss": 0.4614, "step": 24564 }, { "epoch": 1.6486359518136975, "grad_norm": 0.9890289902687073, "learning_rate": 7.873873697116129e-06, "loss": 0.4435, "step": 24566 }, { "epoch": 1.6487701754974666, "grad_norm": 0.9436256885528564, "learning_rate": 7.868020358334139e-06, "loss": 0.4559, "step": 24568 }, { "epoch": 1.6489043991812355, "grad_norm": 1.0989784002304077, "learning_rate": 7.862169010205306e-06, "loss": 0.456, "step": 24570 }, { "epoch": 1.6490386228650045, "grad_norm": 1.0259724855422974, "learning_rate": 7.856319653006144e-06, "loss": 0.4262, "step": 24572 }, { "epoch": 1.6491728465487734, "grad_norm": 1.0732414722442627, "learning_rate": 7.850472287012988e-06, "loss": 0.4381, "step": 24574 }, { "epoch": 1.6493070702325425, "grad_norm": 1.0366039276123047, "learning_rate": 7.84462691250215e-06, "loss": 0.5088, "step": 24576 }, { "epoch": 1.6494412939163117, "grad_norm": 1.1138050556182861, "learning_rate": 7.8387835297498e-06, "loss": 0.4757, "step": 24578 }, { "epoch": 1.6495755176000806, "grad_norm": 1.0513672828674316, "learning_rate": 7.832942139032012e-06, "loss": 0.4919, "step": 24580 }, { "epoch": 1.6497097412838495, "grad_norm": 0.8169397711753845, "learning_rate": 7.827102740624798e-06, "loss": 0.4306, "step": 24582 }, { "epoch": 1.6498439649676184, "grad_norm": 1.0516151189804077, "learning_rate": 7.821265334804028e-06, "loss": 0.458, "step": 24584 }, { "epoch": 1.6499781886513876, "grad_norm": 0.8732035160064697, "learning_rate": 7.815429921845553e-06, "loss": 0.3974, "step": 24586 }, { "epoch": 1.6501124123351565, "grad_norm": 0.8667218685150146, "learning_rate": 7.809596502025057e-06, "loss": 0.4782, "step": 24588 }, { "epoch": 1.6502466360189256, "grad_norm": 0.9133517742156982, "learning_rate": 7.803765075618164e-06, "loss": 0.4389, "step": 24590 }, { "epoch": 1.6503808597026945, "grad_norm": 1.0821890830993652, "learning_rate": 7.797935642900378e-06, "loss": 0.465, "step": 24592 }, { "epoch": 1.6505150833864635, "grad_norm": 0.9128795862197876, "learning_rate": 7.792108204147164e-06, "loss": 0.4588, "step": 24594 }, { "epoch": 1.6506493070702324, "grad_norm": 1.2393850088119507, "learning_rate": 7.786282759633839e-06, "loss": 0.4522, "step": 24596 }, { "epoch": 1.6507835307540015, "grad_norm": 1.0295946598052979, "learning_rate": 7.78045930963564e-06, "loss": 0.511, "step": 24598 }, { "epoch": 1.6509177544377707, "grad_norm": 1.0016932487487793, "learning_rate": 7.774637854427707e-06, "loss": 0.443, "step": 24600 }, { "epoch": 1.6510519781215396, "grad_norm": 0.9255552887916565, "learning_rate": 7.768818394285116e-06, "loss": 0.4247, "step": 24602 }, { "epoch": 1.6511862018053085, "grad_norm": 1.1796424388885498, "learning_rate": 7.7630009294828e-06, "loss": 0.4073, "step": 24604 }, { "epoch": 1.6513204254890774, "grad_norm": 0.8748919367790222, "learning_rate": 7.757185460295657e-06, "loss": 0.3719, "step": 24606 }, { "epoch": 1.6514546491728466, "grad_norm": 1.1171672344207764, "learning_rate": 7.751371986998412e-06, "loss": 0.4661, "step": 24608 }, { "epoch": 1.6515888728566157, "grad_norm": 1.030329704284668, "learning_rate": 7.745560509865773e-06, "loss": 0.4457, "step": 24610 }, { "epoch": 1.6517230965403846, "grad_norm": 1.0265859365463257, "learning_rate": 7.739751029172298e-06, "loss": 0.4772, "step": 24612 }, { "epoch": 1.6518573202241535, "grad_norm": 1.0171829462051392, "learning_rate": 7.7339435451925e-06, "loss": 0.4653, "step": 24614 }, { "epoch": 1.6519915439079225, "grad_norm": 1.1498104333877563, "learning_rate": 7.728138058200757e-06, "loss": 0.4804, "step": 24616 }, { "epoch": 1.6521257675916916, "grad_norm": 1.0023661851882935, "learning_rate": 7.722334568471362e-06, "loss": 0.4265, "step": 24618 }, { "epoch": 1.6522599912754605, "grad_norm": 0.9971412420272827, "learning_rate": 7.716533076278515e-06, "loss": 0.44, "step": 24620 }, { "epoch": 1.6523942149592297, "grad_norm": 1.0238125324249268, "learning_rate": 7.710733581896345e-06, "loss": 0.4527, "step": 24622 }, { "epoch": 1.6525284386429986, "grad_norm": 0.9874955415725708, "learning_rate": 7.704936085598852e-06, "loss": 0.466, "step": 24624 }, { "epoch": 1.6526626623267675, "grad_norm": 1.1266205310821533, "learning_rate": 7.699140587659965e-06, "loss": 0.5204, "step": 24626 }, { "epoch": 1.6527968860105364, "grad_norm": 0.9751604795455933, "learning_rate": 7.693347088353487e-06, "loss": 0.4418, "step": 24628 }, { "epoch": 1.6529311096943056, "grad_norm": 0.9777958393096924, "learning_rate": 7.687555587953176e-06, "loss": 0.452, "step": 24630 }, { "epoch": 1.6530653333780747, "grad_norm": 1.0757713317871094, "learning_rate": 7.681766086732655e-06, "loss": 0.4294, "step": 24632 }, { "epoch": 1.6531995570618436, "grad_norm": 0.9893507957458496, "learning_rate": 7.675978584965482e-06, "loss": 0.4502, "step": 24634 }, { "epoch": 1.6533337807456125, "grad_norm": 1.0750385522842407, "learning_rate": 7.670193082925092e-06, "loss": 0.474, "step": 24636 }, { "epoch": 1.6534680044293815, "grad_norm": 0.9043234586715698, "learning_rate": 7.66440958088484e-06, "loss": 0.4602, "step": 24638 }, { "epoch": 1.6536022281131506, "grad_norm": 1.0703434944152832, "learning_rate": 7.658628079117974e-06, "loss": 0.4676, "step": 24640 }, { "epoch": 1.6537364517969195, "grad_norm": 0.8634180426597595, "learning_rate": 7.652848577897681e-06, "loss": 0.4362, "step": 24642 }, { "epoch": 1.6538706754806887, "grad_norm": 1.051302433013916, "learning_rate": 7.647071077497025e-06, "loss": 0.5006, "step": 24644 }, { "epoch": 1.6540048991644576, "grad_norm": 1.3736884593963623, "learning_rate": 7.641295578188973e-06, "loss": 0.5256, "step": 24646 }, { "epoch": 1.6541391228482265, "grad_norm": 0.9008198380470276, "learning_rate": 7.635522080246404e-06, "loss": 0.4089, "step": 24648 }, { "epoch": 1.6542733465319954, "grad_norm": 0.9291314482688904, "learning_rate": 7.62975058394212e-06, "loss": 0.4096, "step": 24650 }, { "epoch": 1.6544075702157646, "grad_norm": 1.099732756614685, "learning_rate": 7.623981089548798e-06, "loss": 0.4814, "step": 24652 }, { "epoch": 1.6545417938995337, "grad_norm": 1.053756833076477, "learning_rate": 7.618213597339046e-06, "loss": 0.5251, "step": 24654 }, { "epoch": 1.6546760175833026, "grad_norm": 1.060849905014038, "learning_rate": 7.612448107585363e-06, "loss": 0.4829, "step": 24656 }, { "epoch": 1.6548102412670715, "grad_norm": 1.047171950340271, "learning_rate": 7.6066846205601396e-06, "loss": 0.4898, "step": 24658 }, { "epoch": 1.6549444649508405, "grad_norm": 0.9501131772994995, "learning_rate": 7.600923136535715e-06, "loss": 0.5317, "step": 24660 }, { "epoch": 1.6550786886346096, "grad_norm": 1.0560083389282227, "learning_rate": 7.595163655784293e-06, "loss": 0.5048, "step": 24662 }, { "epoch": 1.6552129123183785, "grad_norm": 1.015169620513916, "learning_rate": 7.589406178578018e-06, "loss": 0.4694, "step": 24664 }, { "epoch": 1.6553471360021477, "grad_norm": 0.9952278137207031, "learning_rate": 7.5836507051888885e-06, "loss": 0.445, "step": 24666 }, { "epoch": 1.6554813596859166, "grad_norm": 0.9039380550384521, "learning_rate": 7.577897235888865e-06, "loss": 0.4914, "step": 24668 }, { "epoch": 1.6556155833696855, "grad_norm": 1.1089794635772705, "learning_rate": 7.5721457709497625e-06, "loss": 0.6241, "step": 24670 }, { "epoch": 1.6557498070534544, "grad_norm": 1.0138074159622192, "learning_rate": 7.566396310643353e-06, "loss": 0.4944, "step": 24672 }, { "epoch": 1.6558840307372236, "grad_norm": 1.1758661270141602, "learning_rate": 7.560648855241276e-06, "loss": 0.4968, "step": 24674 }, { "epoch": 1.6560182544209927, "grad_norm": 0.9619575142860413, "learning_rate": 7.5549034050150906e-06, "loss": 0.4951, "step": 24676 }, { "epoch": 1.6561524781047616, "grad_norm": 0.926916778087616, "learning_rate": 7.549159960236241e-06, "loss": 0.437, "step": 24678 }, { "epoch": 1.6562867017885305, "grad_norm": 0.969592273235321, "learning_rate": 7.543418521176121e-06, "loss": 0.4218, "step": 24680 }, { "epoch": 1.6564209254722995, "grad_norm": 1.0959103107452393, "learning_rate": 7.537679088105987e-06, "loss": 0.4791, "step": 24682 }, { "epoch": 1.6565551491560686, "grad_norm": 1.1442915201187134, "learning_rate": 7.531941661297015e-06, "loss": 0.546, "step": 24684 }, { "epoch": 1.6566893728398377, "grad_norm": 1.033138394355774, "learning_rate": 7.526206241020278e-06, "loss": 0.478, "step": 24686 }, { "epoch": 1.6568235965236067, "grad_norm": 1.0051465034484863, "learning_rate": 7.520472827546793e-06, "loss": 0.4685, "step": 24688 }, { "epoch": 1.6569578202073756, "grad_norm": 1.0380859375, "learning_rate": 7.514741421147415e-06, "loss": 0.4148, "step": 24690 }, { "epoch": 1.6570920438911445, "grad_norm": 1.1546108722686768, "learning_rate": 7.509012022092993e-06, "loss": 0.5254, "step": 24692 }, { "epoch": 1.6572262675749136, "grad_norm": 1.1542062759399414, "learning_rate": 7.503284630654167e-06, "loss": 0.5089, "step": 24694 }, { "epoch": 1.6573604912586826, "grad_norm": 1.0694364309310913, "learning_rate": 7.497559247101593e-06, "loss": 0.4592, "step": 24696 }, { "epoch": 1.6574947149424517, "grad_norm": 0.8872901201248169, "learning_rate": 7.491835871705749e-06, "loss": 0.4382, "step": 24698 }, { "epoch": 1.6576289386262206, "grad_norm": 1.0182710886001587, "learning_rate": 7.486114504737091e-06, "loss": 0.5335, "step": 24700 }, { "epoch": 1.6577631623099895, "grad_norm": 1.0655882358551025, "learning_rate": 7.4803951464659185e-06, "loss": 0.4844, "step": 24702 }, { "epoch": 1.6578973859937585, "grad_norm": 1.042083740234375, "learning_rate": 7.4746777971624605e-06, "loss": 0.4664, "step": 24704 }, { "epoch": 1.6580316096775276, "grad_norm": 0.9128807187080383, "learning_rate": 7.468962457096845e-06, "loss": 0.4326, "step": 24706 }, { "epoch": 1.6581658333612967, "grad_norm": 0.8535100817680359, "learning_rate": 7.4632491265391344e-06, "loss": 0.4522, "step": 24708 }, { "epoch": 1.6583000570450657, "grad_norm": 0.9464491605758667, "learning_rate": 7.4575378057592574e-06, "loss": 0.4201, "step": 24710 }, { "epoch": 1.6584342807288346, "grad_norm": 1.014621615409851, "learning_rate": 7.451828495027058e-06, "loss": 0.472, "step": 24712 }, { "epoch": 1.6585685044126035, "grad_norm": 0.9806382656097412, "learning_rate": 7.446121194612288e-06, "loss": 0.4546, "step": 24714 }, { "epoch": 1.6587027280963726, "grad_norm": 0.9317238330841064, "learning_rate": 7.440415904784625e-06, "loss": 0.5104, "step": 24716 }, { "epoch": 1.6588369517801416, "grad_norm": 1.0521879196166992, "learning_rate": 7.434712625813605e-06, "loss": 0.5019, "step": 24718 }, { "epoch": 1.6589711754639107, "grad_norm": 1.0747883319854736, "learning_rate": 7.4290113579687405e-06, "loss": 0.4667, "step": 24720 }, { "epoch": 1.6591053991476796, "grad_norm": 1.1273607015609741, "learning_rate": 7.4233121015193586e-06, "loss": 0.4902, "step": 24722 }, { "epoch": 1.6592396228314485, "grad_norm": 0.9528072476387024, "learning_rate": 7.417614856734751e-06, "loss": 0.4835, "step": 24724 }, { "epoch": 1.6593738465152175, "grad_norm": 1.1074789762496948, "learning_rate": 7.411919623884117e-06, "loss": 0.4096, "step": 24726 }, { "epoch": 1.6595080701989866, "grad_norm": 0.948462188243866, "learning_rate": 7.406226403236522e-06, "loss": 0.4344, "step": 24728 }, { "epoch": 1.6596422938827557, "grad_norm": 0.9927400350570679, "learning_rate": 7.400535195060998e-06, "loss": 0.4515, "step": 24730 }, { "epoch": 1.6597765175665247, "grad_norm": 0.9371539354324341, "learning_rate": 7.394845999626393e-06, "loss": 0.3937, "step": 24732 }, { "epoch": 1.6599107412502936, "grad_norm": 1.1667990684509277, "learning_rate": 7.389158817201542e-06, "loss": 0.5165, "step": 24734 }, { "epoch": 1.6600449649340625, "grad_norm": 1.1483110189437866, "learning_rate": 7.383473648055139e-06, "loss": 0.424, "step": 24736 }, { "epoch": 1.6601791886178316, "grad_norm": 1.0563877820968628, "learning_rate": 7.377790492455816e-06, "loss": 0.5473, "step": 24738 }, { "epoch": 1.6603134123016006, "grad_norm": 1.0585153102874756, "learning_rate": 7.3721093506720795e-06, "loss": 0.42, "step": 24740 }, { "epoch": 1.6604476359853697, "grad_norm": 0.9234572052955627, "learning_rate": 7.366430222972353e-06, "loss": 0.4157, "step": 24742 }, { "epoch": 1.6605818596691386, "grad_norm": 0.918400764465332, "learning_rate": 7.360753109624952e-06, "loss": 0.4457, "step": 24744 }, { "epoch": 1.6607160833529075, "grad_norm": 0.9188838005065918, "learning_rate": 7.355078010898131e-06, "loss": 0.5161, "step": 24746 }, { "epoch": 1.6608503070366765, "grad_norm": 1.0078390836715698, "learning_rate": 7.349404927060011e-06, "loss": 0.4592, "step": 24748 }, { "epoch": 1.6609845307204456, "grad_norm": 1.0019540786743164, "learning_rate": 7.343733858378643e-06, "loss": 0.4762, "step": 24750 }, { "epoch": 1.6611187544042147, "grad_norm": 0.9697957634925842, "learning_rate": 7.338064805121964e-06, "loss": 0.4374, "step": 24752 }, { "epoch": 1.6612529780879837, "grad_norm": 1.0653742551803589, "learning_rate": 7.332397767557836e-06, "loss": 0.4935, "step": 24754 }, { "epoch": 1.6613872017717526, "grad_norm": 0.9930585622787476, "learning_rate": 7.3267327459540015e-06, "loss": 0.4151, "step": 24756 }, { "epoch": 1.6615214254555215, "grad_norm": 1.0316599607467651, "learning_rate": 7.321069740578157e-06, "loss": 0.4034, "step": 24758 }, { "epoch": 1.6616556491392906, "grad_norm": 0.927063524723053, "learning_rate": 7.315408751697817e-06, "loss": 0.4999, "step": 24760 }, { "epoch": 1.6617898728230598, "grad_norm": 1.085348129272461, "learning_rate": 7.309749779580494e-06, "loss": 0.5474, "step": 24762 }, { "epoch": 1.6619240965068287, "grad_norm": 0.9205760955810547, "learning_rate": 7.30409282449353e-06, "loss": 0.4622, "step": 24764 }, { "epoch": 1.6620583201905976, "grad_norm": 0.935708224773407, "learning_rate": 7.298437886704235e-06, "loss": 0.4502, "step": 24766 }, { "epoch": 1.6621925438743665, "grad_norm": 0.9033759832382202, "learning_rate": 7.292784966479782e-06, "loss": 0.4189, "step": 24768 }, { "epoch": 1.6623267675581357, "grad_norm": 0.9785996079444885, "learning_rate": 7.287134064087259e-06, "loss": 0.4686, "step": 24770 }, { "epoch": 1.6624609912419046, "grad_norm": 1.0184147357940674, "learning_rate": 7.2814851797936455e-06, "loss": 0.4644, "step": 24772 }, { "epoch": 1.6625952149256737, "grad_norm": 0.9698885679244995, "learning_rate": 7.275838313865874e-06, "loss": 0.472, "step": 24774 }, { "epoch": 1.6627294386094427, "grad_norm": 0.9263590574264526, "learning_rate": 7.270193466570724e-06, "loss": 0.4197, "step": 24776 }, { "epoch": 1.6628636622932116, "grad_norm": 1.024530053138733, "learning_rate": 7.264550638174905e-06, "loss": 0.4021, "step": 24778 }, { "epoch": 1.6629978859769805, "grad_norm": 0.9894106388092041, "learning_rate": 7.258909828945027e-06, "loss": 0.4817, "step": 24780 }, { "epoch": 1.6631321096607496, "grad_norm": 0.9556596279144287, "learning_rate": 7.2532710391476185e-06, "loss": 0.4675, "step": 24782 }, { "epoch": 1.6632663333445188, "grad_norm": 0.9056533575057983, "learning_rate": 7.247634269049092e-06, "loss": 0.4585, "step": 24784 }, { "epoch": 1.6634005570282877, "grad_norm": 1.0582910776138306, "learning_rate": 7.241999518915793e-06, "loss": 0.4851, "step": 24786 }, { "epoch": 1.6635347807120566, "grad_norm": 1.0909756422042847, "learning_rate": 7.2363667890139265e-06, "loss": 0.4517, "step": 24788 }, { "epoch": 1.6636690043958255, "grad_norm": 1.0948517322540283, "learning_rate": 7.230736079609629e-06, "loss": 0.4894, "step": 24790 }, { "epoch": 1.6638032280795947, "grad_norm": 0.9200692772865295, "learning_rate": 7.22510739096896e-06, "loss": 0.5111, "step": 24792 }, { "epoch": 1.6639374517633636, "grad_norm": 1.0163975954055786, "learning_rate": 7.219480723357841e-06, "loss": 0.4542, "step": 24794 }, { "epoch": 1.6640716754471327, "grad_norm": 1.4413422346115112, "learning_rate": 7.213856077042147e-06, "loss": 0.4463, "step": 24796 }, { "epoch": 1.6642058991309017, "grad_norm": 1.1777487993240356, "learning_rate": 7.208233452287616e-06, "loss": 0.4462, "step": 24798 }, { "epoch": 1.6643401228146706, "grad_norm": 1.193180799484253, "learning_rate": 7.202612849359913e-06, "loss": 0.4186, "step": 24800 }, { "epoch": 1.6644743464984395, "grad_norm": 1.006433129310608, "learning_rate": 7.196994268524576e-06, "loss": 0.4139, "step": 24802 }, { "epoch": 1.6646085701822086, "grad_norm": 0.9738898277282715, "learning_rate": 7.1913777100471096e-06, "loss": 0.4432, "step": 24804 }, { "epoch": 1.6647427938659778, "grad_norm": 1.082888126373291, "learning_rate": 7.185763174192861e-06, "loss": 0.4252, "step": 24806 }, { "epoch": 1.6648770175497467, "grad_norm": 1.2457809448242188, "learning_rate": 7.18015066122712e-06, "loss": 0.5175, "step": 24808 }, { "epoch": 1.6650112412335156, "grad_norm": 0.948826014995575, "learning_rate": 7.174540171415039e-06, "loss": 0.4626, "step": 24810 }, { "epoch": 1.6651454649172845, "grad_norm": 0.8969342112541199, "learning_rate": 7.168931705021737e-06, "loss": 0.4844, "step": 24812 }, { "epoch": 1.6652796886010537, "grad_norm": 0.9707522988319397, "learning_rate": 7.163325262312176e-06, "loss": 0.4855, "step": 24814 }, { "epoch": 1.6654139122848226, "grad_norm": 0.9882151484489441, "learning_rate": 7.157720843551286e-06, "loss": 0.4367, "step": 24816 }, { "epoch": 1.6655481359685917, "grad_norm": 1.0631093978881836, "learning_rate": 7.152118449003819e-06, "loss": 0.478, "step": 24818 }, { "epoch": 1.6656823596523607, "grad_norm": 1.2123793363571167, "learning_rate": 7.146518078934505e-06, "loss": 0.527, "step": 24820 }, { "epoch": 1.6658165833361296, "grad_norm": 0.9780756831169128, "learning_rate": 7.140919733607937e-06, "loss": 0.3846, "step": 24822 }, { "epoch": 1.6659508070198985, "grad_norm": 1.3662919998168945, "learning_rate": 7.135323413288641e-06, "loss": 0.4805, "step": 24824 }, { "epoch": 1.6660850307036676, "grad_norm": 0.8742290139198303, "learning_rate": 7.1297291182410246e-06, "loss": 0.4277, "step": 24826 }, { "epoch": 1.6662192543874368, "grad_norm": 1.0050386190414429, "learning_rate": 7.124136848729407e-06, "loss": 0.476, "step": 24828 }, { "epoch": 1.6663534780712057, "grad_norm": 1.0886958837509155, "learning_rate": 7.118546605018e-06, "loss": 0.4916, "step": 24830 }, { "epoch": 1.6664877017549746, "grad_norm": 1.1330604553222656, "learning_rate": 7.11295838737096e-06, "loss": 0.4312, "step": 24832 }, { "epoch": 1.6666219254387435, "grad_norm": 1.0930204391479492, "learning_rate": 7.107372196052298e-06, "loss": 0.5102, "step": 24834 }, { "epoch": 1.6667561491225127, "grad_norm": 1.0614120960235596, "learning_rate": 7.101788031325951e-06, "loss": 0.4445, "step": 24836 }, { "epoch": 1.6668903728062818, "grad_norm": 0.9910305738449097, "learning_rate": 7.09620589345576e-06, "loss": 0.4942, "step": 24838 }, { "epoch": 1.6670245964900507, "grad_norm": 0.9669861197471619, "learning_rate": 7.0906257827054786e-06, "loss": 0.4726, "step": 24840 }, { "epoch": 1.6671588201738197, "grad_norm": 1.4137396812438965, "learning_rate": 7.085047699338742e-06, "loss": 0.5131, "step": 24842 }, { "epoch": 1.6672930438575886, "grad_norm": 0.9933016300201416, "learning_rate": 7.079471643619135e-06, "loss": 0.4637, "step": 24844 }, { "epoch": 1.6674272675413577, "grad_norm": 1.0396409034729004, "learning_rate": 7.0738976158100715e-06, "loss": 0.5158, "step": 24846 }, { "epoch": 1.6675614912251266, "grad_norm": 1.098038673400879, "learning_rate": 7.068325616174948e-06, "loss": 0.5014, "step": 24848 }, { "epoch": 1.6676957149088958, "grad_norm": 1.0721778869628906, "learning_rate": 7.062755644977004e-06, "loss": 0.4014, "step": 24850 }, { "epoch": 1.6678299385926647, "grad_norm": 1.0025501251220703, "learning_rate": 7.05718770247944e-06, "loss": 0.4526, "step": 24852 }, { "epoch": 1.6679641622764336, "grad_norm": 1.0482416152954102, "learning_rate": 7.051621788945306e-06, "loss": 0.4261, "step": 24854 }, { "epoch": 1.6680983859602025, "grad_norm": 1.5061668157577515, "learning_rate": 7.046057904637593e-06, "loss": 0.4502, "step": 24856 }, { "epoch": 1.6682326096439717, "grad_norm": 1.000950813293457, "learning_rate": 7.040496049819178e-06, "loss": 0.4135, "step": 24858 }, { "epoch": 1.6683668333277408, "grad_norm": 1.0769424438476562, "learning_rate": 7.034936224752836e-06, "loss": 0.5046, "step": 24860 }, { "epoch": 1.6685010570115097, "grad_norm": 1.056839942932129, "learning_rate": 7.029378429701278e-06, "loss": 0.4646, "step": 24862 }, { "epoch": 1.6686352806952787, "grad_norm": 0.93470698595047, "learning_rate": 7.023822664927099e-06, "loss": 0.5146, "step": 24864 }, { "epoch": 1.6687695043790476, "grad_norm": 1.0354546308517456, "learning_rate": 7.018268930692784e-06, "loss": 0.4397, "step": 24866 }, { "epoch": 1.6689037280628167, "grad_norm": 0.8594359159469604, "learning_rate": 7.012717227260734e-06, "loss": 0.5207, "step": 24868 }, { "epoch": 1.6690379517465856, "grad_norm": 1.2969167232513428, "learning_rate": 7.007167554893274e-06, "loss": 0.4622, "step": 24870 }, { "epoch": 1.6691721754303548, "grad_norm": 1.0026137828826904, "learning_rate": 7.001619913852603e-06, "loss": 0.4298, "step": 24872 }, { "epoch": 1.6693063991141237, "grad_norm": 1.0497329235076904, "learning_rate": 6.99607430440084e-06, "loss": 0.4532, "step": 24874 }, { "epoch": 1.6694406227978926, "grad_norm": 1.0968456268310547, "learning_rate": 6.990530726799993e-06, "loss": 0.4998, "step": 24876 }, { "epoch": 1.6695748464816615, "grad_norm": 1.0032005310058594, "learning_rate": 6.984989181312007e-06, "loss": 0.5401, "step": 24878 }, { "epoch": 1.6697090701654307, "grad_norm": 1.1241943836212158, "learning_rate": 6.979449668198679e-06, "loss": 0.486, "step": 24880 }, { "epoch": 1.6698432938491998, "grad_norm": 1.1395469903945923, "learning_rate": 6.973912187721787e-06, "loss": 0.4375, "step": 24882 }, { "epoch": 1.6699775175329687, "grad_norm": 1.0104833841323853, "learning_rate": 6.968376740142912e-06, "loss": 0.4592, "step": 24884 }, { "epoch": 1.6701117412167377, "grad_norm": 1.0128306150436401, "learning_rate": 6.962843325723628e-06, "loss": 0.492, "step": 24886 }, { "epoch": 1.6702459649005066, "grad_norm": 1.1043764352798462, "learning_rate": 6.957311944725359e-06, "loss": 0.4791, "step": 24888 }, { "epoch": 1.6703801885842757, "grad_norm": 0.9430697560310364, "learning_rate": 6.951782597409473e-06, "loss": 0.3964, "step": 24890 }, { "epoch": 1.6705144122680446, "grad_norm": 1.0144459009170532, "learning_rate": 6.946255284037207e-06, "loss": 0.407, "step": 24892 }, { "epoch": 1.6706486359518138, "grad_norm": 1.1235939264297485, "learning_rate": 6.940730004869717e-06, "loss": 0.4974, "step": 24894 }, { "epoch": 1.6707828596355827, "grad_norm": 0.9710730910301208, "learning_rate": 6.9352067601680506e-06, "loss": 0.4397, "step": 24896 }, { "epoch": 1.6709170833193516, "grad_norm": 0.9254615902900696, "learning_rate": 6.929685550193193e-06, "loss": 0.4578, "step": 24898 }, { "epoch": 1.6710513070031205, "grad_norm": 1.1053893566131592, "learning_rate": 6.924166375205998e-06, "loss": 0.4728, "step": 24900 }, { "epoch": 1.6711855306868897, "grad_norm": 0.9544780850410461, "learning_rate": 6.918649235467234e-06, "loss": 0.5049, "step": 24902 }, { "epoch": 1.6713197543706588, "grad_norm": 0.983664333820343, "learning_rate": 6.913134131237575e-06, "loss": 0.4404, "step": 24904 }, { "epoch": 1.6714539780544277, "grad_norm": 1.0222305059432983, "learning_rate": 6.9076210627776085e-06, "loss": 0.4881, "step": 24906 }, { "epoch": 1.6715882017381967, "grad_norm": 1.106665015220642, "learning_rate": 6.902110030347797e-06, "loss": 0.5302, "step": 24908 }, { "epoch": 1.6717224254219656, "grad_norm": 0.9527423977851868, "learning_rate": 6.896601034208561e-06, "loss": 0.4498, "step": 24910 }, { "epoch": 1.6718566491057347, "grad_norm": 1.1243098974227905, "learning_rate": 6.891094074620141e-06, "loss": 0.4221, "step": 24912 }, { "epoch": 1.6719908727895039, "grad_norm": 1.0397752523422241, "learning_rate": 6.88558915184277e-06, "loss": 0.4972, "step": 24914 }, { "epoch": 1.6721250964732728, "grad_norm": 0.9265273809432983, "learning_rate": 6.880086266136521e-06, "loss": 0.4661, "step": 24916 }, { "epoch": 1.6722593201570417, "grad_norm": 1.0777788162231445, "learning_rate": 6.874585417761414e-06, "loss": 0.452, "step": 24918 }, { "epoch": 1.6723935438408106, "grad_norm": 1.089951992034912, "learning_rate": 6.869086606977349e-06, "loss": 0.4876, "step": 24920 }, { "epoch": 1.6725277675245798, "grad_norm": 0.9507853984832764, "learning_rate": 6.863589834044121e-06, "loss": 0.4291, "step": 24922 }, { "epoch": 1.6726619912083487, "grad_norm": 1.1321830749511719, "learning_rate": 6.8580950992214395e-06, "loss": 0.4382, "step": 24924 }, { "epoch": 1.6727962148921178, "grad_norm": 0.9954093098640442, "learning_rate": 6.852602402768943e-06, "loss": 0.4493, "step": 24926 }, { "epoch": 1.6729304385758867, "grad_norm": 1.0333951711654663, "learning_rate": 6.847111744946139e-06, "loss": 0.5084, "step": 24928 }, { "epoch": 1.6730646622596557, "grad_norm": 0.9585345387458801, "learning_rate": 6.841623126012442e-06, "loss": 0.4525, "step": 24930 }, { "epoch": 1.6731988859434246, "grad_norm": 1.0446515083312988, "learning_rate": 6.83613654622719e-06, "loss": 0.5496, "step": 24932 }, { "epoch": 1.6733331096271937, "grad_norm": 1.0466399192810059, "learning_rate": 6.830652005849597e-06, "loss": 0.5162, "step": 24934 }, { "epoch": 1.6734673333109629, "grad_norm": 0.9950196743011475, "learning_rate": 6.825169505138818e-06, "loss": 0.4763, "step": 24936 }, { "epoch": 1.6736015569947318, "grad_norm": 0.9856505393981934, "learning_rate": 6.819689044353877e-06, "loss": 0.5168, "step": 24938 }, { "epoch": 1.6737357806785007, "grad_norm": 1.1162148714065552, "learning_rate": 6.8142106237537255e-06, "loss": 0.4407, "step": 24940 }, { "epoch": 1.6738700043622696, "grad_norm": 1.1244986057281494, "learning_rate": 6.8087342435971815e-06, "loss": 0.4785, "step": 24942 }, { "epoch": 1.6740042280460388, "grad_norm": 0.8998849391937256, "learning_rate": 6.803259904143027e-06, "loss": 0.4196, "step": 24944 }, { "epoch": 1.6741384517298077, "grad_norm": 1.0995714664459229, "learning_rate": 6.797787605649891e-06, "loss": 0.5114, "step": 24946 }, { "epoch": 1.6742726754135768, "grad_norm": 0.9137680530548096, "learning_rate": 6.792317348376348e-06, "loss": 0.422, "step": 24948 }, { "epoch": 1.6744068990973457, "grad_norm": 0.9242937564849854, "learning_rate": 6.786849132580842e-06, "loss": 0.521, "step": 24950 }, { "epoch": 1.6745411227811147, "grad_norm": 0.8841946125030518, "learning_rate": 6.781382958521743e-06, "loss": 0.4557, "step": 24952 }, { "epoch": 1.6746753464648836, "grad_norm": 0.9266219139099121, "learning_rate": 6.7759188264573005e-06, "loss": 0.4404, "step": 24954 }, { "epoch": 1.6748095701486527, "grad_norm": 0.9643049836158752, "learning_rate": 6.770456736645708e-06, "loss": 0.456, "step": 24956 }, { "epoch": 1.6749437938324219, "grad_norm": 1.0734564065933228, "learning_rate": 6.764996689345027e-06, "loss": 0.4664, "step": 24958 }, { "epoch": 1.6750780175161908, "grad_norm": 1.0203007459640503, "learning_rate": 6.7595386848132325e-06, "loss": 0.4589, "step": 24960 }, { "epoch": 1.6752122411999597, "grad_norm": 1.0745856761932373, "learning_rate": 6.754082723308203e-06, "loss": 0.536, "step": 24962 }, { "epoch": 1.6753464648837286, "grad_norm": 1.0680867433547974, "learning_rate": 6.748628805087731e-06, "loss": 0.5385, "step": 24964 }, { "epoch": 1.6754806885674978, "grad_norm": 1.0987399816513062, "learning_rate": 6.743176930409495e-06, "loss": 0.5095, "step": 24966 }, { "epoch": 1.6756149122512667, "grad_norm": 1.0225719213485718, "learning_rate": 6.737727099531094e-06, "loss": 0.4711, "step": 24968 }, { "epoch": 1.6757491359350358, "grad_norm": 1.1248780488967896, "learning_rate": 6.732279312709999e-06, "loss": 0.4256, "step": 24970 }, { "epoch": 1.6758833596188047, "grad_norm": 1.1062440872192383, "learning_rate": 6.726833570203639e-06, "loss": 0.4524, "step": 24972 }, { "epoch": 1.6760175833025737, "grad_norm": 1.0149747133255005, "learning_rate": 6.721389872269291e-06, "loss": 0.4915, "step": 24974 }, { "epoch": 1.6761518069863426, "grad_norm": 1.0408467054367065, "learning_rate": 6.715948219164175e-06, "loss": 0.4601, "step": 24976 }, { "epoch": 1.6762860306701117, "grad_norm": 0.9350197911262512, "learning_rate": 6.7105086111453944e-06, "loss": 0.437, "step": 24978 }, { "epoch": 1.6764202543538809, "grad_norm": 0.9445664286613464, "learning_rate": 6.705071048469952e-06, "loss": 0.4797, "step": 24980 }, { "epoch": 1.6765544780376498, "grad_norm": 1.073475956916809, "learning_rate": 6.69963553139476e-06, "loss": 0.4725, "step": 24982 }, { "epoch": 1.6766887017214187, "grad_norm": 1.0421797037124634, "learning_rate": 6.694202060176652e-06, "loss": 0.4289, "step": 24984 }, { "epoch": 1.6768229254051876, "grad_norm": 2.2069389820098877, "learning_rate": 6.688770635072339e-06, "loss": 0.4576, "step": 24986 }, { "epoch": 1.6769571490889568, "grad_norm": 0.9449515342712402, "learning_rate": 6.683341256338444e-06, "loss": 0.4131, "step": 24988 }, { "epoch": 1.677091372772726, "grad_norm": 0.94171142578125, "learning_rate": 6.677913924231488e-06, "loss": 0.4726, "step": 24990 }, { "epoch": 1.6772255964564948, "grad_norm": 1.0046991109848022, "learning_rate": 6.672488639007918e-06, "loss": 0.4717, "step": 24992 }, { "epoch": 1.6773598201402637, "grad_norm": 0.9732492566108704, "learning_rate": 6.667065400924066e-06, "loss": 0.4163, "step": 24994 }, { "epoch": 1.6774940438240327, "grad_norm": 0.9594755172729492, "learning_rate": 6.661644210236156e-06, "loss": 0.4575, "step": 24996 }, { "epoch": 1.6776282675078018, "grad_norm": 1.0889017581939697, "learning_rate": 6.656225067200339e-06, "loss": 0.4944, "step": 24998 }, { "epoch": 1.6777624911915707, "grad_norm": 1.1145284175872803, "learning_rate": 6.650807972072648e-06, "loss": 0.4343, "step": 25000 }, { "epoch": 1.6778967148753399, "grad_norm": 1.0912227630615234, "learning_rate": 6.645392925109045e-06, "loss": 0.4931, "step": 25002 }, { "epoch": 1.6780309385591088, "grad_norm": 1.1241024732589722, "learning_rate": 6.639979926565359e-06, "loss": 0.4578, "step": 25004 }, { "epoch": 1.6781651622428777, "grad_norm": 1.0750898122787476, "learning_rate": 6.634568976697386e-06, "loss": 0.4653, "step": 25006 }, { "epoch": 1.6782993859266466, "grad_norm": 1.055790901184082, "learning_rate": 6.629160075760726e-06, "loss": 0.5403, "step": 25008 }, { "epoch": 1.6784336096104158, "grad_norm": 1.0737273693084717, "learning_rate": 6.623753224010987e-06, "loss": 0.5264, "step": 25010 }, { "epoch": 1.678567833294185, "grad_norm": 1.0886459350585938, "learning_rate": 6.618348421703596e-06, "loss": 0.4444, "step": 25012 }, { "epoch": 1.6787020569779538, "grad_norm": 0.9762519001960754, "learning_rate": 6.6129456690939455e-06, "loss": 0.4607, "step": 25014 }, { "epoch": 1.6788362806617227, "grad_norm": 1.086861491203308, "learning_rate": 6.607544966437296e-06, "loss": 0.4689, "step": 25016 }, { "epoch": 1.6789705043454917, "grad_norm": 1.0281087160110474, "learning_rate": 6.602146313988822e-06, "loss": 0.4209, "step": 25018 }, { "epoch": 1.6791047280292608, "grad_norm": 0.918687641620636, "learning_rate": 6.59674971200358e-06, "loss": 0.4127, "step": 25020 }, { "epoch": 1.6792389517130297, "grad_norm": 1.1233757734298706, "learning_rate": 6.591355160736578e-06, "loss": 0.4873, "step": 25022 }, { "epoch": 1.6793731753967989, "grad_norm": 0.9721964001655579, "learning_rate": 6.5859626604426905e-06, "loss": 0.4487, "step": 25024 }, { "epoch": 1.6795073990805678, "grad_norm": 1.129554033279419, "learning_rate": 6.580572211376695e-06, "loss": 0.5156, "step": 25026 }, { "epoch": 1.6796416227643367, "grad_norm": 0.9984642863273621, "learning_rate": 6.575183813793267e-06, "loss": 0.4914, "step": 25028 }, { "epoch": 1.6797758464481056, "grad_norm": 0.9877324104309082, "learning_rate": 6.569797467947026e-06, "loss": 0.4615, "step": 25030 }, { "epoch": 1.6799100701318748, "grad_norm": 1.579860806465149, "learning_rate": 6.564413174092443e-06, "loss": 0.5255, "step": 25032 }, { "epoch": 1.680044293815644, "grad_norm": 1.0377464294433594, "learning_rate": 6.559030932483945e-06, "loss": 0.5348, "step": 25034 }, { "epoch": 1.6801785174994128, "grad_norm": 1.0144003629684448, "learning_rate": 6.553650743375794e-06, "loss": 0.4118, "step": 25036 }, { "epoch": 1.6803127411831817, "grad_norm": 1.0023353099822998, "learning_rate": 6.548272607022221e-06, "loss": 0.4661, "step": 25038 }, { "epoch": 1.6804469648669507, "grad_norm": 0.9578472971916199, "learning_rate": 6.542896523677311e-06, "loss": 0.5483, "step": 25040 }, { "epoch": 1.6805811885507198, "grad_norm": 1.0725754499435425, "learning_rate": 6.537522493595094e-06, "loss": 0.511, "step": 25042 }, { "epoch": 1.6807154122344887, "grad_norm": 1.1264251470565796, "learning_rate": 6.532150517029478e-06, "loss": 0.5275, "step": 25044 }, { "epoch": 1.6808496359182579, "grad_norm": 1.1566630601882935, "learning_rate": 6.526780594234272e-06, "loss": 0.4925, "step": 25046 }, { "epoch": 1.6809838596020268, "grad_norm": 1.072445034980774, "learning_rate": 6.521412725463188e-06, "loss": 0.4378, "step": 25048 }, { "epoch": 1.6811180832857957, "grad_norm": 1.4630780220031738, "learning_rate": 6.516046910969864e-06, "loss": 0.4661, "step": 25050 }, { "epoch": 1.6812523069695646, "grad_norm": 0.9673110842704773, "learning_rate": 6.510683151007818e-06, "loss": 0.4452, "step": 25052 }, { "epoch": 1.6813865306533338, "grad_norm": 1.0803834199905396, "learning_rate": 6.505321445830476e-06, "loss": 0.4096, "step": 25054 }, { "epoch": 1.681520754337103, "grad_norm": 1.049075961112976, "learning_rate": 6.499961795691151e-06, "loss": 0.4339, "step": 25056 }, { "epoch": 1.6816549780208718, "grad_norm": 1.1449967622756958, "learning_rate": 6.494604200843107e-06, "loss": 0.4876, "step": 25058 }, { "epoch": 1.6817892017046407, "grad_norm": 0.913049042224884, "learning_rate": 6.489248661539454e-06, "loss": 0.5163, "step": 25060 }, { "epoch": 1.6819234253884097, "grad_norm": 1.0686430931091309, "learning_rate": 6.483895178033261e-06, "loss": 0.5323, "step": 25062 }, { "epoch": 1.6820576490721788, "grad_norm": 1.064252257347107, "learning_rate": 6.478543750577442e-06, "loss": 0.478, "step": 25064 }, { "epoch": 1.682191872755948, "grad_norm": 0.95737624168396, "learning_rate": 6.4731943794248355e-06, "loss": 0.4617, "step": 25066 }, { "epoch": 1.6823260964397169, "grad_norm": 1.0545599460601807, "learning_rate": 6.467847064828214e-06, "loss": 0.4684, "step": 25068 }, { "epoch": 1.6824603201234858, "grad_norm": 1.0754735469818115, "learning_rate": 6.462501807040211e-06, "loss": 0.484, "step": 25070 }, { "epoch": 1.6825945438072547, "grad_norm": 1.1124367713928223, "learning_rate": 6.457158606313402e-06, "loss": 0.439, "step": 25072 }, { "epoch": 1.6827287674910236, "grad_norm": 1.129748821258545, "learning_rate": 6.451817462900206e-06, "loss": 0.4889, "step": 25074 }, { "epoch": 1.6828629911747928, "grad_norm": 1.002609133720398, "learning_rate": 6.4464783770530055e-06, "loss": 0.4862, "step": 25076 }, { "epoch": 1.682997214858562, "grad_norm": 1.1611278057098389, "learning_rate": 6.441141349024055e-06, "loss": 0.5409, "step": 25078 }, { "epoch": 1.6831314385423308, "grad_norm": 1.128123164176941, "learning_rate": 6.435806379065529e-06, "loss": 0.5094, "step": 25080 }, { "epoch": 1.6832656622260997, "grad_norm": 0.9699921607971191, "learning_rate": 6.430473467429482e-06, "loss": 0.4155, "step": 25082 }, { "epoch": 1.6833998859098687, "grad_norm": 1.0676101446151733, "learning_rate": 6.4251426143678904e-06, "loss": 0.4512, "step": 25084 }, { "epoch": 1.6835341095936378, "grad_norm": 0.9882549047470093, "learning_rate": 6.4198138201326145e-06, "loss": 0.4518, "step": 25086 }, { "epoch": 1.683668333277407, "grad_norm": 0.9625896215438843, "learning_rate": 6.414487084975451e-06, "loss": 0.42, "step": 25088 }, { "epoch": 1.6838025569611759, "grad_norm": 0.9984647631645203, "learning_rate": 6.409162409148062e-06, "loss": 0.5404, "step": 25090 }, { "epoch": 1.6839367806449448, "grad_norm": 0.9463441371917725, "learning_rate": 6.403839792902033e-06, "loss": 0.4435, "step": 25092 }, { "epoch": 1.6840710043287137, "grad_norm": 1.0103236436843872, "learning_rate": 6.398519236488837e-06, "loss": 0.4823, "step": 25094 }, { "epoch": 1.6842052280124828, "grad_norm": 0.8966814875602722, "learning_rate": 6.393200740159877e-06, "loss": 0.3919, "step": 25096 }, { "epoch": 1.6843394516962518, "grad_norm": 1.2721905708312988, "learning_rate": 6.387884304166419e-06, "loss": 0.5342, "step": 25098 }, { "epoch": 1.684473675380021, "grad_norm": 1.1116913557052612, "learning_rate": 6.382569928759685e-06, "loss": 0.4427, "step": 25100 }, { "epoch": 1.6846078990637898, "grad_norm": 1.022383213043213, "learning_rate": 6.3772576141907456e-06, "loss": 0.5117, "step": 25102 }, { "epoch": 1.6847421227475587, "grad_norm": 1.0255234241485596, "learning_rate": 6.371947360710606e-06, "loss": 0.4271, "step": 25104 }, { "epoch": 1.6848763464313277, "grad_norm": 1.0771558284759521, "learning_rate": 6.366639168570154e-06, "loss": 0.4597, "step": 25106 }, { "epoch": 1.6850105701150968, "grad_norm": 0.9802014827728271, "learning_rate": 6.361333038020212e-06, "loss": 0.4234, "step": 25108 }, { "epoch": 1.685144793798866, "grad_norm": 1.0145399570465088, "learning_rate": 6.356028969311467e-06, "loss": 0.5103, "step": 25110 }, { "epoch": 1.6852790174826349, "grad_norm": 1.097533941268921, "learning_rate": 6.350726962694537e-06, "loss": 0.4937, "step": 25112 }, { "epoch": 1.6854132411664038, "grad_norm": 1.1500582695007324, "learning_rate": 6.345427018419908e-06, "loss": 0.4659, "step": 25114 }, { "epoch": 1.6855474648501727, "grad_norm": 0.9849839806556702, "learning_rate": 6.340129136738027e-06, "loss": 0.4085, "step": 25116 }, { "epoch": 1.6856816885339418, "grad_norm": 0.755287230014801, "learning_rate": 6.334833317899186e-06, "loss": 0.4087, "step": 25118 }, { "epoch": 1.6858159122177108, "grad_norm": 1.116926670074463, "learning_rate": 6.329539562153608e-06, "loss": 0.4955, "step": 25120 }, { "epoch": 1.68595013590148, "grad_norm": 0.950871467590332, "learning_rate": 6.324247869751398e-06, "loss": 0.4735, "step": 25122 }, { "epoch": 1.6860843595852488, "grad_norm": 0.9359391927719116, "learning_rate": 6.318958240942607e-06, "loss": 0.4395, "step": 25124 }, { "epoch": 1.6862185832690177, "grad_norm": 1.0384999513626099, "learning_rate": 6.313670675977129e-06, "loss": 0.4541, "step": 25126 }, { "epoch": 1.6863528069527867, "grad_norm": 1.1261101961135864, "learning_rate": 6.308385175104819e-06, "loss": 0.4507, "step": 25128 }, { "epoch": 1.6864870306365558, "grad_norm": 1.0034666061401367, "learning_rate": 6.303101738575395e-06, "loss": 0.4504, "step": 25130 }, { "epoch": 1.686621254320325, "grad_norm": 1.0501008033752441, "learning_rate": 6.297820366638485e-06, "loss": 0.4718, "step": 25132 }, { "epoch": 1.6867554780040939, "grad_norm": 1.0580476522445679, "learning_rate": 6.292541059543628e-06, "loss": 0.4951, "step": 25134 }, { "epoch": 1.6868897016878628, "grad_norm": 0.9894029498100281, "learning_rate": 6.2872638175402455e-06, "loss": 0.4317, "step": 25136 }, { "epoch": 1.6870239253716317, "grad_norm": 1.027328610420227, "learning_rate": 6.281988640877706e-06, "loss": 0.499, "step": 25138 }, { "epoch": 1.6871581490554008, "grad_norm": 1.1530015468597412, "learning_rate": 6.276715529805233e-06, "loss": 0.4692, "step": 25140 }, { "epoch": 1.68729237273917, "grad_norm": 0.9684929251670837, "learning_rate": 6.27144448457197e-06, "loss": 0.4657, "step": 25142 }, { "epoch": 1.687426596422939, "grad_norm": 0.9814417958259583, "learning_rate": 6.266175505426958e-06, "loss": 0.5096, "step": 25144 }, { "epoch": 1.6875608201067078, "grad_norm": 1.3349909782409668, "learning_rate": 6.260908592619169e-06, "loss": 0.4892, "step": 25146 }, { "epoch": 1.6876950437904767, "grad_norm": 0.9152083992958069, "learning_rate": 6.255643746397433e-06, "loss": 0.455, "step": 25148 }, { "epoch": 1.6878292674742457, "grad_norm": 0.9364927411079407, "learning_rate": 6.250380967010516e-06, "loss": 0.4415, "step": 25150 }, { "epoch": 1.6879634911580148, "grad_norm": 1.0783220529556274, "learning_rate": 6.2451202547070545e-06, "loss": 0.4086, "step": 25152 }, { "epoch": 1.688097714841784, "grad_norm": 1.1346849203109741, "learning_rate": 6.239861609735631e-06, "loss": 0.4136, "step": 25154 }, { "epoch": 1.6882319385255529, "grad_norm": 1.0446324348449707, "learning_rate": 6.2346050323446865e-06, "loss": 0.5312, "step": 25156 }, { "epoch": 1.6883661622093218, "grad_norm": 7.313377857208252, "learning_rate": 6.229350522782612e-06, "loss": 0.4643, "step": 25158 }, { "epoch": 1.6885003858930907, "grad_norm": 1.1398403644561768, "learning_rate": 6.224098081297636e-06, "loss": 0.4778, "step": 25160 }, { "epoch": 1.6886346095768598, "grad_norm": 1.1007752418518066, "learning_rate": 6.218847708137954e-06, "loss": 0.4406, "step": 25162 }, { "epoch": 1.688768833260629, "grad_norm": 0.9332924485206604, "learning_rate": 6.213599403551617e-06, "loss": 0.434, "step": 25164 }, { "epoch": 1.688903056944398, "grad_norm": 1.010879397392273, "learning_rate": 6.208353167786612e-06, "loss": 0.4313, "step": 25166 }, { "epoch": 1.6890372806281668, "grad_norm": 0.9699853658676147, "learning_rate": 6.203109001090812e-06, "loss": 0.5114, "step": 25168 }, { "epoch": 1.6891715043119357, "grad_norm": 0.9748437404632568, "learning_rate": 6.197866903711985e-06, "loss": 0.4725, "step": 25170 }, { "epoch": 1.6893057279957049, "grad_norm": 0.8466166853904724, "learning_rate": 6.192626875897806e-06, "loss": 0.4495, "step": 25172 }, { "epoch": 1.6894399516794738, "grad_norm": 1.0146995782852173, "learning_rate": 6.18738891789587e-06, "loss": 0.4434, "step": 25174 }, { "epoch": 1.689574175363243, "grad_norm": 0.8807005286216736, "learning_rate": 6.182153029953658e-06, "loss": 0.4049, "step": 25176 }, { "epoch": 1.6897083990470119, "grad_norm": 0.9264736771583557, "learning_rate": 6.1769192123185495e-06, "loss": 0.3903, "step": 25178 }, { "epoch": 1.6898426227307808, "grad_norm": 1.0460312366485596, "learning_rate": 6.171687465237824e-06, "loss": 0.4415, "step": 25180 }, { "epoch": 1.6899768464145497, "grad_norm": 0.9783272743225098, "learning_rate": 6.1664577889586905e-06, "loss": 0.4874, "step": 25182 }, { "epoch": 1.6901110700983188, "grad_norm": 1.0166213512420654, "learning_rate": 6.1612301837282225e-06, "loss": 0.4665, "step": 25184 }, { "epoch": 1.690245293782088, "grad_norm": 1.0550141334533691, "learning_rate": 6.156004649793446e-06, "loss": 0.4403, "step": 25186 }, { "epoch": 1.690379517465857, "grad_norm": 1.0054047107696533, "learning_rate": 6.150781187401211e-06, "loss": 0.4583, "step": 25188 }, { "epoch": 1.6905137411496258, "grad_norm": 0.877746045589447, "learning_rate": 6.145559796798345e-06, "loss": 0.4329, "step": 25190 }, { "epoch": 1.6906479648333947, "grad_norm": 1.2742961645126343, "learning_rate": 6.140340478231537e-06, "loss": 0.5484, "step": 25192 }, { "epoch": 1.6907821885171639, "grad_norm": 1.074097752571106, "learning_rate": 6.135123231947404e-06, "loss": 0.4559, "step": 25194 }, { "epoch": 1.6909164122009328, "grad_norm": 1.1236523389816284, "learning_rate": 6.1299080581924465e-06, "loss": 0.5046, "step": 25196 }, { "epoch": 1.691050635884702, "grad_norm": 1.113170862197876, "learning_rate": 6.124694957213062e-06, "loss": 0.4822, "step": 25198 }, { "epoch": 1.6911848595684709, "grad_norm": 1.0551742315292358, "learning_rate": 6.119483929255559e-06, "loss": 0.4929, "step": 25200 }, { "epoch": 1.6913190832522398, "grad_norm": 0.9076232314109802, "learning_rate": 6.1142749745661536e-06, "loss": 0.4541, "step": 25202 }, { "epoch": 1.6914533069360087, "grad_norm": 1.0314046144485474, "learning_rate": 6.109068093390957e-06, "loss": 0.4655, "step": 25204 }, { "epoch": 1.6915875306197778, "grad_norm": 1.0429438352584839, "learning_rate": 6.103863285975992e-06, "loss": 0.4468, "step": 25206 }, { "epoch": 1.691721754303547, "grad_norm": 1.0242527723312378, "learning_rate": 6.0986605525671705e-06, "loss": 0.3975, "step": 25208 }, { "epoch": 1.691855977987316, "grad_norm": 0.9596940875053406, "learning_rate": 6.093459893410297e-06, "loss": 0.523, "step": 25210 }, { "epoch": 1.6919902016710848, "grad_norm": 1.0623986721038818, "learning_rate": 6.0882613087511146e-06, "loss": 0.3849, "step": 25212 }, { "epoch": 1.6921244253548537, "grad_norm": 0.9892853498458862, "learning_rate": 6.083064798835236e-06, "loss": 0.4769, "step": 25214 }, { "epoch": 1.6922586490386229, "grad_norm": 1.0125302076339722, "learning_rate": 6.077870363908184e-06, "loss": 0.4852, "step": 25216 }, { "epoch": 1.692392872722392, "grad_norm": 1.0274652242660522, "learning_rate": 6.0726780042153816e-06, "loss": 0.5281, "step": 25218 }, { "epoch": 1.692527096406161, "grad_norm": 1.1266363859176636, "learning_rate": 6.067487720002174e-06, "loss": 0.4616, "step": 25220 }, { "epoch": 1.6926613200899299, "grad_norm": 1.0416055917739868, "learning_rate": 6.0622995115137695e-06, "loss": 0.4729, "step": 25222 }, { "epoch": 1.6927955437736988, "grad_norm": 1.0067485570907593, "learning_rate": 6.057113378995338e-06, "loss": 0.4349, "step": 25224 }, { "epoch": 1.6929297674574677, "grad_norm": 0.973070502281189, "learning_rate": 6.051929322691868e-06, "loss": 0.4329, "step": 25226 }, { "epoch": 1.6930639911412368, "grad_norm": 0.9753096699714661, "learning_rate": 6.046747342848325e-06, "loss": 0.4652, "step": 25228 }, { "epoch": 1.693198214825006, "grad_norm": 0.9495730400085449, "learning_rate": 6.041567439709533e-06, "loss": 0.4571, "step": 25230 }, { "epoch": 1.693332438508775, "grad_norm": 0.9078459739685059, "learning_rate": 6.036389613520243e-06, "loss": 0.4677, "step": 25232 }, { "epoch": 1.6934666621925438, "grad_norm": 1.0763219594955444, "learning_rate": 6.031213864525098e-06, "loss": 0.4892, "step": 25234 }, { "epoch": 1.6936008858763127, "grad_norm": 1.1541110277175903, "learning_rate": 6.02604019296864e-06, "loss": 0.4815, "step": 25236 }, { "epoch": 1.6937351095600819, "grad_norm": 0.9476569294929504, "learning_rate": 6.0208685990952975e-06, "loss": 0.3649, "step": 25238 }, { "epoch": 1.693869333243851, "grad_norm": 0.9725757241249084, "learning_rate": 6.0156990831494395e-06, "loss": 0.4933, "step": 25240 }, { "epoch": 1.69400355692762, "grad_norm": 0.9956469535827637, "learning_rate": 6.010531645375312e-06, "loss": 0.4538, "step": 25242 }, { "epoch": 1.6941377806113889, "grad_norm": 0.9989961981773376, "learning_rate": 6.005366286017061e-06, "loss": 0.4813, "step": 25244 }, { "epoch": 1.6942720042951578, "grad_norm": 1.0470447540283203, "learning_rate": 6.000203005318733e-06, "loss": 0.4527, "step": 25246 }, { "epoch": 1.694406227978927, "grad_norm": 0.8920106291770935, "learning_rate": 5.9950418035243036e-06, "loss": 0.4608, "step": 25248 }, { "epoch": 1.6945404516626958, "grad_norm": 1.0534292459487915, "learning_rate": 5.9898826808776006e-06, "loss": 0.4716, "step": 25250 }, { "epoch": 1.694674675346465, "grad_norm": 1.0337680578231812, "learning_rate": 5.984725637622424e-06, "loss": 0.467, "step": 25252 }, { "epoch": 1.694808899030234, "grad_norm": 0.9041849374771118, "learning_rate": 5.979570674002388e-06, "loss": 0.4682, "step": 25254 }, { "epoch": 1.6949431227140028, "grad_norm": 1.0136576890945435, "learning_rate": 5.974417790261083e-06, "loss": 0.4264, "step": 25256 }, { "epoch": 1.6950773463977717, "grad_norm": 1.0099191665649414, "learning_rate": 5.969266986641953e-06, "loss": 0.4767, "step": 25258 }, { "epoch": 1.6952115700815409, "grad_norm": 1.1609923839569092, "learning_rate": 5.964118263388391e-06, "loss": 0.4584, "step": 25260 }, { "epoch": 1.69534579376531, "grad_norm": 0.9966214895248413, "learning_rate": 5.9589716207436475e-06, "loss": 0.4714, "step": 25262 }, { "epoch": 1.695480017449079, "grad_norm": 1.0159577131271362, "learning_rate": 5.9538270589508895e-06, "loss": 0.5211, "step": 25264 }, { "epoch": 1.6956142411328479, "grad_norm": 1.0560600757598877, "learning_rate": 5.948684578253177e-06, "loss": 0.4752, "step": 25266 }, { "epoch": 1.6957484648166168, "grad_norm": 1.02834153175354, "learning_rate": 5.943544178893506e-06, "loss": 0.4604, "step": 25268 }, { "epoch": 1.695882688500386, "grad_norm": 1.0763250589370728, "learning_rate": 5.938405861114743e-06, "loss": 0.5098, "step": 25270 }, { "epoch": 1.6960169121841548, "grad_norm": 1.0647401809692383, "learning_rate": 5.93326962515966e-06, "loss": 0.4427, "step": 25272 }, { "epoch": 1.696151135867924, "grad_norm": 0.8827753067016602, "learning_rate": 5.928135471270929e-06, "loss": 0.4371, "step": 25274 }, { "epoch": 1.696285359551693, "grad_norm": 1.181510329246521, "learning_rate": 5.92300339969113e-06, "loss": 0.4472, "step": 25276 }, { "epoch": 1.6964195832354618, "grad_norm": 1.0063621997833252, "learning_rate": 5.9178734106627495e-06, "loss": 0.4782, "step": 25278 }, { "epoch": 1.6965538069192307, "grad_norm": 0.9924083352088928, "learning_rate": 5.9127455044281684e-06, "loss": 0.4601, "step": 25280 }, { "epoch": 1.6966880306029999, "grad_norm": 0.8651800751686096, "learning_rate": 5.90761968122967e-06, "loss": 0.4234, "step": 25282 }, { "epoch": 1.696822254286769, "grad_norm": 0.956924319267273, "learning_rate": 5.902495941309427e-06, "loss": 0.4557, "step": 25284 }, { "epoch": 1.696956477970538, "grad_norm": 1.0364547967910767, "learning_rate": 5.897374284909546e-06, "loss": 0.5135, "step": 25286 }, { "epoch": 1.6970907016543069, "grad_norm": 1.1208486557006836, "learning_rate": 5.892254712272e-06, "loss": 0.4717, "step": 25288 }, { "epoch": 1.6972249253380758, "grad_norm": 0.8544895648956299, "learning_rate": 5.887137223638689e-06, "loss": 0.3906, "step": 25290 }, { "epoch": 1.697359149021845, "grad_norm": 1.0679324865341187, "learning_rate": 5.882021819251405e-06, "loss": 0.4137, "step": 25292 }, { "epoch": 1.697493372705614, "grad_norm": 0.9690061211585999, "learning_rate": 5.8769084993518355e-06, "loss": 0.4583, "step": 25294 }, { "epoch": 1.697627596389383, "grad_norm": 1.0798450708389282, "learning_rate": 5.871797264181561e-06, "loss": 0.4734, "step": 25296 }, { "epoch": 1.697761820073152, "grad_norm": 1.0062329769134521, "learning_rate": 5.866688113982105e-06, "loss": 0.3716, "step": 25298 }, { "epoch": 1.6978960437569208, "grad_norm": 1.0647355318069458, "learning_rate": 5.861581048994847e-06, "loss": 0.4397, "step": 25300 }, { "epoch": 1.6980302674406897, "grad_norm": 1.480678915977478, "learning_rate": 5.856476069461092e-06, "loss": 0.5071, "step": 25302 }, { "epoch": 1.6981644911244589, "grad_norm": 1.0006400346755981, "learning_rate": 5.851373175622032e-06, "loss": 0.475, "step": 25304 }, { "epoch": 1.698298714808228, "grad_norm": 1.111854910850525, "learning_rate": 5.84627236771878e-06, "loss": 0.4143, "step": 25306 }, { "epoch": 1.698432938491997, "grad_norm": 1.1463509798049927, "learning_rate": 5.841173645992326e-06, "loss": 0.4626, "step": 25308 }, { "epoch": 1.6985671621757659, "grad_norm": 0.9406129717826843, "learning_rate": 5.836077010683599e-06, "loss": 0.3964, "step": 25310 }, { "epoch": 1.6987013858595348, "grad_norm": 1.0047578811645508, "learning_rate": 5.830982462033374e-06, "loss": 0.4708, "step": 25312 }, { "epoch": 1.698835609543304, "grad_norm": 0.8719502091407776, "learning_rate": 5.82589000028238e-06, "loss": 0.4133, "step": 25314 }, { "epoch": 1.698969833227073, "grad_norm": 1.0489619970321655, "learning_rate": 5.8207996256712084e-06, "loss": 0.5194, "step": 25316 }, { "epoch": 1.699104056910842, "grad_norm": 1.0129246711730957, "learning_rate": 5.815711338440394e-06, "loss": 0.4481, "step": 25318 }, { "epoch": 1.699238280594611, "grad_norm": 1.257003664970398, "learning_rate": 5.810625138830333e-06, "loss": 0.4173, "step": 25320 }, { "epoch": 1.6993725042783798, "grad_norm": 1.0962368249893188, "learning_rate": 5.805541027081335e-06, "loss": 0.5, "step": 25322 }, { "epoch": 1.699506727962149, "grad_norm": 0.9793606400489807, "learning_rate": 5.800459003433612e-06, "loss": 0.4969, "step": 25324 }, { "epoch": 1.6996409516459179, "grad_norm": 1.0478161573410034, "learning_rate": 5.795379068127299e-06, "loss": 0.5096, "step": 25326 }, { "epoch": 1.699775175329687, "grad_norm": 1.0728329420089722, "learning_rate": 5.790301221402394e-06, "loss": 0.4675, "step": 25328 }, { "epoch": 1.699909399013456, "grad_norm": 0.9906491041183472, "learning_rate": 5.785225463498828e-06, "loss": 0.4783, "step": 25330 }, { "epoch": 1.7000436226972249, "grad_norm": 1.124284267425537, "learning_rate": 5.780151794656397e-06, "loss": 0.4918, "step": 25332 }, { "epoch": 1.7001778463809938, "grad_norm": 1.0835182666778564, "learning_rate": 5.775080215114853e-06, "loss": 0.4822, "step": 25334 }, { "epoch": 1.700312070064763, "grad_norm": 0.9419242739677429, "learning_rate": 5.770010725113794e-06, "loss": 0.4287, "step": 25336 }, { "epoch": 1.700446293748532, "grad_norm": 0.8902066349983215, "learning_rate": 5.76494332489278e-06, "loss": 0.4638, "step": 25338 }, { "epoch": 1.700580517432301, "grad_norm": 1.1273128986358643, "learning_rate": 5.759878014691189e-06, "loss": 0.5564, "step": 25340 }, { "epoch": 1.70071474111607, "grad_norm": 1.0304105281829834, "learning_rate": 5.754814794748364e-06, "loss": 0.4371, "step": 25342 }, { "epoch": 1.7008489647998388, "grad_norm": 1.0378000736236572, "learning_rate": 5.749753665303542e-06, "loss": 0.4676, "step": 25344 }, { "epoch": 1.700983188483608, "grad_norm": 1.0077403783798218, "learning_rate": 5.744694626595837e-06, "loss": 0.4614, "step": 25346 }, { "epoch": 1.7011174121673769, "grad_norm": 1.2381305694580078, "learning_rate": 5.739637678864307e-06, "loss": 0.5158, "step": 25348 }, { "epoch": 1.701251635851146, "grad_norm": 0.9145381450653076, "learning_rate": 5.734582822347839e-06, "loss": 0.3904, "step": 25350 }, { "epoch": 1.701385859534915, "grad_norm": 1.000479817390442, "learning_rate": 5.729530057285304e-06, "loss": 0.5064, "step": 25352 }, { "epoch": 1.7015200832186839, "grad_norm": 1.0135163068771362, "learning_rate": 5.724479383915404e-06, "loss": 0.4627, "step": 25354 }, { "epoch": 1.7016543069024528, "grad_norm": 0.9599625468254089, "learning_rate": 5.719430802476805e-06, "loss": 0.4497, "step": 25356 }, { "epoch": 1.701788530586222, "grad_norm": 1.050414800643921, "learning_rate": 5.714384313208021e-06, "loss": 0.4701, "step": 25358 }, { "epoch": 1.701922754269991, "grad_norm": 1.0206334590911865, "learning_rate": 5.7093399163474945e-06, "loss": 0.5321, "step": 25360 }, { "epoch": 1.70205697795376, "grad_norm": 1.0668940544128418, "learning_rate": 5.704297612133558e-06, "loss": 0.4434, "step": 25362 }, { "epoch": 1.702191201637529, "grad_norm": 0.978858470916748, "learning_rate": 5.699257400804464e-06, "loss": 0.4902, "step": 25364 }, { "epoch": 1.7023254253212978, "grad_norm": 1.117708683013916, "learning_rate": 5.69421928259834e-06, "loss": 0.5651, "step": 25366 }, { "epoch": 1.702459649005067, "grad_norm": 0.9856138229370117, "learning_rate": 5.689183257753234e-06, "loss": 0.445, "step": 25368 }, { "epoch": 1.702593872688836, "grad_norm": 1.003671646118164, "learning_rate": 5.684149326507077e-06, "loss": 0.488, "step": 25370 }, { "epoch": 1.702728096372605, "grad_norm": 1.1078087091445923, "learning_rate": 5.67911748909773e-06, "loss": 0.4242, "step": 25372 }, { "epoch": 1.702862320056374, "grad_norm": 0.9358893632888794, "learning_rate": 5.674087745762918e-06, "loss": 0.4414, "step": 25374 }, { "epoch": 1.7029965437401429, "grad_norm": 1.1128106117248535, "learning_rate": 5.669060096740314e-06, "loss": 0.459, "step": 25376 }, { "epoch": 1.7031307674239118, "grad_norm": 1.0639312267303467, "learning_rate": 5.6640345422674325e-06, "loss": 0.4877, "step": 25378 }, { "epoch": 1.703264991107681, "grad_norm": 0.9991609454154968, "learning_rate": 5.6590110825817445e-06, "loss": 0.4993, "step": 25380 }, { "epoch": 1.70339921479145, "grad_norm": 0.8029139041900635, "learning_rate": 5.653989717920577e-06, "loss": 0.3592, "step": 25382 }, { "epoch": 1.703533438475219, "grad_norm": 1.1822975873947144, "learning_rate": 5.648970448521207e-06, "loss": 0.4926, "step": 25384 }, { "epoch": 1.703667662158988, "grad_norm": 1.2122814655303955, "learning_rate": 5.643953274620767e-06, "loss": 0.5176, "step": 25386 }, { "epoch": 1.7038018858427568, "grad_norm": 1.0605030059814453, "learning_rate": 5.6389381964563115e-06, "loss": 0.4726, "step": 25388 }, { "epoch": 1.703936109526526, "grad_norm": 1.047315239906311, "learning_rate": 5.63392521426479e-06, "loss": 0.5145, "step": 25390 }, { "epoch": 1.704070333210295, "grad_norm": 0.8657220602035522, "learning_rate": 5.628914328283064e-06, "loss": 0.4303, "step": 25392 }, { "epoch": 1.704204556894064, "grad_norm": 0.9523367881774902, "learning_rate": 5.623905538747887e-06, "loss": 0.4382, "step": 25394 }, { "epoch": 1.704338780577833, "grad_norm": 0.9913052916526794, "learning_rate": 5.618898845895909e-06, "loss": 0.4347, "step": 25396 }, { "epoch": 1.7044730042616019, "grad_norm": 1.1487425565719604, "learning_rate": 5.613894249963681e-06, "loss": 0.4715, "step": 25398 }, { "epoch": 1.704607227945371, "grad_norm": 1.1615405082702637, "learning_rate": 5.608891751187678e-06, "loss": 0.4241, "step": 25400 }, { "epoch": 1.70474145162914, "grad_norm": 1.0483362674713135, "learning_rate": 5.603891349804241e-06, "loss": 0.4625, "step": 25402 }, { "epoch": 1.704875675312909, "grad_norm": 0.9386702179908752, "learning_rate": 5.5988930460496525e-06, "loss": 0.4794, "step": 25404 }, { "epoch": 1.705009898996678, "grad_norm": 1.0929003953933716, "learning_rate": 5.5938968401600455e-06, "loss": 0.4878, "step": 25406 }, { "epoch": 1.705144122680447, "grad_norm": 1.1455669403076172, "learning_rate": 5.588902732371487e-06, "loss": 0.4722, "step": 25408 }, { "epoch": 1.7052783463642158, "grad_norm": 1.1001158952713013, "learning_rate": 5.5839107229199495e-06, "loss": 0.4982, "step": 25410 }, { "epoch": 1.705412570047985, "grad_norm": 1.0804669857025146, "learning_rate": 5.5789208120412824e-06, "loss": 0.3945, "step": 25412 }, { "epoch": 1.705546793731754, "grad_norm": 1.0996463298797607, "learning_rate": 5.573932999971266e-06, "loss": 0.4517, "step": 25414 }, { "epoch": 1.705681017415523, "grad_norm": 0.9030675888061523, "learning_rate": 5.568947286945553e-06, "loss": 0.4267, "step": 25416 }, { "epoch": 1.705815241099292, "grad_norm": 0.9615384340286255, "learning_rate": 5.563963673199713e-06, "loss": 0.4534, "step": 25418 }, { "epoch": 1.7059494647830609, "grad_norm": 0.9762392640113831, "learning_rate": 5.558982158969195e-06, "loss": 0.4562, "step": 25420 }, { "epoch": 1.70608368846683, "grad_norm": 0.8485289812088013, "learning_rate": 5.554002744489395e-06, "loss": 0.4173, "step": 25422 }, { "epoch": 1.706217912150599, "grad_norm": 1.0777990818023682, "learning_rate": 5.549025429995569e-06, "loss": 0.4533, "step": 25424 }, { "epoch": 1.706352135834368, "grad_norm": 1.0341979265213013, "learning_rate": 5.5440502157228794e-06, "loss": 0.4627, "step": 25426 }, { "epoch": 1.706486359518137, "grad_norm": 1.2203598022460938, "learning_rate": 5.539077101906387e-06, "loss": 0.4815, "step": 25428 }, { "epoch": 1.706620583201906, "grad_norm": 0.9701406359672546, "learning_rate": 5.534106088781082e-06, "loss": 0.5055, "step": 25430 }, { "epoch": 1.7067548068856748, "grad_norm": 0.9804043769836426, "learning_rate": 5.529137176581828e-06, "loss": 0.4936, "step": 25432 }, { "epoch": 1.706889030569444, "grad_norm": 0.9713636636734009, "learning_rate": 5.5241703655433966e-06, "loss": 0.4967, "step": 25434 }, { "epoch": 1.707023254253213, "grad_norm": 0.9633068442344666, "learning_rate": 5.519205655900439e-06, "loss": 0.4242, "step": 25436 }, { "epoch": 1.707157477936982, "grad_norm": 1.0099884271621704, "learning_rate": 5.514243047887563e-06, "loss": 0.5068, "step": 25438 }, { "epoch": 1.707291701620751, "grad_norm": 1.053117036819458, "learning_rate": 5.509282541739213e-06, "loss": 0.4916, "step": 25440 }, { "epoch": 1.7074259253045199, "grad_norm": 1.0143020153045654, "learning_rate": 5.504324137689793e-06, "loss": 0.4918, "step": 25442 }, { "epoch": 1.707560148988289, "grad_norm": 1.090571641921997, "learning_rate": 5.4993678359735514e-06, "loss": 0.4509, "step": 25444 }, { "epoch": 1.7076943726720581, "grad_norm": 1.0744800567626953, "learning_rate": 5.494413636824675e-06, "loss": 0.4603, "step": 25446 }, { "epoch": 1.707828596355827, "grad_norm": 1.0884594917297363, "learning_rate": 5.489461540477231e-06, "loss": 0.4951, "step": 25448 }, { "epoch": 1.707962820039596, "grad_norm": 1.0418914556503296, "learning_rate": 5.484511547165211e-06, "loss": 0.48, "step": 25450 }, { "epoch": 1.708097043723365, "grad_norm": 1.0547127723693848, "learning_rate": 5.479563657122483e-06, "loss": 0.4791, "step": 25452 }, { "epoch": 1.7082312674071338, "grad_norm": 1.3083374500274658, "learning_rate": 5.474617870582826e-06, "loss": 0.4474, "step": 25454 }, { "epoch": 1.708365491090903, "grad_norm": 1.7928715944290161, "learning_rate": 5.469674187779911e-06, "loss": 0.522, "step": 25456 }, { "epoch": 1.708499714774672, "grad_norm": 1.0545799732208252, "learning_rate": 5.464732608947337e-06, "loss": 0.4532, "step": 25458 }, { "epoch": 1.708633938458441, "grad_norm": 1.0372275114059448, "learning_rate": 5.45979313431857e-06, "loss": 0.4123, "step": 25460 }, { "epoch": 1.70876816214221, "grad_norm": 0.9443169236183167, "learning_rate": 5.454855764126992e-06, "loss": 0.4913, "step": 25462 }, { "epoch": 1.7089023858259789, "grad_norm": 1.266764760017395, "learning_rate": 5.4499204986058815e-06, "loss": 0.4059, "step": 25464 }, { "epoch": 1.709036609509748, "grad_norm": 1.0108635425567627, "learning_rate": 5.444987337988428e-06, "loss": 0.4567, "step": 25466 }, { "epoch": 1.7091708331935171, "grad_norm": 1.1910806894302368, "learning_rate": 5.440056282507699e-06, "loss": 0.4899, "step": 25468 }, { "epoch": 1.709305056877286, "grad_norm": 1.028946876525879, "learning_rate": 5.435127332396695e-06, "loss": 0.465, "step": 25470 }, { "epoch": 1.709439280561055, "grad_norm": 0.9755556583404541, "learning_rate": 5.4302004878882986e-06, "loss": 0.4738, "step": 25472 }, { "epoch": 1.709573504244824, "grad_norm": 0.9384263753890991, "learning_rate": 5.425275749215281e-06, "loss": 0.4892, "step": 25474 }, { "epoch": 1.709707727928593, "grad_norm": 0.9885052442550659, "learning_rate": 5.4203531166103325e-06, "loss": 0.4537, "step": 25476 }, { "epoch": 1.709841951612362, "grad_norm": 0.8746045827865601, "learning_rate": 5.415432590306024e-06, "loss": 0.4216, "step": 25478 }, { "epoch": 1.709976175296131, "grad_norm": 1.0783562660217285, "learning_rate": 5.41051417053487e-06, "loss": 0.4666, "step": 25480 }, { "epoch": 1.7101103989799, "grad_norm": 1.1761975288391113, "learning_rate": 5.405597857529238e-06, "loss": 0.4829, "step": 25482 }, { "epoch": 1.710244622663669, "grad_norm": 1.053392767906189, "learning_rate": 5.400683651521416e-06, "loss": 0.4962, "step": 25484 }, { "epoch": 1.7103788463474379, "grad_norm": 0.8936684727668762, "learning_rate": 5.395771552743578e-06, "loss": 0.4194, "step": 25486 }, { "epoch": 1.710513070031207, "grad_norm": 0.8621320128440857, "learning_rate": 5.390861561427835e-06, "loss": 0.3909, "step": 25488 }, { "epoch": 1.7106472937149761, "grad_norm": 0.9584031701087952, "learning_rate": 5.3859536778061655e-06, "loss": 0.4953, "step": 25490 }, { "epoch": 1.710781517398745, "grad_norm": 1.0502445697784424, "learning_rate": 5.381047902110453e-06, "loss": 0.4979, "step": 25492 }, { "epoch": 1.710915741082514, "grad_norm": 1.0048887729644775, "learning_rate": 5.376144234572478e-06, "loss": 0.3925, "step": 25494 }, { "epoch": 1.711049964766283, "grad_norm": 0.9571940898895264, "learning_rate": 5.371242675423949e-06, "loss": 0.3957, "step": 25496 }, { "epoch": 1.711184188450052, "grad_norm": 0.9335609078407288, "learning_rate": 5.366343224896436e-06, "loss": 0.434, "step": 25498 }, { "epoch": 1.711318412133821, "grad_norm": 1.1270277500152588, "learning_rate": 5.361445883221456e-06, "loss": 0.5048, "step": 25500 }, { "epoch": 1.71145263581759, "grad_norm": 1.101144790649414, "learning_rate": 5.356550650630359e-06, "loss": 0.4865, "step": 25502 }, { "epoch": 1.711586859501359, "grad_norm": 0.9321258664131165, "learning_rate": 5.35165752735447e-06, "loss": 0.421, "step": 25504 }, { "epoch": 1.711721083185128, "grad_norm": 1.087894082069397, "learning_rate": 5.346766513624951e-06, "loss": 0.453, "step": 25506 }, { "epoch": 1.7118553068688969, "grad_norm": 0.9895612597465515, "learning_rate": 5.341877609672919e-06, "loss": 0.4835, "step": 25508 }, { "epoch": 1.711989530552666, "grad_norm": 1.0466296672821045, "learning_rate": 5.336990815729354e-06, "loss": 0.4136, "step": 25510 }, { "epoch": 1.7121237542364351, "grad_norm": 1.0212011337280273, "learning_rate": 5.33210613202515e-06, "loss": 0.5037, "step": 25512 }, { "epoch": 1.712257977920204, "grad_norm": 1.071318507194519, "learning_rate": 5.327223558791084e-06, "loss": 0.5119, "step": 25514 }, { "epoch": 1.712392201603973, "grad_norm": 1.0949668884277344, "learning_rate": 5.322343096257864e-06, "loss": 0.5161, "step": 25516 }, { "epoch": 1.712526425287742, "grad_norm": 0.8076902627944946, "learning_rate": 5.317464744656081e-06, "loss": 0.4839, "step": 25518 }, { "epoch": 1.712660648971511, "grad_norm": 1.0798898935317993, "learning_rate": 5.312588504216226e-06, "loss": 0.5646, "step": 25520 }, { "epoch": 1.7127948726552802, "grad_norm": 0.9206319451332092, "learning_rate": 5.307714375168682e-06, "loss": 0.5107, "step": 25522 }, { "epoch": 1.712929096339049, "grad_norm": 1.0472077131271362, "learning_rate": 5.302842357743754e-06, "loss": 0.5383, "step": 25524 }, { "epoch": 1.713063320022818, "grad_norm": 1.1249159574508667, "learning_rate": 5.297972452171629e-06, "loss": 0.4088, "step": 25526 }, { "epoch": 1.713197543706587, "grad_norm": 1.0618162155151367, "learning_rate": 5.293104658682419e-06, "loss": 0.4658, "step": 25528 }, { "epoch": 1.7133317673903559, "grad_norm": 1.0651888847351074, "learning_rate": 5.288238977506077e-06, "loss": 0.5678, "step": 25530 }, { "epoch": 1.713465991074125, "grad_norm": 1.0403186082839966, "learning_rate": 5.283375408872537e-06, "loss": 0.4787, "step": 25532 }, { "epoch": 1.7136002147578941, "grad_norm": 1.0864567756652832, "learning_rate": 5.278513953011566e-06, "loss": 0.5136, "step": 25534 }, { "epoch": 1.713734438441663, "grad_norm": 0.9364561438560486, "learning_rate": 5.273654610152884e-06, "loss": 0.4665, "step": 25536 }, { "epoch": 1.713868662125432, "grad_norm": 1.0810825824737549, "learning_rate": 5.268797380526064e-06, "loss": 0.4828, "step": 25538 }, { "epoch": 1.714002885809201, "grad_norm": 0.7876555919647217, "learning_rate": 5.26394226436061e-06, "loss": 0.4377, "step": 25540 }, { "epoch": 1.71413710949297, "grad_norm": 1.094198226928711, "learning_rate": 5.259089261885908e-06, "loss": 0.4901, "step": 25542 }, { "epoch": 1.7142713331767392, "grad_norm": 1.0900894403457642, "learning_rate": 5.254238373331266e-06, "loss": 0.4378, "step": 25544 }, { "epoch": 1.714405556860508, "grad_norm": 0.9810633659362793, "learning_rate": 5.249389598925869e-06, "loss": 0.4832, "step": 25546 }, { "epoch": 1.714539780544277, "grad_norm": 0.9120228290557861, "learning_rate": 5.244542938898822e-06, "loss": 0.4448, "step": 25548 }, { "epoch": 1.714674004228046, "grad_norm": 0.9897778630256653, "learning_rate": 5.2396983934791085e-06, "loss": 0.4572, "step": 25550 }, { "epoch": 1.714808227911815, "grad_norm": 0.949037492275238, "learning_rate": 5.2348559628956185e-06, "loss": 0.4812, "step": 25552 }, { "epoch": 1.714942451595584, "grad_norm": 1.08169686794281, "learning_rate": 5.230015647377168e-06, "loss": 0.4856, "step": 25554 }, { "epoch": 1.7150766752793531, "grad_norm": 0.8839130997657776, "learning_rate": 5.225177447152446e-06, "loss": 0.5099, "step": 25556 }, { "epoch": 1.715210898963122, "grad_norm": 1.1443039178848267, "learning_rate": 5.220341362450038e-06, "loss": 0.4887, "step": 25558 }, { "epoch": 1.715345122646891, "grad_norm": 1.0248621702194214, "learning_rate": 5.215507393498437e-06, "loss": 0.4513, "step": 25560 }, { "epoch": 1.71547934633066, "grad_norm": 1.023232340812683, "learning_rate": 5.2106755405260555e-06, "loss": 0.4827, "step": 25562 }, { "epoch": 1.715613570014429, "grad_norm": 1.0187675952911377, "learning_rate": 5.205845803761172e-06, "loss": 0.422, "step": 25564 }, { "epoch": 1.7157477936981982, "grad_norm": 0.9954760670661926, "learning_rate": 5.201018183432005e-06, "loss": 0.4455, "step": 25566 }, { "epoch": 1.715882017381967, "grad_norm": 0.9201191067695618, "learning_rate": 5.196192679766626e-06, "loss": 0.4457, "step": 25568 }, { "epoch": 1.716016241065736, "grad_norm": 1.1385751962661743, "learning_rate": 5.191369292993048e-06, "loss": 0.4852, "step": 25570 }, { "epoch": 1.716150464749505, "grad_norm": 1.0543646812438965, "learning_rate": 5.186548023339144e-06, "loss": 0.5159, "step": 25572 }, { "epoch": 1.716284688433274, "grad_norm": 1.015798807144165, "learning_rate": 5.181728871032737e-06, "loss": 0.4604, "step": 25574 }, { "epoch": 1.716418912117043, "grad_norm": 1.0484914779663086, "learning_rate": 5.176911836301507e-06, "loss": 0.4596, "step": 25576 }, { "epoch": 1.7165531358008121, "grad_norm": 1.0813695192337036, "learning_rate": 5.172096919373048e-06, "loss": 0.5332, "step": 25578 }, { "epoch": 1.716687359484581, "grad_norm": 0.9233369827270508, "learning_rate": 5.167284120474858e-06, "loss": 0.4424, "step": 25580 }, { "epoch": 1.71682158316835, "grad_norm": 0.9958009719848633, "learning_rate": 5.162473439834337e-06, "loss": 0.4603, "step": 25582 }, { "epoch": 1.716955806852119, "grad_norm": 1.0955454111099243, "learning_rate": 5.157664877678781e-06, "loss": 0.4336, "step": 25584 }, { "epoch": 1.717090030535888, "grad_norm": 0.8009112477302551, "learning_rate": 5.15285843423538e-06, "loss": 0.3836, "step": 25586 }, { "epoch": 1.7172242542196572, "grad_norm": 1.1768845319747925, "learning_rate": 5.148054109731215e-06, "loss": 0.4783, "step": 25588 }, { "epoch": 1.717358477903426, "grad_norm": 1.0453424453735352, "learning_rate": 5.143251904393309e-06, "loss": 0.5188, "step": 25590 }, { "epoch": 1.717492701587195, "grad_norm": 1.0332114696502686, "learning_rate": 5.1384518184485365e-06, "loss": 0.4412, "step": 25592 }, { "epoch": 1.717626925270964, "grad_norm": 1.0137203931808472, "learning_rate": 5.1336538521237045e-06, "loss": 0.4908, "step": 25594 }, { "epoch": 1.717761148954733, "grad_norm": 1.1090179681777954, "learning_rate": 5.128858005645504e-06, "loss": 0.4465, "step": 25596 }, { "epoch": 1.7178953726385022, "grad_norm": 1.002066969871521, "learning_rate": 5.1240642792405314e-06, "loss": 0.4907, "step": 25598 }, { "epoch": 1.7180295963222711, "grad_norm": 1.0695041418075562, "learning_rate": 5.1192726731352616e-06, "loss": 0.4986, "step": 25600 }, { "epoch": 1.71816382000604, "grad_norm": 1.0685627460479736, "learning_rate": 5.114483187556113e-06, "loss": 0.4662, "step": 25602 }, { "epoch": 1.718298043689809, "grad_norm": 1.0577702522277832, "learning_rate": 5.1096958227293765e-06, "loss": 0.5022, "step": 25604 }, { "epoch": 1.718432267373578, "grad_norm": 1.0659515857696533, "learning_rate": 5.104910578881234e-06, "loss": 0.4163, "step": 25606 }, { "epoch": 1.718566491057347, "grad_norm": 1.0222342014312744, "learning_rate": 5.100127456237774e-06, "loss": 0.5014, "step": 25608 }, { "epoch": 1.7187007147411162, "grad_norm": 0.907884955406189, "learning_rate": 5.095346455025013e-06, "loss": 0.439, "step": 25610 }, { "epoch": 1.718834938424885, "grad_norm": 0.9424964189529419, "learning_rate": 5.09056757546883e-06, "loss": 0.4704, "step": 25612 }, { "epoch": 1.718969162108654, "grad_norm": 1.0116387605667114, "learning_rate": 5.085790817795016e-06, "loss": 0.45, "step": 25614 }, { "epoch": 1.719103385792423, "grad_norm": 1.033856749534607, "learning_rate": 5.08101618222927e-06, "loss": 0.46, "step": 25616 }, { "epoch": 1.719237609476192, "grad_norm": 1.0117535591125488, "learning_rate": 5.076243668997166e-06, "loss": 0.4289, "step": 25618 }, { "epoch": 1.7193718331599612, "grad_norm": 1.018268346786499, "learning_rate": 5.071473278324223e-06, "loss": 0.4865, "step": 25620 }, { "epoch": 1.7195060568437301, "grad_norm": 0.9708705544471741, "learning_rate": 5.066705010435807e-06, "loss": 0.4655, "step": 25622 }, { "epoch": 1.719640280527499, "grad_norm": 1.0578694343566895, "learning_rate": 5.061938865557242e-06, "loss": 0.4237, "step": 25624 }, { "epoch": 1.719774504211268, "grad_norm": 0.9732657074928284, "learning_rate": 5.057174843913681e-06, "loss": 0.4777, "step": 25626 }, { "epoch": 1.7199087278950371, "grad_norm": 1.0057907104492188, "learning_rate": 5.05241294573024e-06, "loss": 0.4298, "step": 25628 }, { "epoch": 1.720042951578806, "grad_norm": 0.9988004565238953, "learning_rate": 5.047653171231892e-06, "loss": 0.4104, "step": 25630 }, { "epoch": 1.7201771752625752, "grad_norm": 1.0792229175567627, "learning_rate": 5.042895520643548e-06, "loss": 0.4588, "step": 25632 }, { "epoch": 1.720311398946344, "grad_norm": 1.048791766166687, "learning_rate": 5.038139994189983e-06, "loss": 0.432, "step": 25634 }, { "epoch": 1.720445622630113, "grad_norm": 0.9940659403800964, "learning_rate": 5.03338659209589e-06, "loss": 0.4821, "step": 25636 }, { "epoch": 1.720579846313882, "grad_norm": 1.1075587272644043, "learning_rate": 5.02863531458585e-06, "loss": 0.4728, "step": 25638 }, { "epoch": 1.720714069997651, "grad_norm": 0.9461389780044556, "learning_rate": 5.023886161884372e-06, "loss": 0.4511, "step": 25640 }, { "epoch": 1.7208482936814202, "grad_norm": 1.0536267757415771, "learning_rate": 5.019139134215828e-06, "loss": 0.4944, "step": 25642 }, { "epoch": 1.7209825173651891, "grad_norm": 1.0569151639938354, "learning_rate": 5.014394231804503e-06, "loss": 0.458, "step": 25644 }, { "epoch": 1.721116741048958, "grad_norm": 1.0429084300994873, "learning_rate": 5.009651454874587e-06, "loss": 0.4244, "step": 25646 }, { "epoch": 1.721250964732727, "grad_norm": 1.1130216121673584, "learning_rate": 5.004910803650181e-06, "loss": 0.4691, "step": 25648 }, { "epoch": 1.7213851884164961, "grad_norm": 0.9673048257827759, "learning_rate": 5.000172278355248e-06, "loss": 0.4656, "step": 25650 }, { "epoch": 1.721519412100265, "grad_norm": 1.0673524141311646, "learning_rate": 4.9954358792137054e-06, "loss": 0.4935, "step": 25652 }, { "epoch": 1.7216536357840342, "grad_norm": 1.0728918313980103, "learning_rate": 4.9907016064493e-06, "loss": 0.4575, "step": 25654 }, { "epoch": 1.721787859467803, "grad_norm": 1.2493226528167725, "learning_rate": 4.98596946028575e-06, "loss": 0.4618, "step": 25656 }, { "epoch": 1.721922083151572, "grad_norm": 2.444298505783081, "learning_rate": 4.981239440946612e-06, "loss": 0.4587, "step": 25658 }, { "epoch": 1.722056306835341, "grad_norm": 1.0520304441452026, "learning_rate": 4.976511548655399e-06, "loss": 0.4246, "step": 25660 }, { "epoch": 1.72219053051911, "grad_norm": 1.1098594665527344, "learning_rate": 4.97178578363548e-06, "loss": 0.5034, "step": 25662 }, { "epoch": 1.7223247542028792, "grad_norm": 1.1211425065994263, "learning_rate": 4.967062146110135e-06, "loss": 0.5041, "step": 25664 }, { "epoch": 1.7224589778866481, "grad_norm": 0.9872859120368958, "learning_rate": 4.962340636302543e-06, "loss": 0.4887, "step": 25666 }, { "epoch": 1.722593201570417, "grad_norm": 1.1875641345977783, "learning_rate": 4.957621254435801e-06, "loss": 0.5114, "step": 25668 }, { "epoch": 1.722727425254186, "grad_norm": 0.8748855590820312, "learning_rate": 4.952904000732883e-06, "loss": 0.4223, "step": 25670 }, { "epoch": 1.7228616489379551, "grad_norm": 1.0315881967544556, "learning_rate": 4.948188875416671e-06, "loss": 0.49, "step": 25672 }, { "epoch": 1.7229958726217243, "grad_norm": 1.328240156173706, "learning_rate": 4.943475878709936e-06, "loss": 0.4946, "step": 25674 }, { "epoch": 1.7231300963054932, "grad_norm": 1.059990644454956, "learning_rate": 4.938765010835373e-06, "loss": 0.5015, "step": 25676 }, { "epoch": 1.723264319989262, "grad_norm": 1.1780343055725098, "learning_rate": 4.934056272015541e-06, "loss": 0.45, "step": 25678 }, { "epoch": 1.723398543673031, "grad_norm": 1.0624507665634155, "learning_rate": 4.92934966247296e-06, "loss": 0.4655, "step": 25680 }, { "epoch": 1.7235327673568, "grad_norm": 1.0479305982589722, "learning_rate": 4.924645182429966e-06, "loss": 0.4233, "step": 25682 }, { "epoch": 1.723666991040569, "grad_norm": 1.0353846549987793, "learning_rate": 4.919942832108837e-06, "loss": 0.5271, "step": 25684 }, { "epoch": 1.7238012147243382, "grad_norm": 1.0945695638656616, "learning_rate": 4.915242611731774e-06, "loss": 0.501, "step": 25686 }, { "epoch": 1.7239354384081071, "grad_norm": 0.9750905632972717, "learning_rate": 4.910544521520838e-06, "loss": 0.399, "step": 25688 }, { "epoch": 1.724069662091876, "grad_norm": 0.9547551274299622, "learning_rate": 4.905848561698023e-06, "loss": 0.4843, "step": 25690 }, { "epoch": 1.724203885775645, "grad_norm": 1.0283516645431519, "learning_rate": 4.901154732485169e-06, "loss": 0.435, "step": 25692 }, { "epoch": 1.7243381094594141, "grad_norm": 1.0085430145263672, "learning_rate": 4.896463034104082e-06, "loss": 0.4541, "step": 25694 }, { "epoch": 1.7244723331431833, "grad_norm": 1.0000320672988892, "learning_rate": 4.891773466776417e-06, "loss": 0.441, "step": 25696 }, { "epoch": 1.7246065568269522, "grad_norm": 1.0053753852844238, "learning_rate": 4.887086030723764e-06, "loss": 0.459, "step": 25698 }, { "epoch": 1.724740780510721, "grad_norm": 0.9564087986946106, "learning_rate": 4.88240072616758e-06, "loss": 0.495, "step": 25700 }, { "epoch": 1.72487500419449, "grad_norm": 1.08132803440094, "learning_rate": 4.877717553329247e-06, "loss": 0.4802, "step": 25702 }, { "epoch": 1.7250092278782592, "grad_norm": 0.9329249262809753, "learning_rate": 4.873036512430018e-06, "loss": 0.4915, "step": 25704 }, { "epoch": 1.725143451562028, "grad_norm": 0.9538440704345703, "learning_rate": 4.868357603691087e-06, "loss": 0.4306, "step": 25706 }, { "epoch": 1.7252776752457972, "grad_norm": 1.119462013244629, "learning_rate": 4.863680827333511e-06, "loss": 0.4666, "step": 25708 }, { "epoch": 1.7254118989295661, "grad_norm": 1.0777448415756226, "learning_rate": 4.859006183578263e-06, "loss": 0.4519, "step": 25710 }, { "epoch": 1.725546122613335, "grad_norm": 1.002044677734375, "learning_rate": 4.854333672646188e-06, "loss": 0.3923, "step": 25712 }, { "epoch": 1.725680346297104, "grad_norm": 1.0467420816421509, "learning_rate": 4.849663294758089e-06, "loss": 0.4285, "step": 25714 }, { "epoch": 1.7258145699808731, "grad_norm": 1.0576964616775513, "learning_rate": 4.844995050134604e-06, "loss": 0.4917, "step": 25716 }, { "epoch": 1.7259487936646423, "grad_norm": 1.1266741752624512, "learning_rate": 4.840328938996325e-06, "loss": 0.4802, "step": 25718 }, { "epoch": 1.7260830173484112, "grad_norm": 1.1078001260757446, "learning_rate": 4.835664961563685e-06, "loss": 0.4634, "step": 25720 }, { "epoch": 1.72621724103218, "grad_norm": 1.1022974252700806, "learning_rate": 4.831003118057076e-06, "loss": 0.4226, "step": 25722 }, { "epoch": 1.726351464715949, "grad_norm": 1.0248111486434937, "learning_rate": 4.826343408696732e-06, "loss": 0.4809, "step": 25724 }, { "epoch": 1.7264856883997182, "grad_norm": 0.9810677170753479, "learning_rate": 4.821685833702849e-06, "loss": 0.4922, "step": 25726 }, { "epoch": 1.726619912083487, "grad_norm": 0.9958468675613403, "learning_rate": 4.817030393295463e-06, "loss": 0.4347, "step": 25728 }, { "epoch": 1.7267541357672562, "grad_norm": 0.9662964344024658, "learning_rate": 4.812377087694547e-06, "loss": 0.4466, "step": 25730 }, { "epoch": 1.7268883594510251, "grad_norm": 1.0769606828689575, "learning_rate": 4.807725917119949e-06, "loss": 0.4559, "step": 25732 }, { "epoch": 1.727022583134794, "grad_norm": 1.0293160676956177, "learning_rate": 4.803076881791441e-06, "loss": 0.4372, "step": 25734 }, { "epoch": 1.727156806818563, "grad_norm": 1.045389175415039, "learning_rate": 4.798429981928676e-06, "loss": 0.4758, "step": 25736 }, { "epoch": 1.7272910305023321, "grad_norm": 1.0050643682479858, "learning_rate": 4.793785217751206e-06, "loss": 0.4356, "step": 25738 }, { "epoch": 1.7274252541861013, "grad_norm": 1.9420435428619385, "learning_rate": 4.7891425894784845e-06, "loss": 0.4696, "step": 25740 }, { "epoch": 1.7275594778698702, "grad_norm": 0.9572662115097046, "learning_rate": 4.7845020973298795e-06, "loss": 0.504, "step": 25742 }, { "epoch": 1.727693701553639, "grad_norm": 0.9887237548828125, "learning_rate": 4.779863741524626e-06, "loss": 0.5097, "step": 25744 }, { "epoch": 1.727827925237408, "grad_norm": 0.9231210350990295, "learning_rate": 4.775227522281911e-06, "loss": 0.4272, "step": 25746 }, { "epoch": 1.7279621489211772, "grad_norm": 1.1272821426391602, "learning_rate": 4.770593439820747e-06, "loss": 0.4523, "step": 25748 }, { "epoch": 1.7280963726049463, "grad_norm": 1.0660916566848755, "learning_rate": 4.76596149436011e-06, "loss": 0.5215, "step": 25750 }, { "epoch": 1.7282305962887152, "grad_norm": 1.0930113792419434, "learning_rate": 4.761331686118848e-06, "loss": 0.4823, "step": 25752 }, { "epoch": 1.7283648199724841, "grad_norm": 1.0196325778961182, "learning_rate": 4.756704015315694e-06, "loss": 0.4392, "step": 25754 }, { "epoch": 1.728499043656253, "grad_norm": 1.0132194757461548, "learning_rate": 4.7520784821693146e-06, "loss": 0.441, "step": 25756 }, { "epoch": 1.728633267340022, "grad_norm": 0.9595327377319336, "learning_rate": 4.74745508689825e-06, "loss": 0.3998, "step": 25758 }, { "epoch": 1.7287674910237911, "grad_norm": 1.6109375953674316, "learning_rate": 4.742833829720955e-06, "loss": 0.4632, "step": 25760 }, { "epoch": 1.7289017147075603, "grad_norm": 1.0205250978469849, "learning_rate": 4.738214710855748e-06, "loss": 0.4755, "step": 25762 }, { "epoch": 1.7290359383913292, "grad_norm": 1.1388388872146606, "learning_rate": 4.733597730520911e-06, "loss": 0.4542, "step": 25764 }, { "epoch": 1.729170162075098, "grad_norm": 0.931294322013855, "learning_rate": 4.728982888934563e-06, "loss": 0.502, "step": 25766 }, { "epoch": 1.729304385758867, "grad_norm": 1.1117701530456543, "learning_rate": 4.7243701863147525e-06, "loss": 0.4964, "step": 25768 }, { "epoch": 1.7294386094426362, "grad_norm": 1.090018630027771, "learning_rate": 4.71975962287941e-06, "loss": 0.5132, "step": 25770 }, { "epoch": 1.7295728331264053, "grad_norm": 1.0194764137268066, "learning_rate": 4.7151511988463955e-06, "loss": 0.439, "step": 25772 }, { "epoch": 1.7297070568101742, "grad_norm": 1.075696587562561, "learning_rate": 4.7105449144334345e-06, "loss": 0.5526, "step": 25774 }, { "epoch": 1.7298412804939431, "grad_norm": 1.0920310020446777, "learning_rate": 4.70594076985818e-06, "loss": 0.4999, "step": 25776 }, { "epoch": 1.729975504177712, "grad_norm": 0.9902180433273315, "learning_rate": 4.7013387653381424e-06, "loss": 0.4246, "step": 25778 }, { "epoch": 1.7301097278614812, "grad_norm": 1.0894775390625, "learning_rate": 4.696738901090781e-06, "loss": 0.4508, "step": 25780 }, { "epoch": 1.7302439515452501, "grad_norm": 0.9822218418121338, "learning_rate": 4.692141177333409e-06, "loss": 0.4512, "step": 25782 }, { "epoch": 1.7303781752290193, "grad_norm": 0.9878543615341187, "learning_rate": 4.687545594283282e-06, "loss": 0.4415, "step": 25784 }, { "epoch": 1.7305123989127882, "grad_norm": 1.0516366958618164, "learning_rate": 4.682952152157522e-06, "loss": 0.4311, "step": 25786 }, { "epoch": 1.730646622596557, "grad_norm": 1.0247827768325806, "learning_rate": 4.6783608511731635e-06, "loss": 0.4547, "step": 25788 }, { "epoch": 1.730780846280326, "grad_norm": 1.1121560335159302, "learning_rate": 4.673771691547124e-06, "loss": 0.5038, "step": 25790 }, { "epoch": 1.7309150699640952, "grad_norm": 1.0238251686096191, "learning_rate": 4.669184673496252e-06, "loss": 0.3958, "step": 25792 }, { "epoch": 1.7310492936478643, "grad_norm": 0.9520905017852783, "learning_rate": 4.664599797237263e-06, "loss": 0.4223, "step": 25794 }, { "epoch": 1.7311835173316332, "grad_norm": 1.1433610916137695, "learning_rate": 4.660017062986782e-06, "loss": 0.4488, "step": 25796 }, { "epoch": 1.7313177410154021, "grad_norm": 1.0909218788146973, "learning_rate": 4.655436470961333e-06, "loss": 0.514, "step": 25798 }, { "epoch": 1.731451964699171, "grad_norm": 1.041691780090332, "learning_rate": 4.650858021377352e-06, "loss": 0.4272, "step": 25800 }, { "epoch": 1.7315861883829402, "grad_norm": 0.889392077922821, "learning_rate": 4.64628171445115e-06, "loss": 0.4671, "step": 25802 }, { "epoch": 1.7317204120667091, "grad_norm": 0.9627600908279419, "learning_rate": 4.641707550398966e-06, "loss": 0.4055, "step": 25804 }, { "epoch": 1.7318546357504783, "grad_norm": 1.0190232992172241, "learning_rate": 4.6371355294368865e-06, "loss": 0.5, "step": 25806 }, { "epoch": 1.7319888594342472, "grad_norm": 1.1649503707885742, "learning_rate": 4.632565651780968e-06, "loss": 0.4778, "step": 25808 }, { "epoch": 1.732123083118016, "grad_norm": 1.000589370727539, "learning_rate": 4.627997917647098e-06, "loss": 0.428, "step": 25810 }, { "epoch": 1.732257306801785, "grad_norm": 1.0200680494308472, "learning_rate": 4.623432327251109e-06, "loss": 0.5022, "step": 25812 }, { "epoch": 1.7323915304855542, "grad_norm": 0.9837169051170349, "learning_rate": 4.618868880808725e-06, "loss": 0.4593, "step": 25814 }, { "epoch": 1.7325257541693233, "grad_norm": 1.0563697814941406, "learning_rate": 4.614307578535537e-06, "loss": 0.3936, "step": 25816 }, { "epoch": 1.7326599778530922, "grad_norm": 1.0165019035339355, "learning_rate": 4.6097484206470756e-06, "loss": 0.4896, "step": 25818 }, { "epoch": 1.7327942015368611, "grad_norm": 1.0807873010635376, "learning_rate": 4.605191407358733e-06, "loss": 0.4371, "step": 25820 }, { "epoch": 1.73292842522063, "grad_norm": 1.0654383897781372, "learning_rate": 4.600636538885844e-06, "loss": 0.4853, "step": 25822 }, { "epoch": 1.7330626489043992, "grad_norm": 1.1478780508041382, "learning_rate": 4.596083815443602e-06, "loss": 0.5305, "step": 25824 }, { "epoch": 1.7331968725881683, "grad_norm": 1.028700590133667, "learning_rate": 4.5915332372471195e-06, "loss": 0.4778, "step": 25826 }, { "epoch": 1.7333310962719373, "grad_norm": 1.0302728414535522, "learning_rate": 4.586984804511385e-06, "loss": 0.4407, "step": 25828 }, { "epoch": 1.7334653199557062, "grad_norm": 1.2134575843811035, "learning_rate": 4.5824385174513316e-06, "loss": 0.523, "step": 25830 }, { "epoch": 1.733599543639475, "grad_norm": 0.983025074005127, "learning_rate": 4.577894376281744e-06, "loss": 0.413, "step": 25832 }, { "epoch": 1.733733767323244, "grad_norm": 1.09456467628479, "learning_rate": 4.573352381217333e-06, "loss": 0.459, "step": 25834 }, { "epoch": 1.7338679910070132, "grad_norm": 1.0942330360412598, "learning_rate": 4.568812532472683e-06, "loss": 0.4829, "step": 25836 }, { "epoch": 1.7340022146907823, "grad_norm": 1.0838247537612915, "learning_rate": 4.564274830262316e-06, "loss": 0.4836, "step": 25838 }, { "epoch": 1.7341364383745512, "grad_norm": 1.0898464918136597, "learning_rate": 4.559739274800606e-06, "loss": 0.4746, "step": 25840 }, { "epoch": 1.7342706620583201, "grad_norm": 1.0721734762191772, "learning_rate": 4.555205866301876e-06, "loss": 0.416, "step": 25842 }, { "epoch": 1.734404885742089, "grad_norm": 1.0424118041992188, "learning_rate": 4.5506746049802925e-06, "loss": 0.5083, "step": 25844 }, { "epoch": 1.7345391094258582, "grad_norm": 1.02657151222229, "learning_rate": 4.546145491049969e-06, "loss": 0.4488, "step": 25846 }, { "epoch": 1.7346733331096273, "grad_norm": 1.0207029581069946, "learning_rate": 4.541618524724878e-06, "loss": 0.4448, "step": 25848 }, { "epoch": 1.7348075567933963, "grad_norm": 1.104272723197937, "learning_rate": 4.537093706218937e-06, "loss": 0.4415, "step": 25850 }, { "epoch": 1.7349417804771652, "grad_norm": 0.89890056848526, "learning_rate": 4.532571035745914e-06, "loss": 0.4176, "step": 25852 }, { "epoch": 1.735076004160934, "grad_norm": 0.9957574009895325, "learning_rate": 4.528050513519505e-06, "loss": 0.4607, "step": 25854 }, { "epoch": 1.7352102278447032, "grad_norm": 0.9569705724716187, "learning_rate": 4.523532139753278e-06, "loss": 0.4531, "step": 25856 }, { "epoch": 1.7353444515284722, "grad_norm": 0.9871608018875122, "learning_rate": 4.519015914660746e-06, "loss": 0.4515, "step": 25858 }, { "epoch": 1.7354786752122413, "grad_norm": 1.1361002922058105, "learning_rate": 4.514501838455276e-06, "loss": 0.4474, "step": 25860 }, { "epoch": 1.7356128988960102, "grad_norm": 1.1392649412155151, "learning_rate": 4.5099899113501545e-06, "loss": 0.4414, "step": 25862 }, { "epoch": 1.7357471225797791, "grad_norm": 1.0350909233093262, "learning_rate": 4.505480133558543e-06, "loss": 0.4694, "step": 25864 }, { "epoch": 1.735881346263548, "grad_norm": 1.1096014976501465, "learning_rate": 4.500972505293544e-06, "loss": 0.4315, "step": 25866 }, { "epoch": 1.7360155699473172, "grad_norm": 0.9327784776687622, "learning_rate": 4.4964670267681146e-06, "loss": 0.4408, "step": 25868 }, { "epoch": 1.7361497936310863, "grad_norm": 0.9927994608879089, "learning_rate": 4.491963698195162e-06, "loss": 0.4367, "step": 25870 }, { "epoch": 1.7362840173148553, "grad_norm": 1.0583723783493042, "learning_rate": 4.487462519787416e-06, "loss": 0.4304, "step": 25872 }, { "epoch": 1.7364182409986242, "grad_norm": 1.0991770029067993, "learning_rate": 4.482963491757574e-06, "loss": 0.5365, "step": 25874 }, { "epoch": 1.736552464682393, "grad_norm": 1.037761926651001, "learning_rate": 4.4784666143181996e-06, "loss": 0.4912, "step": 25876 }, { "epoch": 1.7366866883661622, "grad_norm": 1.0798840522766113, "learning_rate": 4.473971887681772e-06, "loss": 0.4917, "step": 25878 }, { "epoch": 1.7368209120499312, "grad_norm": 1.3390872478485107, "learning_rate": 4.469479312060643e-06, "loss": 0.4258, "step": 25880 }, { "epoch": 1.7369551357337003, "grad_norm": 1.0587348937988281, "learning_rate": 4.464988887667087e-06, "loss": 0.4164, "step": 25882 }, { "epoch": 1.7370893594174692, "grad_norm": 1.1341004371643066, "learning_rate": 4.460500614713259e-06, "loss": 0.4923, "step": 25884 }, { "epoch": 1.7372235831012381, "grad_norm": 0.9715561866760254, "learning_rate": 4.456014493411237e-06, "loss": 0.4884, "step": 25886 }, { "epoch": 1.737357806785007, "grad_norm": 0.9592590928077698, "learning_rate": 4.4515305239729685e-06, "loss": 0.483, "step": 25888 }, { "epoch": 1.7374920304687762, "grad_norm": 1.0191220045089722, "learning_rate": 4.447048706610318e-06, "loss": 0.4264, "step": 25890 }, { "epoch": 1.7376262541525453, "grad_norm": 1.0349215269088745, "learning_rate": 4.442569041535039e-06, "loss": 0.4431, "step": 25892 }, { "epoch": 1.7377604778363143, "grad_norm": 1.1754120588302612, "learning_rate": 4.4380915289587825e-06, "loss": 0.5029, "step": 25894 }, { "epoch": 1.7378947015200832, "grad_norm": 1.0315639972686768, "learning_rate": 4.433616169093113e-06, "loss": 0.4927, "step": 25896 }, { "epoch": 1.738028925203852, "grad_norm": 1.0900938510894775, "learning_rate": 4.429142962149474e-06, "loss": 0.5482, "step": 25898 }, { "epoch": 1.7381631488876212, "grad_norm": 0.9947560429573059, "learning_rate": 4.424671908339223e-06, "loss": 0.4212, "step": 25900 }, { "epoch": 1.7382973725713904, "grad_norm": 0.999907910823822, "learning_rate": 4.420203007873597e-06, "loss": 0.5176, "step": 25902 }, { "epoch": 1.7384315962551593, "grad_norm": 1.1585700511932373, "learning_rate": 4.415736260963755e-06, "loss": 0.4646, "step": 25904 }, { "epoch": 1.7385658199389282, "grad_norm": 1.0315663814544678, "learning_rate": 4.411271667820727e-06, "loss": 0.4106, "step": 25906 }, { "epoch": 1.7387000436226971, "grad_norm": 0.9721665978431702, "learning_rate": 4.406809228655479e-06, "loss": 0.4395, "step": 25908 }, { "epoch": 1.738834267306466, "grad_norm": 0.9676013588905334, "learning_rate": 4.4023489436788355e-06, "loss": 0.5162, "step": 25910 }, { "epoch": 1.7389684909902352, "grad_norm": 0.998012363910675, "learning_rate": 4.397890813101546e-06, "loss": 0.5001, "step": 25912 }, { "epoch": 1.7391027146740043, "grad_norm": 1.006813406944275, "learning_rate": 4.39343483713423e-06, "loss": 0.4572, "step": 25914 }, { "epoch": 1.7392369383577733, "grad_norm": 0.9706727862358093, "learning_rate": 4.388981015987448e-06, "loss": 0.4709, "step": 25916 }, { "epoch": 1.7393711620415422, "grad_norm": 1.0576826333999634, "learning_rate": 4.384529349871625e-06, "loss": 0.4457, "step": 25918 }, { "epoch": 1.739505385725311, "grad_norm": 1.560713291168213, "learning_rate": 4.380079838997086e-06, "loss": 0.42, "step": 25920 }, { "epoch": 1.7396396094090802, "grad_norm": 1.1104066371917725, "learning_rate": 4.37563248357406e-06, "loss": 0.5065, "step": 25922 }, { "epoch": 1.7397738330928494, "grad_norm": 0.9899181127548218, "learning_rate": 4.371187283812689e-06, "loss": 0.4128, "step": 25924 }, { "epoch": 1.7399080567766183, "grad_norm": 0.9700545072555542, "learning_rate": 4.366744239922998e-06, "loss": 0.4631, "step": 25926 }, { "epoch": 1.7400422804603872, "grad_norm": 1.1491373777389526, "learning_rate": 4.362303352114905e-06, "loss": 0.4641, "step": 25928 }, { "epoch": 1.7401765041441561, "grad_norm": 1.3101502656936646, "learning_rate": 4.357864620598229e-06, "loss": 0.4574, "step": 25930 }, { "epoch": 1.7403107278279253, "grad_norm": 0.8173026442527771, "learning_rate": 4.3534280455827024e-06, "loss": 0.4014, "step": 25932 }, { "epoch": 1.7404449515116942, "grad_norm": 1.0041718482971191, "learning_rate": 4.348993627277936e-06, "loss": 0.4621, "step": 25934 }, { "epoch": 1.7405791751954633, "grad_norm": 0.9188247919082642, "learning_rate": 4.3445613658934624e-06, "loss": 0.4776, "step": 25936 }, { "epoch": 1.7407133988792323, "grad_norm": 1.0293664932250977, "learning_rate": 4.340131261638686e-06, "loss": 0.4991, "step": 25938 }, { "epoch": 1.7408476225630012, "grad_norm": 0.984596848487854, "learning_rate": 4.335703314722916e-06, "loss": 0.4099, "step": 25940 }, { "epoch": 1.74098184624677, "grad_norm": 1.1050786972045898, "learning_rate": 4.331277525355365e-06, "loss": 0.441, "step": 25942 }, { "epoch": 1.7411160699305392, "grad_norm": 1.0016874074935913, "learning_rate": 4.326853893745153e-06, "loss": 0.413, "step": 25944 }, { "epoch": 1.7412502936143084, "grad_norm": 1.119035005569458, "learning_rate": 4.3224324201012854e-06, "loss": 0.44, "step": 25946 }, { "epoch": 1.7413845172980773, "grad_norm": 0.9717772006988525, "learning_rate": 4.3180131046326626e-06, "loss": 0.4439, "step": 25948 }, { "epoch": 1.7415187409818462, "grad_norm": 1.0132006406784058, "learning_rate": 4.313595947548082e-06, "loss": 0.4931, "step": 25950 }, { "epoch": 1.7416529646656151, "grad_norm": 1.0019216537475586, "learning_rate": 4.309180949056269e-06, "loss": 0.4531, "step": 25952 }, { "epoch": 1.7417871883493843, "grad_norm": 1.0533734560012817, "learning_rate": 4.304768109365792e-06, "loss": 0.43, "step": 25954 }, { "epoch": 1.7419214120331532, "grad_norm": 1.076453447341919, "learning_rate": 4.300357428685187e-06, "loss": 0.5029, "step": 25956 }, { "epoch": 1.7420556357169223, "grad_norm": 1.091066598892212, "learning_rate": 4.295948907222824e-06, "loss": 0.4205, "step": 25958 }, { "epoch": 1.7421898594006913, "grad_norm": 1.061374545097351, "learning_rate": 4.2915425451869884e-06, "loss": 0.4575, "step": 25960 }, { "epoch": 1.7423240830844602, "grad_norm": 1.1788666248321533, "learning_rate": 4.287138342785896e-06, "loss": 0.4443, "step": 25962 }, { "epoch": 1.742458306768229, "grad_norm": 0.9478891491889954, "learning_rate": 4.2827363002276135e-06, "loss": 0.4862, "step": 25964 }, { "epoch": 1.7425925304519982, "grad_norm": 1.106594204902649, "learning_rate": 4.278336417720163e-06, "loss": 0.513, "step": 25966 }, { "epoch": 1.7427267541357674, "grad_norm": 1.0672873258590698, "learning_rate": 4.273938695471391e-06, "loss": 0.5416, "step": 25968 }, { "epoch": 1.7428609778195363, "grad_norm": 0.9814993143081665, "learning_rate": 4.269543133689108e-06, "loss": 0.4372, "step": 25970 }, { "epoch": 1.7429952015033052, "grad_norm": 1.0161956548690796, "learning_rate": 4.265149732580981e-06, "loss": 0.4437, "step": 25972 }, { "epoch": 1.7431294251870741, "grad_norm": 1.0389659404754639, "learning_rate": 4.2607584923546e-06, "loss": 0.4365, "step": 25974 }, { "epoch": 1.7432636488708433, "grad_norm": 1.1154420375823975, "learning_rate": 4.256369413217442e-06, "loss": 0.5105, "step": 25976 }, { "epoch": 1.7433978725546124, "grad_norm": 1.0543137788772583, "learning_rate": 4.251982495376872e-06, "loss": 0.4155, "step": 25978 }, { "epoch": 1.7435320962383813, "grad_norm": 1.1536834239959717, "learning_rate": 4.247597739040166e-06, "loss": 0.4518, "step": 25980 }, { "epoch": 1.7436663199221503, "grad_norm": 1.071794867515564, "learning_rate": 4.2432151444145085e-06, "loss": 0.4511, "step": 25982 }, { "epoch": 1.7438005436059192, "grad_norm": 1.0777956247329712, "learning_rate": 4.238834711706952e-06, "loss": 0.4817, "step": 25984 }, { "epoch": 1.743934767289688, "grad_norm": 0.9815927147865295, "learning_rate": 4.234456441124474e-06, "loss": 0.4578, "step": 25986 }, { "epoch": 1.7440689909734572, "grad_norm": 1.0425565242767334, "learning_rate": 4.230080332873926e-06, "loss": 0.4444, "step": 25988 }, { "epoch": 1.7442032146572264, "grad_norm": 1.0789756774902344, "learning_rate": 4.225706387162087e-06, "loss": 0.4435, "step": 25990 }, { "epoch": 1.7443374383409953, "grad_norm": 1.0805772542953491, "learning_rate": 4.221334604195604e-06, "loss": 0.4818, "step": 25992 }, { "epoch": 1.7444716620247642, "grad_norm": 1.0187358856201172, "learning_rate": 4.21696498418106e-06, "loss": 0.399, "step": 25994 }, { "epoch": 1.7446058857085331, "grad_norm": 0.9971460103988647, "learning_rate": 4.212597527324869e-06, "loss": 0.4628, "step": 25996 }, { "epoch": 1.7447401093923023, "grad_norm": 1.0509072542190552, "learning_rate": 4.2082322338334204e-06, "loss": 0.4876, "step": 25998 }, { "epoch": 1.7448743330760714, "grad_norm": 0.9012466073036194, "learning_rate": 4.203869103912944e-06, "loss": 0.418, "step": 26000 }, { "epoch": 1.7450085567598403, "grad_norm": 1.104740023612976, "learning_rate": 4.1995081377696035e-06, "loss": 0.4798, "step": 26002 }, { "epoch": 1.7451427804436093, "grad_norm": 1.0096579790115356, "learning_rate": 4.195149335609444e-06, "loss": 0.4943, "step": 26004 }, { "epoch": 1.7452770041273782, "grad_norm": 0.9526199698448181, "learning_rate": 4.190792697638407e-06, "loss": 0.4269, "step": 26006 }, { "epoch": 1.7454112278111473, "grad_norm": 1.027766466140747, "learning_rate": 4.186438224062322e-06, "loss": 0.4646, "step": 26008 }, { "epoch": 1.7455454514949162, "grad_norm": 1.0538530349731445, "learning_rate": 4.182085915086958e-06, "loss": 0.4501, "step": 26010 }, { "epoch": 1.7456796751786854, "grad_norm": 0.9318313002586365, "learning_rate": 4.1777357709179345e-06, "loss": 0.4863, "step": 26012 }, { "epoch": 1.7458138988624543, "grad_norm": 1.1026418209075928, "learning_rate": 4.1733877917607914e-06, "loss": 0.5779, "step": 26014 }, { "epoch": 1.7459481225462232, "grad_norm": 1.0496286153793335, "learning_rate": 4.169041977820948e-06, "loss": 0.4837, "step": 26016 }, { "epoch": 1.7460823462299921, "grad_norm": 1.0120539665222168, "learning_rate": 4.164698329303762e-06, "loss": 0.469, "step": 26018 }, { "epoch": 1.7462165699137613, "grad_norm": 0.9936702251434326, "learning_rate": 4.1603568464144375e-06, "loss": 0.4565, "step": 26020 }, { "epoch": 1.7463507935975304, "grad_norm": 0.9050431847572327, "learning_rate": 4.156017529358131e-06, "loss": 0.4575, "step": 26022 }, { "epoch": 1.7464850172812993, "grad_norm": 0.9547222852706909, "learning_rate": 4.151680378339834e-06, "loss": 0.4462, "step": 26024 }, { "epoch": 1.7466192409650683, "grad_norm": 0.9970137476921082, "learning_rate": 4.147345393564478e-06, "loss": 0.459, "step": 26026 }, { "epoch": 1.7467534646488372, "grad_norm": 1.0239909887313843, "learning_rate": 4.1430125752368986e-06, "loss": 0.4064, "step": 26028 }, { "epoch": 1.7468876883326063, "grad_norm": 0.9084471464157104, "learning_rate": 4.138681923561788e-06, "loss": 0.4702, "step": 26030 }, { "epoch": 1.7470219120163752, "grad_norm": 1.1111291646957397, "learning_rate": 4.134353438743793e-06, "loss": 0.4944, "step": 26032 }, { "epoch": 1.7471561357001444, "grad_norm": 0.9989820718765259, "learning_rate": 4.13002712098739e-06, "loss": 0.4798, "step": 26034 }, { "epoch": 1.7472903593839133, "grad_norm": 1.0311503410339355, "learning_rate": 4.125702970497014e-06, "loss": 0.4669, "step": 26036 }, { "epoch": 1.7474245830676822, "grad_norm": 0.9694811105728149, "learning_rate": 4.121380987476953e-06, "loss": 0.4531, "step": 26038 }, { "epoch": 1.7475588067514511, "grad_norm": 1.0406783819198608, "learning_rate": 4.117061172131437e-06, "loss": 0.4443, "step": 26040 }, { "epoch": 1.7476930304352203, "grad_norm": 0.9970607161521912, "learning_rate": 4.112743524664553e-06, "loss": 0.3971, "step": 26042 }, { "epoch": 1.7478272541189894, "grad_norm": 1.0036909580230713, "learning_rate": 4.108428045280305e-06, "loss": 0.4603, "step": 26044 }, { "epoch": 1.7479614778027583, "grad_norm": 0.939334511756897, "learning_rate": 4.104114734182584e-06, "loss": 0.4816, "step": 26046 }, { "epoch": 1.7480957014865273, "grad_norm": 1.4095072746276855, "learning_rate": 4.0998035915751965e-06, "loss": 0.4273, "step": 26048 }, { "epoch": 1.7482299251702962, "grad_norm": 1.0417141914367676, "learning_rate": 4.095494617661833e-06, "loss": 0.4415, "step": 26050 }, { "epoch": 1.7483641488540653, "grad_norm": 1.396823525428772, "learning_rate": 4.0911878126460805e-06, "loss": 0.4924, "step": 26052 }, { "epoch": 1.7484983725378345, "grad_norm": 0.9825583100318909, "learning_rate": 4.086883176731415e-06, "loss": 0.439, "step": 26054 }, { "epoch": 1.7486325962216034, "grad_norm": 1.0633548498153687, "learning_rate": 4.082580710121248e-06, "loss": 0.3985, "step": 26056 }, { "epoch": 1.7487668199053723, "grad_norm": 1.0380185842514038, "learning_rate": 4.078280413018843e-06, "loss": 0.4545, "step": 26058 }, { "epoch": 1.7489010435891412, "grad_norm": 0.9857343435287476, "learning_rate": 4.0739822856273925e-06, "loss": 0.429, "step": 26060 }, { "epoch": 1.7490352672729101, "grad_norm": 1.0646899938583374, "learning_rate": 4.069686328149969e-06, "loss": 0.4804, "step": 26062 }, { "epoch": 1.7491694909566793, "grad_norm": 1.0434508323669434, "learning_rate": 4.065392540789553e-06, "loss": 0.428, "step": 26064 }, { "epoch": 1.7493037146404484, "grad_norm": 1.1091803312301636, "learning_rate": 4.061100923749001e-06, "loss": 0.4761, "step": 26066 }, { "epoch": 1.7494379383242173, "grad_norm": 1.0577932596206665, "learning_rate": 4.0568114772311035e-06, "loss": 0.4923, "step": 26068 }, { "epoch": 1.7495721620079863, "grad_norm": 1.0199114084243774, "learning_rate": 4.052524201438523e-06, "loss": 0.4548, "step": 26070 }, { "epoch": 1.7497063856917552, "grad_norm": 1.0450282096862793, "learning_rate": 4.048239096573819e-06, "loss": 0.5336, "step": 26072 }, { "epoch": 1.7498406093755243, "grad_norm": 0.8555244207382202, "learning_rate": 4.043956162839452e-06, "loss": 0.4252, "step": 26074 }, { "epoch": 1.7499748330592935, "grad_norm": 1.0349127054214478, "learning_rate": 4.0396754004377925e-06, "loss": 0.4433, "step": 26076 }, { "epoch": 1.7501090567430624, "grad_norm": 1.0625731945037842, "learning_rate": 4.035396809571096e-06, "loss": 0.4539, "step": 26078 }, { "epoch": 1.7502432804268313, "grad_norm": 1.057253360748291, "learning_rate": 4.0311203904415116e-06, "loss": 0.4649, "step": 26080 }, { "epoch": 1.7503775041106002, "grad_norm": 1.0560394525527954, "learning_rate": 4.02684614325109e-06, "loss": 0.4382, "step": 26082 }, { "epoch": 1.7505117277943694, "grad_norm": 1.0100197792053223, "learning_rate": 4.02257406820179e-06, "loss": 0.4892, "step": 26084 }, { "epoch": 1.7506459514781383, "grad_norm": 1.061964988708496, "learning_rate": 4.0183041654954515e-06, "loss": 0.4295, "step": 26086 }, { "epoch": 1.7507801751619074, "grad_norm": 1.0337510108947754, "learning_rate": 4.0140364353338286e-06, "loss": 0.4432, "step": 26088 }, { "epoch": 1.7509143988456763, "grad_norm": 0.8317381739616394, "learning_rate": 4.009770877918556e-06, "loss": 0.4072, "step": 26090 }, { "epoch": 1.7510486225294453, "grad_norm": 1.0885119438171387, "learning_rate": 4.005507493451171e-06, "loss": 0.5211, "step": 26092 }, { "epoch": 1.7511828462132142, "grad_norm": 1.3589690923690796, "learning_rate": 4.0012462821331145e-06, "loss": 0.4578, "step": 26094 }, { "epoch": 1.7513170698969833, "grad_norm": 0.9653885364532471, "learning_rate": 3.996987244165712e-06, "loss": 0.4643, "step": 26096 }, { "epoch": 1.7514512935807525, "grad_norm": 2.2596049308776855, "learning_rate": 3.9927303797502096e-06, "loss": 0.4178, "step": 26098 }, { "epoch": 1.7515855172645214, "grad_norm": 1.04008150100708, "learning_rate": 3.988475689087723e-06, "loss": 0.4817, "step": 26100 }, { "epoch": 1.7517197409482903, "grad_norm": 1.0107393264770508, "learning_rate": 3.984223172379287e-06, "loss": 0.4844, "step": 26102 }, { "epoch": 1.7518539646320592, "grad_norm": 0.9756828546524048, "learning_rate": 3.979972829825807e-06, "loss": 0.4817, "step": 26104 }, { "epoch": 1.7519881883158284, "grad_norm": 1.2110542058944702, "learning_rate": 3.975724661628128e-06, "loss": 0.4734, "step": 26106 }, { "epoch": 1.7521224119995973, "grad_norm": 1.1260565519332886, "learning_rate": 3.971478667986955e-06, "loss": 0.4805, "step": 26108 }, { "epoch": 1.7522566356833664, "grad_norm": 1.2594767808914185, "learning_rate": 3.967234849102907e-06, "loss": 0.5398, "step": 26110 }, { "epoch": 1.7523908593671353, "grad_norm": 0.9751372337341309, "learning_rate": 3.962993205176479e-06, "loss": 0.4528, "step": 26112 }, { "epoch": 1.7525250830509043, "grad_norm": 1.1253643035888672, "learning_rate": 3.958753736408105e-06, "loss": 0.5152, "step": 26114 }, { "epoch": 1.7526593067346732, "grad_norm": 1.102760910987854, "learning_rate": 3.954516442998074e-06, "loss": 0.4306, "step": 26116 }, { "epoch": 1.7527935304184423, "grad_norm": 1.0021374225616455, "learning_rate": 3.9502813251466096e-06, "loss": 0.4496, "step": 26118 }, { "epoch": 1.7529277541022115, "grad_norm": 1.0082281827926636, "learning_rate": 3.946048383053786e-06, "loss": 0.5659, "step": 26120 }, { "epoch": 1.7530619777859804, "grad_norm": 0.8066741824150085, "learning_rate": 3.941817616919624e-06, "loss": 0.4263, "step": 26122 }, { "epoch": 1.7531962014697493, "grad_norm": 0.8860553503036499, "learning_rate": 3.937589026944e-06, "loss": 0.5009, "step": 26124 }, { "epoch": 1.7533304251535182, "grad_norm": 1.0312938690185547, "learning_rate": 3.9333626133267244e-06, "loss": 0.5285, "step": 26126 }, { "epoch": 1.7534646488372874, "grad_norm": 0.9705163836479187, "learning_rate": 3.929138376267477e-06, "loss": 0.5149, "step": 26128 }, { "epoch": 1.7535988725210565, "grad_norm": 0.9016320705413818, "learning_rate": 3.924916315965854e-06, "loss": 0.4818, "step": 26130 }, { "epoch": 1.7537330962048254, "grad_norm": 1.1750874519348145, "learning_rate": 3.920696432621318e-06, "loss": 0.4783, "step": 26132 }, { "epoch": 1.7538673198885943, "grad_norm": 1.048884630203247, "learning_rate": 3.916478726433276e-06, "loss": 0.5275, "step": 26134 }, { "epoch": 1.7540015435723633, "grad_norm": 1.061607003211975, "learning_rate": 3.91226319760099e-06, "loss": 0.4817, "step": 26136 }, { "epoch": 1.7541357672561322, "grad_norm": 0.9778168201446533, "learning_rate": 3.908049846323647e-06, "loss": 0.5421, "step": 26138 }, { "epoch": 1.7542699909399013, "grad_norm": 1.0087391138076782, "learning_rate": 3.903838672800297e-06, "loss": 0.4944, "step": 26140 }, { "epoch": 1.7544042146236705, "grad_norm": 1.061140775680542, "learning_rate": 3.899629677229943e-06, "loss": 0.5601, "step": 26142 }, { "epoch": 1.7545384383074394, "grad_norm": 1.113389015197754, "learning_rate": 3.89542285981142e-06, "loss": 0.4302, "step": 26144 }, { "epoch": 1.7546726619912083, "grad_norm": 1.1354550123214722, "learning_rate": 3.89121822074352e-06, "loss": 0.4703, "step": 26146 }, { "epoch": 1.7548068856749772, "grad_norm": 0.956482470035553, "learning_rate": 3.887015760224877e-06, "loss": 0.4745, "step": 26148 }, { "epoch": 1.7549411093587464, "grad_norm": 1.0160974264144897, "learning_rate": 3.8828154784540714e-06, "loss": 0.4272, "step": 26150 }, { "epoch": 1.7550753330425155, "grad_norm": 1.0287387371063232, "learning_rate": 3.87861737562954e-06, "loss": 0.5053, "step": 26152 }, { "epoch": 1.7552095567262844, "grad_norm": 1.0463069677352905, "learning_rate": 3.87442145194965e-06, "loss": 0.541, "step": 26154 }, { "epoch": 1.7553437804100533, "grad_norm": 0.9868055582046509, "learning_rate": 3.87022770761265e-06, "loss": 0.4387, "step": 26156 }, { "epoch": 1.7554780040938223, "grad_norm": 1.0721690654754639, "learning_rate": 3.866036142816676e-06, "loss": 0.4508, "step": 26158 }, { "epoch": 1.7556122277775914, "grad_norm": 0.9048008918762207, "learning_rate": 3.861846757759768e-06, "loss": 0.45, "step": 26160 }, { "epoch": 1.7557464514613603, "grad_norm": 1.0534013509750366, "learning_rate": 3.857659552639881e-06, "loss": 0.4851, "step": 26162 }, { "epoch": 1.7558806751451295, "grad_norm": 1.102059006690979, "learning_rate": 3.853474527654849e-06, "loss": 0.4907, "step": 26164 }, { "epoch": 1.7560148988288984, "grad_norm": 0.8822693824768066, "learning_rate": 3.849291683002398e-06, "loss": 0.4201, "step": 26166 }, { "epoch": 1.7561491225126673, "grad_norm": 0.9882935285568237, "learning_rate": 3.845111018880165e-06, "loss": 0.4663, "step": 26168 }, { "epoch": 1.7562833461964362, "grad_norm": 0.9759923219680786, "learning_rate": 3.84093253548567e-06, "loss": 0.4165, "step": 26170 }, { "epoch": 1.7564175698802054, "grad_norm": 0.9933663606643677, "learning_rate": 3.836756233016348e-06, "loss": 0.4877, "step": 26172 }, { "epoch": 1.7565517935639745, "grad_norm": 1.0421911478042603, "learning_rate": 3.832582111669525e-06, "loss": 0.4357, "step": 26174 }, { "epoch": 1.7566860172477434, "grad_norm": 0.9907096028327942, "learning_rate": 3.828410171642404e-06, "loss": 0.4631, "step": 26176 }, { "epoch": 1.7568202409315123, "grad_norm": 0.8957405090332031, "learning_rate": 3.824240413132107e-06, "loss": 0.4537, "step": 26178 }, { "epoch": 1.7569544646152813, "grad_norm": 1.045711874961853, "learning_rate": 3.820072836335658e-06, "loss": 0.4105, "step": 26180 }, { "epoch": 1.7570886882990504, "grad_norm": 1.179186224937439, "learning_rate": 3.815907441449951e-06, "loss": 0.4719, "step": 26182 }, { "epoch": 1.7572229119828193, "grad_norm": 1.2131788730621338, "learning_rate": 3.811744228671815e-06, "loss": 0.4861, "step": 26184 }, { "epoch": 1.7573571356665885, "grad_norm": 1.085301160812378, "learning_rate": 3.807583198197923e-06, "loss": 0.4765, "step": 26186 }, { "epoch": 1.7574913593503574, "grad_norm": 1.029398798942566, "learning_rate": 3.8034243502249045e-06, "loss": 0.4611, "step": 26188 }, { "epoch": 1.7576255830341263, "grad_norm": 1.011246919631958, "learning_rate": 3.799267684949226e-06, "loss": 0.5085, "step": 26190 }, { "epoch": 1.7577598067178952, "grad_norm": 1.0009100437164307, "learning_rate": 3.795113202567313e-06, "loss": 0.4058, "step": 26192 }, { "epoch": 1.7578940304016644, "grad_norm": 0.7475231289863586, "learning_rate": 3.7909609032754424e-06, "loss": 0.4094, "step": 26194 }, { "epoch": 1.7580282540854335, "grad_norm": 1.1544121503829956, "learning_rate": 3.786810787269807e-06, "loss": 0.4569, "step": 26196 }, { "epoch": 1.7581624777692024, "grad_norm": 0.9361860752105713, "learning_rate": 3.782662854746477e-06, "loss": 0.5096, "step": 26198 }, { "epoch": 1.7582967014529713, "grad_norm": 1.1230194568634033, "learning_rate": 3.7785171059014468e-06, "loss": 0.4443, "step": 26200 }, { "epoch": 1.7584309251367403, "grad_norm": 0.8865429162979126, "learning_rate": 3.7743735409305982e-06, "loss": 0.4462, "step": 26202 }, { "epoch": 1.7585651488205094, "grad_norm": 1.0899479389190674, "learning_rate": 3.7702321600297017e-06, "loss": 0.4709, "step": 26204 }, { "epoch": 1.7586993725042785, "grad_norm": 0.9332501292228699, "learning_rate": 3.7660929633944186e-06, "loss": 0.4255, "step": 26206 }, { "epoch": 1.7588335961880475, "grad_norm": 1.083202838897705, "learning_rate": 3.7619559512203363e-06, "loss": 0.4086, "step": 26208 }, { "epoch": 1.7589678198718164, "grad_norm": 1.1723979711532593, "learning_rate": 3.757821123702904e-06, "loss": 0.5048, "step": 26210 }, { "epoch": 1.7591020435555853, "grad_norm": 0.9791699647903442, "learning_rate": 3.7536884810375094e-06, "loss": 0.4624, "step": 26212 }, { "epoch": 1.7592362672393542, "grad_norm": 1.0547990798950195, "learning_rate": 3.7495580234193804e-06, "loss": 0.4836, "step": 26214 }, { "epoch": 1.7593704909231234, "grad_norm": 1.0521347522735596, "learning_rate": 3.7454297510436885e-06, "loss": 0.4835, "step": 26216 }, { "epoch": 1.7595047146068925, "grad_norm": 0.9637994170188904, "learning_rate": 3.7413036641054834e-06, "loss": 0.4173, "step": 26218 }, { "epoch": 1.7596389382906614, "grad_norm": 1.0964466333389282, "learning_rate": 3.7371797627997194e-06, "loss": 0.4228, "step": 26220 }, { "epoch": 1.7597731619744303, "grad_norm": 0.9575738906860352, "learning_rate": 3.7330580473212473e-06, "loss": 0.4499, "step": 26222 }, { "epoch": 1.7599073856581993, "grad_norm": 1.091765284538269, "learning_rate": 3.728938517864794e-06, "loss": 0.4908, "step": 26224 }, { "epoch": 1.7600416093419684, "grad_norm": 1.1609419584274292, "learning_rate": 3.724821174625004e-06, "loss": 0.4288, "step": 26226 }, { "epoch": 1.7601758330257375, "grad_norm": 0.9989089965820312, "learning_rate": 3.720706017796427e-06, "loss": 0.433, "step": 26228 }, { "epoch": 1.7603100567095065, "grad_norm": 1.0468043088912964, "learning_rate": 3.71659304757348e-06, "loss": 0.4575, "step": 26230 }, { "epoch": 1.7604442803932754, "grad_norm": 1.0158153772354126, "learning_rate": 3.7124822641505017e-06, "loss": 0.3912, "step": 26232 }, { "epoch": 1.7605785040770443, "grad_norm": 1.0709056854248047, "learning_rate": 3.7083736677217206e-06, "loss": 0.4414, "step": 26234 }, { "epoch": 1.7607127277608134, "grad_norm": 1.1032527685165405, "learning_rate": 3.704267258481242e-06, "loss": 0.4658, "step": 26236 }, { "epoch": 1.7608469514445824, "grad_norm": 0.9846126437187195, "learning_rate": 3.7001630366231112e-06, "loss": 0.419, "step": 26238 }, { "epoch": 1.7609811751283515, "grad_norm": 1.036513090133667, "learning_rate": 3.696061002341217e-06, "loss": 0.4931, "step": 26240 }, { "epoch": 1.7611153988121204, "grad_norm": 1.1536115407943726, "learning_rate": 3.6919611558294098e-06, "loss": 0.4925, "step": 26242 }, { "epoch": 1.7612496224958893, "grad_norm": 0.8856650590896606, "learning_rate": 3.6878634972813576e-06, "loss": 0.3997, "step": 26244 }, { "epoch": 1.7613838461796583, "grad_norm": 1.0974348783493042, "learning_rate": 3.6837680268906992e-06, "loss": 0.4774, "step": 26246 }, { "epoch": 1.7615180698634274, "grad_norm": 0.9839081764221191, "learning_rate": 3.679674744850914e-06, "loss": 0.4007, "step": 26248 }, { "epoch": 1.7616522935471965, "grad_norm": 1.161922574043274, "learning_rate": 3.6755836513554185e-06, "loss": 0.4659, "step": 26250 }, { "epoch": 1.7617865172309655, "grad_norm": 0.9850311279296875, "learning_rate": 3.6714947465975035e-06, "loss": 0.442, "step": 26252 }, { "epoch": 1.7619207409147344, "grad_norm": 1.0139830112457275, "learning_rate": 3.6674080307703586e-06, "loss": 0.4257, "step": 26254 }, { "epoch": 1.7620549645985033, "grad_norm": 1.1162172555923462, "learning_rate": 3.6633235040670675e-06, "loss": 0.4695, "step": 26256 }, { "epoch": 1.7621891882822724, "grad_norm": 1.002410888671875, "learning_rate": 3.6592411666806326e-06, "loss": 0.5363, "step": 26258 }, { "epoch": 1.7623234119660414, "grad_norm": 1.111937403678894, "learning_rate": 3.655161018803932e-06, "loss": 0.4447, "step": 26260 }, { "epoch": 1.7624576356498105, "grad_norm": 1.1555240154266357, "learning_rate": 3.6510830606297343e-06, "loss": 0.5071, "step": 26262 }, { "epoch": 1.7625918593335794, "grad_norm": 1.0110962390899658, "learning_rate": 3.6470072923507125e-06, "loss": 0.4434, "step": 26264 }, { "epoch": 1.7627260830173483, "grad_norm": 1.3436285257339478, "learning_rate": 3.6429337141594578e-06, "loss": 0.5013, "step": 26266 }, { "epoch": 1.7628603067011173, "grad_norm": 1.1013902425765991, "learning_rate": 3.638862326248421e-06, "loss": 0.4359, "step": 26268 }, { "epoch": 1.7629945303848864, "grad_norm": 1.0049915313720703, "learning_rate": 3.634793128809988e-06, "loss": 0.4744, "step": 26270 }, { "epoch": 1.7631287540686555, "grad_norm": 1.0206397771835327, "learning_rate": 3.6307261220363876e-06, "loss": 0.4962, "step": 26272 }, { "epoch": 1.7632629777524245, "grad_norm": 1.2429076433181763, "learning_rate": 3.626661306119805e-06, "loss": 0.4501, "step": 26274 }, { "epoch": 1.7633972014361934, "grad_norm": 1.0040055513381958, "learning_rate": 3.6225986812522816e-06, "loss": 0.4602, "step": 26276 }, { "epoch": 1.7635314251199623, "grad_norm": 0.9989480972290039, "learning_rate": 3.61853824762578e-06, "loss": 0.487, "step": 26278 }, { "epoch": 1.7636656488037314, "grad_norm": 0.9974848628044128, "learning_rate": 3.6144800054321415e-06, "loss": 0.4237, "step": 26280 }, { "epoch": 1.7637998724875004, "grad_norm": 1.0944793224334717, "learning_rate": 3.6104239548631127e-06, "loss": 0.4444, "step": 26282 }, { "epoch": 1.7639340961712695, "grad_norm": 1.0131146907806396, "learning_rate": 3.6063700961103174e-06, "loss": 0.5055, "step": 26284 }, { "epoch": 1.7640683198550384, "grad_norm": 1.0708969831466675, "learning_rate": 3.6023184293653143e-06, "loss": 0.4939, "step": 26286 }, { "epoch": 1.7642025435388073, "grad_norm": 0.9781100153923035, "learning_rate": 3.5982689548195338e-06, "loss": 0.432, "step": 26288 }, { "epoch": 1.7643367672225763, "grad_norm": 1.0512197017669678, "learning_rate": 3.594221672664294e-06, "loss": 0.4748, "step": 26290 }, { "epoch": 1.7644709909063454, "grad_norm": 1.1227929592132568, "learning_rate": 3.590176583090821e-06, "loss": 0.4332, "step": 26292 }, { "epoch": 1.7646052145901145, "grad_norm": 1.1276625394821167, "learning_rate": 3.586133686290255e-06, "loss": 0.4524, "step": 26294 }, { "epoch": 1.7647394382738835, "grad_norm": 1.1209701299667358, "learning_rate": 3.582092982453589e-06, "loss": 0.4821, "step": 26296 }, { "epoch": 1.7648736619576524, "grad_norm": 1.0472911596298218, "learning_rate": 3.5780544717717747e-06, "loss": 0.4664, "step": 26298 }, { "epoch": 1.7650078856414213, "grad_norm": 1.0919382572174072, "learning_rate": 3.5740181544355875e-06, "loss": 0.4555, "step": 26300 }, { "epoch": 1.7651421093251904, "grad_norm": 1.0900133848190308, "learning_rate": 3.5699840306357478e-06, "loss": 0.5195, "step": 26302 }, { "epoch": 1.7652763330089596, "grad_norm": 0.995741069316864, "learning_rate": 3.565952100562869e-06, "loss": 0.4058, "step": 26304 }, { "epoch": 1.7654105566927285, "grad_norm": 1.073737621307373, "learning_rate": 3.561922364407433e-06, "loss": 0.563, "step": 26306 }, { "epoch": 1.7655447803764974, "grad_norm": 1.0843747854232788, "learning_rate": 3.557894822359864e-06, "loss": 0.4034, "step": 26308 }, { "epoch": 1.7656790040602663, "grad_norm": 1.110669732093811, "learning_rate": 3.5538694746104274e-06, "loss": 0.4691, "step": 26310 }, { "epoch": 1.7658132277440355, "grad_norm": 0.9983861446380615, "learning_rate": 3.5498463213493372e-06, "loss": 0.459, "step": 26312 }, { "epoch": 1.7659474514278044, "grad_norm": 0.9595610499382019, "learning_rate": 3.545825362766653e-06, "loss": 0.4353, "step": 26314 }, { "epoch": 1.7660816751115735, "grad_norm": 1.0454193353652954, "learning_rate": 3.541806599052383e-06, "loss": 0.424, "step": 26316 }, { "epoch": 1.7662158987953425, "grad_norm": 0.9348208904266357, "learning_rate": 3.5377900303963986e-06, "loss": 0.5193, "step": 26318 }, { "epoch": 1.7663501224791114, "grad_norm": 1.0392781496047974, "learning_rate": 3.533775656988464e-06, "loss": 0.4424, "step": 26320 }, { "epoch": 1.7664843461628803, "grad_norm": 1.0983844995498657, "learning_rate": 3.5297634790182555e-06, "loss": 0.4687, "step": 26322 }, { "epoch": 1.7666185698466494, "grad_norm": 1.1866750717163086, "learning_rate": 3.525753496675349e-06, "loss": 0.5033, "step": 26324 }, { "epoch": 1.7667527935304186, "grad_norm": 1.07992422580719, "learning_rate": 3.5217457101492045e-06, "loss": 0.4064, "step": 26326 }, { "epoch": 1.7668870172141875, "grad_norm": 1.0780870914459229, "learning_rate": 3.517740119629176e-06, "loss": 0.4615, "step": 26328 }, { "epoch": 1.7670212408979564, "grad_norm": 1.0299321413040161, "learning_rate": 3.5137367253045173e-06, "loss": 0.4651, "step": 26330 }, { "epoch": 1.7671554645817253, "grad_norm": 0.8918436169624329, "learning_rate": 3.509735527364394e-06, "loss": 0.4433, "step": 26332 }, { "epoch": 1.7672896882654945, "grad_norm": 1.1658223867416382, "learning_rate": 3.5057365259978383e-06, "loss": 0.5088, "step": 26334 }, { "epoch": 1.7674239119492634, "grad_norm": 1.009101390838623, "learning_rate": 3.501739721393826e-06, "loss": 0.4667, "step": 26336 }, { "epoch": 1.7675581356330325, "grad_norm": 1.0059616565704346, "learning_rate": 3.4977451137411577e-06, "loss": 0.4776, "step": 26338 }, { "epoch": 1.7676923593168015, "grad_norm": 0.931182324886322, "learning_rate": 3.4937527032285976e-06, "loss": 0.3969, "step": 26340 }, { "epoch": 1.7678265830005704, "grad_norm": 0.9413557052612305, "learning_rate": 3.4897624900447624e-06, "loss": 0.4209, "step": 26342 }, { "epoch": 1.7679608066843393, "grad_norm": 0.9050574898719788, "learning_rate": 3.4857744743782007e-06, "loss": 0.4548, "step": 26344 }, { "epoch": 1.7680950303681084, "grad_norm": 0.9665685892105103, "learning_rate": 3.4817886564173287e-06, "loss": 0.4743, "step": 26346 }, { "epoch": 1.7682292540518776, "grad_norm": 1.056808590888977, "learning_rate": 3.4778050363504675e-06, "loss": 0.4282, "step": 26348 }, { "epoch": 1.7683634777356465, "grad_norm": 1.005738615989685, "learning_rate": 3.473823614365829e-06, "loss": 0.4223, "step": 26350 }, { "epoch": 1.7684977014194154, "grad_norm": 1.014094352722168, "learning_rate": 3.469844390651544e-06, "loss": 0.4461, "step": 26352 }, { "epoch": 1.7686319251031843, "grad_norm": 1.0013539791107178, "learning_rate": 3.465867365395614e-06, "loss": 0.4602, "step": 26354 }, { "epoch": 1.7687661487869535, "grad_norm": 0.9541317224502563, "learning_rate": 3.4618925387859437e-06, "loss": 0.478, "step": 26356 }, { "epoch": 1.7689003724707224, "grad_norm": 1.037687063217163, "learning_rate": 3.457919911010332e-06, "loss": 0.486, "step": 26358 }, { "epoch": 1.7690345961544915, "grad_norm": 0.9743841290473938, "learning_rate": 3.453949482256491e-06, "loss": 0.4508, "step": 26360 }, { "epoch": 1.7691688198382605, "grad_norm": 1.5829144716262817, "learning_rate": 3.449981252711998e-06, "loss": 0.4721, "step": 26362 }, { "epoch": 1.7693030435220294, "grad_norm": 0.9538119435310364, "learning_rate": 3.4460152225643692e-06, "loss": 0.472, "step": 26364 }, { "epoch": 1.7694372672057983, "grad_norm": 0.9716622829437256, "learning_rate": 3.442051392000967e-06, "loss": 0.4669, "step": 26366 }, { "epoch": 1.7695714908895674, "grad_norm": 0.9730298519134521, "learning_rate": 3.4380897612090843e-06, "loss": 0.4532, "step": 26368 }, { "epoch": 1.7697057145733366, "grad_norm": 1.224579930305481, "learning_rate": 3.4341303303759064e-06, "loss": 0.4159, "step": 26370 }, { "epoch": 1.7698399382571055, "grad_norm": 1.0344945192337036, "learning_rate": 3.4301730996884884e-06, "loss": 0.5461, "step": 26372 }, { "epoch": 1.7699741619408744, "grad_norm": 1.0257664918899536, "learning_rate": 3.4262180693338306e-06, "loss": 0.5275, "step": 26374 }, { "epoch": 1.7701083856246433, "grad_norm": 0.9448689818382263, "learning_rate": 3.4222652394987843e-06, "loss": 0.3993, "step": 26376 }, { "epoch": 1.7702426093084125, "grad_norm": 1.0021743774414062, "learning_rate": 3.4183146103701104e-06, "loss": 0.4299, "step": 26378 }, { "epoch": 1.7703768329921816, "grad_norm": 1.0137749910354614, "learning_rate": 3.4143661821344654e-06, "loss": 0.4464, "step": 26380 }, { "epoch": 1.7705110566759505, "grad_norm": 1.0789955854415894, "learning_rate": 3.4104199549784223e-06, "loss": 0.4875, "step": 26382 }, { "epoch": 1.7706452803597195, "grad_norm": 0.9993133544921875, "learning_rate": 3.4064759290884207e-06, "loss": 0.4478, "step": 26384 }, { "epoch": 1.7707795040434884, "grad_norm": 1.0137730836868286, "learning_rate": 3.4025341046508064e-06, "loss": 0.5015, "step": 26386 }, { "epoch": 1.7709137277272575, "grad_norm": 1.077858328819275, "learning_rate": 3.398594481851819e-06, "loss": 0.4951, "step": 26388 }, { "epoch": 1.7710479514110264, "grad_norm": 0.948650062084198, "learning_rate": 3.394657060877615e-06, "loss": 0.4374, "step": 26390 }, { "epoch": 1.7711821750947956, "grad_norm": 1.0998425483703613, "learning_rate": 3.3907218419142182e-06, "loss": 0.4774, "step": 26392 }, { "epoch": 1.7713163987785645, "grad_norm": 1.0868417024612427, "learning_rate": 3.3867888251475577e-06, "loss": 0.4282, "step": 26394 }, { "epoch": 1.7714506224623334, "grad_norm": 0.9734367728233337, "learning_rate": 3.3828580107634623e-06, "loss": 0.4114, "step": 26396 }, { "epoch": 1.7715848461461023, "grad_norm": 1.0528538227081299, "learning_rate": 3.378929398947661e-06, "loss": 0.4872, "step": 26398 }, { "epoch": 1.7717190698298715, "grad_norm": 1.1084693670272827, "learning_rate": 3.3750029898857614e-06, "loss": 0.4852, "step": 26400 }, { "epoch": 1.7718532935136406, "grad_norm": 0.9850924015045166, "learning_rate": 3.3710787837632982e-06, "loss": 0.4211, "step": 26402 }, { "epoch": 1.7719875171974095, "grad_norm": 0.9967346787452698, "learning_rate": 3.367156780765668e-06, "loss": 0.4618, "step": 26404 }, { "epoch": 1.7721217408811785, "grad_norm": 1.0487021207809448, "learning_rate": 3.3632369810781774e-06, "loss": 0.4789, "step": 26406 }, { "epoch": 1.7722559645649474, "grad_norm": 0.9948101043701172, "learning_rate": 3.3593193848860284e-06, "loss": 0.4703, "step": 26408 }, { "epoch": 1.7723901882487165, "grad_norm": 1.0135157108306885, "learning_rate": 3.3554039923743286e-06, "loss": 0.4501, "step": 26410 }, { "epoch": 1.7725244119324854, "grad_norm": 0.9719632863998413, "learning_rate": 3.351490803728069e-06, "loss": 0.4677, "step": 26412 }, { "epoch": 1.7726586356162546, "grad_norm": 1.0641772747039795, "learning_rate": 3.3475798191321406e-06, "loss": 0.5544, "step": 26414 }, { "epoch": 1.7727928593000235, "grad_norm": 0.8491989970207214, "learning_rate": 3.3436710387713176e-06, "loss": 0.3951, "step": 26416 }, { "epoch": 1.7729270829837924, "grad_norm": 1.0544626712799072, "learning_rate": 3.3397644628303026e-06, "loss": 0.5106, "step": 26418 }, { "epoch": 1.7730613066675613, "grad_norm": 1.0775730609893799, "learning_rate": 3.335860091493653e-06, "loss": 0.4315, "step": 26420 }, { "epoch": 1.7731955303513305, "grad_norm": 1.0925928354263306, "learning_rate": 3.3319579249458667e-06, "loss": 0.435, "step": 26422 }, { "epoch": 1.7733297540350996, "grad_norm": 1.0120404958724976, "learning_rate": 3.328057963371284e-06, "loss": 0.4855, "step": 26424 }, { "epoch": 1.7734639777188685, "grad_norm": 1.024065375328064, "learning_rate": 3.3241602069541967e-06, "loss": 0.4473, "step": 26426 }, { "epoch": 1.7735982014026375, "grad_norm": 1.0965101718902588, "learning_rate": 3.3202646558787463e-06, "loss": 0.5376, "step": 26428 }, { "epoch": 1.7737324250864064, "grad_norm": 1.0976194143295288, "learning_rate": 3.3163713103290084e-06, "loss": 0.4475, "step": 26430 }, { "epoch": 1.7738666487701755, "grad_norm": 0.9388175010681152, "learning_rate": 3.3124801704889298e-06, "loss": 0.5204, "step": 26432 }, { "epoch": 1.7740008724539444, "grad_norm": 0.9413573741912842, "learning_rate": 3.308591236542352e-06, "loss": 0.4106, "step": 26434 }, { "epoch": 1.7741350961377136, "grad_norm": 0.9693027138710022, "learning_rate": 3.3047045086730233e-06, "loss": 0.4768, "step": 26436 }, { "epoch": 1.7742693198214825, "grad_norm": 1.11866295337677, "learning_rate": 3.300819987064574e-06, "loss": 0.4673, "step": 26438 }, { "epoch": 1.7744035435052514, "grad_norm": 1.123133659362793, "learning_rate": 3.296937671900563e-06, "loss": 0.5068, "step": 26440 }, { "epoch": 1.7745377671890203, "grad_norm": 0.9647020101547241, "learning_rate": 3.2930575633644103e-06, "loss": 0.4149, "step": 26442 }, { "epoch": 1.7746719908727895, "grad_norm": 0.9570835828781128, "learning_rate": 3.2891796616394353e-06, "loss": 0.4342, "step": 26444 }, { "epoch": 1.7748062145565586, "grad_norm": 0.906319797039032, "learning_rate": 3.285303966908865e-06, "loss": 0.4271, "step": 26446 }, { "epoch": 1.7749404382403275, "grad_norm": 0.9696779251098633, "learning_rate": 3.2814304793558294e-06, "loss": 0.4509, "step": 26448 }, { "epoch": 1.7750746619240965, "grad_norm": 1.0504570007324219, "learning_rate": 3.277559199163338e-06, "loss": 0.4424, "step": 26450 }, { "epoch": 1.7752088856078654, "grad_norm": 1.1588096618652344, "learning_rate": 3.2736901265142948e-06, "loss": 0.4405, "step": 26452 }, { "epoch": 1.7753431092916345, "grad_norm": 1.0212655067443848, "learning_rate": 3.2698232615915034e-06, "loss": 0.474, "step": 26454 }, { "epoch": 1.7754773329754037, "grad_norm": 1.0489578247070312, "learning_rate": 3.265958604577679e-06, "loss": 0.4647, "step": 26456 }, { "epoch": 1.7756115566591726, "grad_norm": 0.8789849281311035, "learning_rate": 3.262096155655403e-06, "loss": 0.4147, "step": 26458 }, { "epoch": 1.7757457803429415, "grad_norm": 1.1079792976379395, "learning_rate": 3.2582359150071963e-06, "loss": 0.4621, "step": 26460 }, { "epoch": 1.7758800040267104, "grad_norm": 1.1104118824005127, "learning_rate": 3.2543778828154125e-06, "loss": 0.4943, "step": 26462 }, { "epoch": 1.7760142277104796, "grad_norm": 0.8925808668136597, "learning_rate": 3.250522059262362e-06, "loss": 0.3928, "step": 26464 }, { "epoch": 1.7761484513942485, "grad_norm": 1.026313304901123, "learning_rate": 3.246668444530204e-06, "loss": 0.4922, "step": 26466 }, { "epoch": 1.7762826750780176, "grad_norm": 1.1014049053192139, "learning_rate": 3.2428170388010326e-06, "loss": 0.5189, "step": 26468 }, { "epoch": 1.7764168987617865, "grad_norm": 1.0588924884796143, "learning_rate": 3.238967842256818e-06, "loss": 0.4708, "step": 26470 }, { "epoch": 1.7765511224455555, "grad_norm": 1.07292902469635, "learning_rate": 3.2351208550794154e-06, "loss": 0.4514, "step": 26472 }, { "epoch": 1.7766853461293244, "grad_norm": 0.9535008072853088, "learning_rate": 3.2312760774505845e-06, "loss": 0.3932, "step": 26474 }, { "epoch": 1.7768195698130935, "grad_norm": 1.1209310293197632, "learning_rate": 3.2274335095520026e-06, "loss": 0.4718, "step": 26476 }, { "epoch": 1.7769537934968627, "grad_norm": 0.9832744002342224, "learning_rate": 3.223593151565213e-06, "loss": 0.437, "step": 26478 }, { "epoch": 1.7770880171806316, "grad_norm": 1.0584068298339844, "learning_rate": 3.219755003671665e-06, "loss": 0.4624, "step": 26480 }, { "epoch": 1.7772222408644005, "grad_norm": 1.0272608995437622, "learning_rate": 3.2159190660526972e-06, "loss": 0.4485, "step": 26482 }, { "epoch": 1.7773564645481694, "grad_norm": 1.1383942365646362, "learning_rate": 3.2120853388895643e-06, "loss": 0.4513, "step": 26484 }, { "epoch": 1.7774906882319386, "grad_norm": 1.079269289970398, "learning_rate": 3.2082538223633884e-06, "loss": 0.4312, "step": 26486 }, { "epoch": 1.7776249119157075, "grad_norm": 1.074112892150879, "learning_rate": 3.204424516655219e-06, "loss": 0.4639, "step": 26488 }, { "epoch": 1.7777591355994766, "grad_norm": 1.0962426662445068, "learning_rate": 3.2005974219459556e-06, "loss": 0.4773, "step": 26490 }, { "epoch": 1.7778933592832455, "grad_norm": 0.909841775894165, "learning_rate": 3.196772538416448e-06, "loss": 0.4395, "step": 26492 }, { "epoch": 1.7780275829670145, "grad_norm": 1.0619467496871948, "learning_rate": 3.1929498662473965e-06, "loss": 0.5161, "step": 26494 }, { "epoch": 1.7781618066507834, "grad_norm": 1.096822738647461, "learning_rate": 3.1891294056194233e-06, "loss": 0.4501, "step": 26496 }, { "epoch": 1.7782960303345525, "grad_norm": 1.1792314052581787, "learning_rate": 3.1853111567130387e-06, "loss": 0.477, "step": 26498 }, { "epoch": 1.7784302540183217, "grad_norm": 0.9513255953788757, "learning_rate": 3.1814951197086495e-06, "loss": 0.4912, "step": 26500 }, { "epoch": 1.7785644777020906, "grad_norm": 0.9621514081954956, "learning_rate": 3.1776812947865385e-06, "loss": 0.4396, "step": 26502 }, { "epoch": 1.7786987013858595, "grad_norm": 1.0372976064682007, "learning_rate": 3.173869682126923e-06, "loss": 0.5395, "step": 26504 }, { "epoch": 1.7788329250696284, "grad_norm": 0.98854660987854, "learning_rate": 3.1700602819098868e-06, "loss": 0.4634, "step": 26506 }, { "epoch": 1.7789671487533976, "grad_norm": 1.076910376548767, "learning_rate": 3.1662530943154134e-06, "loss": 0.5016, "step": 26508 }, { "epoch": 1.7791013724371665, "grad_norm": 1.0470190048217773, "learning_rate": 3.162448119523387e-06, "loss": 0.486, "step": 26510 }, { "epoch": 1.7792355961209356, "grad_norm": 1.1797438859939575, "learning_rate": 3.1586453577135798e-06, "loss": 0.5459, "step": 26512 }, { "epoch": 1.7793698198047045, "grad_norm": 1.1045492887496948, "learning_rate": 3.154844809065677e-06, "loss": 0.491, "step": 26514 }, { "epoch": 1.7795040434884735, "grad_norm": 0.9104241728782654, "learning_rate": 3.1510464737592336e-06, "loss": 0.4016, "step": 26516 }, { "epoch": 1.7796382671722424, "grad_norm": 1.12786865234375, "learning_rate": 3.1472503519737295e-06, "loss": 0.4704, "step": 26518 }, { "epoch": 1.7797724908560115, "grad_norm": 1.0569543838500977, "learning_rate": 3.1434564438884983e-06, "loss": 0.4124, "step": 26520 }, { "epoch": 1.7799067145397807, "grad_norm": 1.2627952098846436, "learning_rate": 3.1396647496828247e-06, "loss": 0.4108, "step": 26522 }, { "epoch": 1.7800409382235496, "grad_norm": 0.9939997792243958, "learning_rate": 3.1358752695358316e-06, "loss": 0.4242, "step": 26524 }, { "epoch": 1.7801751619073185, "grad_norm": 0.9446654915809631, "learning_rate": 3.132088003626588e-06, "loss": 0.4203, "step": 26526 }, { "epoch": 1.7803093855910874, "grad_norm": 1.1433025598526, "learning_rate": 3.1283029521340212e-06, "loss": 0.4893, "step": 26528 }, { "epoch": 1.7804436092748566, "grad_norm": 0.9577170014381409, "learning_rate": 3.1245201152369783e-06, "loss": 0.515, "step": 26530 }, { "epoch": 1.7805778329586257, "grad_norm": 0.9782341122627258, "learning_rate": 3.1207394931141665e-06, "loss": 0.4448, "step": 26532 }, { "epoch": 1.7807120566423946, "grad_norm": 0.9431700706481934, "learning_rate": 3.116961085944242e-06, "loss": 0.4242, "step": 26534 }, { "epoch": 1.7808462803261635, "grad_norm": 1.0530515909194946, "learning_rate": 3.1131848939057186e-06, "loss": 0.4992, "step": 26536 }, { "epoch": 1.7809805040099325, "grad_norm": 1.1985836029052734, "learning_rate": 3.1094109171770027e-06, "loss": 0.5346, "step": 26538 }, { "epoch": 1.7811147276937016, "grad_norm": 0.9395465850830078, "learning_rate": 3.105639155936413e-06, "loss": 0.4576, "step": 26540 }, { "epoch": 1.7812489513774705, "grad_norm": 1.0075362920761108, "learning_rate": 3.1018696103621625e-06, "loss": 0.4682, "step": 26542 }, { "epoch": 1.7813831750612397, "grad_norm": 0.9399588108062744, "learning_rate": 3.0981022806323536e-06, "loss": 0.4313, "step": 26544 }, { "epoch": 1.7815173987450086, "grad_norm": 1.0964415073394775, "learning_rate": 3.094337166924982e-06, "loss": 0.4771, "step": 26546 }, { "epoch": 1.7816516224287775, "grad_norm": 0.9933387041091919, "learning_rate": 3.090574269417934e-06, "loss": 0.427, "step": 26548 }, { "epoch": 1.7817858461125464, "grad_norm": 1.0570873022079468, "learning_rate": 3.0868135882890224e-06, "loss": 0.492, "step": 26550 }, { "epoch": 1.7819200697963156, "grad_norm": 0.946513295173645, "learning_rate": 3.0830551237159056e-06, "loss": 0.4297, "step": 26552 }, { "epoch": 1.7820542934800847, "grad_norm": 1.0778554677963257, "learning_rate": 3.0792988758761854e-06, "loss": 0.5005, "step": 26554 }, { "epoch": 1.7821885171638536, "grad_norm": 0.9276990294456482, "learning_rate": 3.0755448449473255e-06, "loss": 0.4806, "step": 26556 }, { "epoch": 1.7823227408476225, "grad_norm": 1.0314631462097168, "learning_rate": 3.071793031106701e-06, "loss": 0.4806, "step": 26558 }, { "epoch": 1.7824569645313915, "grad_norm": 0.9608511924743652, "learning_rate": 3.0680434345315643e-06, "loss": 0.4765, "step": 26560 }, { "epoch": 1.7825911882151606, "grad_norm": 1.042948842048645, "learning_rate": 3.0642960553990963e-06, "loss": 0.4899, "step": 26562 }, { "epoch": 1.7827254118989295, "grad_norm": 0.9586542248725891, "learning_rate": 3.0605508938863493e-06, "loss": 0.4453, "step": 26564 }, { "epoch": 1.7828596355826987, "grad_norm": 1.0348963737487793, "learning_rate": 3.056807950170265e-06, "loss": 0.484, "step": 26566 }, { "epoch": 1.7829938592664676, "grad_norm": 1.127644658088684, "learning_rate": 3.0530672244276912e-06, "loss": 0.467, "step": 26568 }, { "epoch": 1.7831280829502365, "grad_norm": 1.0233221054077148, "learning_rate": 3.049328716835376e-06, "loss": 0.5077, "step": 26570 }, { "epoch": 1.7832623066340054, "grad_norm": 1.0783993005752563, "learning_rate": 3.04559242756996e-06, "loss": 0.4298, "step": 26572 }, { "epoch": 1.7833965303177746, "grad_norm": 1.0966519117355347, "learning_rate": 3.0418583568079697e-06, "loss": 0.4818, "step": 26574 }, { "epoch": 1.7835307540015437, "grad_norm": 1.0691412687301636, "learning_rate": 3.03812650472583e-06, "loss": 0.5221, "step": 26576 }, { "epoch": 1.7836649776853126, "grad_norm": 1.1239451169967651, "learning_rate": 3.034396871499856e-06, "loss": 0.5107, "step": 26578 }, { "epoch": 1.7837992013690815, "grad_norm": 1.1886765956878662, "learning_rate": 3.0306694573062898e-06, "loss": 0.4383, "step": 26580 }, { "epoch": 1.7839334250528505, "grad_norm": 0.9689651131629944, "learning_rate": 3.0269442623212186e-06, "loss": 0.5116, "step": 26582 }, { "epoch": 1.7840676487366196, "grad_norm": 1.0736305713653564, "learning_rate": 3.023221286720679e-06, "loss": 0.4788, "step": 26584 }, { "epoch": 1.7842018724203885, "grad_norm": 1.016939401626587, "learning_rate": 3.019500530680547e-06, "loss": 0.462, "step": 26586 }, { "epoch": 1.7843360961041577, "grad_norm": 1.0670020580291748, "learning_rate": 3.015781994376632e-06, "loss": 0.4268, "step": 26588 }, { "epoch": 1.7844703197879266, "grad_norm": 0.9900224804878235, "learning_rate": 3.0120656779846214e-06, "loss": 0.5255, "step": 26590 }, { "epoch": 1.7846045434716955, "grad_norm": 0.9999427199363708, "learning_rate": 3.008351581680119e-06, "loss": 0.481, "step": 26592 }, { "epoch": 1.7847387671554644, "grad_norm": 1.0749753713607788, "learning_rate": 3.004639705638601e-06, "loss": 0.4785, "step": 26594 }, { "epoch": 1.7848729908392336, "grad_norm": 0.9687768816947937, "learning_rate": 3.0009300500354444e-06, "loss": 0.4166, "step": 26596 }, { "epoch": 1.7850072145230027, "grad_norm": 0.9108654856681824, "learning_rate": 2.9972226150459137e-06, "loss": 0.4143, "step": 26598 }, { "epoch": 1.7851414382067716, "grad_norm": 0.9863176941871643, "learning_rate": 2.9935174008451917e-06, "loss": 0.4059, "step": 26600 }, { "epoch": 1.7852756618905405, "grad_norm": 0.9309257864952087, "learning_rate": 2.9898144076083433e-06, "loss": 0.4821, "step": 26602 }, { "epoch": 1.7854098855743095, "grad_norm": 1.1409417390823364, "learning_rate": 2.9861136355103235e-06, "loss": 0.4612, "step": 26604 }, { "epoch": 1.7855441092580786, "grad_norm": 0.9512568116188049, "learning_rate": 2.982415084725976e-06, "loss": 0.4737, "step": 26606 }, { "epoch": 1.7856783329418477, "grad_norm": 0.8105921149253845, "learning_rate": 2.9787187554300656e-06, "loss": 0.4305, "step": 26608 }, { "epoch": 1.7858125566256167, "grad_norm": 1.0694352388381958, "learning_rate": 2.97502464779722e-06, "loss": 0.483, "step": 26610 }, { "epoch": 1.7859467803093856, "grad_norm": 1.0770924091339111, "learning_rate": 2.9713327620020103e-06, "loss": 0.5138, "step": 26612 }, { "epoch": 1.7860810039931545, "grad_norm": 1.0274219512939453, "learning_rate": 2.967643098218831e-06, "loss": 0.5058, "step": 26614 }, { "epoch": 1.7862152276769236, "grad_norm": 1.0504615306854248, "learning_rate": 2.9639556566220415e-06, "loss": 0.4684, "step": 26616 }, { "epoch": 1.7863494513606926, "grad_norm": 0.9464894533157349, "learning_rate": 2.9602704373858426e-06, "loss": 0.4762, "step": 26618 }, { "epoch": 1.7864836750444617, "grad_norm": 1.0615830421447754, "learning_rate": 2.9565874406843776e-06, "loss": 0.5122, "step": 26620 }, { "epoch": 1.7866178987282306, "grad_norm": 1.1229662895202637, "learning_rate": 2.952906666691646e-06, "loss": 0.4577, "step": 26622 }, { "epoch": 1.7867521224119995, "grad_norm": 0.9584335088729858, "learning_rate": 2.949228115581565e-06, "loss": 0.4124, "step": 26624 }, { "epoch": 1.7868863460957685, "grad_norm": 1.3853611946105957, "learning_rate": 2.9455517875279225e-06, "loss": 0.4759, "step": 26626 }, { "epoch": 1.7870205697795376, "grad_norm": 1.0542583465576172, "learning_rate": 2.9418776827044357e-06, "loss": 0.503, "step": 26628 }, { "epoch": 1.7871547934633067, "grad_norm": 1.0837205648422241, "learning_rate": 2.9382058012846936e-06, "loss": 0.471, "step": 26630 }, { "epoch": 1.7872890171470757, "grad_norm": 1.0488024950027466, "learning_rate": 2.93453614344219e-06, "loss": 0.4463, "step": 26632 }, { "epoch": 1.7874232408308446, "grad_norm": 1.056859016418457, "learning_rate": 2.930868709350287e-06, "loss": 0.4371, "step": 26634 }, { "epoch": 1.7875574645146135, "grad_norm": 1.1180368661880493, "learning_rate": 2.927203499182296e-06, "loss": 0.5127, "step": 26636 }, { "epoch": 1.7876916881983826, "grad_norm": 1.02019464969635, "learning_rate": 2.9235405131113615e-06, "loss": 0.4248, "step": 26638 }, { "epoch": 1.7878259118821516, "grad_norm": 0.9511210918426514, "learning_rate": 2.9198797513105834e-06, "loss": 0.4121, "step": 26640 }, { "epoch": 1.7879601355659207, "grad_norm": 0.9936085939407349, "learning_rate": 2.9162212139528967e-06, "loss": 0.4594, "step": 26642 }, { "epoch": 1.7880943592496896, "grad_norm": 0.9715378880500793, "learning_rate": 2.9125649012111678e-06, "loss": 0.4199, "step": 26644 }, { "epoch": 1.7882285829334585, "grad_norm": 1.149639368057251, "learning_rate": 2.9089108132581587e-06, "loss": 0.4676, "step": 26646 }, { "epoch": 1.7883628066172275, "grad_norm": 0.9170129895210266, "learning_rate": 2.9052589502665095e-06, "loss": 0.4516, "step": 26648 }, { "epoch": 1.7884970303009966, "grad_norm": 0.9058769345283508, "learning_rate": 2.9016093124087817e-06, "loss": 0.4454, "step": 26650 }, { "epoch": 1.7886312539847657, "grad_norm": 1.0133819580078125, "learning_rate": 2.897961899857382e-06, "loss": 0.4251, "step": 26652 }, { "epoch": 1.7887654776685347, "grad_norm": 1.010780692100525, "learning_rate": 2.894316712784667e-06, "loss": 0.4627, "step": 26654 }, { "epoch": 1.7888997013523036, "grad_norm": 1.0257115364074707, "learning_rate": 2.890673751362849e-06, "loss": 0.4628, "step": 26656 }, { "epoch": 1.7890339250360725, "grad_norm": 1.0811824798583984, "learning_rate": 2.887033015764068e-06, "loss": 0.4746, "step": 26658 }, { "epoch": 1.7891681487198416, "grad_norm": 1.054804801940918, "learning_rate": 2.883394506160336e-06, "loss": 0.445, "step": 26660 }, { "epoch": 1.7893023724036106, "grad_norm": 1.0269854068756104, "learning_rate": 2.8797582227235608e-06, "loss": 0.4701, "step": 26662 }, { "epoch": 1.7894365960873797, "grad_norm": 1.1081819534301758, "learning_rate": 2.876124165625549e-06, "loss": 0.4196, "step": 26664 }, { "epoch": 1.7895708197711486, "grad_norm": 0.932039737701416, "learning_rate": 2.8724923350380075e-06, "loss": 0.4012, "step": 26666 }, { "epoch": 1.7897050434549175, "grad_norm": 1.0136710405349731, "learning_rate": 2.868862731132532e-06, "loss": 0.4776, "step": 26668 }, { "epoch": 1.7898392671386865, "grad_norm": 0.9252318739891052, "learning_rate": 2.8652353540806086e-06, "loss": 0.452, "step": 26670 }, { "epoch": 1.7899734908224556, "grad_norm": 1.0677413940429688, "learning_rate": 2.8616102040536274e-06, "loss": 0.4439, "step": 26672 }, { "epoch": 1.7901077145062247, "grad_norm": 1.121899962425232, "learning_rate": 2.8579872812228738e-06, "loss": 0.4599, "step": 26674 }, { "epoch": 1.7902419381899937, "grad_norm": 1.0476349592208862, "learning_rate": 2.8543665857595158e-06, "loss": 0.5122, "step": 26676 }, { "epoch": 1.7903761618737626, "grad_norm": 1.0133148431777954, "learning_rate": 2.850748117834645e-06, "loss": 0.5094, "step": 26678 }, { "epoch": 1.7905103855575315, "grad_norm": 1.0524612665176392, "learning_rate": 2.8471318776191914e-06, "loss": 0.4234, "step": 26680 }, { "epoch": 1.7906446092413006, "grad_norm": 1.100936770439148, "learning_rate": 2.8435178652840456e-06, "loss": 0.4973, "step": 26682 }, { "epoch": 1.7907788329250698, "grad_norm": 0.9793028235435486, "learning_rate": 2.839906080999938e-06, "loss": 0.4941, "step": 26684 }, { "epoch": 1.7909130566088387, "grad_norm": 1.0425536632537842, "learning_rate": 2.8362965249375485e-06, "loss": 0.4424, "step": 26686 }, { "epoch": 1.7910472802926076, "grad_norm": 1.0285500288009644, "learning_rate": 2.8326891972673965e-06, "loss": 0.4544, "step": 26688 }, { "epoch": 1.7911815039763765, "grad_norm": 1.0473531484603882, "learning_rate": 2.829084098159934e-06, "loss": 0.4808, "step": 26690 }, { "epoch": 1.7913157276601457, "grad_norm": 1.0424187183380127, "learning_rate": 2.8254812277854813e-06, "loss": 0.4182, "step": 26692 }, { "epoch": 1.7914499513439146, "grad_norm": 1.0150495767593384, "learning_rate": 2.8218805863142794e-06, "loss": 0.541, "step": 26694 }, { "epoch": 1.7915841750276837, "grad_norm": 0.9561358094215393, "learning_rate": 2.8182821739164534e-06, "loss": 0.414, "step": 26696 }, { "epoch": 1.7917183987114527, "grad_norm": 1.0548745393753052, "learning_rate": 2.8146859907620172e-06, "loss": 0.3886, "step": 26698 }, { "epoch": 1.7918526223952216, "grad_norm": 1.0386500358581543, "learning_rate": 2.8110920370208682e-06, "loss": 0.4123, "step": 26700 }, { "epoch": 1.7919868460789905, "grad_norm": 1.0239613056182861, "learning_rate": 2.8075003128628374e-06, "loss": 0.5219, "step": 26702 }, { "epoch": 1.7921210697627596, "grad_norm": 1.0330843925476074, "learning_rate": 2.8039108184576113e-06, "loss": 0.442, "step": 26704 }, { "epoch": 1.7922552934465288, "grad_norm": 1.0979597568511963, "learning_rate": 2.8003235539747984e-06, "loss": 0.4461, "step": 26706 }, { "epoch": 1.7923895171302977, "grad_norm": 0.9252837300300598, "learning_rate": 2.796738519583886e-06, "loss": 0.4517, "step": 26708 }, { "epoch": 1.7925237408140666, "grad_norm": 1.0468565225601196, "learning_rate": 2.7931557154542487e-06, "loss": 0.472, "step": 26710 }, { "epoch": 1.7926579644978355, "grad_norm": 1.0020323991775513, "learning_rate": 2.7895751417551852e-06, "loss": 0.5184, "step": 26712 }, { "epoch": 1.7927921881816047, "grad_norm": 0.9411514401435852, "learning_rate": 2.785996798655849e-06, "loss": 0.45, "step": 26714 }, { "epoch": 1.7929264118653736, "grad_norm": 0.7674515247344971, "learning_rate": 2.7824206863253264e-06, "loss": 0.4257, "step": 26716 }, { "epoch": 1.7930606355491427, "grad_norm": 1.0416252613067627, "learning_rate": 2.778846804932583e-06, "loss": 0.4851, "step": 26718 }, { "epoch": 1.7931948592329117, "grad_norm": 1.0742361545562744, "learning_rate": 2.775275154646467e-06, "loss": 0.4777, "step": 26720 }, { "epoch": 1.7933290829166806, "grad_norm": 1.0092241764068604, "learning_rate": 2.7717057356357323e-06, "loss": 0.5018, "step": 26722 }, { "epoch": 1.7934633066004495, "grad_norm": 0.8847936391830444, "learning_rate": 2.768138548069038e-06, "loss": 0.4576, "step": 26724 }, { "epoch": 1.7935975302842186, "grad_norm": 1.039747714996338, "learning_rate": 2.7645735921149217e-06, "loss": 0.5439, "step": 26726 }, { "epoch": 1.7937317539679878, "grad_norm": 0.8865048289299011, "learning_rate": 2.7610108679418156e-06, "loss": 0.4585, "step": 26728 }, { "epoch": 1.7938659776517567, "grad_norm": 1.035498023033142, "learning_rate": 2.757450375718046e-06, "loss": 0.4834, "step": 26730 }, { "epoch": 1.7940002013355256, "grad_norm": 1.0822051763534546, "learning_rate": 2.753892115611856e-06, "loss": 0.4815, "step": 26732 }, { "epoch": 1.7941344250192945, "grad_norm": 0.8330800533294678, "learning_rate": 2.7503360877913497e-06, "loss": 0.4315, "step": 26734 }, { "epoch": 1.7942686487030637, "grad_norm": 0.9999399185180664, "learning_rate": 2.7467822924245713e-06, "loss": 0.433, "step": 26736 }, { "epoch": 1.7944028723868326, "grad_norm": 1.141569972038269, "learning_rate": 2.7432307296793914e-06, "loss": 0.4499, "step": 26738 }, { "epoch": 1.7945370960706017, "grad_norm": 1.006667971611023, "learning_rate": 2.739681399723637e-06, "loss": 0.4665, "step": 26740 }, { "epoch": 1.7946713197543707, "grad_norm": 0.9503446221351624, "learning_rate": 2.736134302725002e-06, "loss": 0.435, "step": 26742 }, { "epoch": 1.7948055434381396, "grad_norm": 1.1470330953598022, "learning_rate": 2.7325894388510855e-06, "loss": 0.4996, "step": 26744 }, { "epoch": 1.7949397671219085, "grad_norm": 1.1522454023361206, "learning_rate": 2.72904680826937e-06, "loss": 0.5004, "step": 26746 }, { "epoch": 1.7950739908056776, "grad_norm": 0.937041699886322, "learning_rate": 2.7255064111472385e-06, "loss": 0.4256, "step": 26748 }, { "epoch": 1.7952082144894468, "grad_norm": 1.0546565055847168, "learning_rate": 2.721968247651957e-06, "loss": 0.3999, "step": 26750 }, { "epoch": 1.7953424381732157, "grad_norm": 0.8641650080680847, "learning_rate": 2.71843231795072e-06, "loss": 0.3858, "step": 26752 }, { "epoch": 1.7954766618569846, "grad_norm": 1.1361602544784546, "learning_rate": 2.7148986222105765e-06, "loss": 0.5382, "step": 26754 }, { "epoch": 1.7956108855407535, "grad_norm": 0.9405742287635803, "learning_rate": 2.7113671605984934e-06, "loss": 0.4319, "step": 26756 }, { "epoch": 1.7957451092245227, "grad_norm": 1.0715886354446411, "learning_rate": 2.707837933281321e-06, "loss": 0.5071, "step": 26758 }, { "epoch": 1.7958793329082918, "grad_norm": 0.885618269443512, "learning_rate": 2.704310940425808e-06, "loss": 0.4731, "step": 26760 }, { "epoch": 1.7960135565920607, "grad_norm": 1.038923740386963, "learning_rate": 2.7007861821986e-06, "loss": 0.4156, "step": 26762 }, { "epoch": 1.7961477802758297, "grad_norm": 1.0108329057693481, "learning_rate": 2.697263658766247e-06, "loss": 0.5048, "step": 26764 }, { "epoch": 1.7962820039595986, "grad_norm": 1.0313938856124878, "learning_rate": 2.6937433702951543e-06, "loss": 0.4768, "step": 26766 }, { "epoch": 1.7964162276433677, "grad_norm": 1.0988893508911133, "learning_rate": 2.690225316951672e-06, "loss": 0.4781, "step": 26768 }, { "epoch": 1.7965504513271366, "grad_norm": 0.91923588514328, "learning_rate": 2.6867094989020016e-06, "loss": 0.4576, "step": 26770 }, { "epoch": 1.7966846750109058, "grad_norm": 1.169795036315918, "learning_rate": 2.683195916312281e-06, "loss": 0.5209, "step": 26772 }, { "epoch": 1.7968188986946747, "grad_norm": 1.0356361865997314, "learning_rate": 2.6796845693485064e-06, "loss": 0.4319, "step": 26774 }, { "epoch": 1.7969531223784436, "grad_norm": 1.1351392269134521, "learning_rate": 2.6761754581765886e-06, "loss": 0.4949, "step": 26776 }, { "epoch": 1.7970873460622125, "grad_norm": 1.1579749584197998, "learning_rate": 2.672668582962312e-06, "loss": 0.5209, "step": 26778 }, { "epoch": 1.7972215697459817, "grad_norm": 1.3297175168991089, "learning_rate": 2.6691639438713834e-06, "loss": 0.5057, "step": 26780 }, { "epoch": 1.7973557934297508, "grad_norm": 1.0900532007217407, "learning_rate": 2.6656615410693918e-06, "loss": 0.4534, "step": 26782 }, { "epoch": 1.7974900171135197, "grad_norm": 1.1601152420043945, "learning_rate": 2.662161374721811e-06, "loss": 0.4267, "step": 26784 }, { "epoch": 1.7976242407972887, "grad_norm": 0.9871609210968018, "learning_rate": 2.6586634449940194e-06, "loss": 0.5198, "step": 26786 }, { "epoch": 1.7977584644810576, "grad_norm": 1.0594075918197632, "learning_rate": 2.6551677520512797e-06, "loss": 0.5015, "step": 26788 }, { "epoch": 1.7978926881648267, "grad_norm": 0.9030765295028687, "learning_rate": 2.651674296058776e-06, "loss": 0.4498, "step": 26790 }, { "epoch": 1.7980269118485956, "grad_norm": 1.0988365411758423, "learning_rate": 2.6481830771815486e-06, "loss": 0.4167, "step": 26792 }, { "epoch": 1.7981611355323648, "grad_norm": 1.1187043190002441, "learning_rate": 2.64469409558456e-06, "loss": 0.5317, "step": 26794 }, { "epoch": 1.7982953592161337, "grad_norm": 1.0270341634750366, "learning_rate": 2.6412073514326508e-06, "loss": 0.467, "step": 26796 }, { "epoch": 1.7984295828999026, "grad_norm": 1.0437359809875488, "learning_rate": 2.637722844890572e-06, "loss": 0.4379, "step": 26798 }, { "epoch": 1.7985638065836715, "grad_norm": 1.043473243713379, "learning_rate": 2.6342405761229485e-06, "loss": 0.4877, "step": 26800 }, { "epoch": 1.7986980302674407, "grad_norm": 1.1618503332138062, "learning_rate": 2.630760545294325e-06, "loss": 0.4708, "step": 26802 }, { "epoch": 1.7988322539512098, "grad_norm": 1.0864089727401733, "learning_rate": 2.6272827525691104e-06, "loss": 0.4528, "step": 26804 }, { "epoch": 1.7989664776349787, "grad_norm": 0.9820147752761841, "learning_rate": 2.6238071981116383e-06, "loss": 0.4048, "step": 26806 }, { "epoch": 1.7991007013187477, "grad_norm": 1.0986045598983765, "learning_rate": 2.620333882086107e-06, "loss": 0.5094, "step": 26808 }, { "epoch": 1.7992349250025166, "grad_norm": 1.055762767791748, "learning_rate": 2.616862804656639e-06, "loss": 0.4684, "step": 26810 }, { "epoch": 1.7993691486862857, "grad_norm": 1.0238934755325317, "learning_rate": 2.6133939659872265e-06, "loss": 0.3908, "step": 26812 }, { "epoch": 1.7995033723700546, "grad_norm": 1.0113105773925781, "learning_rate": 2.6099273662417713e-06, "loss": 0.4477, "step": 26814 }, { "epoch": 1.7996375960538238, "grad_norm": 1.0518611669540405, "learning_rate": 2.6064630055840477e-06, "loss": 0.4438, "step": 26816 }, { "epoch": 1.7997718197375927, "grad_norm": 1.034557580947876, "learning_rate": 2.6030008841777585e-06, "loss": 0.4016, "step": 26818 }, { "epoch": 1.7999060434213616, "grad_norm": 1.0221138000488281, "learning_rate": 2.5995410021864787e-06, "loss": 0.4985, "step": 26820 }, { "epoch": 1.8000402671051305, "grad_norm": 1.2014293670654297, "learning_rate": 2.596083359773677e-06, "loss": 0.5002, "step": 26822 }, { "epoch": 1.8001744907888997, "grad_norm": 1.0326858758926392, "learning_rate": 2.5926279571027113e-06, "loss": 0.5282, "step": 26824 }, { "epoch": 1.8003087144726688, "grad_norm": 1.0132777690887451, "learning_rate": 2.589174794336863e-06, "loss": 0.3885, "step": 26826 }, { "epoch": 1.8004429381564377, "grad_norm": 0.9559783339500427, "learning_rate": 2.585723871639267e-06, "loss": 0.47, "step": 26828 }, { "epoch": 1.8005771618402067, "grad_norm": 1.0535095930099487, "learning_rate": 2.5822751891729945e-06, "loss": 0.4312, "step": 26830 }, { "epoch": 1.8007113855239756, "grad_norm": 1.1592003107070923, "learning_rate": 2.578828747100964e-06, "loss": 0.4777, "step": 26832 }, { "epoch": 1.8008456092077447, "grad_norm": 1.245162844657898, "learning_rate": 2.5753845455860347e-06, "loss": 0.5542, "step": 26834 }, { "epoch": 1.8009798328915139, "grad_norm": 0.9433565735816956, "learning_rate": 2.5719425847909206e-06, "loss": 0.4472, "step": 26836 }, { "epoch": 1.8011140565752828, "grad_norm": 1.0376611948013306, "learning_rate": 2.5685028648782638e-06, "loss": 0.4262, "step": 26838 }, { "epoch": 1.8012482802590517, "grad_norm": 1.1596550941467285, "learning_rate": 2.5650653860105733e-06, "loss": 0.4543, "step": 26840 }, { "epoch": 1.8013825039428206, "grad_norm": 1.075486421585083, "learning_rate": 2.561630148350269e-06, "loss": 0.4355, "step": 26842 }, { "epoch": 1.8015167276265898, "grad_norm": 0.9907163977622986, "learning_rate": 2.558197152059649e-06, "loss": 0.4969, "step": 26844 }, { "epoch": 1.8016509513103587, "grad_norm": 1.024023413658142, "learning_rate": 2.5547663973009284e-06, "loss": 0.5127, "step": 26846 }, { "epoch": 1.8017851749941278, "grad_norm": 0.9813905954360962, "learning_rate": 2.551337884236199e-06, "loss": 0.4463, "step": 26848 }, { "epoch": 1.8019193986778967, "grad_norm": 1.15963613986969, "learning_rate": 2.547911613027454e-06, "loss": 0.4663, "step": 26850 }, { "epoch": 1.8020536223616657, "grad_norm": 1.0435926914215088, "learning_rate": 2.5444875838365745e-06, "loss": 0.4895, "step": 26852 }, { "epoch": 1.8021878460454346, "grad_norm": 1.1256290674209595, "learning_rate": 2.5410657968253317e-06, "loss": 0.5043, "step": 26854 }, { "epoch": 1.8023220697292037, "grad_norm": 1.3820325136184692, "learning_rate": 2.5376462521554125e-06, "loss": 0.4759, "step": 26856 }, { "epoch": 1.8024562934129729, "grad_norm": 1.0215343236923218, "learning_rate": 2.5342289499883766e-06, "loss": 0.4456, "step": 26858 }, { "epoch": 1.8025905170967418, "grad_norm": 0.9229339957237244, "learning_rate": 2.5308138904856838e-06, "loss": 0.4317, "step": 26860 }, { "epoch": 1.8027247407805107, "grad_norm": 0.9657912850379944, "learning_rate": 2.527401073808683e-06, "loss": 0.5191, "step": 26862 }, { "epoch": 1.8028589644642796, "grad_norm": 1.0605182647705078, "learning_rate": 2.5239905001186394e-06, "loss": 0.4714, "step": 26864 }, { "epoch": 1.8029931881480488, "grad_norm": 1.1260061264038086, "learning_rate": 2.520582169576685e-06, "loss": 0.506, "step": 26866 }, { "epoch": 1.8031274118318177, "grad_norm": 1.3092259168624878, "learning_rate": 2.517176082343858e-06, "loss": 0.4284, "step": 26868 }, { "epoch": 1.8032616355155868, "grad_norm": 1.0388407707214355, "learning_rate": 2.513772238581097e-06, "loss": 0.4028, "step": 26870 }, { "epoch": 1.8033958591993557, "grad_norm": 0.997028112411499, "learning_rate": 2.5103706384492164e-06, "loss": 0.4643, "step": 26872 }, { "epoch": 1.8035300828831247, "grad_norm": 1.0246165990829468, "learning_rate": 2.5069712821089277e-06, "loss": 0.5137, "step": 26874 }, { "epoch": 1.8036643065668936, "grad_norm": 0.8851560950279236, "learning_rate": 2.503574169720868e-06, "loss": 0.4453, "step": 26876 }, { "epoch": 1.8037985302506627, "grad_norm": 0.9529306888580322, "learning_rate": 2.5001793014455266e-06, "loss": 0.4621, "step": 26878 }, { "epoch": 1.8039327539344319, "grad_norm": 0.9867005348205566, "learning_rate": 2.4967866774433136e-06, "loss": 0.3798, "step": 26880 }, { "epoch": 1.8040669776182008, "grad_norm": 0.9776849150657654, "learning_rate": 2.4933962978745117e-06, "loss": 0.4723, "step": 26882 }, { "epoch": 1.8042012013019697, "grad_norm": 0.9732565879821777, "learning_rate": 2.490008162899321e-06, "loss": 0.4714, "step": 26884 }, { "epoch": 1.8043354249857386, "grad_norm": 0.9912878274917603, "learning_rate": 2.486622272677813e-06, "loss": 0.4294, "step": 26886 }, { "epoch": 1.8044696486695078, "grad_norm": 1.0751785039901733, "learning_rate": 2.483238627369988e-06, "loss": 0.4929, "step": 26888 }, { "epoch": 1.8046038723532767, "grad_norm": 1.1649205684661865, "learning_rate": 2.4798572271356846e-06, "loss": 0.4399, "step": 26890 }, { "epoch": 1.8047380960370458, "grad_norm": 0.9231991767883301, "learning_rate": 2.4764780721346914e-06, "loss": 0.5099, "step": 26892 }, { "epoch": 1.8048723197208147, "grad_norm": 1.1894398927688599, "learning_rate": 2.4731011625266477e-06, "loss": 0.4499, "step": 26894 }, { "epoch": 1.8050065434045837, "grad_norm": 0.9900402426719666, "learning_rate": 2.4697264984711257e-06, "loss": 0.4933, "step": 26896 }, { "epoch": 1.8051407670883526, "grad_norm": 1.1936390399932861, "learning_rate": 2.466354080127564e-06, "loss": 0.4654, "step": 26898 }, { "epoch": 1.8052749907721217, "grad_norm": 1.0529814958572388, "learning_rate": 2.4629839076552974e-06, "loss": 0.4659, "step": 26900 }, { "epoch": 1.8054092144558909, "grad_norm": 0.9130473136901855, "learning_rate": 2.459615981213559e-06, "loss": 0.4362, "step": 26902 }, { "epoch": 1.8055434381396598, "grad_norm": 1.0363892316818237, "learning_rate": 2.456250300961488e-06, "loss": 0.4582, "step": 26904 }, { "epoch": 1.8056776618234287, "grad_norm": 1.144631266593933, "learning_rate": 2.452886867058102e-06, "loss": 0.4553, "step": 26906 }, { "epoch": 1.8058118855071976, "grad_norm": 1.192743182182312, "learning_rate": 2.449525679662312e-06, "loss": 0.4986, "step": 26908 }, { "epoch": 1.8059461091909668, "grad_norm": 0.9952451586723328, "learning_rate": 2.44616673893292e-06, "loss": 0.5182, "step": 26910 }, { "epoch": 1.806080332874736, "grad_norm": 1.0130771398544312, "learning_rate": 2.442810045028654e-06, "loss": 0.5121, "step": 26912 }, { "epoch": 1.8062145565585048, "grad_norm": 0.8701079487800598, "learning_rate": 2.439455598108081e-06, "loss": 0.4631, "step": 26914 }, { "epoch": 1.8063487802422737, "grad_norm": 0.9158850312232971, "learning_rate": 2.4361033983297255e-06, "loss": 0.4168, "step": 26916 }, { "epoch": 1.8064830039260427, "grad_norm": 1.0420955419540405, "learning_rate": 2.432753445851943e-06, "loss": 0.4193, "step": 26918 }, { "epoch": 1.8066172276098118, "grad_norm": 1.009220004081726, "learning_rate": 2.4294057408330184e-06, "loss": 0.4435, "step": 26920 }, { "epoch": 1.8067514512935807, "grad_norm": 1.074559211730957, "learning_rate": 2.4260602834311364e-06, "loss": 0.4526, "step": 26922 }, { "epoch": 1.8068856749773499, "grad_norm": 1.014679193496704, "learning_rate": 2.422717073804348e-06, "loss": 0.4674, "step": 26924 }, { "epoch": 1.8070198986611188, "grad_norm": 1.091285228729248, "learning_rate": 2.4193761121106384e-06, "loss": 0.4736, "step": 26926 }, { "epoch": 1.8071541223448877, "grad_norm": 0.9874877333641052, "learning_rate": 2.4160373985078256e-06, "loss": 0.4707, "step": 26928 }, { "epoch": 1.8072883460286566, "grad_norm": 1.039380669593811, "learning_rate": 2.4127009331536832e-06, "loss": 0.5182, "step": 26930 }, { "epoch": 1.8074225697124258, "grad_norm": 1.1278899908065796, "learning_rate": 2.409366716205841e-06, "loss": 0.481, "step": 26932 }, { "epoch": 1.807556793396195, "grad_norm": 1.0794038772583008, "learning_rate": 2.4060347478218446e-06, "loss": 0.5274, "step": 26934 }, { "epoch": 1.8076910170799638, "grad_norm": 1.056154727935791, "learning_rate": 2.402705028159119e-06, "loss": 0.4559, "step": 26936 }, { "epoch": 1.8078252407637327, "grad_norm": 0.8898627758026123, "learning_rate": 2.399377557374982e-06, "loss": 0.44, "step": 26938 }, { "epoch": 1.8079594644475017, "grad_norm": 0.9880284667015076, "learning_rate": 2.3960523356266475e-06, "loss": 0.4151, "step": 26940 }, { "epoch": 1.8080936881312708, "grad_norm": 0.878262996673584, "learning_rate": 2.3927293630712332e-06, "loss": 0.4972, "step": 26942 }, { "epoch": 1.8082279118150397, "grad_norm": 0.9728204607963562, "learning_rate": 2.3894086398657478e-06, "loss": 0.4705, "step": 26944 }, { "epoch": 1.8083621354988089, "grad_norm": 0.9428135752677917, "learning_rate": 2.3860901661670764e-06, "loss": 0.4006, "step": 26946 }, { "epoch": 1.8084963591825778, "grad_norm": 0.9841963648796082, "learning_rate": 2.3827739421320105e-06, "loss": 0.4725, "step": 26948 }, { "epoch": 1.8086305828663467, "grad_norm": 1.0408563613891602, "learning_rate": 2.379459967917247e-06, "loss": 0.4695, "step": 26950 }, { "epoch": 1.8087648065501156, "grad_norm": 1.026936411857605, "learning_rate": 2.376148243679355e-06, "loss": 0.5047, "step": 26952 }, { "epoch": 1.8088990302338848, "grad_norm": 1.1736093759536743, "learning_rate": 2.37283876957482e-06, "loss": 0.5216, "step": 26954 }, { "epoch": 1.809033253917654, "grad_norm": 1.0411231517791748, "learning_rate": 2.369531545759984e-06, "loss": 0.4811, "step": 26956 }, { "epoch": 1.8091674776014228, "grad_norm": 1.0473124980926514, "learning_rate": 2.3662265723911337e-06, "loss": 0.4326, "step": 26958 }, { "epoch": 1.8093017012851917, "grad_norm": 0.8875226974487305, "learning_rate": 2.3629238496243987e-06, "loss": 0.4732, "step": 26960 }, { "epoch": 1.8094359249689607, "grad_norm": 0.9540635347366333, "learning_rate": 2.359623377615844e-06, "loss": 0.4236, "step": 26962 }, { "epoch": 1.8095701486527298, "grad_norm": 1.1037060022354126, "learning_rate": 2.356325156521405e-06, "loss": 0.4242, "step": 26964 }, { "epoch": 1.8097043723364987, "grad_norm": 0.9551806449890137, "learning_rate": 2.353029186496919e-06, "loss": 0.4801, "step": 26966 }, { "epoch": 1.8098385960202679, "grad_norm": 1.0264946222305298, "learning_rate": 2.349735467698094e-06, "loss": 0.4097, "step": 26968 }, { "epoch": 1.8099728197040368, "grad_norm": 0.9813930988311768, "learning_rate": 2.3464440002805844e-06, "loss": 0.4443, "step": 26970 }, { "epoch": 1.8101070433878057, "grad_norm": 1.1017805337905884, "learning_rate": 2.3431547843998815e-06, "loss": 0.4758, "step": 26972 }, { "epoch": 1.8102412670715746, "grad_norm": 0.9897599816322327, "learning_rate": 2.3398678202114054e-06, "loss": 0.4848, "step": 26974 }, { "epoch": 1.8103754907553438, "grad_norm": 1.0189718008041382, "learning_rate": 2.336583107870449e-06, "loss": 0.4375, "step": 26976 }, { "epoch": 1.810509714439113, "grad_norm": 1.0726696252822876, "learning_rate": 2.333300647532222e-06, "loss": 0.4518, "step": 26978 }, { "epoch": 1.8106439381228818, "grad_norm": 0.9022804498672485, "learning_rate": 2.330020439351799e-06, "loss": 0.4779, "step": 26980 }, { "epoch": 1.8107781618066507, "grad_norm": 1.4354796409606934, "learning_rate": 2.3267424834841845e-06, "loss": 0.5082, "step": 26982 }, { "epoch": 1.8109123854904197, "grad_norm": 0.8356832265853882, "learning_rate": 2.323466780084227e-06, "loss": 0.3892, "step": 26984 }, { "epoch": 1.8110466091741888, "grad_norm": 1.0602096319198608, "learning_rate": 2.3201933293067247e-06, "loss": 0.4176, "step": 26986 }, { "epoch": 1.811180832857958, "grad_norm": 1.0468652248382568, "learning_rate": 2.3169221313063207e-06, "loss": 0.468, "step": 26988 }, { "epoch": 1.8113150565417269, "grad_norm": 1.003930687904358, "learning_rate": 2.31365318623758e-06, "loss": 0.4576, "step": 26990 }, { "epoch": 1.8114492802254958, "grad_norm": 1.1603525876998901, "learning_rate": 2.3103864942549623e-06, "loss": 0.493, "step": 26992 }, { "epoch": 1.8115835039092647, "grad_norm": 1.0017002820968628, "learning_rate": 2.307122055512806e-06, "loss": 0.4167, "step": 26994 }, { "epoch": 1.8117177275930338, "grad_norm": 1.084191083908081, "learning_rate": 2.3038598701653425e-06, "loss": 0.4877, "step": 26996 }, { "epoch": 1.8118519512768028, "grad_norm": 0.9252565503120422, "learning_rate": 2.30059993836671e-06, "loss": 0.415, "step": 26998 }, { "epoch": 1.811986174960572, "grad_norm": 1.047005534172058, "learning_rate": 2.297342260270935e-06, "loss": 0.4355, "step": 27000 }, { "epoch": 1.8121203986443408, "grad_norm": 1.072208285331726, "learning_rate": 2.294086836031939e-06, "loss": 0.4838, "step": 27002 }, { "epoch": 1.8122546223281097, "grad_norm": 1.0251212120056152, "learning_rate": 2.2908336658035266e-06, "loss": 0.3947, "step": 27004 }, { "epoch": 1.8123888460118787, "grad_norm": 1.0754388570785522, "learning_rate": 2.287582749739403e-06, "loss": 0.4685, "step": 27006 }, { "epoch": 1.8125230696956478, "grad_norm": 1.0514887571334839, "learning_rate": 2.2843340879931783e-06, "loss": 0.4582, "step": 27008 }, { "epoch": 1.812657293379417, "grad_norm": 1.3117904663085938, "learning_rate": 2.2810876807183414e-06, "loss": 0.4585, "step": 27010 }, { "epoch": 1.8127915170631859, "grad_norm": 1.0328198671340942, "learning_rate": 2.277843528068274e-06, "loss": 0.4514, "step": 27012 }, { "epoch": 1.8129257407469548, "grad_norm": 0.9663453102111816, "learning_rate": 2.274601630196249e-06, "loss": 0.4671, "step": 27014 }, { "epoch": 1.8130599644307237, "grad_norm": 1.099232792854309, "learning_rate": 2.2713619872554603e-06, "loss": 0.4821, "step": 27016 }, { "epoch": 1.8131941881144928, "grad_norm": 1.1244513988494873, "learning_rate": 2.268124599398952e-06, "loss": 0.4422, "step": 27018 }, { "epoch": 1.8133284117982618, "grad_norm": 1.174409031867981, "learning_rate": 2.264889466779707e-06, "loss": 0.4618, "step": 27020 }, { "epoch": 1.813462635482031, "grad_norm": 1.0032176971435547, "learning_rate": 2.261656589550565e-06, "loss": 0.4631, "step": 27022 }, { "epoch": 1.8135968591657998, "grad_norm": 1.0601005554199219, "learning_rate": 2.258425967864275e-06, "loss": 0.4899, "step": 27024 }, { "epoch": 1.8137310828495687, "grad_norm": 1.045218825340271, "learning_rate": 2.2551976018734766e-06, "loss": 0.5004, "step": 27026 }, { "epoch": 1.8138653065333377, "grad_norm": 1.0657036304473877, "learning_rate": 2.251971491730709e-06, "loss": 0.4616, "step": 27028 }, { "epoch": 1.8139995302171068, "grad_norm": 4.555449485778809, "learning_rate": 2.248747637588394e-06, "loss": 0.4756, "step": 27030 }, { "epoch": 1.814133753900876, "grad_norm": 1.0749096870422363, "learning_rate": 2.2455260395988553e-06, "loss": 0.4449, "step": 27032 }, { "epoch": 1.8142679775846449, "grad_norm": 1.120177984237671, "learning_rate": 2.242306697914298e-06, "loss": 0.4682, "step": 27034 }, { "epoch": 1.8144022012684138, "grad_norm": 1.0862208604812622, "learning_rate": 2.2390896126868453e-06, "loss": 0.5205, "step": 27036 }, { "epoch": 1.8145364249521827, "grad_norm": 1.0489317178726196, "learning_rate": 2.2358747840684924e-06, "loss": 0.4493, "step": 27038 }, { "epoch": 1.8146706486359518, "grad_norm": 0.9261298775672913, "learning_rate": 2.232662212211134e-06, "loss": 0.4632, "step": 27040 }, { "epoch": 1.8148048723197208, "grad_norm": 0.7837489247322083, "learning_rate": 2.2294518972665434e-06, "loss": 0.4459, "step": 27042 }, { "epoch": 1.81493909600349, "grad_norm": 0.9737602472305298, "learning_rate": 2.2262438393864214e-06, "loss": 0.4512, "step": 27044 }, { "epoch": 1.8150733196872588, "grad_norm": 1.214451551437378, "learning_rate": 2.2230380387223305e-06, "loss": 0.5024, "step": 27046 }, { "epoch": 1.8152075433710277, "grad_norm": 0.860659122467041, "learning_rate": 2.219834495425749e-06, "loss": 0.377, "step": 27048 }, { "epoch": 1.8153417670547967, "grad_norm": 0.9898740649223328, "learning_rate": 2.2166332096480337e-06, "loss": 0.4767, "step": 27050 }, { "epoch": 1.8154759907385658, "grad_norm": 1.0652974843978882, "learning_rate": 2.213434181540436e-06, "loss": 0.4188, "step": 27052 }, { "epoch": 1.815610214422335, "grad_norm": 1.0848276615142822, "learning_rate": 2.2102374112541013e-06, "loss": 0.4586, "step": 27054 }, { "epoch": 1.8157444381061039, "grad_norm": 0.9306811094284058, "learning_rate": 2.207042898940076e-06, "loss": 0.447, "step": 27056 }, { "epoch": 1.8158786617898728, "grad_norm": 1.0598939657211304, "learning_rate": 2.2038506447492945e-06, "loss": 0.4723, "step": 27058 }, { "epoch": 1.8160128854736417, "grad_norm": 1.1055607795715332, "learning_rate": 2.200660648832581e-06, "loss": 0.4762, "step": 27060 }, { "epoch": 1.8161471091574108, "grad_norm": 1.04423189163208, "learning_rate": 2.197472911340659e-06, "loss": 0.486, "step": 27062 }, { "epoch": 1.81628133284118, "grad_norm": 1.1375607252120972, "learning_rate": 2.1942874324241357e-06, "loss": 0.5542, "step": 27064 }, { "epoch": 1.816415556524949, "grad_norm": 1.0168747901916504, "learning_rate": 2.1911042122335356e-06, "loss": 0.571, "step": 27066 }, { "epoch": 1.8165497802087178, "grad_norm": 1.0431084632873535, "learning_rate": 2.1879232509192494e-06, "loss": 0.4328, "step": 27068 }, { "epoch": 1.8166840038924867, "grad_norm": 1.0373541116714478, "learning_rate": 2.1847445486315675e-06, "loss": 0.4037, "step": 27070 }, { "epoch": 1.8168182275762559, "grad_norm": 1.02206289768219, "learning_rate": 2.181568105520676e-06, "loss": 0.4972, "step": 27072 }, { "epoch": 1.8169524512600248, "grad_norm": 1.0911699533462524, "learning_rate": 2.1783939217366655e-06, "loss": 0.4703, "step": 27074 }, { "epoch": 1.817086674943794, "grad_norm": 0.9556540846824646, "learning_rate": 2.1752219974294995e-06, "loss": 0.4669, "step": 27076 }, { "epoch": 1.8172208986275629, "grad_norm": 1.0169297456741333, "learning_rate": 2.1720523327490636e-06, "loss": 0.4742, "step": 27078 }, { "epoch": 1.8173551223113318, "grad_norm": 0.8982275724411011, "learning_rate": 2.168884927845094e-06, "loss": 0.4801, "step": 27080 }, { "epoch": 1.8174893459951007, "grad_norm": 0.9359860420227051, "learning_rate": 2.165719782867259e-06, "loss": 0.488, "step": 27082 }, { "epoch": 1.8176235696788698, "grad_norm": 0.9763116836547852, "learning_rate": 2.1625568979651014e-06, "loss": 0.4066, "step": 27084 }, { "epoch": 1.817757793362639, "grad_norm": 0.8585172891616821, "learning_rate": 2.1593962732880615e-06, "loss": 0.4162, "step": 27086 }, { "epoch": 1.817892017046408, "grad_norm": 1.0699492692947388, "learning_rate": 2.1562379089854755e-06, "loss": 0.4776, "step": 27088 }, { "epoch": 1.8180262407301768, "grad_norm": 1.0023192167282104, "learning_rate": 2.1530818052065695e-06, "loss": 0.4617, "step": 27090 }, { "epoch": 1.8181604644139457, "grad_norm": 1.0295124053955078, "learning_rate": 2.1499279621004564e-06, "loss": 0.484, "step": 27092 }, { "epoch": 1.8182946880977149, "grad_norm": 1.0506134033203125, "learning_rate": 2.146776379816157e-06, "loss": 0.4966, "step": 27094 }, { "epoch": 1.8184289117814838, "grad_norm": 1.12366783618927, "learning_rate": 2.1436270585025732e-06, "loss": 0.4895, "step": 27096 }, { "epoch": 1.818563135465253, "grad_norm": 1.0606402158737183, "learning_rate": 2.140479998308509e-06, "loss": 0.546, "step": 27098 }, { "epoch": 1.8186973591490219, "grad_norm": 1.046035885810852, "learning_rate": 2.137335199382645e-06, "loss": 0.4516, "step": 27100 }, { "epoch": 1.8188315828327908, "grad_norm": 0.9801777601242065, "learning_rate": 2.1341926618735796e-06, "loss": 0.48, "step": 27102 }, { "epoch": 1.8189658065165597, "grad_norm": 0.9874941110610962, "learning_rate": 2.1310523859297828e-06, "loss": 0.4771, "step": 27104 }, { "epoch": 1.8191000302003288, "grad_norm": 0.9163682460784912, "learning_rate": 2.1279143716996464e-06, "loss": 0.4498, "step": 27106 }, { "epoch": 1.819234253884098, "grad_norm": 0.8990226984024048, "learning_rate": 2.1247786193314025e-06, "loss": 0.4924, "step": 27108 }, { "epoch": 1.819368477567867, "grad_norm": 1.0113532543182373, "learning_rate": 2.121645128973232e-06, "loss": 0.4578, "step": 27110 }, { "epoch": 1.8195027012516358, "grad_norm": 1.1612234115600586, "learning_rate": 2.118513900773178e-06, "loss": 0.4687, "step": 27112 }, { "epoch": 1.8196369249354047, "grad_norm": 1.055635929107666, "learning_rate": 2.1153849348791887e-06, "loss": 0.468, "step": 27114 }, { "epoch": 1.8197711486191739, "grad_norm": 0.9631549119949341, "learning_rate": 2.1122582314391013e-06, "loss": 0.4729, "step": 27116 }, { "epoch": 1.8199053723029428, "grad_norm": 0.9034852981567383, "learning_rate": 2.1091337906006482e-06, "loss": 0.4397, "step": 27118 }, { "epoch": 1.820039595986712, "grad_norm": 1.051687240600586, "learning_rate": 2.1060116125114436e-06, "loss": 0.5044, "step": 27120 }, { "epoch": 1.8201738196704809, "grad_norm": 0.9925054907798767, "learning_rate": 2.102891697319015e-06, "loss": 0.4694, "step": 27122 }, { "epoch": 1.8203080433542498, "grad_norm": 1.147497296333313, "learning_rate": 2.099774045170766e-06, "loss": 0.4803, "step": 27124 }, { "epoch": 1.8204422670380187, "grad_norm": 0.9984075427055359, "learning_rate": 2.096658656214007e-06, "loss": 0.5095, "step": 27126 }, { "epoch": 1.8205764907217878, "grad_norm": 0.8068634271621704, "learning_rate": 2.093545530595925e-06, "loss": 0.4256, "step": 27128 }, { "epoch": 1.820710714405557, "grad_norm": 1.0053539276123047, "learning_rate": 2.090434668463609e-06, "loss": 0.4349, "step": 27130 }, { "epoch": 1.820844938089326, "grad_norm": 0.77165287733078, "learning_rate": 2.0873260699640462e-06, "loss": 0.423, "step": 27132 }, { "epoch": 1.8209791617730948, "grad_norm": 1.0810253620147705, "learning_rate": 2.084219735244114e-06, "loss": 0.4758, "step": 27134 }, { "epoch": 1.8211133854568637, "grad_norm": 1.009852409362793, "learning_rate": 2.081115664450578e-06, "loss": 0.4755, "step": 27136 }, { "epoch": 1.8212476091406329, "grad_norm": 1.0758378505706787, "learning_rate": 2.078013857730088e-06, "loss": 0.4598, "step": 27138 }, { "epoch": 1.821381832824402, "grad_norm": 1.057146668434143, "learning_rate": 2.0749143152292204e-06, "loss": 0.4423, "step": 27140 }, { "epoch": 1.821516056508171, "grad_norm": 0.950028121471405, "learning_rate": 2.071817037094398e-06, "loss": 0.4749, "step": 27142 }, { "epoch": 1.8216502801919399, "grad_norm": 0.9723566770553589, "learning_rate": 2.068722023471992e-06, "loss": 0.4895, "step": 27144 }, { "epoch": 1.8217845038757088, "grad_norm": 1.006791114807129, "learning_rate": 2.065629274508202e-06, "loss": 0.5504, "step": 27146 }, { "epoch": 1.821918727559478, "grad_norm": 1.1130188703536987, "learning_rate": 2.062538790349178e-06, "loss": 0.4283, "step": 27148 }, { "epoch": 1.8220529512432468, "grad_norm": 0.988906741142273, "learning_rate": 2.059450571140925e-06, "loss": 0.4908, "step": 27150 }, { "epoch": 1.822187174927016, "grad_norm": 1.1045299768447876, "learning_rate": 2.056364617029366e-06, "loss": 0.5038, "step": 27152 }, { "epoch": 1.822321398610785, "grad_norm": 1.1224522590637207, "learning_rate": 2.0532809281603062e-06, "loss": 0.4549, "step": 27154 }, { "epoch": 1.8224556222945538, "grad_norm": 1.1844680309295654, "learning_rate": 2.050199504679434e-06, "loss": 0.4976, "step": 27156 }, { "epoch": 1.8225898459783227, "grad_norm": 0.9185380935668945, "learning_rate": 2.0471203467323398e-06, "loss": 0.4312, "step": 27158 }, { "epoch": 1.8227240696620919, "grad_norm": 0.8855293989181519, "learning_rate": 2.0440434544645227e-06, "loss": 0.435, "step": 27160 }, { "epoch": 1.822858293345861, "grad_norm": 1.1739426851272583, "learning_rate": 2.0409688280213503e-06, "loss": 0.4522, "step": 27162 }, { "epoch": 1.82299251702963, "grad_norm": 1.1525919437408447, "learning_rate": 2.037896467548095e-06, "loss": 0.537, "step": 27164 }, { "epoch": 1.8231267407133989, "grad_norm": 0.9660004377365112, "learning_rate": 2.0348263731899076e-06, "loss": 0.4421, "step": 27166 }, { "epoch": 1.8232609643971678, "grad_norm": 1.1061416864395142, "learning_rate": 2.0317585450918663e-06, "loss": 0.5127, "step": 27168 }, { "epoch": 1.823395188080937, "grad_norm": 1.0320607423782349, "learning_rate": 2.0286929833988943e-06, "loss": 0.4894, "step": 27170 }, { "epoch": 1.8235294117647058, "grad_norm": 0.9689834713935852, "learning_rate": 2.025629688255859e-06, "loss": 0.4341, "step": 27172 }, { "epoch": 1.823663635448475, "grad_norm": 1.2229254245758057, "learning_rate": 2.0225686598074835e-06, "loss": 0.4455, "step": 27174 }, { "epoch": 1.823797859132244, "grad_norm": 0.9925333857536316, "learning_rate": 2.0195098981983963e-06, "loss": 0.4776, "step": 27176 }, { "epoch": 1.8239320828160128, "grad_norm": 1.0733305215835571, "learning_rate": 2.01645340357311e-06, "loss": 0.4905, "step": 27178 }, { "epoch": 1.8240663064997817, "grad_norm": 0.9591812491416931, "learning_rate": 2.0133991760760475e-06, "loss": 0.451, "step": 27180 }, { "epoch": 1.8242005301835509, "grad_norm": 0.9299895763397217, "learning_rate": 2.0103472158515158e-06, "loss": 0.4195, "step": 27182 }, { "epoch": 1.82433475386732, "grad_norm": 0.9533336758613586, "learning_rate": 2.007297523043711e-06, "loss": 0.4337, "step": 27184 }, { "epoch": 1.824468977551089, "grad_norm": 0.9015142321586609, "learning_rate": 2.0042500977967172e-06, "loss": 0.4188, "step": 27186 }, { "epoch": 1.8246032012348579, "grad_norm": 1.0596815347671509, "learning_rate": 2.0012049402545306e-06, "loss": 0.4273, "step": 27188 }, { "epoch": 1.8247374249186268, "grad_norm": 1.067183256149292, "learning_rate": 1.9981620505610255e-06, "loss": 0.4149, "step": 27190 }, { "epoch": 1.824871648602396, "grad_norm": 1.2220438718795776, "learning_rate": 1.995121428859975e-06, "loss": 0.5013, "step": 27192 }, { "epoch": 1.8250058722861648, "grad_norm": 1.0703747272491455, "learning_rate": 1.992083075295037e-06, "loss": 0.4664, "step": 27194 }, { "epoch": 1.825140095969934, "grad_norm": 1.1133596897125244, "learning_rate": 1.989046990009763e-06, "loss": 0.4828, "step": 27196 }, { "epoch": 1.825274319653703, "grad_norm": 1.0013623237609863, "learning_rate": 1.9860131731476217e-06, "loss": 0.4463, "step": 27198 }, { "epoch": 1.8254085433374718, "grad_norm": 1.1642730236053467, "learning_rate": 1.982981624851932e-06, "loss": 0.45, "step": 27200 }, { "epoch": 1.8255427670212407, "grad_norm": 1.0820966958999634, "learning_rate": 1.979952345265951e-06, "loss": 0.4992, "step": 27202 }, { "epoch": 1.8256769907050099, "grad_norm": 1.0612512826919556, "learning_rate": 1.976925334532781e-06, "loss": 0.4568, "step": 27204 }, { "epoch": 1.825811214388779, "grad_norm": 1.078147292137146, "learning_rate": 1.9739005927954633e-06, "loss": 0.4685, "step": 27206 }, { "epoch": 1.825945438072548, "grad_norm": 1.1312330961227417, "learning_rate": 1.9708781201968952e-06, "loss": 0.5071, "step": 27208 }, { "epoch": 1.8260796617563169, "grad_norm": 1.0434460639953613, "learning_rate": 1.9678579168799004e-06, "loss": 0.4322, "step": 27210 }, { "epoch": 1.8262138854400858, "grad_norm": 1.106642723083496, "learning_rate": 1.9648399829871654e-06, "loss": 0.453, "step": 27212 }, { "epoch": 1.826348109123855, "grad_norm": 1.0751092433929443, "learning_rate": 1.9618243186612815e-06, "loss": 0.4817, "step": 27214 }, { "epoch": 1.826482332807624, "grad_norm": 0.8817471265792847, "learning_rate": 1.958810924044735e-06, "loss": 0.4875, "step": 27216 }, { "epoch": 1.826616556491393, "grad_norm": 0.8890557289123535, "learning_rate": 1.955799799279906e-06, "loss": 0.4036, "step": 27218 }, { "epoch": 1.826750780175162, "grad_norm": 0.9986615180969238, "learning_rate": 1.952790944509064e-06, "loss": 0.458, "step": 27220 }, { "epoch": 1.8268850038589308, "grad_norm": 0.9373031258583069, "learning_rate": 1.949784359874368e-06, "loss": 0.4374, "step": 27222 }, { "epoch": 1.8270192275427, "grad_norm": 1.0496248006820679, "learning_rate": 1.9467800455178654e-06, "loss": 0.497, "step": 27224 }, { "epoch": 1.8271534512264689, "grad_norm": 1.1269254684448242, "learning_rate": 1.9437780015815254e-06, "loss": 0.5425, "step": 27226 }, { "epoch": 1.827287674910238, "grad_norm": 1.1114847660064697, "learning_rate": 1.940778228207163e-06, "loss": 0.4086, "step": 27228 }, { "epoch": 1.827421898594007, "grad_norm": 1.167726993560791, "learning_rate": 1.937780725536542e-06, "loss": 0.4717, "step": 27230 }, { "epoch": 1.8275561222777759, "grad_norm": 0.7843297123908997, "learning_rate": 1.9347854937112606e-06, "loss": 0.4127, "step": 27232 }, { "epoch": 1.8276903459615448, "grad_norm": 0.9357795119285583, "learning_rate": 1.9317925328728504e-06, "loss": 0.466, "step": 27234 }, { "epoch": 1.827824569645314, "grad_norm": 0.9309419393539429, "learning_rate": 1.9288018431627143e-06, "loss": 0.3674, "step": 27236 }, { "epoch": 1.827958793329083, "grad_norm": 1.0073319673538208, "learning_rate": 1.9258134247221725e-06, "loss": 0.5228, "step": 27238 }, { "epoch": 1.828093017012852, "grad_norm": 1.0502758026123047, "learning_rate": 1.922827277692413e-06, "loss": 0.4475, "step": 27240 }, { "epoch": 1.828227240696621, "grad_norm": 1.1033629179000854, "learning_rate": 1.9198434022145216e-06, "loss": 0.4783, "step": 27242 }, { "epoch": 1.8283614643803898, "grad_norm": 1.001257300376892, "learning_rate": 1.916861798429481e-06, "loss": 0.5018, "step": 27244 }, { "epoch": 1.828495688064159, "grad_norm": 1.077662706375122, "learning_rate": 1.913882466478173e-06, "loss": 0.4509, "step": 27246 }, { "epoch": 1.8286299117479279, "grad_norm": 1.1423128843307495, "learning_rate": 1.910905406501362e-06, "loss": 0.5072, "step": 27248 }, { "epoch": 1.828764135431697, "grad_norm": 1.0833154916763306, "learning_rate": 1.907930618639703e-06, "loss": 0.5248, "step": 27250 }, { "epoch": 1.828898359115466, "grad_norm": 1.0415693521499634, "learning_rate": 1.9049581030337548e-06, "loss": 0.4609, "step": 27252 }, { "epoch": 1.8290325827992349, "grad_norm": 1.0075165033340454, "learning_rate": 1.9019878598239615e-06, "loss": 0.4583, "step": 27254 }, { "epoch": 1.8291668064830038, "grad_norm": 0.9416204690933228, "learning_rate": 1.8990198891506605e-06, "loss": 0.4052, "step": 27256 }, { "epoch": 1.829301030166773, "grad_norm": 1.0448124408721924, "learning_rate": 1.8960541911540896e-06, "loss": 0.4682, "step": 27258 }, { "epoch": 1.829435253850542, "grad_norm": 1.0680210590362549, "learning_rate": 1.8930907659743646e-06, "loss": 0.4813, "step": 27260 }, { "epoch": 1.829569477534311, "grad_norm": 0.9063601493835449, "learning_rate": 1.89012961375149e-06, "loss": 0.5119, "step": 27262 }, { "epoch": 1.82970370121808, "grad_norm": 1.0432617664337158, "learning_rate": 1.8871707346253986e-06, "loss": 0.4449, "step": 27264 }, { "epoch": 1.8298379249018488, "grad_norm": 0.9900143146514893, "learning_rate": 1.8842141287358727e-06, "loss": 0.4451, "step": 27266 }, { "epoch": 1.829972148585618, "grad_norm": 1.103989839553833, "learning_rate": 1.8812597962226286e-06, "loss": 0.4721, "step": 27268 }, { "epoch": 1.8301063722693869, "grad_norm": 1.019983172416687, "learning_rate": 1.878307737225221e-06, "loss": 0.5002, "step": 27270 }, { "epoch": 1.830240595953156, "grad_norm": 1.1201612949371338, "learning_rate": 1.8753579518831554e-06, "loss": 0.5069, "step": 27272 }, { "epoch": 1.830374819636925, "grad_norm": 1.0612053871154785, "learning_rate": 1.8724104403357868e-06, "loss": 0.4793, "step": 27274 }, { "epoch": 1.8305090433206939, "grad_norm": 0.7627202272415161, "learning_rate": 1.8694652027223925e-06, "loss": 0.4046, "step": 27276 }, { "epoch": 1.8306432670044628, "grad_norm": 0.9914665222167969, "learning_rate": 1.8665222391821169e-06, "loss": 0.4982, "step": 27278 }, { "epoch": 1.830777490688232, "grad_norm": 0.8946560621261597, "learning_rate": 1.863581549854021e-06, "loss": 0.4352, "step": 27280 }, { "epoch": 1.830911714372001, "grad_norm": 1.0527199506759644, "learning_rate": 1.8606431348770325e-06, "loss": 0.4364, "step": 27282 }, { "epoch": 1.83104593805577, "grad_norm": 1.009649634361267, "learning_rate": 1.8577069943900017e-06, "loss": 0.4558, "step": 27284 }, { "epoch": 1.831180161739539, "grad_norm": 1.0558478832244873, "learning_rate": 1.854773128531645e-06, "loss": 0.4794, "step": 27286 }, { "epoch": 1.8313143854233078, "grad_norm": 0.9969227910041809, "learning_rate": 1.851841537440585e-06, "loss": 0.4659, "step": 27288 }, { "epoch": 1.831448609107077, "grad_norm": 1.1174561977386475, "learning_rate": 1.8489122212553279e-06, "loss": 0.4979, "step": 27290 }, { "epoch": 1.831582832790846, "grad_norm": 1.0655980110168457, "learning_rate": 1.8459851801142902e-06, "loss": 0.4702, "step": 27292 }, { "epoch": 1.831717056474615, "grad_norm": 1.2060457468032837, "learning_rate": 1.8430604141557562e-06, "loss": 0.4752, "step": 27294 }, { "epoch": 1.831851280158384, "grad_norm": 0.9956331253051758, "learning_rate": 1.8401379235179316e-06, "loss": 0.4351, "step": 27296 }, { "epoch": 1.8319855038421529, "grad_norm": 1.0068447589874268, "learning_rate": 1.8372177083388786e-06, "loss": 0.4986, "step": 27298 }, { "epoch": 1.832119727525922, "grad_norm": 1.085825800895691, "learning_rate": 1.834299768756581e-06, "loss": 0.5099, "step": 27300 }, { "epoch": 1.832253951209691, "grad_norm": 1.0742210149765015, "learning_rate": 1.8313841049089065e-06, "loss": 0.4737, "step": 27302 }, { "epoch": 1.83238817489346, "grad_norm": 1.0905416011810303, "learning_rate": 1.8284707169336169e-06, "loss": 0.4125, "step": 27304 }, { "epoch": 1.832522398577229, "grad_norm": 1.0812736749649048, "learning_rate": 1.8255596049683576e-06, "loss": 0.4896, "step": 27306 }, { "epoch": 1.832656622260998, "grad_norm": 1.1849606037139893, "learning_rate": 1.82265076915068e-06, "loss": 0.4589, "step": 27308 }, { "epoch": 1.8327908459447668, "grad_norm": 1.093017339706421, "learning_rate": 1.8197442096180072e-06, "loss": 0.4888, "step": 27310 }, { "epoch": 1.832925069628536, "grad_norm": 1.0499763488769531, "learning_rate": 1.8168399265076852e-06, "loss": 0.4531, "step": 27312 }, { "epoch": 1.833059293312305, "grad_norm": 0.9453720450401306, "learning_rate": 1.8139379199569318e-06, "loss": 0.4253, "step": 27314 }, { "epoch": 1.833193516996074, "grad_norm": 1.0987991094589233, "learning_rate": 1.8110381901028539e-06, "loss": 0.44, "step": 27316 }, { "epoch": 1.833327740679843, "grad_norm": 1.2033222913742065, "learning_rate": 1.8081407370824588e-06, "loss": 0.4005, "step": 27318 }, { "epoch": 1.8334619643636119, "grad_norm": 1.0011298656463623, "learning_rate": 1.8052455610326534e-06, "loss": 0.4492, "step": 27320 }, { "epoch": 1.833596188047381, "grad_norm": 1.1182427406311035, "learning_rate": 1.8023526620902176e-06, "loss": 0.4544, "step": 27322 }, { "epoch": 1.83373041173115, "grad_norm": 1.1529039144515991, "learning_rate": 1.7994620403918527e-06, "loss": 0.4499, "step": 27324 }, { "epoch": 1.833864635414919, "grad_norm": 1.0249526500701904, "learning_rate": 1.7965736960741164e-06, "loss": 0.4595, "step": 27326 }, { "epoch": 1.833998859098688, "grad_norm": 1.0299060344696045, "learning_rate": 1.7936876292734883e-06, "loss": 0.4258, "step": 27328 }, { "epoch": 1.834133082782457, "grad_norm": 1.0333058834075928, "learning_rate": 1.790803840126326e-06, "loss": 0.5368, "step": 27330 }, { "epoch": 1.8342673064662258, "grad_norm": 0.9318089485168457, "learning_rate": 1.7879223287688761e-06, "loss": 0.4958, "step": 27332 }, { "epoch": 1.834401530149995, "grad_norm": 0.9352691769599915, "learning_rate": 1.7850430953372966e-06, "loss": 0.4329, "step": 27334 }, { "epoch": 1.834535753833764, "grad_norm": 0.9868640899658203, "learning_rate": 1.7821661399676226e-06, "loss": 0.4515, "step": 27336 }, { "epoch": 1.834669977517533, "grad_norm": 1.107014536857605, "learning_rate": 1.779291462795779e-06, "loss": 0.4447, "step": 27338 }, { "epoch": 1.834804201201302, "grad_norm": 1.046768069267273, "learning_rate": 1.7764190639575907e-06, "loss": 0.4571, "step": 27340 }, { "epoch": 1.8349384248850709, "grad_norm": 1.2242913246154785, "learning_rate": 1.7735489435887764e-06, "loss": 0.4358, "step": 27342 }, { "epoch": 1.83507264856884, "grad_norm": 1.113647699356079, "learning_rate": 1.7706811018249447e-06, "loss": 0.521, "step": 27344 }, { "epoch": 1.835206872252609, "grad_norm": 1.1053078174591064, "learning_rate": 1.7678155388015872e-06, "loss": 0.4808, "step": 27346 }, { "epoch": 1.835341095936378, "grad_norm": 1.020797610282898, "learning_rate": 1.764952254654101e-06, "loss": 0.4903, "step": 27348 }, { "epoch": 1.835475319620147, "grad_norm": 1.1306965351104736, "learning_rate": 1.7620912495177721e-06, "loss": 0.419, "step": 27350 }, { "epoch": 1.835609543303916, "grad_norm": 1.0897547006607056, "learning_rate": 1.7592325235277762e-06, "loss": 0.4437, "step": 27352 }, { "epoch": 1.8357437669876848, "grad_norm": 1.2764606475830078, "learning_rate": 1.7563760768191828e-06, "loss": 0.5219, "step": 27354 }, { "epoch": 1.835877990671454, "grad_norm": 0.944107711315155, "learning_rate": 1.7535219095269451e-06, "loss": 0.4352, "step": 27356 }, { "epoch": 1.836012214355223, "grad_norm": 1.0962316989898682, "learning_rate": 1.7506700217859329e-06, "loss": 0.4248, "step": 27358 }, { "epoch": 1.836146438038992, "grad_norm": 1.0919740200042725, "learning_rate": 1.7478204137308773e-06, "loss": 0.512, "step": 27360 }, { "epoch": 1.836280661722761, "grad_norm": 1.1897571086883545, "learning_rate": 1.744973085496432e-06, "loss": 0.4814, "step": 27362 }, { "epoch": 1.8364148854065299, "grad_norm": 1.0135506391525269, "learning_rate": 1.7421280372171167e-06, "loss": 0.4684, "step": 27364 }, { "epoch": 1.836549109090299, "grad_norm": 1.0432535409927368, "learning_rate": 1.739285269027352e-06, "loss": 0.4677, "step": 27366 }, { "epoch": 1.8366833327740681, "grad_norm": 1.1191753149032593, "learning_rate": 1.736444781061458e-06, "loss": 0.4813, "step": 27368 }, { "epoch": 1.836817556457837, "grad_norm": 1.058532476425171, "learning_rate": 1.7336065734536444e-06, "loss": 0.4499, "step": 27370 }, { "epoch": 1.836951780141606, "grad_norm": 1.0622410774230957, "learning_rate": 1.7307706463380092e-06, "loss": 0.451, "step": 27372 }, { "epoch": 1.837086003825375, "grad_norm": 0.9336556196212769, "learning_rate": 1.727936999848545e-06, "loss": 0.4338, "step": 27374 }, { "epoch": 1.837220227509144, "grad_norm": 0.8977116346359253, "learning_rate": 1.7251056341191285e-06, "loss": 0.4844, "step": 27376 }, { "epoch": 1.837354451192913, "grad_norm": 1.1026514768600464, "learning_rate": 1.7222765492835468e-06, "loss": 0.5071, "step": 27378 }, { "epoch": 1.837488674876682, "grad_norm": 1.0812137126922607, "learning_rate": 1.7194497454754599e-06, "loss": 0.4951, "step": 27380 }, { "epoch": 1.837622898560451, "grad_norm": 0.9849303364753723, "learning_rate": 1.716625222828444e-06, "loss": 0.4815, "step": 27382 }, { "epoch": 1.83775712224422, "grad_norm": 1.177372694015503, "learning_rate": 1.7138029814759316e-06, "loss": 0.4735, "step": 27384 }, { "epoch": 1.8378913459279889, "grad_norm": 1.0768537521362305, "learning_rate": 1.7109830215512767e-06, "loss": 0.45, "step": 27386 }, { "epoch": 1.838025569611758, "grad_norm": 1.1244202852249146, "learning_rate": 1.7081653431877175e-06, "loss": 0.4956, "step": 27388 }, { "epoch": 1.8381597932955271, "grad_norm": 1.0745270252227783, "learning_rate": 1.7053499465183864e-06, "loss": 0.4951, "step": 27390 }, { "epoch": 1.838294016979296, "grad_norm": 0.9530426859855652, "learning_rate": 1.7025368316763047e-06, "loss": 0.4391, "step": 27392 }, { "epoch": 1.838428240663065, "grad_norm": 0.9638584852218628, "learning_rate": 1.699725998794388e-06, "loss": 0.4233, "step": 27394 }, { "epoch": 1.838562464346834, "grad_norm": 0.9115172624588013, "learning_rate": 1.6969174480054305e-06, "loss": 0.4646, "step": 27396 }, { "epoch": 1.838696688030603, "grad_norm": 1.073967456817627, "learning_rate": 1.6941111794421482e-06, "loss": 0.4027, "step": 27398 }, { "epoch": 1.838830911714372, "grad_norm": 1.0610753297805786, "learning_rate": 1.6913071932371184e-06, "loss": 0.5404, "step": 27400 }, { "epoch": 1.838965135398141, "grad_norm": 1.1293768882751465, "learning_rate": 1.688505489522829e-06, "loss": 0.4975, "step": 27402 }, { "epoch": 1.83909935908191, "grad_norm": 1.069830060005188, "learning_rate": 1.6857060684316584e-06, "loss": 0.442, "step": 27404 }, { "epoch": 1.839233582765679, "grad_norm": 1.0448309183120728, "learning_rate": 1.682908930095861e-06, "loss": 0.4767, "step": 27406 }, { "epoch": 1.8393678064494479, "grad_norm": 1.0019116401672363, "learning_rate": 1.6801140746476152e-06, "loss": 0.4503, "step": 27408 }, { "epoch": 1.839502030133217, "grad_norm": 1.1428416967391968, "learning_rate": 1.6773215022189592e-06, "loss": 0.4754, "step": 27410 }, { "epoch": 1.8396362538169861, "grad_norm": 1.1250988245010376, "learning_rate": 1.6745312129418434e-06, "loss": 0.4478, "step": 27412 }, { "epoch": 1.839770477500755, "grad_norm": 1.0673909187316895, "learning_rate": 1.6717432069480897e-06, "loss": 0.4561, "step": 27414 }, { "epoch": 1.839904701184524, "grad_norm": 0.9419642686843872, "learning_rate": 1.6689574843694433e-06, "loss": 0.442, "step": 27416 }, { "epoch": 1.840038924868293, "grad_norm": 1.28042733669281, "learning_rate": 1.6661740453375096e-06, "loss": 0.5354, "step": 27418 }, { "epoch": 1.840173148552062, "grad_norm": 1.0464138984680176, "learning_rate": 1.6633928899838224e-06, "loss": 0.5032, "step": 27420 }, { "epoch": 1.840307372235831, "grad_norm": 0.9780840873718262, "learning_rate": 1.6606140184397546e-06, "loss": 0.4566, "step": 27422 }, { "epoch": 1.8404415959196, "grad_norm": 1.0669537782669067, "learning_rate": 1.6578374308366285e-06, "loss": 0.4608, "step": 27424 }, { "epoch": 1.840575819603369, "grad_norm": 0.8034111261367798, "learning_rate": 1.6550631273056171e-06, "loss": 0.4243, "step": 27426 }, { "epoch": 1.840710043287138, "grad_norm": 1.0807321071624756, "learning_rate": 1.6522911079778158e-06, "loss": 0.4601, "step": 27428 }, { "epoch": 1.8408442669709069, "grad_norm": 0.9290803670883179, "learning_rate": 1.6495213729841808e-06, "loss": 0.443, "step": 27430 }, { "epoch": 1.840978490654676, "grad_norm": 0.8990726470947266, "learning_rate": 1.6467539224555906e-06, "loss": 0.5122, "step": 27432 }, { "epoch": 1.8411127143384451, "grad_norm": 1.0240440368652344, "learning_rate": 1.643988756522785e-06, "loss": 0.4162, "step": 27434 }, { "epoch": 1.841246938022214, "grad_norm": 1.0276577472686768, "learning_rate": 1.6412258753164322e-06, "loss": 0.533, "step": 27436 }, { "epoch": 1.841381161705983, "grad_norm": 0.9784525036811829, "learning_rate": 1.6384652789670608e-06, "loss": 0.5061, "step": 27438 }, { "epoch": 1.841515385389752, "grad_norm": 0.9746937155723572, "learning_rate": 1.6357069676051051e-06, "loss": 0.4498, "step": 27440 }, { "epoch": 1.841649609073521, "grad_norm": 1.075529932975769, "learning_rate": 1.6329509413608834e-06, "loss": 0.4486, "step": 27442 }, { "epoch": 1.8417838327572902, "grad_norm": 1.12813401222229, "learning_rate": 1.6301972003646304e-06, "loss": 0.4737, "step": 27444 }, { "epoch": 1.841918056441059, "grad_norm": 0.9420235753059387, "learning_rate": 1.6274457447464363e-06, "loss": 0.5351, "step": 27446 }, { "epoch": 1.842052280124828, "grad_norm": 0.9531462788581848, "learning_rate": 1.6246965746363197e-06, "loss": 0.4584, "step": 27448 }, { "epoch": 1.842186503808597, "grad_norm": 0.8998656272888184, "learning_rate": 1.6219496901641596e-06, "loss": 0.429, "step": 27450 }, { "epoch": 1.8423207274923659, "grad_norm": 1.0480685234069824, "learning_rate": 1.6192050914597467e-06, "loss": 0.4507, "step": 27452 }, { "epoch": 1.842454951176135, "grad_norm": 1.145783543586731, "learning_rate": 1.6164627786527498e-06, "loss": 0.5235, "step": 27454 }, { "epoch": 1.8425891748599041, "grad_norm": 1.1258286237716675, "learning_rate": 1.6137227518727482e-06, "loss": 0.4591, "step": 27456 }, { "epoch": 1.842723398543673, "grad_norm": 0.9835421442985535, "learning_rate": 1.6109850112492054e-06, "loss": 0.5163, "step": 27458 }, { "epoch": 1.842857622227442, "grad_norm": 1.0142710208892822, "learning_rate": 1.608249556911462e-06, "loss": 0.4324, "step": 27460 }, { "epoch": 1.842991845911211, "grad_norm": 1.022153377532959, "learning_rate": 1.6055163889887647e-06, "loss": 0.5142, "step": 27462 }, { "epoch": 1.84312606959498, "grad_norm": 1.1606864929199219, "learning_rate": 1.6027855076102605e-06, "loss": 0.5038, "step": 27464 }, { "epoch": 1.8432602932787492, "grad_norm": 0.8721981048583984, "learning_rate": 1.6000569129049735e-06, "loss": 0.4326, "step": 27466 }, { "epoch": 1.843394516962518, "grad_norm": 1.065577507019043, "learning_rate": 1.5973306050018178e-06, "loss": 0.4931, "step": 27468 }, { "epoch": 1.843528740646287, "grad_norm": 0.9497195482254028, "learning_rate": 1.5946065840296177e-06, "loss": 0.4524, "step": 27470 }, { "epoch": 1.843662964330056, "grad_norm": 0.9861379265785217, "learning_rate": 1.5918848501170647e-06, "loss": 0.4944, "step": 27472 }, { "epoch": 1.843797188013825, "grad_norm": 0.9308890104293823, "learning_rate": 1.5891654033927617e-06, "loss": 0.4061, "step": 27474 }, { "epoch": 1.843931411697594, "grad_norm": 1.0119004249572754, "learning_rate": 1.5864482439852058e-06, "loss": 0.4442, "step": 27476 }, { "epoch": 1.8440656353813631, "grad_norm": 1.034685492515564, "learning_rate": 1.583733372022761e-06, "loss": 0.4174, "step": 27478 }, { "epoch": 1.844199859065132, "grad_norm": 0.9737145900726318, "learning_rate": 1.581020787633708e-06, "loss": 0.4782, "step": 27480 }, { "epoch": 1.844334082748901, "grad_norm": 0.9639759063720703, "learning_rate": 1.5783104909462166e-06, "loss": 0.3882, "step": 27482 }, { "epoch": 1.84446830643267, "grad_norm": 1.7538946866989136, "learning_rate": 1.5756024820883287e-06, "loss": 0.4969, "step": 27484 }, { "epoch": 1.844602530116439, "grad_norm": 1.1073065996170044, "learning_rate": 1.5728967611880085e-06, "loss": 0.48, "step": 27486 }, { "epoch": 1.8447367538002082, "grad_norm": 0.9614431262016296, "learning_rate": 1.5701933283730875e-06, "loss": 0.453, "step": 27488 }, { "epoch": 1.844870977483977, "grad_norm": 0.9399886727333069, "learning_rate": 1.567492183771302e-06, "loss": 0.5411, "step": 27490 }, { "epoch": 1.845005201167746, "grad_norm": 1.0651518106460571, "learning_rate": 1.564793327510261e-06, "loss": 0.464, "step": 27492 }, { "epoch": 1.845139424851515, "grad_norm": 0.9036216735839844, "learning_rate": 1.5620967597175017e-06, "loss": 0.4539, "step": 27494 }, { "epoch": 1.845273648535284, "grad_norm": 1.1416680812835693, "learning_rate": 1.559402480520422e-06, "loss": 0.5153, "step": 27496 }, { "epoch": 1.845407872219053, "grad_norm": 1.1369332075119019, "learning_rate": 1.5567104900463147e-06, "loss": 0.4874, "step": 27498 }, { "epoch": 1.8455420959028221, "grad_norm": 0.9781507849693298, "learning_rate": 1.554020788422378e-06, "loss": 0.4902, "step": 27500 }, { "epoch": 1.845676319586591, "grad_norm": 1.0934257507324219, "learning_rate": 1.5513333757756987e-06, "loss": 0.4657, "step": 27502 }, { "epoch": 1.84581054327036, "grad_norm": 1.0390064716339111, "learning_rate": 1.5486482522332424e-06, "loss": 0.4627, "step": 27504 }, { "epoch": 1.845944766954129, "grad_norm": 1.1115378141403198, "learning_rate": 1.5459654179218852e-06, "loss": 0.4134, "step": 27506 }, { "epoch": 1.846078990637898, "grad_norm": 1.0079962015151978, "learning_rate": 1.5432848729683758e-06, "loss": 0.4602, "step": 27508 }, { "epoch": 1.8462132143216672, "grad_norm": 0.9150170683860779, "learning_rate": 1.5406066174993739e-06, "loss": 0.4671, "step": 27510 }, { "epoch": 1.846347438005436, "grad_norm": 1.2085232734680176, "learning_rate": 1.5379306516414115e-06, "loss": 0.5562, "step": 27512 }, { "epoch": 1.846481661689205, "grad_norm": 1.037790298461914, "learning_rate": 1.5352569755209378e-06, "loss": 0.479, "step": 27514 }, { "epoch": 1.846615885372974, "grad_norm": 0.9107427000999451, "learning_rate": 1.5325855892642682e-06, "loss": 0.4289, "step": 27516 }, { "epoch": 1.846750109056743, "grad_norm": 1.0650413036346436, "learning_rate": 1.529916492997624e-06, "loss": 0.5249, "step": 27518 }, { "epoch": 1.8468843327405122, "grad_norm": 1.061478853225708, "learning_rate": 1.5272496868471043e-06, "loss": 0.5132, "step": 27520 }, { "epoch": 1.8470185564242811, "grad_norm": 0.9944947361946106, "learning_rate": 1.5245851709387249e-06, "loss": 0.4591, "step": 27522 }, { "epoch": 1.84715278010805, "grad_norm": 0.9647601842880249, "learning_rate": 1.5219229453983796e-06, "loss": 0.4661, "step": 27524 }, { "epoch": 1.847287003791819, "grad_norm": 0.9864873886108398, "learning_rate": 1.5192630103518402e-06, "loss": 0.4881, "step": 27526 }, { "epoch": 1.847421227475588, "grad_norm": 1.0107039213180542, "learning_rate": 1.5166053659247892e-06, "loss": 0.4718, "step": 27528 }, { "epoch": 1.847555451159357, "grad_norm": 1.0673303604125977, "learning_rate": 1.5139500122427986e-06, "loss": 0.4393, "step": 27530 }, { "epoch": 1.8476896748431262, "grad_norm": 0.9937414526939392, "learning_rate": 1.511296949431329e-06, "loss": 0.4249, "step": 27532 }, { "epoch": 1.847823898526895, "grad_norm": 1.0674530267715454, "learning_rate": 1.5086461776157245e-06, "loss": 0.5207, "step": 27534 }, { "epoch": 1.847958122210664, "grad_norm": 0.8959336280822754, "learning_rate": 1.5059976969212408e-06, "loss": 0.4826, "step": 27536 }, { "epoch": 1.848092345894433, "grad_norm": 0.9730433225631714, "learning_rate": 1.5033515074729942e-06, "loss": 0.4278, "step": 27538 }, { "epoch": 1.848226569578202, "grad_norm": 1.0837477445602417, "learning_rate": 1.5007076093960348e-06, "loss": 0.4376, "step": 27540 }, { "epoch": 1.8483607932619712, "grad_norm": 1.0965659618377686, "learning_rate": 1.4980660028152627e-06, "loss": 0.4846, "step": 27542 }, { "epoch": 1.8484950169457401, "grad_norm": 0.9107245802879333, "learning_rate": 1.4954266878555112e-06, "loss": 0.4312, "step": 27544 }, { "epoch": 1.848629240629509, "grad_norm": 1.0831120014190674, "learning_rate": 1.4927896646414529e-06, "loss": 0.4578, "step": 27546 }, { "epoch": 1.848763464313278, "grad_norm": 1.0490244626998901, "learning_rate": 1.4901549332977105e-06, "loss": 0.4303, "step": 27548 }, { "epoch": 1.8488976879970471, "grad_norm": 1.0324196815490723, "learning_rate": 1.4875224939487454e-06, "loss": 0.3801, "step": 27550 }, { "epoch": 1.849031911680816, "grad_norm": 1.0803989171981812, "learning_rate": 1.4848923467189524e-06, "loss": 0.4548, "step": 27552 }, { "epoch": 1.8491661353645852, "grad_norm": 1.1710008382797241, "learning_rate": 1.482264491732599e-06, "loss": 0.4808, "step": 27554 }, { "epoch": 1.849300359048354, "grad_norm": 1.099749207496643, "learning_rate": 1.4796389291138358e-06, "loss": 0.4845, "step": 27556 }, { "epoch": 1.849434582732123, "grad_norm": 0.9039605259895325, "learning_rate": 1.4770156589867246e-06, "loss": 0.4377, "step": 27558 }, { "epoch": 1.849568806415892, "grad_norm": 1.1397470235824585, "learning_rate": 1.4743946814752053e-06, "loss": 0.5262, "step": 27560 }, { "epoch": 1.849703030099661, "grad_norm": 0.8652995228767395, "learning_rate": 1.471775996703123e-06, "loss": 0.4527, "step": 27562 }, { "epoch": 1.8498372537834302, "grad_norm": 1.0195953845977783, "learning_rate": 1.4691596047941902e-06, "loss": 0.4406, "step": 27564 }, { "epoch": 1.8499714774671991, "grad_norm": 1.1417049169540405, "learning_rate": 1.4665455058720357e-06, "loss": 0.4805, "step": 27566 }, { "epoch": 1.850105701150968, "grad_norm": 0.9819300770759583, "learning_rate": 1.4639337000601717e-06, "loss": 0.4426, "step": 27568 }, { "epoch": 1.850239924834737, "grad_norm": 1.0256808996200562, "learning_rate": 1.4613241874819938e-06, "loss": 0.493, "step": 27570 }, { "epoch": 1.8503741485185061, "grad_norm": 1.0297558307647705, "learning_rate": 1.4587169682608093e-06, "loss": 0.5481, "step": 27572 }, { "epoch": 1.850508372202275, "grad_norm": 1.0053192377090454, "learning_rate": 1.456112042519786e-06, "loss": 0.4182, "step": 27574 }, { "epoch": 1.8506425958860442, "grad_norm": 1.1509085893630981, "learning_rate": 1.45350941038202e-06, "loss": 0.4912, "step": 27576 }, { "epoch": 1.850776819569813, "grad_norm": 1.0100876092910767, "learning_rate": 1.4509090719704631e-06, "loss": 0.4583, "step": 27578 }, { "epoch": 1.850911043253582, "grad_norm": 0.8980428576469421, "learning_rate": 1.4483110274079891e-06, "loss": 0.415, "step": 27580 }, { "epoch": 1.851045266937351, "grad_norm": 1.1096354722976685, "learning_rate": 1.4457152768173444e-06, "loss": 0.4795, "step": 27582 }, { "epoch": 1.85117949062112, "grad_norm": 1.0847545862197876, "learning_rate": 1.4431218203211804e-06, "loss": 0.4038, "step": 27584 }, { "epoch": 1.8513137143048892, "grad_norm": 1.030122995376587, "learning_rate": 1.440530658042022e-06, "loss": 0.4204, "step": 27586 }, { "epoch": 1.8514479379886581, "grad_norm": 1.0669091939926147, "learning_rate": 1.437941790102304e-06, "loss": 0.483, "step": 27588 }, { "epoch": 1.851582161672427, "grad_norm": 1.0459216833114624, "learning_rate": 1.4353552166243457e-06, "loss": 0.4687, "step": 27590 }, { "epoch": 1.851716385356196, "grad_norm": 1.1613476276397705, "learning_rate": 1.4327709377303544e-06, "loss": 0.4149, "step": 27592 }, { "epoch": 1.8518506090399651, "grad_norm": 1.0539007186889648, "learning_rate": 1.4301889535424218e-06, "loss": 0.4758, "step": 27594 }, { "epoch": 1.8519848327237343, "grad_norm": 1.0329196453094482, "learning_rate": 1.4276092641825668e-06, "loss": 0.4155, "step": 27596 }, { "epoch": 1.8521190564075032, "grad_norm": 1.1315053701400757, "learning_rate": 1.4250318697726529e-06, "loss": 0.4739, "step": 27598 }, { "epoch": 1.852253280091272, "grad_norm": 1.0232601165771484, "learning_rate": 1.4224567704344716e-06, "loss": 0.4584, "step": 27600 }, { "epoch": 1.852387503775041, "grad_norm": 1.107422947883606, "learning_rate": 1.4198839662896758e-06, "loss": 0.5413, "step": 27602 }, { "epoch": 1.85252172745881, "grad_norm": 1.064710021018982, "learning_rate": 1.4173134574598402e-06, "loss": 0.4463, "step": 27604 }, { "epoch": 1.852655951142579, "grad_norm": 1.0158103704452515, "learning_rate": 1.4147452440664065e-06, "loss": 0.4464, "step": 27606 }, { "epoch": 1.8527901748263482, "grad_norm": 1.0038433074951172, "learning_rate": 1.4121793262307226e-06, "loss": 0.5025, "step": 27608 }, { "epoch": 1.8529243985101171, "grad_norm": 1.045138955116272, "learning_rate": 1.4096157040740244e-06, "loss": 0.4454, "step": 27610 }, { "epoch": 1.853058622193886, "grad_norm": 1.006141185760498, "learning_rate": 1.4070543777174317e-06, "loss": 0.4525, "step": 27612 }, { "epoch": 1.853192845877655, "grad_norm": 1.0381324291229248, "learning_rate": 1.4044953472819644e-06, "loss": 0.4833, "step": 27614 }, { "epoch": 1.8533270695614241, "grad_norm": 1.0100772380828857, "learning_rate": 1.4019386128885314e-06, "loss": 0.4649, "step": 27616 }, { "epoch": 1.8534612932451933, "grad_norm": 1.0271670818328857, "learning_rate": 1.3993841746579416e-06, "loss": 0.4925, "step": 27618 }, { "epoch": 1.8535955169289622, "grad_norm": 1.0052720308303833, "learning_rate": 1.396832032710882e-06, "loss": 0.4151, "step": 27620 }, { "epoch": 1.853729740612731, "grad_norm": 0.9382982850074768, "learning_rate": 1.3942821871679279e-06, "loss": 0.5207, "step": 27622 }, { "epoch": 1.8538639642965, "grad_norm": 1.1375670433044434, "learning_rate": 1.391734638149561e-06, "loss": 0.4553, "step": 27624 }, { "epoch": 1.8539981879802692, "grad_norm": 1.0256612300872803, "learning_rate": 1.3891893857761519e-06, "loss": 0.4518, "step": 27626 }, { "epoch": 1.854132411664038, "grad_norm": 1.0710864067077637, "learning_rate": 1.3866464301679593e-06, "loss": 0.5361, "step": 27628 }, { "epoch": 1.8542666353478072, "grad_norm": 1.1199383735656738, "learning_rate": 1.3841057714451212e-06, "loss": 0.5286, "step": 27630 }, { "epoch": 1.8544008590315761, "grad_norm": 1.0359209775924683, "learning_rate": 1.3815674097276854e-06, "loss": 0.4916, "step": 27632 }, { "epoch": 1.854535082715345, "grad_norm": 0.8993666768074036, "learning_rate": 1.379031345135595e-06, "loss": 0.5123, "step": 27634 }, { "epoch": 1.854669306399114, "grad_norm": 1.0495448112487793, "learning_rate": 1.3764975777886547e-06, "loss": 0.4336, "step": 27636 }, { "epoch": 1.8548035300828831, "grad_norm": 0.7166053056716919, "learning_rate": 1.3739661078065957e-06, "loss": 0.4121, "step": 27638 }, { "epoch": 1.8549377537666523, "grad_norm": 1.0230076313018799, "learning_rate": 1.3714369353090173e-06, "loss": 0.4041, "step": 27640 }, { "epoch": 1.8550719774504212, "grad_norm": 1.0174587965011597, "learning_rate": 1.3689100604154182e-06, "loss": 0.3957, "step": 27642 }, { "epoch": 1.85520620113419, "grad_norm": 1.0715550184249878, "learning_rate": 1.3663854832451916e-06, "loss": 0.4528, "step": 27644 }, { "epoch": 1.855340424817959, "grad_norm": 0.9875494241714478, "learning_rate": 1.3638632039176147e-06, "loss": 0.5179, "step": 27646 }, { "epoch": 1.8554746485017282, "grad_norm": 0.9951808452606201, "learning_rate": 1.3613432225518696e-06, "loss": 0.4658, "step": 27648 }, { "epoch": 1.855608872185497, "grad_norm": 0.9630551338195801, "learning_rate": 1.3588255392670058e-06, "loss": 0.5425, "step": 27650 }, { "epoch": 1.8557430958692662, "grad_norm": 1.0404685735702515, "learning_rate": 1.3563101541819889e-06, "loss": 0.4919, "step": 27652 }, { "epoch": 1.8558773195530351, "grad_norm": 1.045067548751831, "learning_rate": 1.3537970674156631e-06, "loss": 0.5117, "step": 27654 }, { "epoch": 1.856011543236804, "grad_norm": 1.054994821548462, "learning_rate": 1.351286279086772e-06, "loss": 0.5023, "step": 27656 }, { "epoch": 1.856145766920573, "grad_norm": 1.0288803577423096, "learning_rate": 1.3487777893139374e-06, "loss": 0.5293, "step": 27658 }, { "epoch": 1.8562799906043421, "grad_norm": 0.9933587908744812, "learning_rate": 1.346271598215676e-06, "loss": 0.5119, "step": 27660 }, { "epoch": 1.8564142142881113, "grad_norm": 0.9720452427864075, "learning_rate": 1.3437677059104147e-06, "loss": 0.4397, "step": 27662 }, { "epoch": 1.8565484379718802, "grad_norm": 1.1074039936065674, "learning_rate": 1.3412661125164483e-06, "loss": 0.5094, "step": 27664 }, { "epoch": 1.856682661655649, "grad_norm": 1.0138037204742432, "learning_rate": 1.338766818151982e-06, "loss": 0.4095, "step": 27666 }, { "epoch": 1.856816885339418, "grad_norm": 1.1227648258209229, "learning_rate": 1.3362698229350935e-06, "loss": 0.4334, "step": 27668 }, { "epoch": 1.8569511090231872, "grad_norm": 1.0338616371154785, "learning_rate": 1.3337751269837606e-06, "loss": 0.5013, "step": 27670 }, { "epoch": 1.8570853327069563, "grad_norm": 1.1134554147720337, "learning_rate": 1.331282730415856e-06, "loss": 0.4636, "step": 27672 }, { "epoch": 1.8572195563907252, "grad_norm": 1.0271615982055664, "learning_rate": 1.3287926333491353e-06, "loss": 0.4261, "step": 27674 }, { "epoch": 1.8573537800744941, "grad_norm": 0.9932689666748047, "learning_rate": 1.3263048359012543e-06, "loss": 0.501, "step": 27676 }, { "epoch": 1.857488003758263, "grad_norm": 1.0491055250167847, "learning_rate": 1.3238193381897635e-06, "loss": 0.4436, "step": 27678 }, { "epoch": 1.857622227442032, "grad_norm": 0.8929378390312195, "learning_rate": 1.321336140332091e-06, "loss": 0.4561, "step": 27680 }, { "epoch": 1.8577564511258011, "grad_norm": 2.278045177459717, "learning_rate": 1.3188552424455546e-06, "loss": 0.5313, "step": 27682 }, { "epoch": 1.8578906748095703, "grad_norm": 0.8573141694068909, "learning_rate": 1.316376644647388e-06, "loss": 0.4151, "step": 27684 }, { "epoch": 1.8580248984933392, "grad_norm": 0.8995969891548157, "learning_rate": 1.3139003470546918e-06, "loss": 0.392, "step": 27686 }, { "epoch": 1.858159122177108, "grad_norm": 1.1379754543304443, "learning_rate": 1.3114263497844669e-06, "loss": 0.406, "step": 27688 }, { "epoch": 1.858293345860877, "grad_norm": 1.0205707550048828, "learning_rate": 1.3089546529536034e-06, "loss": 0.4731, "step": 27690 }, { "epoch": 1.8584275695446462, "grad_norm": 1.0011777877807617, "learning_rate": 1.306485256678891e-06, "loss": 0.4883, "step": 27692 }, { "epoch": 1.8585617932284153, "grad_norm": 1.0003676414489746, "learning_rate": 1.3040181610769865e-06, "loss": 0.4204, "step": 27694 }, { "epoch": 1.8586960169121842, "grad_norm": 1.0474743843078613, "learning_rate": 1.3015533662644852e-06, "loss": 0.4178, "step": 27696 }, { "epoch": 1.8588302405959531, "grad_norm": 1.0012686252593994, "learning_rate": 1.299090872357811e-06, "loss": 0.4525, "step": 27698 }, { "epoch": 1.858964464279722, "grad_norm": 0.9653547406196594, "learning_rate": 1.2966306794733318e-06, "loss": 0.4891, "step": 27700 }, { "epoch": 1.8590986879634912, "grad_norm": 0.8683306574821472, "learning_rate": 1.2941727877272825e-06, "loss": 0.4006, "step": 27702 }, { "epoch": 1.8592329116472601, "grad_norm": 1.1569554805755615, "learning_rate": 1.2917171972357922e-06, "loss": 0.5005, "step": 27704 }, { "epoch": 1.8593671353310293, "grad_norm": 1.2232944965362549, "learning_rate": 1.289263908114885e-06, "loss": 0.4743, "step": 27706 }, { "epoch": 1.8595013590147982, "grad_norm": 0.9982020258903503, "learning_rate": 1.2868129204804735e-06, "loss": 0.4907, "step": 27708 }, { "epoch": 1.859635582698567, "grad_norm": 1.1773412227630615, "learning_rate": 1.2843642344483542e-06, "loss": 0.4654, "step": 27710 }, { "epoch": 1.859769806382336, "grad_norm": 1.1091125011444092, "learning_rate": 1.2819178501342343e-06, "loss": 0.4409, "step": 27712 }, { "epoch": 1.8599040300661052, "grad_norm": 1.0534645318984985, "learning_rate": 1.2794737676536994e-06, "loss": 0.5205, "step": 27714 }, { "epoch": 1.8600382537498743, "grad_norm": 0.9437594413757324, "learning_rate": 1.2770319871222236e-06, "loss": 0.4711, "step": 27716 }, { "epoch": 1.8601724774336432, "grad_norm": 0.958147406578064, "learning_rate": 1.2745925086551702e-06, "loss": 0.4477, "step": 27718 }, { "epoch": 1.8603067011174121, "grad_norm": 1.0931216478347778, "learning_rate": 1.2721553323678137e-06, "loss": 0.4576, "step": 27720 }, { "epoch": 1.860440924801181, "grad_norm": 1.0033011436462402, "learning_rate": 1.2697204583752898e-06, "loss": 0.4928, "step": 27722 }, { "epoch": 1.8605751484849502, "grad_norm": 1.105586290359497, "learning_rate": 1.267287886792662e-06, "loss": 0.4718, "step": 27724 }, { "epoch": 1.8607093721687191, "grad_norm": 1.0070650577545166, "learning_rate": 1.2648576177348437e-06, "loss": 0.4843, "step": 27726 }, { "epoch": 1.8608435958524883, "grad_norm": 1.1382700204849243, "learning_rate": 1.2624296513166712e-06, "loss": 0.4592, "step": 27728 }, { "epoch": 1.8609778195362572, "grad_norm": 1.1092818975448608, "learning_rate": 1.260003987652858e-06, "loss": 0.43, "step": 27730 }, { "epoch": 1.861112043220026, "grad_norm": 1.1581449508666992, "learning_rate": 1.2575806268580182e-06, "loss": 0.465, "step": 27732 }, { "epoch": 1.861246266903795, "grad_norm": 0.9862481355667114, "learning_rate": 1.2551595690466434e-06, "loss": 0.4387, "step": 27734 }, { "epoch": 1.8613804905875642, "grad_norm": 1.0038784742355347, "learning_rate": 1.252740814333131e-06, "loss": 0.4707, "step": 27736 }, { "epoch": 1.8615147142713333, "grad_norm": 1.1910805702209473, "learning_rate": 1.250324362831745e-06, "loss": 0.5406, "step": 27738 }, { "epoch": 1.8616489379551022, "grad_norm": 1.095921277999878, "learning_rate": 1.2479102146566834e-06, "loss": 0.4791, "step": 27740 }, { "epoch": 1.8617831616388711, "grad_norm": 1.1198900938034058, "learning_rate": 1.2454983699219936e-06, "loss": 0.4982, "step": 27742 }, { "epoch": 1.86191738532264, "grad_norm": 0.9647154211997986, "learning_rate": 1.2430888287416342e-06, "loss": 0.469, "step": 27744 }, { "epoch": 1.8620516090064092, "grad_norm": 0.9845030903816223, "learning_rate": 1.2406815912294535e-06, "loss": 0.399, "step": 27746 }, { "epoch": 1.8621858326901783, "grad_norm": 1.0828144550323486, "learning_rate": 1.2382766574991766e-06, "loss": 0.4334, "step": 27748 }, { "epoch": 1.8623200563739473, "grad_norm": 1.013677716255188, "learning_rate": 1.235874027664452e-06, "loss": 0.4792, "step": 27750 }, { "epoch": 1.8624542800577162, "grad_norm": 0.9506924748420715, "learning_rate": 1.2334737018387887e-06, "loss": 0.4674, "step": 27752 }, { "epoch": 1.862588503741485, "grad_norm": 1.0657380819320679, "learning_rate": 1.231075680135596e-06, "loss": 0.457, "step": 27754 }, { "epoch": 1.862722727425254, "grad_norm": 0.9487188458442688, "learning_rate": 1.2286799626681721e-06, "loss": 0.463, "step": 27756 }, { "epoch": 1.8628569511090232, "grad_norm": 1.0851175785064697, "learning_rate": 1.226286549549721e-06, "loss": 0.4614, "step": 27758 }, { "epoch": 1.8629911747927923, "grad_norm": 0.949742317199707, "learning_rate": 1.2238954408933134e-06, "loss": 0.4029, "step": 27760 }, { "epoch": 1.8631253984765612, "grad_norm": 0.99049311876297, "learning_rate": 1.2215066368119476e-06, "loss": 0.4526, "step": 27762 }, { "epoch": 1.8632596221603301, "grad_norm": 0.9609589576721191, "learning_rate": 1.2191201374184614e-06, "loss": 0.4424, "step": 27764 }, { "epoch": 1.863393845844099, "grad_norm": 1.0745807886123657, "learning_rate": 1.2167359428256253e-06, "loss": 0.4545, "step": 27766 }, { "epoch": 1.8635280695278682, "grad_norm": 1.0423164367675781, "learning_rate": 1.2143540531460885e-06, "loss": 0.4838, "step": 27768 }, { "epoch": 1.8636622932116373, "grad_norm": 1.070115327835083, "learning_rate": 1.211974468492394e-06, "loss": 0.4114, "step": 27770 }, { "epoch": 1.8637965168954063, "grad_norm": 1.0345423221588135, "learning_rate": 1.2095971889769686e-06, "loss": 0.4636, "step": 27772 }, { "epoch": 1.8639307405791752, "grad_norm": 1.089133381843567, "learning_rate": 1.2072222147121338e-06, "loss": 0.491, "step": 27774 }, { "epoch": 1.864064964262944, "grad_norm": 1.0854707956314087, "learning_rate": 1.2048495458100995e-06, "loss": 0.4679, "step": 27776 }, { "epoch": 1.8641991879467132, "grad_norm": 0.9781652688980103, "learning_rate": 1.2024791823829762e-06, "loss": 0.387, "step": 27778 }, { "epoch": 1.8643334116304822, "grad_norm": 1.0830963850021362, "learning_rate": 1.200111124542752e-06, "loss": 0.5194, "step": 27780 }, { "epoch": 1.8644676353142513, "grad_norm": 1.0142366886138916, "learning_rate": 1.1977453724013154e-06, "loss": 0.4902, "step": 27782 }, { "epoch": 1.8646018589980202, "grad_norm": 1.1209266185760498, "learning_rate": 1.1953819260704436e-06, "loss": 0.5397, "step": 27784 }, { "epoch": 1.8647360826817891, "grad_norm": 1.0151091814041138, "learning_rate": 1.1930207856618137e-06, "loss": 0.4546, "step": 27786 }, { "epoch": 1.864870306365558, "grad_norm": 1.0850834846496582, "learning_rate": 1.1906619512869644e-06, "loss": 0.4781, "step": 27788 }, { "epoch": 1.8650045300493272, "grad_norm": 1.0318243503570557, "learning_rate": 1.1883054230573731e-06, "loss": 0.4647, "step": 27790 }, { "epoch": 1.8651387537330963, "grad_norm": 0.9039753079414368, "learning_rate": 1.1859512010843565e-06, "loss": 0.4377, "step": 27792 }, { "epoch": 1.8652729774168653, "grad_norm": 1.061113953590393, "learning_rate": 1.183599285479159e-06, "loss": 0.4675, "step": 27794 }, { "epoch": 1.8654072011006342, "grad_norm": 1.1509170532226562, "learning_rate": 1.1812496763528968e-06, "loss": 0.5651, "step": 27796 }, { "epoch": 1.865541424784403, "grad_norm": 1.0290863513946533, "learning_rate": 1.1789023738165983e-06, "loss": 0.4957, "step": 27798 }, { "epoch": 1.8656756484681722, "grad_norm": 0.8547778725624084, "learning_rate": 1.1765573779811578e-06, "loss": 0.402, "step": 27800 }, { "epoch": 1.8658098721519412, "grad_norm": 1.064781904220581, "learning_rate": 1.1742146889573758e-06, "loss": 0.4567, "step": 27802 }, { "epoch": 1.8659440958357103, "grad_norm": 1.036177635192871, "learning_rate": 1.17187430685593e-06, "loss": 0.4933, "step": 27804 }, { "epoch": 1.8660783195194792, "grad_norm": 1.0641292333602905, "learning_rate": 1.1695362317874158e-06, "loss": 0.512, "step": 27806 }, { "epoch": 1.8662125432032481, "grad_norm": 1.0304546356201172, "learning_rate": 1.1672004638622892e-06, "loss": 0.4732, "step": 27808 }, { "epoch": 1.866346766887017, "grad_norm": 1.0697804689407349, "learning_rate": 1.164867003190917e-06, "loss": 0.4858, "step": 27810 }, { "epoch": 1.8664809905707862, "grad_norm": 1.2139830589294434, "learning_rate": 1.1625358498835505e-06, "loss": 0.5135, "step": 27812 }, { "epoch": 1.8666152142545553, "grad_norm": 1.0166290998458862, "learning_rate": 1.1602070040503232e-06, "loss": 0.4203, "step": 27814 }, { "epoch": 1.8667494379383243, "grad_norm": 0.9351518750190735, "learning_rate": 1.157880465801281e-06, "loss": 0.4898, "step": 27816 }, { "epoch": 1.8668836616220932, "grad_norm": 0.9220516681671143, "learning_rate": 1.155556235246341e-06, "loss": 0.3922, "step": 27818 }, { "epoch": 1.867017885305862, "grad_norm": 0.9939961433410645, "learning_rate": 1.1532343124953216e-06, "loss": 0.4737, "step": 27820 }, { "epoch": 1.8671521089896312, "grad_norm": 1.5443549156188965, "learning_rate": 1.1509146976579232e-06, "loss": 0.4162, "step": 27822 }, { "epoch": 1.8672863326734004, "grad_norm": 1.0167597532272339, "learning_rate": 1.148597390843753e-06, "loss": 0.3998, "step": 27824 }, { "epoch": 1.8674205563571693, "grad_norm": 1.2478851079940796, "learning_rate": 1.14628239216229e-06, "loss": 0.5228, "step": 27826 }, { "epoch": 1.8675547800409382, "grad_norm": 0.9998477101325989, "learning_rate": 1.143969701722919e-06, "loss": 0.4396, "step": 27828 }, { "epoch": 1.8676890037247071, "grad_norm": 1.055120587348938, "learning_rate": 1.1416593196349134e-06, "loss": 0.4476, "step": 27830 }, { "epoch": 1.867823227408476, "grad_norm": 1.0692424774169922, "learning_rate": 1.139351246007425e-06, "loss": 0.4335, "step": 27832 }, { "epoch": 1.8679574510922452, "grad_norm": 0.9727033376693726, "learning_rate": 1.1370454809495056e-06, "loss": 0.4036, "step": 27834 }, { "epoch": 1.8680916747760143, "grad_norm": 0.8969436287879944, "learning_rate": 1.1347420245701068e-06, "loss": 0.45, "step": 27836 }, { "epoch": 1.8682258984597833, "grad_norm": 0.9650223255157471, "learning_rate": 1.1324408769780636e-06, "loss": 0.4541, "step": 27838 }, { "epoch": 1.8683601221435522, "grad_norm": 1.0078452825546265, "learning_rate": 1.1301420382820893e-06, "loss": 0.5195, "step": 27840 }, { "epoch": 1.868494345827321, "grad_norm": 1.0130294561386108, "learning_rate": 1.1278455085908025e-06, "loss": 0.4784, "step": 27842 }, { "epoch": 1.8686285695110902, "grad_norm": 1.0932484865188599, "learning_rate": 1.125551288012716e-06, "loss": 0.4415, "step": 27844 }, { "epoch": 1.8687627931948594, "grad_norm": 1.096427321434021, "learning_rate": 1.123259376656216e-06, "loss": 0.4845, "step": 27846 }, { "epoch": 1.8688970168786283, "grad_norm": 0.9929922223091125, "learning_rate": 1.12096977462961e-06, "loss": 0.4352, "step": 27848 }, { "epoch": 1.8690312405623972, "grad_norm": 1.0440667867660522, "learning_rate": 1.1186824820410614e-06, "loss": 0.4269, "step": 27850 }, { "epoch": 1.8691654642461661, "grad_norm": 0.9539794325828552, "learning_rate": 1.1163974989986447e-06, "loss": 0.449, "step": 27852 }, { "epoch": 1.8692996879299353, "grad_norm": 1.1850839853286743, "learning_rate": 1.1141148256103128e-06, "loss": 0.5031, "step": 27854 }, { "epoch": 1.8694339116137042, "grad_norm": 1.045531988143921, "learning_rate": 1.111834461983935e-06, "loss": 0.5045, "step": 27856 }, { "epoch": 1.8695681352974733, "grad_norm": 1.4767205715179443, "learning_rate": 1.1095564082272469e-06, "loss": 0.4553, "step": 27858 }, { "epoch": 1.8697023589812423, "grad_norm": 1.1055418252944946, "learning_rate": 1.1072806644478739e-06, "loss": 0.4786, "step": 27860 }, { "epoch": 1.8698365826650112, "grad_norm": 0.6489827036857605, "learning_rate": 1.105007230753341e-06, "loss": 0.4137, "step": 27862 }, { "epoch": 1.86997080634878, "grad_norm": 1.0843795537948608, "learning_rate": 1.1027361072510788e-06, "loss": 0.4643, "step": 27864 }, { "epoch": 1.8701050300325492, "grad_norm": 1.3519821166992188, "learning_rate": 1.1004672940483796e-06, "loss": 0.5044, "step": 27866 }, { "epoch": 1.8702392537163184, "grad_norm": 0.9754189848899841, "learning_rate": 1.0982007912524405e-06, "loss": 0.3895, "step": 27868 }, { "epoch": 1.8703734774000873, "grad_norm": 0.9900445938110352, "learning_rate": 1.0959365989703541e-06, "loss": 0.4306, "step": 27870 }, { "epoch": 1.8705077010838562, "grad_norm": 1.0068804025650024, "learning_rate": 1.093674717309101e-06, "loss": 0.4488, "step": 27872 }, { "epoch": 1.8706419247676251, "grad_norm": 0.9073607921600342, "learning_rate": 1.0914151463755407e-06, "loss": 0.4347, "step": 27874 }, { "epoch": 1.8707761484513943, "grad_norm": 1.0566726922988892, "learning_rate": 1.089157886276454e-06, "loss": 0.4534, "step": 27876 }, { "epoch": 1.8709103721351632, "grad_norm": 0.9509255886077881, "learning_rate": 1.086902937118467e-06, "loss": 0.4083, "step": 27878 }, { "epoch": 1.8710445958189323, "grad_norm": 1.119420051574707, "learning_rate": 1.084650299008133e-06, "loss": 0.4667, "step": 27880 }, { "epoch": 1.8711788195027013, "grad_norm": 1.0077191591262817, "learning_rate": 1.082399972051884e-06, "loss": 0.4978, "step": 27882 }, { "epoch": 1.8713130431864702, "grad_norm": 0.8377633094787598, "learning_rate": 1.080151956356046e-06, "loss": 0.4594, "step": 27884 }, { "epoch": 1.871447266870239, "grad_norm": 0.9869218468666077, "learning_rate": 1.0779062520268335e-06, "loss": 0.4655, "step": 27886 }, { "epoch": 1.8715814905540082, "grad_norm": 0.9637986421585083, "learning_rate": 1.0756628591703455e-06, "loss": 0.464, "step": 27888 }, { "epoch": 1.8717157142377774, "grad_norm": 1.01296865940094, "learning_rate": 1.0734217778925858e-06, "loss": 0.427, "step": 27890 }, { "epoch": 1.8718499379215463, "grad_norm": 0.9736398458480835, "learning_rate": 1.0711830082994312e-06, "loss": 0.4895, "step": 27892 }, { "epoch": 1.8719841616053152, "grad_norm": 0.9469650387763977, "learning_rate": 1.068946550496669e-06, "loss": 0.3917, "step": 27894 }, { "epoch": 1.8721183852890841, "grad_norm": 1.0767292976379395, "learning_rate": 1.0667124045899646e-06, "loss": 0.4398, "step": 27896 }, { "epoch": 1.8722526089728533, "grad_norm": 1.0303303003311157, "learning_rate": 1.0644805706848782e-06, "loss": 0.471, "step": 27898 }, { "epoch": 1.8723868326566224, "grad_norm": 1.0257786512374878, "learning_rate": 1.0622510488868476e-06, "loss": 0.4258, "step": 27900 }, { "epoch": 1.8725210563403913, "grad_norm": 1.0219508409500122, "learning_rate": 1.0600238393012329e-06, "loss": 0.4765, "step": 27902 }, { "epoch": 1.8726552800241603, "grad_norm": 1.1240330934524536, "learning_rate": 1.05779894203325e-06, "loss": 0.5023, "step": 27904 }, { "epoch": 1.8727895037079292, "grad_norm": 1.1029444932937622, "learning_rate": 1.0555763571880318e-06, "loss": 0.4776, "step": 27906 }, { "epoch": 1.872923727391698, "grad_norm": 1.269639015197754, "learning_rate": 1.053356084870577e-06, "loss": 0.4947, "step": 27908 }, { "epoch": 1.8730579510754672, "grad_norm": 0.9267687201499939, "learning_rate": 1.051138125185802e-06, "loss": 0.4427, "step": 27910 }, { "epoch": 1.8731921747592364, "grad_norm": 1.052469253540039, "learning_rate": 1.0489224782384955e-06, "loss": 0.4915, "step": 27912 }, { "epoch": 1.8733263984430053, "grad_norm": 0.903032660484314, "learning_rate": 1.046709144133351e-06, "loss": 0.4941, "step": 27914 }, { "epoch": 1.8734606221267742, "grad_norm": 0.958709716796875, "learning_rate": 1.0444981229749295e-06, "loss": 0.4109, "step": 27916 }, { "epoch": 1.8735948458105431, "grad_norm": 0.9494258761405945, "learning_rate": 1.0422894148677087e-06, "loss": 0.4648, "step": 27918 }, { "epoch": 1.8737290694943123, "grad_norm": 1.0680596828460693, "learning_rate": 1.0400830199160328e-06, "loss": 0.625, "step": 27920 }, { "epoch": 1.8738632931780814, "grad_norm": 1.104437232017517, "learning_rate": 1.0378789382241682e-06, "loss": 0.4726, "step": 27922 }, { "epoch": 1.8739975168618503, "grad_norm": 1.0886648893356323, "learning_rate": 1.035677169896243e-06, "loss": 0.4234, "step": 27924 }, { "epoch": 1.8741317405456193, "grad_norm": 1.0351425409317017, "learning_rate": 1.033477715036285e-06, "loss": 0.479, "step": 27926 }, { "epoch": 1.8742659642293882, "grad_norm": 1.007570743560791, "learning_rate": 1.0312805737482167e-06, "loss": 0.4489, "step": 27928 }, { "epoch": 1.8744001879131573, "grad_norm": 1.0454297065734863, "learning_rate": 1.0290857461358495e-06, "loss": 0.4559, "step": 27930 }, { "epoch": 1.8745344115969262, "grad_norm": 1.1414828300476074, "learning_rate": 1.0268932323028834e-06, "loss": 0.4754, "step": 27932 }, { "epoch": 1.8746686352806954, "grad_norm": 0.9029853343963623, "learning_rate": 1.0247030323529138e-06, "loss": 0.4742, "step": 27934 }, { "epoch": 1.8748028589644643, "grad_norm": 1.0764508247375488, "learning_rate": 1.0225151463894133e-06, "loss": 0.4821, "step": 27936 }, { "epoch": 1.8749370826482332, "grad_norm": 1.073437213897705, "learning_rate": 1.020329574515766e-06, "loss": 0.5068, "step": 27938 }, { "epoch": 1.8750713063320021, "grad_norm": 1.0652241706848145, "learning_rate": 1.018146316835228e-06, "loss": 0.5957, "step": 27940 }, { "epoch": 1.8752055300157713, "grad_norm": 1.028754711151123, "learning_rate": 1.0159653734509666e-06, "loss": 0.4757, "step": 27942 }, { "epoch": 1.8753397536995404, "grad_norm": 1.2118598222732544, "learning_rate": 1.0137867444660055e-06, "loss": 0.5381, "step": 27944 }, { "epoch": 1.8754739773833093, "grad_norm": 0.9155813455581665, "learning_rate": 1.0116104299833062e-06, "loss": 0.4686, "step": 27946 }, { "epoch": 1.8756082010670783, "grad_norm": 0.9844268560409546, "learning_rate": 1.0094364301056759e-06, "loss": 0.4699, "step": 27948 }, { "epoch": 1.8757424247508472, "grad_norm": 1.0277265310287476, "learning_rate": 1.0072647449358375e-06, "loss": 0.4672, "step": 27950 }, { "epoch": 1.8758766484346163, "grad_norm": 0.9917311668395996, "learning_rate": 1.0050953745764037e-06, "loss": 0.4056, "step": 27952 }, { "epoch": 1.8760108721183852, "grad_norm": 0.9991815090179443, "learning_rate": 1.0029283191298644e-06, "loss": 0.488, "step": 27954 }, { "epoch": 1.8761450958021544, "grad_norm": 1.0271320343017578, "learning_rate": 1.0007635786986214e-06, "loss": 0.4535, "step": 27956 }, { "epoch": 1.8762793194859233, "grad_norm": 0.9581825733184814, "learning_rate": 9.98601153384937e-07, "loss": 0.475, "step": 27958 }, { "epoch": 1.8764135431696922, "grad_norm": 1.1085233688354492, "learning_rate": 9.964410432909965e-07, "loss": 0.4434, "step": 27960 }, { "epoch": 1.8765477668534611, "grad_norm": 1.0936341285705566, "learning_rate": 9.94283248518857e-07, "loss": 0.5027, "step": 27962 }, { "epoch": 1.8766819905372303, "grad_norm": 1.6282494068145752, "learning_rate": 9.921277691704644e-07, "loss": 0.4618, "step": 27964 }, { "epoch": 1.8768162142209994, "grad_norm": 0.9852066040039062, "learning_rate": 9.899746053476655e-07, "loss": 0.4552, "step": 27966 }, { "epoch": 1.8769504379047683, "grad_norm": 1.048144817352295, "learning_rate": 9.878237571521898e-07, "loss": 0.451, "step": 27968 }, { "epoch": 1.8770846615885373, "grad_norm": 0.959335446357727, "learning_rate": 9.85675224685667e-07, "loss": 0.5187, "step": 27970 }, { "epoch": 1.8772188852723062, "grad_norm": 1.0953601598739624, "learning_rate": 9.835290080496107e-07, "loss": 0.4798, "step": 27972 }, { "epoch": 1.8773531089560753, "grad_norm": 1.058064579963684, "learning_rate": 9.813851073454117e-07, "loss": 0.4687, "step": 27974 }, { "epoch": 1.8774873326398445, "grad_norm": 0.9155431389808655, "learning_rate": 9.792435226743835e-07, "loss": 0.4683, "step": 27976 }, { "epoch": 1.8776215563236134, "grad_norm": 0.9273266196250916, "learning_rate": 9.771042541377008e-07, "loss": 0.4774, "step": 27978 }, { "epoch": 1.8777557800073823, "grad_norm": 0.9717110991477966, "learning_rate": 9.749673018364436e-07, "loss": 0.4711, "step": 27980 }, { "epoch": 1.8778900036911512, "grad_norm": 1.0166131258010864, "learning_rate": 9.728326658715815e-07, "loss": 0.5031, "step": 27982 }, { "epoch": 1.8780242273749201, "grad_norm": 1.1387819051742554, "learning_rate": 9.707003463439668e-07, "loss": 0.4494, "step": 27984 }, { "epoch": 1.8781584510586893, "grad_norm": 1.0542840957641602, "learning_rate": 9.685703433543414e-07, "loss": 0.4705, "step": 27986 }, { "epoch": 1.8782926747424584, "grad_norm": 1.1977485418319702, "learning_rate": 9.66442657003358e-07, "loss": 0.5649, "step": 27988 }, { "epoch": 1.8784268984262273, "grad_norm": 1.185279130935669, "learning_rate": 9.643172873915363e-07, "loss": 0.4437, "step": 27990 }, { "epoch": 1.8785611221099963, "grad_norm": 0.9907231330871582, "learning_rate": 9.621942346193013e-07, "loss": 0.445, "step": 27992 }, { "epoch": 1.8786953457937652, "grad_norm": 1.0656880140304565, "learning_rate": 9.600734987869564e-07, "loss": 0.5222, "step": 27994 }, { "epoch": 1.8788295694775343, "grad_norm": 1.0482569932937622, "learning_rate": 9.579550799947046e-07, "loss": 0.433, "step": 27996 }, { "epoch": 1.8789637931613035, "grad_norm": 1.067169189453125, "learning_rate": 9.558389783426436e-07, "loss": 0.4399, "step": 27998 }, { "epoch": 1.8790980168450724, "grad_norm": 1.0769200325012207, "learning_rate": 9.53725193930749e-07, "loss": 0.4912, "step": 28000 }, { "epoch": 1.8792322405288413, "grad_norm": 1.0300370454788208, "learning_rate": 9.516137268588854e-07, "loss": 0.4552, "step": 28002 }, { "epoch": 1.8793664642126102, "grad_norm": 0.9600032567977905, "learning_rate": 9.495045772268341e-07, "loss": 0.424, "step": 28004 }, { "epoch": 1.8795006878963794, "grad_norm": 1.1331719160079956, "learning_rate": 9.473977451342265e-07, "loss": 0.4525, "step": 28006 }, { "epoch": 1.8796349115801483, "grad_norm": 1.1848702430725098, "learning_rate": 9.452932306806273e-07, "loss": 0.4544, "step": 28008 }, { "epoch": 1.8797691352639174, "grad_norm": 1.0405181646347046, "learning_rate": 9.43191033965457e-07, "loss": 0.4596, "step": 28010 }, { "epoch": 1.8799033589476863, "grad_norm": 1.0564866065979004, "learning_rate": 9.410911550880475e-07, "loss": 0.4541, "step": 28012 }, { "epoch": 1.8800375826314553, "grad_norm": 0.9685310125350952, "learning_rate": 9.38993594147608e-07, "loss": 0.3954, "step": 28014 }, { "epoch": 1.8801718063152242, "grad_norm": 1.0685757398605347, "learning_rate": 9.368983512432483e-07, "loss": 0.4814, "step": 28016 }, { "epoch": 1.8803060299989933, "grad_norm": 1.1232695579528809, "learning_rate": 9.348054264739614e-07, "loss": 0.4593, "step": 28018 }, { "epoch": 1.8804402536827625, "grad_norm": 1.0030947923660278, "learning_rate": 9.327148199386404e-07, "loss": 0.4582, "step": 28020 }, { "epoch": 1.8805744773665314, "grad_norm": 1.0549216270446777, "learning_rate": 9.306265317360507e-07, "loss": 0.5093, "step": 28022 }, { "epoch": 1.8807087010503003, "grad_norm": 0.9294421672821045, "learning_rate": 9.28540561964869e-07, "loss": 0.4286, "step": 28024 }, { "epoch": 1.8808429247340692, "grad_norm": 1.1034150123596191, "learning_rate": 9.264569107236498e-07, "loss": 0.5005, "step": 28026 }, { "epoch": 1.8809771484178384, "grad_norm": 1.177290678024292, "learning_rate": 9.243755781108476e-07, "loss": 0.4942, "step": 28028 }, { "epoch": 1.8811113721016073, "grad_norm": 0.9686825275421143, "learning_rate": 9.222965642247949e-07, "loss": 0.4471, "step": 28030 }, { "epoch": 1.8812455957853764, "grad_norm": 1.0099127292633057, "learning_rate": 9.202198691637131e-07, "loss": 0.4512, "step": 28032 }, { "epoch": 1.8813798194691453, "grad_norm": 1.0454397201538086, "learning_rate": 9.181454930257405e-07, "loss": 0.4566, "step": 28034 }, { "epoch": 1.8815140431529143, "grad_norm": 1.2138104438781738, "learning_rate": 9.160734359088763e-07, "loss": 0.4955, "step": 28036 }, { "epoch": 1.8816482668366832, "grad_norm": 0.9711406230926514, "learning_rate": 9.140036979110256e-07, "loss": 0.4588, "step": 28038 }, { "epoch": 1.8817824905204523, "grad_norm": 1.0114809274673462, "learning_rate": 9.119362791299713e-07, "loss": 0.427, "step": 28040 }, { "epoch": 1.8819167142042215, "grad_norm": 1.0150818824768066, "learning_rate": 9.098711796634018e-07, "loss": 0.4919, "step": 28042 }, { "epoch": 1.8820509378879904, "grad_norm": 0.9563175439834595, "learning_rate": 9.078083996088838e-07, "loss": 0.4625, "step": 28044 }, { "epoch": 1.8821851615717593, "grad_norm": 0.8081610202789307, "learning_rate": 9.05747939063889e-07, "loss": 0.368, "step": 28046 }, { "epoch": 1.8823193852555282, "grad_norm": 1.072668194770813, "learning_rate": 9.036897981257675e-07, "loss": 0.4521, "step": 28048 }, { "epoch": 1.8824536089392974, "grad_norm": 1.107627272605896, "learning_rate": 9.016339768917526e-07, "loss": 0.4439, "step": 28050 }, { "epoch": 1.8825878326230665, "grad_norm": 1.011053442955017, "learning_rate": 8.995804754589832e-07, "loss": 0.493, "step": 28052 }, { "epoch": 1.8827220563068354, "grad_norm": 1.0553630590438843, "learning_rate": 8.975292939244928e-07, "loss": 0.4957, "step": 28054 }, { "epoch": 1.8828562799906043, "grad_norm": 1.025673508644104, "learning_rate": 8.954804323851818e-07, "loss": 0.4545, "step": 28056 }, { "epoch": 1.8829905036743733, "grad_norm": 0.9037055969238281, "learning_rate": 8.934338909378615e-07, "loss": 0.3896, "step": 28058 }, { "epoch": 1.8831247273581422, "grad_norm": 1.1714078187942505, "learning_rate": 8.913896696792212e-07, "loss": 0.4653, "step": 28060 }, { "epoch": 1.8832589510419113, "grad_norm": 0.8923357129096985, "learning_rate": 8.893477687058615e-07, "loss": 0.4543, "step": 28062 }, { "epoch": 1.8833931747256805, "grad_norm": 0.9870643019676208, "learning_rate": 8.873081881142386e-07, "loss": 0.4144, "step": 28064 }, { "epoch": 1.8835273984094494, "grad_norm": 1.1387721300125122, "learning_rate": 8.85270928000731e-07, "loss": 0.4651, "step": 28066 }, { "epoch": 1.8836616220932183, "grad_norm": 1.132055401802063, "learning_rate": 8.832359884615893e-07, "loss": 0.4449, "step": 28068 }, { "epoch": 1.8837958457769872, "grad_norm": 1.0421757698059082, "learning_rate": 8.8120336959297e-07, "loss": 0.4609, "step": 28070 }, { "epoch": 1.8839300694607564, "grad_norm": 0.9961799383163452, "learning_rate": 8.791730714908964e-07, "loss": 0.4653, "step": 28072 }, { "epoch": 1.8840642931445255, "grad_norm": 0.8882201910018921, "learning_rate": 8.771450942513081e-07, "loss": 0.4657, "step": 28074 }, { "epoch": 1.8841985168282944, "grad_norm": 1.0295106172561646, "learning_rate": 8.751194379700179e-07, "loss": 0.4799, "step": 28076 }, { "epoch": 1.8843327405120633, "grad_norm": 0.9758402109146118, "learning_rate": 8.730961027427321e-07, "loss": 0.4765, "step": 28078 }, { "epoch": 1.8844669641958323, "grad_norm": 1.07142972946167, "learning_rate": 8.710750886650465e-07, "loss": 0.5259, "step": 28080 }, { "epoch": 1.8846011878796014, "grad_norm": 1.0539863109588623, "learning_rate": 8.690563958324627e-07, "loss": 0.4452, "step": 28082 }, { "epoch": 1.8847354115633703, "grad_norm": 1.1762025356292725, "learning_rate": 8.670400243403543e-07, "loss": 0.5145, "step": 28084 }, { "epoch": 1.8848696352471395, "grad_norm": 0.940888524055481, "learning_rate": 8.65025974283984e-07, "loss": 0.505, "step": 28086 }, { "epoch": 1.8850038589309084, "grad_norm": 0.9293429255485535, "learning_rate": 8.6301424575852e-07, "loss": 0.4445, "step": 28088 }, { "epoch": 1.8851380826146773, "grad_norm": 1.121043086051941, "learning_rate": 8.610048388590031e-07, "loss": 0.4739, "step": 28090 }, { "epoch": 1.8852723062984462, "grad_norm": 0.9941401481628418, "learning_rate": 8.58997753680385e-07, "loss": 0.4619, "step": 28092 }, { "epoch": 1.8854065299822154, "grad_norm": 0.9810492992401123, "learning_rate": 8.569929903174901e-07, "loss": 0.4199, "step": 28094 }, { "epoch": 1.8855407536659845, "grad_norm": 0.9809359312057495, "learning_rate": 8.549905488650422e-07, "loss": 0.417, "step": 28096 }, { "epoch": 1.8856749773497534, "grad_norm": 1.2116690874099731, "learning_rate": 8.529904294176494e-07, "loss": 0.4513, "step": 28098 }, { "epoch": 1.8858092010335223, "grad_norm": 1.0022441148757935, "learning_rate": 8.509926320698137e-07, "loss": 0.4976, "step": 28100 }, { "epoch": 1.8859434247172913, "grad_norm": 1.0324827432632446, "learning_rate": 8.489971569159261e-07, "loss": 0.478, "step": 28102 }, { "epoch": 1.8860776484010604, "grad_norm": 1.0444756746292114, "learning_rate": 8.470040040502836e-07, "loss": 0.4764, "step": 28104 }, { "epoch": 1.8862118720848293, "grad_norm": 1.0981961488723755, "learning_rate": 8.450131735670386e-07, "loss": 0.5176, "step": 28106 }, { "epoch": 1.8863460957685985, "grad_norm": 0.9942836165428162, "learning_rate": 8.430246655602602e-07, "loss": 0.4403, "step": 28108 }, { "epoch": 1.8864803194523674, "grad_norm": 0.9508644342422485, "learning_rate": 8.410384801239068e-07, "loss": 0.4027, "step": 28110 }, { "epoch": 1.8866145431361363, "grad_norm": 1.0890480279922485, "learning_rate": 8.390546173518143e-07, "loss": 0.4543, "step": 28112 }, { "epoch": 1.8867487668199052, "grad_norm": 1.0575063228607178, "learning_rate": 8.370730773377245e-07, "loss": 0.5199, "step": 28114 }, { "epoch": 1.8868829905036744, "grad_norm": 1.016358733177185, "learning_rate": 8.350938601752567e-07, "loss": 0.4694, "step": 28116 }, { "epoch": 1.8870172141874435, "grad_norm": 3.7425155639648438, "learning_rate": 8.331169659579252e-07, "loss": 0.4485, "step": 28118 }, { "epoch": 1.8871514378712124, "grad_norm": 0.9548282623291016, "learning_rate": 8.31142394779133e-07, "loss": 0.4012, "step": 28120 }, { "epoch": 1.8872856615549813, "grad_norm": 1.1152081489562988, "learning_rate": 8.291701467321778e-07, "loss": 0.4721, "step": 28122 }, { "epoch": 1.8874198852387503, "grad_norm": 1.0232319831848145, "learning_rate": 8.272002219102459e-07, "loss": 0.5133, "step": 28124 }, { "epoch": 1.8875541089225194, "grad_norm": 1.034847378730774, "learning_rate": 8.252326204064021e-07, "loss": 0.4311, "step": 28126 }, { "epoch": 1.8876883326062885, "grad_norm": 1.163466453552246, "learning_rate": 8.232673423136217e-07, "loss": 0.4596, "step": 28128 }, { "epoch": 1.8878225562900575, "grad_norm": 1.1249845027923584, "learning_rate": 8.213043877247528e-07, "loss": 0.4962, "step": 28130 }, { "epoch": 1.8879567799738264, "grad_norm": 1.1233371496200562, "learning_rate": 8.193437567325546e-07, "loss": 0.4717, "step": 28132 }, { "epoch": 1.8880910036575953, "grad_norm": 1.1094117164611816, "learning_rate": 8.173854494296529e-07, "loss": 0.5262, "step": 28134 }, { "epoch": 1.8882252273413642, "grad_norm": 1.107286810874939, "learning_rate": 8.154294659085737e-07, "loss": 0.5168, "step": 28136 }, { "epoch": 1.8883594510251334, "grad_norm": 1.1415817737579346, "learning_rate": 8.13475806261732e-07, "loss": 0.5258, "step": 28138 }, { "epoch": 1.8884936747089025, "grad_norm": 1.0087707042694092, "learning_rate": 8.115244705814429e-07, "loss": 0.4978, "step": 28140 }, { "epoch": 1.8886278983926714, "grad_norm": 0.979401707649231, "learning_rate": 8.095754589598936e-07, "loss": 0.5136, "step": 28142 }, { "epoch": 1.8887621220764403, "grad_norm": 1.1137760877609253, "learning_rate": 8.076287714891773e-07, "loss": 0.5194, "step": 28144 }, { "epoch": 1.8888963457602093, "grad_norm": 0.983637809753418, "learning_rate": 8.056844082612647e-07, "loss": 0.3967, "step": 28146 }, { "epoch": 1.8890305694439784, "grad_norm": 1.125389575958252, "learning_rate": 8.037423693680324e-07, "loss": 0.4757, "step": 28148 }, { "epoch": 1.8891647931277475, "grad_norm": 1.065266728401184, "learning_rate": 8.018026549012292e-07, "loss": 0.4739, "step": 28150 }, { "epoch": 1.8892990168115165, "grad_norm": 1.0932797193527222, "learning_rate": 7.998652649525096e-07, "loss": 0.4626, "step": 28152 }, { "epoch": 1.8894332404952854, "grad_norm": 1.0319827795028687, "learning_rate": 7.979301996134059e-07, "loss": 0.4448, "step": 28154 }, { "epoch": 1.8895674641790543, "grad_norm": 1.1214451789855957, "learning_rate": 7.95997458975345e-07, "loss": 0.4566, "step": 28156 }, { "epoch": 1.8897016878628234, "grad_norm": 1.0158464908599854, "learning_rate": 7.940670431296538e-07, "loss": 0.4573, "step": 28158 }, { "epoch": 1.8898359115465924, "grad_norm": 1.0854089260101318, "learning_rate": 7.921389521675315e-07, "loss": 0.4473, "step": 28160 }, { "epoch": 1.8899701352303615, "grad_norm": 1.1003270149230957, "learning_rate": 7.902131861800888e-07, "loss": 0.4704, "step": 28162 }, { "epoch": 1.8901043589141304, "grad_norm": 1.0552144050598145, "learning_rate": 7.882897452583027e-07, "loss": 0.433, "step": 28164 }, { "epoch": 1.8902385825978993, "grad_norm": 1.0242884159088135, "learning_rate": 7.863686294930506e-07, "loss": 0.4525, "step": 28166 }, { "epoch": 1.8903728062816683, "grad_norm": 1.108998417854309, "learning_rate": 7.844498389751098e-07, "loss": 0.4841, "step": 28168 }, { "epoch": 1.8905070299654374, "grad_norm": 1.0566809177398682, "learning_rate": 7.825333737951357e-07, "loss": 0.5314, "step": 28170 }, { "epoch": 1.8906412536492065, "grad_norm": 1.0766716003417969, "learning_rate": 7.80619234043678e-07, "loss": 0.4626, "step": 28172 }, { "epoch": 1.8907754773329755, "grad_norm": 0.9279155135154724, "learning_rate": 7.787074198111755e-07, "loss": 0.4588, "step": 28174 }, { "epoch": 1.8909097010167444, "grad_norm": 0.9887163043022156, "learning_rate": 7.767979311879559e-07, "loss": 0.4419, "step": 28176 }, { "epoch": 1.8910439247005133, "grad_norm": 0.9838598966598511, "learning_rate": 7.748907682642414e-07, "loss": 0.4378, "step": 28178 }, { "epoch": 1.8911781483842824, "grad_norm": 1.0394545793533325, "learning_rate": 7.729859311301435e-07, "loss": 0.5049, "step": 28180 }, { "epoch": 1.8913123720680514, "grad_norm": 0.8739699125289917, "learning_rate": 7.710834198756622e-07, "loss": 0.423, "step": 28182 }, { "epoch": 1.8914465957518205, "grad_norm": 1.0114548206329346, "learning_rate": 7.691832345906757e-07, "loss": 0.4613, "step": 28184 }, { "epoch": 1.8915808194355894, "grad_norm": 1.0551714897155762, "learning_rate": 7.672853753649789e-07, "loss": 0.4701, "step": 28186 }, { "epoch": 1.8917150431193583, "grad_norm": 1.108930230140686, "learning_rate": 7.653898422882333e-07, "loss": 0.4972, "step": 28188 }, { "epoch": 1.8918492668031273, "grad_norm": 1.1477514505386353, "learning_rate": 7.634966354500117e-07, "loss": 0.4582, "step": 28190 }, { "epoch": 1.8919834904868964, "grad_norm": 0.8802066445350647, "learning_rate": 7.616057549397427e-07, "loss": 0.4003, "step": 28192 }, { "epoch": 1.8921177141706655, "grad_norm": 0.9493222236633301, "learning_rate": 7.597172008467824e-07, "loss": 0.444, "step": 28194 }, { "epoch": 1.8922519378544345, "grad_norm": 1.0440152883529663, "learning_rate": 7.578309732603539e-07, "loss": 0.5023, "step": 28196 }, { "epoch": 1.8923861615382034, "grad_norm": 1.0263266563415527, "learning_rate": 7.559470722695916e-07, "loss": 0.4919, "step": 28198 }, { "epoch": 1.8925203852219723, "grad_norm": 1.7940959930419922, "learning_rate": 7.540654979634909e-07, "loss": 0.4466, "step": 28200 }, { "epoch": 1.8926546089057414, "grad_norm": 1.1088297367095947, "learning_rate": 7.52186250430953e-07, "loss": 0.4774, "step": 28202 }, { "epoch": 1.8927888325895106, "grad_norm": 0.8987371325492859, "learning_rate": 7.503093297607732e-07, "loss": 0.476, "step": 28204 }, { "epoch": 1.8929230562732795, "grad_norm": 0.9417023062705994, "learning_rate": 7.484347360416367e-07, "loss": 0.4354, "step": 28206 }, { "epoch": 1.8930572799570484, "grad_norm": 1.0279747247695923, "learning_rate": 7.465624693621109e-07, "loss": 0.4528, "step": 28208 }, { "epoch": 1.8931915036408173, "grad_norm": 1.2020224332809448, "learning_rate": 7.446925298106532e-07, "loss": 0.5022, "step": 28210 }, { "epoch": 1.8933257273245863, "grad_norm": 1.1889111995697021, "learning_rate": 7.428249174756152e-07, "loss": 0.541, "step": 28212 }, { "epoch": 1.8934599510083554, "grad_norm": 1.0302174091339111, "learning_rate": 7.409596324452428e-07, "loss": 0.465, "step": 28214 }, { "epoch": 1.8935941746921245, "grad_norm": 1.1223982572555542, "learning_rate": 7.390966748076599e-07, "loss": 0.4886, "step": 28216 }, { "epoch": 1.8937283983758935, "grad_norm": 1.1422455310821533, "learning_rate": 7.372360446509019e-07, "loss": 0.4741, "step": 28218 }, { "epoch": 1.8938626220596624, "grad_norm": 0.883145809173584, "learning_rate": 7.353777420628594e-07, "loss": 0.4819, "step": 28220 }, { "epoch": 1.8939968457434313, "grad_norm": 1.1092454195022583, "learning_rate": 7.335217671313455e-07, "loss": 0.5085, "step": 28222 }, { "epoch": 1.8941310694272004, "grad_norm": 1.1661089658737183, "learning_rate": 7.316681199440568e-07, "loss": 0.4922, "step": 28224 }, { "epoch": 1.8942652931109696, "grad_norm": 1.0931086540222168, "learning_rate": 7.298168005885564e-07, "loss": 0.4556, "step": 28226 }, { "epoch": 1.8943995167947385, "grad_norm": 1.04469895362854, "learning_rate": 7.279678091523357e-07, "loss": 0.4748, "step": 28228 }, { "epoch": 1.8945337404785074, "grad_norm": 1.0515769720077515, "learning_rate": 7.261211457227413e-07, "loss": 0.4116, "step": 28230 }, { "epoch": 1.8946679641622763, "grad_norm": 0.9401378035545349, "learning_rate": 7.242768103870312e-07, "loss": 0.4345, "step": 28232 }, { "epoch": 1.8948021878460455, "grad_norm": 1.1771726608276367, "learning_rate": 7.22434803232347e-07, "loss": 0.4912, "step": 28234 }, { "epoch": 1.8949364115298144, "grad_norm": 1.0906013250350952, "learning_rate": 7.205951243457132e-07, "loss": 0.4631, "step": 28236 }, { "epoch": 1.8950706352135835, "grad_norm": 1.1231417655944824, "learning_rate": 7.187577738140605e-07, "loss": 0.5209, "step": 28238 }, { "epoch": 1.8952048588973525, "grad_norm": 0.9925578236579895, "learning_rate": 7.169227517241972e-07, "loss": 0.4699, "step": 28240 }, { "epoch": 1.8953390825811214, "grad_norm": 1.080661416053772, "learning_rate": 7.150900581628206e-07, "loss": 0.4685, "step": 28242 }, { "epoch": 1.8954733062648903, "grad_norm": 0.9531117081642151, "learning_rate": 7.132596932165225e-07, "loss": 0.4411, "step": 28244 }, { "epoch": 1.8956075299486594, "grad_norm": 0.9632938504219055, "learning_rate": 7.114316569717894e-07, "loss": 0.4384, "step": 28246 }, { "epoch": 1.8957417536324286, "grad_norm": 0.9770664572715759, "learning_rate": 7.096059495149854e-07, "loss": 0.4697, "step": 28248 }, { "epoch": 1.8958759773161975, "grad_norm": 1.046528935432434, "learning_rate": 7.077825709323749e-07, "loss": 0.455, "step": 28250 }, { "epoch": 1.8960102009999664, "grad_norm": 1.4299813508987427, "learning_rate": 7.059615213101112e-07, "loss": 0.442, "step": 28252 }, { "epoch": 1.8961444246837353, "grad_norm": 1.1879844665527344, "learning_rate": 7.041428007342254e-07, "loss": 0.4722, "step": 28254 }, { "epoch": 1.8962786483675045, "grad_norm": 1.0131479501724243, "learning_rate": 7.023264092906711e-07, "loss": 0.5223, "step": 28256 }, { "epoch": 1.8964128720512734, "grad_norm": 0.9336449503898621, "learning_rate": 7.005123470652408e-07, "loss": 0.4136, "step": 28258 }, { "epoch": 1.8965470957350425, "grad_norm": 1.1306425333023071, "learning_rate": 6.987006141436659e-07, "loss": 0.4994, "step": 28260 }, { "epoch": 1.8966813194188115, "grad_norm": 1.1440653800964355, "learning_rate": 6.968912106115333e-07, "loss": 0.4489, "step": 28262 }, { "epoch": 1.8968155431025804, "grad_norm": 1.000525951385498, "learning_rate": 6.95084136554347e-07, "loss": 0.4658, "step": 28264 }, { "epoch": 1.8969497667863493, "grad_norm": 1.216410756111145, "learning_rate": 6.932793920574831e-07, "loss": 0.4564, "step": 28266 }, { "epoch": 1.8970839904701184, "grad_norm": 1.0653504133224487, "learning_rate": 6.914769772062069e-07, "loss": 0.4797, "step": 28268 }, { "epoch": 1.8972182141538876, "grad_norm": 1.0873147249221802, "learning_rate": 6.896768920856778e-07, "loss": 0.4857, "step": 28270 }, { "epoch": 1.8973524378376565, "grad_norm": 1.1958423852920532, "learning_rate": 6.878791367809556e-07, "loss": 0.4793, "step": 28272 }, { "epoch": 1.8974866615214254, "grad_norm": 1.1073347330093384, "learning_rate": 6.86083711376978e-07, "loss": 0.4675, "step": 28274 }, { "epoch": 1.8976208852051943, "grad_norm": 1.1181777715682983, "learning_rate": 6.842906159585716e-07, "loss": 0.5045, "step": 28276 }, { "epoch": 1.8977551088889635, "grad_norm": 0.8882530331611633, "learning_rate": 6.824998506104574e-07, "loss": 0.4368, "step": 28278 }, { "epoch": 1.8978893325727326, "grad_norm": 1.0279948711395264, "learning_rate": 6.807114154172456e-07, "loss": 0.54, "step": 28280 }, { "epoch": 1.8980235562565015, "grad_norm": 0.9511619210243225, "learning_rate": 6.789253104634352e-07, "loss": 0.4265, "step": 28282 }, { "epoch": 1.8981577799402705, "grad_norm": 1.027387261390686, "learning_rate": 6.77141535833431e-07, "loss": 0.4772, "step": 28284 }, { "epoch": 1.8982920036240394, "grad_norm": 1.0526982545852661, "learning_rate": 6.753600916114877e-07, "loss": 0.4691, "step": 28286 }, { "epoch": 1.8984262273078083, "grad_norm": 1.0057649612426758, "learning_rate": 6.735809778817881e-07, "loss": 0.4471, "step": 28288 }, { "epoch": 1.8985604509915774, "grad_norm": 1.0767104625701904, "learning_rate": 6.718041947283926e-07, "loss": 0.5003, "step": 28290 }, { "epoch": 1.8986946746753466, "grad_norm": 1.1719970703125, "learning_rate": 6.700297422352508e-07, "loss": 0.4701, "step": 28292 }, { "epoch": 1.8988288983591155, "grad_norm": 0.998748242855072, "learning_rate": 6.682576204862012e-07, "loss": 0.4037, "step": 28294 }, { "epoch": 1.8989631220428844, "grad_norm": 1.0387566089630127, "learning_rate": 6.664878295649713e-07, "loss": 0.467, "step": 28296 }, { "epoch": 1.8990973457266533, "grad_norm": 0.9817729592323303, "learning_rate": 6.647203695551829e-07, "loss": 0.51, "step": 28298 }, { "epoch": 1.8992315694104225, "grad_norm": 1.0320035219192505, "learning_rate": 6.629552405403361e-07, "loss": 0.4597, "step": 28300 }, { "epoch": 1.8993657930941916, "grad_norm": 1.0402313470840454, "learning_rate": 6.611924426038419e-07, "loss": 0.4076, "step": 28302 }, { "epoch": 1.8995000167779605, "grad_norm": 0.920330286026001, "learning_rate": 6.594319758289836e-07, "loss": 0.4829, "step": 28304 }, { "epoch": 1.8996342404617295, "grad_norm": 0.9015540480613708, "learning_rate": 6.576738402989447e-07, "loss": 0.4493, "step": 28306 }, { "epoch": 1.8997684641454984, "grad_norm": 1.0200977325439453, "learning_rate": 6.55918036096781e-07, "loss": 0.4504, "step": 28308 }, { "epoch": 1.8999026878292675, "grad_norm": 0.9403783082962036, "learning_rate": 6.54164563305465e-07, "loss": 0.4972, "step": 28310 }, { "epoch": 1.9000369115130364, "grad_norm": 1.0478039979934692, "learning_rate": 6.52413422007836e-07, "loss": 0.4678, "step": 28312 }, { "epoch": 1.9001711351968056, "grad_norm": 1.0461442470550537, "learning_rate": 6.506646122866445e-07, "loss": 0.3982, "step": 28314 }, { "epoch": 1.9003053588805745, "grad_norm": 1.138066291809082, "learning_rate": 6.489181342244965e-07, "loss": 0.5021, "step": 28316 }, { "epoch": 1.9004395825643434, "grad_norm": 1.0282994508743286, "learning_rate": 6.471739879039262e-07, "loss": 0.4285, "step": 28318 }, { "epoch": 1.9005738062481123, "grad_norm": 1.1668262481689453, "learning_rate": 6.454321734073344e-07, "loss": 0.5443, "step": 28320 }, { "epoch": 1.9007080299318815, "grad_norm": 1.0456217527389526, "learning_rate": 6.43692690817027e-07, "loss": 0.5124, "step": 28322 }, { "epoch": 1.9008422536156506, "grad_norm": 1.0460165739059448, "learning_rate": 6.419555402151777e-07, "loss": 0.543, "step": 28324 }, { "epoch": 1.9009764772994195, "grad_norm": 1.1124913692474365, "learning_rate": 6.402207216838762e-07, "loss": 0.4607, "step": 28326 }, { "epoch": 1.9011107009831885, "grad_norm": 0.9830761551856995, "learning_rate": 6.384882353050791e-07, "loss": 0.5103, "step": 28328 }, { "epoch": 1.9012449246669574, "grad_norm": 1.0155304670333862, "learning_rate": 6.367580811606544e-07, "loss": 0.4781, "step": 28330 }, { "epoch": 1.9013791483507265, "grad_norm": 1.1133677959442139, "learning_rate": 6.350302593323365e-07, "loss": 0.4248, "step": 28332 }, { "epoch": 1.9015133720344954, "grad_norm": 1.009499430656433, "learning_rate": 6.333047699017714e-07, "loss": 0.4323, "step": 28334 }, { "epoch": 1.9016475957182646, "grad_norm": 1.0685688257217407, "learning_rate": 6.315816129504715e-07, "loss": 0.4891, "step": 28336 }, { "epoch": 1.9017818194020335, "grad_norm": 1.0232664346694946, "learning_rate": 6.298607885598718e-07, "loss": 0.4593, "step": 28338 }, { "epoch": 1.9019160430858024, "grad_norm": 1.0040220022201538, "learning_rate": 6.281422968112571e-07, "loss": 0.469, "step": 28340 }, { "epoch": 1.9020502667695713, "grad_norm": 0.9486731886863708, "learning_rate": 6.26426137785846e-07, "loss": 0.435, "step": 28342 }, { "epoch": 1.9021844904533405, "grad_norm": 0.8470131754875183, "learning_rate": 6.247123115647013e-07, "loss": 0.4355, "step": 28344 }, { "epoch": 1.9023187141371096, "grad_norm": 0.9575321078300476, "learning_rate": 6.230008182288083e-07, "loss": 0.4909, "step": 28346 }, { "epoch": 1.9024529378208785, "grad_norm": 1.233555555343628, "learning_rate": 6.212916578590355e-07, "loss": 0.4328, "step": 28348 }, { "epoch": 1.9025871615046475, "grad_norm": 1.0835378170013428, "learning_rate": 6.195848305361296e-07, "loss": 0.5304, "step": 28350 }, { "epoch": 1.9027213851884164, "grad_norm": 1.0284099578857422, "learning_rate": 6.178803363407371e-07, "loss": 0.465, "step": 28352 }, { "epoch": 1.9028556088721855, "grad_norm": 1.1114816665649414, "learning_rate": 6.16178175353399e-07, "loss": 0.4593, "step": 28354 }, { "epoch": 1.9029898325559547, "grad_norm": 1.04660165309906, "learning_rate": 6.144783476545234e-07, "loss": 0.4788, "step": 28356 }, { "epoch": 1.9031240562397236, "grad_norm": 0.9865087270736694, "learning_rate": 6.127808533244406e-07, "loss": 0.478, "step": 28358 }, { "epoch": 1.9032582799234925, "grad_norm": 1.0369333028793335, "learning_rate": 6.110856924433473e-07, "loss": 0.4624, "step": 28360 }, { "epoch": 1.9033925036072614, "grad_norm": 1.0701358318328857, "learning_rate": 6.093928650913294e-07, "loss": 0.4459, "step": 28362 }, { "epoch": 1.9035267272910303, "grad_norm": 0.9676421284675598, "learning_rate": 6.077023713483843e-07, "loss": 0.4204, "step": 28364 }, { "epoch": 1.9036609509747995, "grad_norm": 0.9129562377929688, "learning_rate": 6.060142112943701e-07, "loss": 0.4245, "step": 28366 }, { "epoch": 1.9037951746585686, "grad_norm": 1.0327708721160889, "learning_rate": 6.043283850090564e-07, "loss": 0.5011, "step": 28368 }, { "epoch": 1.9039293983423375, "grad_norm": 1.020492672920227, "learning_rate": 6.026448925720962e-07, "loss": 0.4124, "step": 28370 }, { "epoch": 1.9040636220261065, "grad_norm": 1.1958445310592651, "learning_rate": 6.009637340630258e-07, "loss": 0.5311, "step": 28372 }, { "epoch": 1.9041978457098754, "grad_norm": 1.1057733297348022, "learning_rate": 5.992849095612819e-07, "loss": 0.4411, "step": 28374 }, { "epoch": 1.9043320693936445, "grad_norm": 1.0061222314834595, "learning_rate": 5.97608419146184e-07, "loss": 0.4637, "step": 28376 }, { "epoch": 1.9044662930774137, "grad_norm": 1.0512781143188477, "learning_rate": 5.95934262896941e-07, "loss": 0.4505, "step": 28378 }, { "epoch": 1.9046005167611826, "grad_norm": 1.0482077598571777, "learning_rate": 5.942624408926623e-07, "loss": 0.4323, "step": 28380 }, { "epoch": 1.9047347404449515, "grad_norm": 1.0113775730133057, "learning_rate": 5.925929532123231e-07, "loss": 0.436, "step": 28382 }, { "epoch": 1.9048689641287204, "grad_norm": 1.005273699760437, "learning_rate": 5.909257999348106e-07, "loss": 0.3966, "step": 28384 }, { "epoch": 1.9050031878124896, "grad_norm": 1.0841667652130127, "learning_rate": 5.892609811388949e-07, "loss": 0.4567, "step": 28386 }, { "epoch": 1.9051374114962585, "grad_norm": 1.0270723104476929, "learning_rate": 5.87598496903241e-07, "loss": 0.4843, "step": 28388 }, { "epoch": 1.9052716351800276, "grad_norm": 1.1717411279678345, "learning_rate": 5.859383473063918e-07, "loss": 0.4578, "step": 28390 }, { "epoch": 1.9054058588637965, "grad_norm": 1.0161560773849487, "learning_rate": 5.842805324267897e-07, "loss": 0.4392, "step": 28392 }, { "epoch": 1.9055400825475655, "grad_norm": 0.9141351580619812, "learning_rate": 5.826250523427557e-07, "loss": 0.4215, "step": 28394 }, { "epoch": 1.9056743062313344, "grad_norm": 0.9663956761360168, "learning_rate": 5.809719071325103e-07, "loss": 0.4045, "step": 28396 }, { "epoch": 1.9058085299151035, "grad_norm": 1.2365949153900146, "learning_rate": 5.793210968741691e-07, "loss": 0.4915, "step": 28398 }, { "epoch": 1.9059427535988727, "grad_norm": 1.044782280921936, "learning_rate": 5.776726216457251e-07, "loss": 0.3957, "step": 28400 }, { "epoch": 1.9060769772826416, "grad_norm": 0.9589051604270935, "learning_rate": 5.760264815250605e-07, "loss": 0.4413, "step": 28402 }, { "epoch": 1.9062112009664105, "grad_norm": 0.9788399934768677, "learning_rate": 5.743826765899629e-07, "loss": 0.4861, "step": 28404 }, { "epoch": 1.9063454246501794, "grad_norm": 1.0734039545059204, "learning_rate": 5.727412069180871e-07, "loss": 0.4753, "step": 28406 }, { "epoch": 1.9064796483339486, "grad_norm": 0.8587889075279236, "learning_rate": 5.711020725869986e-07, "loss": 0.4593, "step": 28408 }, { "epoch": 1.9066138720177175, "grad_norm": 1.0568461418151855, "learning_rate": 5.694652736741357e-07, "loss": 0.5035, "step": 28410 }, { "epoch": 1.9067480957014866, "grad_norm": 1.0681660175323486, "learning_rate": 5.678308102568364e-07, "loss": 0.463, "step": 28412 }, { "epoch": 1.9068823193852555, "grad_norm": 1.0138444900512695, "learning_rate": 5.661986824123278e-07, "loss": 0.5085, "step": 28414 }, { "epoch": 1.9070165430690245, "grad_norm": 1.0412541627883911, "learning_rate": 5.645688902177315e-07, "loss": 0.5018, "step": 28416 }, { "epoch": 1.9071507667527934, "grad_norm": 1.213753342628479, "learning_rate": 5.62941433750036e-07, "loss": 0.5085, "step": 28418 }, { "epoch": 1.9072849904365625, "grad_norm": 1.0579547882080078, "learning_rate": 5.613163130861521e-07, "loss": 0.4646, "step": 28420 }, { "epoch": 1.9074192141203317, "grad_norm": 1.0913277864456177, "learning_rate": 5.59693528302846e-07, "loss": 0.4481, "step": 28422 }, { "epoch": 1.9075534378041006, "grad_norm": 0.9536145329475403, "learning_rate": 5.580730794768064e-07, "loss": 0.4465, "step": 28424 }, { "epoch": 1.9076876614878695, "grad_norm": 0.9059435725212097, "learning_rate": 5.564549666845886e-07, "loss": 0.5258, "step": 28426 }, { "epoch": 1.9078218851716384, "grad_norm": 1.2305552959442139, "learning_rate": 5.548391900026484e-07, "loss": 0.5643, "step": 28428 }, { "epoch": 1.9079561088554076, "grad_norm": 0.9224756956100464, "learning_rate": 5.532257495073245e-07, "loss": 0.4555, "step": 28430 }, { "epoch": 1.9080903325391767, "grad_norm": 1.0110727548599243, "learning_rate": 5.516146452748506e-07, "loss": 0.4943, "step": 28432 }, { "epoch": 1.9082245562229456, "grad_norm": 1.033556342124939, "learning_rate": 5.500058773813543e-07, "loss": 0.5134, "step": 28434 }, { "epoch": 1.9083587799067145, "grad_norm": 1.0685513019561768, "learning_rate": 5.483994459028363e-07, "loss": 0.4744, "step": 28436 }, { "epoch": 1.9084930035904835, "grad_norm": 1.018219232559204, "learning_rate": 5.467953509152024e-07, "loss": 0.5194, "step": 28438 }, { "epoch": 1.9086272272742524, "grad_norm": 0.8616458773612976, "learning_rate": 5.45193592494242e-07, "loss": 0.461, "step": 28440 }, { "epoch": 1.9087614509580215, "grad_norm": 1.039685845375061, "learning_rate": 5.435941707156389e-07, "loss": 0.4196, "step": 28442 }, { "epoch": 1.9088956746417907, "grad_norm": 1.1347438097000122, "learning_rate": 5.41997085654955e-07, "loss": 0.4563, "step": 28444 }, { "epoch": 1.9090298983255596, "grad_norm": 1.1098366975784302, "learning_rate": 5.404023373876521e-07, "loss": 0.4252, "step": 28446 }, { "epoch": 1.9091641220093285, "grad_norm": 1.00871741771698, "learning_rate": 5.388099259890867e-07, "loss": 0.4486, "step": 28448 }, { "epoch": 1.9092983456930974, "grad_norm": 0.9640150666236877, "learning_rate": 5.372198515344929e-07, "loss": 0.4002, "step": 28450 }, { "epoch": 1.9094325693768666, "grad_norm": 1.0901308059692383, "learning_rate": 5.356321140989884e-07, "loss": 0.4761, "step": 28452 }, { "epoch": 1.9095667930606357, "grad_norm": 1.0453557968139648, "learning_rate": 5.340467137576022e-07, "loss": 0.4536, "step": 28454 }, { "epoch": 1.9097010167444046, "grad_norm": 1.0754693746566772, "learning_rate": 5.32463650585241e-07, "loss": 0.424, "step": 28456 }, { "epoch": 1.9098352404281735, "grad_norm": 1.1840853691101074, "learning_rate": 5.308829246567004e-07, "loss": 0.4866, "step": 28458 }, { "epoch": 1.9099694641119425, "grad_norm": 0.9412599802017212, "learning_rate": 5.293045360466541e-07, "loss": 0.3965, "step": 28460 }, { "epoch": 1.9101036877957116, "grad_norm": 1.0066570043563843, "learning_rate": 5.277284848296981e-07, "loss": 0.4325, "step": 28462 }, { "epoch": 1.9102379114794805, "grad_norm": 0.9843421578407288, "learning_rate": 5.261547710802894e-07, "loss": 0.4693, "step": 28464 }, { "epoch": 1.9103721351632497, "grad_norm": 1.1081351041793823, "learning_rate": 5.245833948727741e-07, "loss": 0.4465, "step": 28466 }, { "epoch": 1.9105063588470186, "grad_norm": 1.0450243949890137, "learning_rate": 5.230143562814093e-07, "loss": 0.5147, "step": 28468 }, { "epoch": 1.9106405825307875, "grad_norm": 0.8684070706367493, "learning_rate": 5.214476553803193e-07, "loss": 0.4537, "step": 28470 }, { "epoch": 1.9107748062145564, "grad_norm": 1.0927025079727173, "learning_rate": 5.198832922435337e-07, "loss": 0.4677, "step": 28472 }, { "epoch": 1.9109090298983256, "grad_norm": 0.9476495385169983, "learning_rate": 5.183212669449656e-07, "loss": 0.3967, "step": 28474 }, { "epoch": 1.9110432535820947, "grad_norm": 0.9752746224403381, "learning_rate": 5.167615795584169e-07, "loss": 0.482, "step": 28476 }, { "epoch": 1.9111774772658636, "grad_norm": 1.0357590913772583, "learning_rate": 5.152042301575788e-07, "loss": 0.4535, "step": 28478 }, { "epoch": 1.9113117009496325, "grad_norm": 1.0198721885681152, "learning_rate": 5.136492188160313e-07, "loss": 0.4679, "step": 28480 }, { "epoch": 1.9114459246334015, "grad_norm": 0.9641233682632446, "learning_rate": 5.12096545607249e-07, "loss": 0.4682, "step": 28482 }, { "epoch": 1.9115801483171706, "grad_norm": 0.9013912081718445, "learning_rate": 5.105462106045955e-07, "loss": 0.4329, "step": 28484 }, { "epoch": 1.9117143720009395, "grad_norm": 1.004481315612793, "learning_rate": 5.08998213881312e-07, "loss": 0.4515, "step": 28486 }, { "epoch": 1.9118485956847087, "grad_norm": 1.0103305578231812, "learning_rate": 5.074525555105403e-07, "loss": 0.526, "step": 28488 }, { "epoch": 1.9119828193684776, "grad_norm": 1.1251330375671387, "learning_rate": 5.059092355653161e-07, "loss": 0.452, "step": 28490 }, { "epoch": 1.9121170430522465, "grad_norm": 1.0311959981918335, "learning_rate": 5.043682541185479e-07, "loss": 0.48, "step": 28492 }, { "epoch": 1.9122512667360154, "grad_norm": 1.04188871383667, "learning_rate": 5.028296112430608e-07, "loss": 0.4371, "step": 28494 }, { "epoch": 1.9123854904197846, "grad_norm": 0.9846208095550537, "learning_rate": 5.012933070115411e-07, "loss": 0.4802, "step": 28496 }, { "epoch": 1.9125197141035537, "grad_norm": 1.1321674585342407, "learning_rate": 4.997593414965751e-07, "loss": 0.5022, "step": 28498 }, { "epoch": 1.9126539377873226, "grad_norm": 0.8902405500411987, "learning_rate": 4.982277147706382e-07, "loss": 0.426, "step": 28500 }, { "epoch": 1.9127881614710915, "grad_norm": 1.275682806968689, "learning_rate": 4.966984269061059e-07, "loss": 0.5051, "step": 28502 }, { "epoch": 1.9129223851548605, "grad_norm": 1.003994345664978, "learning_rate": 4.951714779752314e-07, "loss": 0.4934, "step": 28504 }, { "epoch": 1.9130566088386296, "grad_norm": 1.0241422653198242, "learning_rate": 4.93646868050146e-07, "loss": 0.4803, "step": 28506 }, { "epoch": 1.9131908325223987, "grad_norm": 0.9359074234962463, "learning_rate": 4.921245972029087e-07, "loss": 0.4462, "step": 28508 }, { "epoch": 1.9133250562061677, "grad_norm": 1.139502763748169, "learning_rate": 4.906046655054231e-07, "loss": 0.4593, "step": 28510 }, { "epoch": 1.9134592798899366, "grad_norm": 1.130993366241455, "learning_rate": 4.89087073029515e-07, "loss": 0.4675, "step": 28512 }, { "epoch": 1.9135935035737055, "grad_norm": 1.132150411605835, "learning_rate": 4.875718198468827e-07, "loss": 0.4903, "step": 28514 }, { "epoch": 1.9137277272574744, "grad_norm": 0.9900282025337219, "learning_rate": 4.860589060291188e-07, "loss": 0.5699, "step": 28516 }, { "epoch": 1.9138619509412436, "grad_norm": 1.066292405128479, "learning_rate": 4.845483316477051e-07, "loss": 0.4572, "step": 28518 }, { "epoch": 1.9139961746250127, "grad_norm": 0.9880896210670471, "learning_rate": 4.830400967740178e-07, "loss": 0.4137, "step": 28520 }, { "epoch": 1.9141303983087816, "grad_norm": 1.0525450706481934, "learning_rate": 4.815342014793167e-07, "loss": 0.4943, "step": 28522 }, { "epoch": 1.9142646219925505, "grad_norm": 1.0320472717285156, "learning_rate": 4.800306458347448e-07, "loss": 0.4619, "step": 28524 }, { "epoch": 1.9143988456763195, "grad_norm": 1.0434107780456543, "learning_rate": 4.785294299113508e-07, "loss": 0.4554, "step": 28526 }, { "epoch": 1.9145330693600886, "grad_norm": 0.9134626984596252, "learning_rate": 4.770305537800613e-07, "loss": 0.4422, "step": 28528 }, { "epoch": 1.9146672930438577, "grad_norm": 1.0702944993972778, "learning_rate": 4.7553401751169735e-07, "loss": 0.4806, "step": 28530 }, { "epoch": 1.9148015167276267, "grad_norm": 0.8887928128242493, "learning_rate": 4.7403982117696923e-07, "loss": 0.4599, "step": 28532 }, { "epoch": 1.9149357404113956, "grad_norm": 0.9969332814216614, "learning_rate": 4.7254796484645925e-07, "loss": 0.4259, "step": 28534 }, { "epoch": 1.9150699640951645, "grad_norm": 0.9118576645851135, "learning_rate": 4.710584485906777e-07, "loss": 0.4251, "step": 28536 }, { "epoch": 1.9152041877789336, "grad_norm": 0.8886921405792236, "learning_rate": 4.695712724799795e-07, "loss": 0.4508, "step": 28538 }, { "epoch": 1.9153384114627026, "grad_norm": 1.0292127132415771, "learning_rate": 4.680864365846471e-07, "loss": 0.4989, "step": 28540 }, { "epoch": 1.9154726351464717, "grad_norm": 0.9974724650382996, "learning_rate": 4.666039409748357e-07, "loss": 0.4317, "step": 28542 }, { "epoch": 1.9156068588302406, "grad_norm": 1.3106881380081177, "learning_rate": 4.651237857205781e-07, "loss": 0.4241, "step": 28544 }, { "epoch": 1.9157410825140095, "grad_norm": 1.1389553546905518, "learning_rate": 4.636459708918128e-07, "loss": 0.508, "step": 28546 }, { "epoch": 1.9158753061977785, "grad_norm": 0.8844836950302124, "learning_rate": 4.6217049655837287e-07, "loss": 0.4577, "step": 28548 }, { "epoch": 1.9160095298815476, "grad_norm": 1.1401379108428955, "learning_rate": 4.606973627899636e-07, "loss": 0.5309, "step": 28550 }, { "epoch": 1.9161437535653167, "grad_norm": 1.0861908197402954, "learning_rate": 4.5922656965618484e-07, "loss": 0.4301, "step": 28552 }, { "epoch": 1.9162779772490857, "grad_norm": 1.0897116661071777, "learning_rate": 4.577581172265366e-07, "loss": 0.4486, "step": 28554 }, { "epoch": 1.9164122009328546, "grad_norm": 1.0773807764053345, "learning_rate": 4.5629200557039674e-07, "loss": 0.5119, "step": 28556 }, { "epoch": 1.9165464246166235, "grad_norm": 0.956876814365387, "learning_rate": 4.5482823475703205e-07, "loss": 0.5197, "step": 28558 }, { "epoch": 1.9166806483003926, "grad_norm": 1.1136553287506104, "learning_rate": 4.53366804855615e-07, "loss": 0.453, "step": 28560 }, { "epoch": 1.9168148719841616, "grad_norm": 0.9848942756652832, "learning_rate": 4.519077159351792e-07, "loss": 0.4496, "step": 28562 }, { "epoch": 1.9169490956679307, "grad_norm": 1.036436915397644, "learning_rate": 4.504509680646751e-07, "loss": 0.4695, "step": 28564 }, { "epoch": 1.9170833193516996, "grad_norm": 1.1097133159637451, "learning_rate": 4.48996561312931e-07, "loss": 0.4585, "step": 28566 }, { "epoch": 1.9172175430354685, "grad_norm": 1.0167475938796997, "learning_rate": 4.47544495748653e-07, "loss": 0.4732, "step": 28568 }, { "epoch": 1.9173517667192375, "grad_norm": 0.9866967797279358, "learning_rate": 4.4609477144046395e-07, "loss": 0.4768, "step": 28570 }, { "epoch": 1.9174859904030066, "grad_norm": 5.733170032501221, "learning_rate": 4.4464738845685363e-07, "loss": 0.4874, "step": 28572 }, { "epoch": 1.9176202140867757, "grad_norm": 1.0686026811599731, "learning_rate": 4.432023468662061e-07, "loss": 0.5011, "step": 28574 }, { "epoch": 1.9177544377705447, "grad_norm": 0.9280849099159241, "learning_rate": 4.417596467367946e-07, "loss": 0.4321, "step": 28576 }, { "epoch": 1.9178886614543136, "grad_norm": 1.0819580554962158, "learning_rate": 4.4031928813679235e-07, "loss": 0.4728, "step": 28578 }, { "epoch": 1.9180228851380825, "grad_norm": 0.971687376499176, "learning_rate": 4.3888127113424496e-07, "loss": 0.4933, "step": 28580 }, { "epoch": 1.9181571088218516, "grad_norm": 1.021970272064209, "learning_rate": 4.3744559579710353e-07, "loss": 0.4983, "step": 28582 }, { "epoch": 1.9182913325056208, "grad_norm": 1.0264430046081543, "learning_rate": 4.360122621931917e-07, "loss": 0.4828, "step": 28584 }, { "epoch": 1.9184255561893897, "grad_norm": 0.9958664178848267, "learning_rate": 4.345812703902441e-07, "loss": 0.4592, "step": 28586 }, { "epoch": 1.9185597798731586, "grad_norm": 0.9951131343841553, "learning_rate": 4.3315262045586224e-07, "loss": 0.4928, "step": 28588 }, { "epoch": 1.9186940035569275, "grad_norm": 0.8685447573661804, "learning_rate": 4.317263124575477e-07, "loss": 0.4654, "step": 28590 }, { "epoch": 1.9188282272406965, "grad_norm": 0.9491552710533142, "learning_rate": 4.3030234646269653e-07, "loss": 0.4621, "step": 28592 }, { "epoch": 1.9189624509244656, "grad_norm": 1.1487892866134644, "learning_rate": 4.288807225385827e-07, "loss": 0.4918, "step": 28594 }, { "epoch": 1.9190966746082347, "grad_norm": 1.0633422136306763, "learning_rate": 4.274614407523747e-07, "loss": 0.4567, "step": 28596 }, { "epoch": 1.9192308982920037, "grad_norm": 1.007346272468567, "learning_rate": 4.2604450117114114e-07, "loss": 0.4498, "step": 28598 }, { "epoch": 1.9193651219757726, "grad_norm": 0.9971712231636047, "learning_rate": 4.2462990386181735e-07, "loss": 0.4672, "step": 28600 }, { "epoch": 1.9194993456595415, "grad_norm": 1.153144121170044, "learning_rate": 4.2321764889124425e-07, "loss": 0.46, "step": 28602 }, { "epoch": 1.9196335693433106, "grad_norm": 1.036102294921875, "learning_rate": 4.2180773632614637e-07, "loss": 0.4401, "step": 28604 }, { "epoch": 1.9197677930270798, "grad_norm": 1.0317116975784302, "learning_rate": 4.2040016623314804e-07, "loss": 0.4512, "step": 28606 }, { "epoch": 1.9199020167108487, "grad_norm": 0.9724912643432617, "learning_rate": 4.189949386787462e-07, "loss": 0.4622, "step": 28608 }, { "epoch": 1.9200362403946176, "grad_norm": 0.9077244997024536, "learning_rate": 4.1759205372933206e-07, "loss": 0.4768, "step": 28610 }, { "epoch": 1.9201704640783865, "grad_norm": 1.0253747701644897, "learning_rate": 4.161915114511972e-07, "loss": 0.4779, "step": 28612 }, { "epoch": 1.9203046877621557, "grad_norm": 1.1545828580856323, "learning_rate": 4.147933119105107e-07, "loss": 0.4816, "step": 28614 }, { "epoch": 1.9204389114459246, "grad_norm": 0.9470486640930176, "learning_rate": 4.133974551733366e-07, "loss": 0.4111, "step": 28616 }, { "epoch": 1.9205731351296937, "grad_norm": 0.9704294800758362, "learning_rate": 4.12003941305622e-07, "loss": 0.5076, "step": 28618 }, { "epoch": 1.9207073588134627, "grad_norm": 0.9441697597503662, "learning_rate": 4.106127703732088e-07, "loss": 0.412, "step": 28620 }, { "epoch": 1.9208415824972316, "grad_norm": 1.0234549045562744, "learning_rate": 4.0922394244183315e-07, "loss": 0.4747, "step": 28622 }, { "epoch": 1.9209758061810005, "grad_norm": 1.0847262144088745, "learning_rate": 4.0783745757710935e-07, "loss": 0.5063, "step": 28624 }, { "epoch": 1.9211100298647696, "grad_norm": 1.040287733078003, "learning_rate": 4.0645331584454606e-07, "loss": 0.506, "step": 28626 }, { "epoch": 1.9212442535485388, "grad_norm": 1.115692377090454, "learning_rate": 4.0507151730954095e-07, "loss": 0.487, "step": 28628 }, { "epoch": 1.9213784772323077, "grad_norm": 1.7195245027542114, "learning_rate": 4.036920620373863e-07, "loss": 0.4531, "step": 28630 }, { "epoch": 1.9215127009160766, "grad_norm": 0.9717667102813721, "learning_rate": 4.0231495009325215e-07, "loss": 0.434, "step": 28632 }, { "epoch": 1.9216469245998455, "grad_norm": 0.8491559624671936, "learning_rate": 4.0094018154220316e-07, "loss": 0.4655, "step": 28634 }, { "epoch": 1.9217811482836147, "grad_norm": 1.1206490993499756, "learning_rate": 3.9956775644920395e-07, "loss": 0.4388, "step": 28636 }, { "epoch": 1.9219153719673836, "grad_norm": 0.9006564021110535, "learning_rate": 3.9819767487909165e-07, "loss": 0.426, "step": 28638 }, { "epoch": 1.9220495956511527, "grad_norm": 0.9929448366165161, "learning_rate": 3.968299368966033e-07, "loss": 0.4574, "step": 28640 }, { "epoch": 1.9221838193349217, "grad_norm": 1.083114743232727, "learning_rate": 3.954645425663539e-07, "loss": 0.4758, "step": 28642 }, { "epoch": 1.9223180430186906, "grad_norm": 1.041306972503662, "learning_rate": 3.9410149195286963e-07, "loss": 0.4512, "step": 28644 }, { "epoch": 1.9224522667024595, "grad_norm": 1.2051905393600464, "learning_rate": 3.927407851205378e-07, "loss": 0.4464, "step": 28646 }, { "epoch": 1.9225864903862286, "grad_norm": 1.0940951108932495, "learning_rate": 3.9138242213365703e-07, "loss": 0.4367, "step": 28648 }, { "epoch": 1.9227207140699978, "grad_norm": 1.0143513679504395, "learning_rate": 3.900264030564038e-07, "loss": 0.4202, "step": 28650 }, { "epoch": 1.9228549377537667, "grad_norm": 1.453702449798584, "learning_rate": 3.8867272795285456e-07, "loss": 0.4457, "step": 28652 }, { "epoch": 1.9229891614375356, "grad_norm": 1.0661327838897705, "learning_rate": 3.8732139688695825e-07, "loss": 0.4789, "step": 28654 }, { "epoch": 1.9231233851213045, "grad_norm": 0.9508505463600159, "learning_rate": 3.8597240992256924e-07, "loss": 0.3891, "step": 28656 }, { "epoch": 1.9232576088050737, "grad_norm": 1.0582705736160278, "learning_rate": 3.846257671234199e-07, "loss": 0.5021, "step": 28658 }, { "epoch": 1.9233918324888426, "grad_norm": 0.9684615731239319, "learning_rate": 3.8328146855314275e-07, "loss": 0.4531, "step": 28660 }, { "epoch": 1.9235260561726117, "grad_norm": 1.162911057472229, "learning_rate": 3.8193951427524243e-07, "loss": 0.4367, "step": 28662 }, { "epoch": 1.9236602798563807, "grad_norm": 1.106896996498108, "learning_rate": 3.8059990435313495e-07, "loss": 0.4696, "step": 28664 }, { "epoch": 1.9237945035401496, "grad_norm": 1.1426647901535034, "learning_rate": 3.792626388501086e-07, "loss": 0.5195, "step": 28666 }, { "epoch": 1.9239287272239185, "grad_norm": 1.0694050788879395, "learning_rate": 3.7792771782934613e-07, "loss": 0.4896, "step": 28668 }, { "epoch": 1.9240629509076876, "grad_norm": 0.9436160326004028, "learning_rate": 3.7659514135391924e-07, "loss": 0.4324, "step": 28670 }, { "epoch": 1.9241971745914568, "grad_norm": 1.0565946102142334, "learning_rate": 3.7526490948679995e-07, "loss": 0.449, "step": 28672 }, { "epoch": 1.9243313982752257, "grad_norm": 1.1161247491836548, "learning_rate": 3.739370222908267e-07, "loss": 0.4936, "step": 28674 }, { "epoch": 1.9244656219589946, "grad_norm": 1.0940635204315186, "learning_rate": 3.726114798287439e-07, "loss": 0.4694, "step": 28676 }, { "epoch": 1.9245998456427635, "grad_norm": 1.0296701192855835, "learning_rate": 3.712882821631736e-07, "loss": 0.4453, "step": 28678 }, { "epoch": 1.9247340693265327, "grad_norm": 1.1180201768875122, "learning_rate": 3.6996742935664906e-07, "loss": 0.4635, "step": 28680 }, { "epoch": 1.9248682930103018, "grad_norm": 1.0563124418258667, "learning_rate": 3.6864892147156496e-07, "loss": 0.3876, "step": 28682 }, { "epoch": 1.9250025166940707, "grad_norm": 0.8866017460823059, "learning_rate": 3.673327585702324e-07, "loss": 0.4398, "step": 28684 }, { "epoch": 1.9251367403778397, "grad_norm": 1.0113062858581543, "learning_rate": 3.6601894071482403e-07, "loss": 0.4645, "step": 28686 }, { "epoch": 1.9252709640616086, "grad_norm": 1.0368983745574951, "learning_rate": 3.6470746796741783e-07, "loss": 0.5158, "step": 28688 }, { "epoch": 1.9254051877453777, "grad_norm": 1.0530612468719482, "learning_rate": 3.6339834038997545e-07, "loss": 0.5536, "step": 28690 }, { "epoch": 1.9255394114291466, "grad_norm": 1.2044345140457153, "learning_rate": 3.62091558044364e-07, "loss": 0.5074, "step": 28692 }, { "epoch": 1.9256736351129158, "grad_norm": 1.0812793970108032, "learning_rate": 3.607871209923175e-07, "loss": 0.4488, "step": 28694 }, { "epoch": 1.9258078587966847, "grad_norm": 0.8661904335021973, "learning_rate": 3.5948502929546433e-07, "loss": 0.4305, "step": 28696 }, { "epoch": 1.9259420824804536, "grad_norm": 1.0397839546203613, "learning_rate": 3.581852830153276e-07, "loss": 0.4291, "step": 28698 }, { "epoch": 1.9260763061642225, "grad_norm": 1.0755635499954224, "learning_rate": 3.5688788221332483e-07, "loss": 0.4847, "step": 28700 }, { "epoch": 1.9262105298479917, "grad_norm": 1.0838686227798462, "learning_rate": 3.555928269507458e-07, "loss": 0.4305, "step": 28702 }, { "epoch": 1.9263447535317608, "grad_norm": 1.0830137729644775, "learning_rate": 3.5430011728879164e-07, "loss": 0.4864, "step": 28704 }, { "epoch": 1.9264789772155297, "grad_norm": 0.9551955461502075, "learning_rate": 3.5300975328853014e-07, "loss": 0.4718, "step": 28706 }, { "epoch": 1.9266132008992987, "grad_norm": 1.0307772159576416, "learning_rate": 3.517217350109236e-07, "loss": 0.5096, "step": 28708 }, { "epoch": 1.9267474245830676, "grad_norm": 1.031486988067627, "learning_rate": 3.504360625168457e-07, "loss": 0.5634, "step": 28710 }, { "epoch": 1.9268816482668367, "grad_norm": 1.1116483211517334, "learning_rate": 3.4915273586702546e-07, "loss": 0.523, "step": 28712 }, { "epoch": 1.9270158719506056, "grad_norm": 0.9586636424064636, "learning_rate": 3.478717551221089e-07, "loss": 0.4818, "step": 28714 }, { "epoch": 1.9271500956343748, "grad_norm": 1.0831985473632812, "learning_rate": 3.465931203426087e-07, "loss": 0.4314, "step": 28716 }, { "epoch": 1.9272843193181437, "grad_norm": 0.9724242687225342, "learning_rate": 3.453168315889488e-07, "loss": 0.4643, "step": 28718 }, { "epoch": 1.9274185430019126, "grad_norm": 1.8084481954574585, "learning_rate": 3.440428889214253e-07, "loss": 0.4437, "step": 28720 }, { "epoch": 1.9275527666856815, "grad_norm": 1.1648857593536377, "learning_rate": 3.427712924002402e-07, "loss": 0.5064, "step": 28722 }, { "epoch": 1.9276869903694507, "grad_norm": 1.189074158668518, "learning_rate": 3.415020420854509e-07, "loss": 0.4949, "step": 28724 }, { "epoch": 1.9278212140532198, "grad_norm": 1.1187207698822021, "learning_rate": 3.402351380370483e-07, "loss": 0.4282, "step": 28726 }, { "epoch": 1.9279554377369887, "grad_norm": 0.9836530089378357, "learning_rate": 3.3897058031487906e-07, "loss": 0.4768, "step": 28728 }, { "epoch": 1.9280896614207577, "grad_norm": 1.033627986907959, "learning_rate": 3.3770836897870086e-07, "loss": 0.4541, "step": 28730 }, { "epoch": 1.9282238851045266, "grad_norm": 1.1201447248458862, "learning_rate": 3.364485040881438e-07, "loss": 0.4658, "step": 28732 }, { "epoch": 1.9283581087882957, "grad_norm": 1.0621964931488037, "learning_rate": 3.3519098570273797e-07, "loss": 0.4132, "step": 28734 }, { "epoch": 1.9284923324720646, "grad_norm": 1.0216847658157349, "learning_rate": 3.3393581388189157e-07, "loss": 0.4506, "step": 28736 }, { "epoch": 1.9286265561558338, "grad_norm": 1.0879138708114624, "learning_rate": 3.326829886849181e-07, "loss": 0.5463, "step": 28738 }, { "epoch": 1.9287607798396027, "grad_norm": 1.0055351257324219, "learning_rate": 3.314325101710036e-07, "loss": 0.4407, "step": 28740 }, { "epoch": 1.9288950035233716, "grad_norm": 1.1964483261108398, "learning_rate": 3.301843783992398e-07, "loss": 0.4547, "step": 28742 }, { "epoch": 1.9290292272071405, "grad_norm": 0.9628129601478577, "learning_rate": 3.289385934285849e-07, "loss": 0.4544, "step": 28744 }, { "epoch": 1.9291634508909097, "grad_norm": 1.0963078737258911, "learning_rate": 3.276951553179086e-07, "loss": 0.4807, "step": 28746 }, { "epoch": 1.9292976745746788, "grad_norm": 0.8719479441642761, "learning_rate": 3.264540641259639e-07, "loss": 0.4312, "step": 28748 }, { "epoch": 1.9294318982584477, "grad_norm": 1.026655912399292, "learning_rate": 3.252153199113872e-07, "loss": 0.5329, "step": 28750 }, { "epoch": 1.9295661219422167, "grad_norm": 0.8928234577178955, "learning_rate": 3.2397892273269835e-07, "loss": 0.3986, "step": 28752 }, { "epoch": 1.9297003456259856, "grad_norm": 1.0384950637817383, "learning_rate": 3.227448726483284e-07, "loss": 0.4794, "step": 28754 }, { "epoch": 1.9298345693097547, "grad_norm": 0.9321753978729248, "learning_rate": 3.2151316971656963e-07, "loss": 0.4443, "step": 28756 }, { "epoch": 1.9299687929935239, "grad_norm": 1.0566009283065796, "learning_rate": 3.2028381399563103e-07, "loss": 0.4563, "step": 28758 }, { "epoch": 1.9301030166772928, "grad_norm": 0.9426655173301697, "learning_rate": 3.190568055435883e-07, "loss": 0.4385, "step": 28760 }, { "epoch": 1.9302372403610617, "grad_norm": 1.0611015558242798, "learning_rate": 3.178321444184229e-07, "loss": 0.4217, "step": 28762 }, { "epoch": 1.9303714640448306, "grad_norm": 0.8264487981796265, "learning_rate": 3.1660983067798856e-07, "loss": 0.4225, "step": 28764 }, { "epoch": 1.9305056877285998, "grad_norm": 1.009493350982666, "learning_rate": 3.15389864380039e-07, "loss": 0.4217, "step": 28766 }, { "epoch": 1.9306399114123687, "grad_norm": 1.0614426136016846, "learning_rate": 3.141722455822227e-07, "loss": 0.4555, "step": 28768 }, { "epoch": 1.9307741350961378, "grad_norm": 1.0587729215621948, "learning_rate": 3.1295697434206573e-07, "loss": 0.4292, "step": 28770 }, { "epoch": 1.9309083587799067, "grad_norm": 1.0446990728378296, "learning_rate": 3.117440507169833e-07, "loss": 0.4865, "step": 28772 }, { "epoch": 1.9310425824636757, "grad_norm": 1.14275062084198, "learning_rate": 3.105334747642852e-07, "loss": 0.4542, "step": 28774 }, { "epoch": 1.9311768061474446, "grad_norm": 1.0680385828018188, "learning_rate": 3.093252465411756e-07, "loss": 0.4283, "step": 28776 }, { "epoch": 1.9313110298312137, "grad_norm": 1.2552454471588135, "learning_rate": 3.0811936610473103e-07, "loss": 0.4425, "step": 28778 }, { "epoch": 1.9314452535149829, "grad_norm": 0.8596842288970947, "learning_rate": 3.0691583351193377e-07, "loss": 0.4169, "step": 28780 }, { "epoch": 1.9315794771987518, "grad_norm": 0.9840183258056641, "learning_rate": 3.0571464881964385e-07, "loss": 0.4169, "step": 28782 }, { "epoch": 1.9317137008825207, "grad_norm": 0.9783838987350464, "learning_rate": 3.0451581208462143e-07, "loss": 0.4212, "step": 28784 }, { "epoch": 1.9318479245662896, "grad_norm": 0.914157509803772, "learning_rate": 3.0331932336349897e-07, "loss": 0.4042, "step": 28786 }, { "epoch": 1.9319821482500588, "grad_norm": 1.0035114288330078, "learning_rate": 3.0212518271281466e-07, "loss": 0.4492, "step": 28788 }, { "epoch": 1.9321163719338277, "grad_norm": 1.0145611763000488, "learning_rate": 3.0093339018899544e-07, "loss": 0.4848, "step": 28790 }, { "epoch": 1.9322505956175968, "grad_norm": 0.8387595415115356, "learning_rate": 2.9974394584834085e-07, "loss": 0.4025, "step": 28792 }, { "epoch": 1.9323848193013657, "grad_norm": 0.9812015295028687, "learning_rate": 2.9855684974705034e-07, "loss": 0.4563, "step": 28794 }, { "epoch": 1.9325190429851347, "grad_norm": 1.0451850891113281, "learning_rate": 2.973721019412179e-07, "loss": 0.4825, "step": 28796 }, { "epoch": 1.9326532666689036, "grad_norm": 1.0891306400299072, "learning_rate": 2.96189702486821e-07, "loss": 0.464, "step": 28798 }, { "epoch": 1.9327874903526727, "grad_norm": 0.9749390482902527, "learning_rate": 2.950096514397149e-07, "loss": 0.4394, "step": 28800 }, { "epoch": 1.9329217140364419, "grad_norm": 0.9677867889404297, "learning_rate": 2.9383194885566623e-07, "loss": 0.4786, "step": 28802 }, { "epoch": 1.9330559377202108, "grad_norm": 0.8928714394569397, "learning_rate": 2.926565947903137e-07, "loss": 0.4969, "step": 28804 }, { "epoch": 1.9331901614039797, "grad_norm": 0.9890260696411133, "learning_rate": 2.9148358929919073e-07, "loss": 0.462, "step": 28806 }, { "epoch": 1.9333243850877486, "grad_norm": 1.1038089990615845, "learning_rate": 2.903129324377252e-07, "loss": 0.4754, "step": 28808 }, { "epoch": 1.9334586087715178, "grad_norm": 1.1482740640640259, "learning_rate": 2.891446242612228e-07, "loss": 0.516, "step": 28810 }, { "epoch": 1.9335928324552867, "grad_norm": 0.973935604095459, "learning_rate": 2.8797866482488387e-07, "loss": 0.4626, "step": 28812 }, { "epoch": 1.9337270561390558, "grad_norm": 0.9728585481643677, "learning_rate": 2.868150541837922e-07, "loss": 0.3713, "step": 28814 }, { "epoch": 1.9338612798228247, "grad_norm": 0.8140623569488525, "learning_rate": 2.8565379239294257e-07, "loss": 0.463, "step": 28816 }, { "epoch": 1.9339955035065937, "grad_norm": 0.9422245025634766, "learning_rate": 2.844948795071856e-07, "loss": 0.5291, "step": 28818 }, { "epoch": 1.9341297271903626, "grad_norm": 1.8759267330169678, "learning_rate": 2.8333831558128856e-07, "loss": 0.3971, "step": 28820 }, { "epoch": 1.9342639508741317, "grad_norm": 1.5151934623718262, "learning_rate": 2.8218410066988554e-07, "loss": 0.4601, "step": 28822 }, { "epoch": 1.9343981745579009, "grad_norm": 1.103763461112976, "learning_rate": 2.8103223482752183e-07, "loss": 0.486, "step": 28824 }, { "epoch": 1.9345323982416698, "grad_norm": 0.9701932668685913, "learning_rate": 2.7988271810862054e-07, "loss": 0.4239, "step": 28826 }, { "epoch": 1.9346666219254387, "grad_norm": 0.9574766159057617, "learning_rate": 2.787355505674882e-07, "loss": 0.4315, "step": 28828 }, { "epoch": 1.9348008456092076, "grad_norm": 0.9307126998901367, "learning_rate": 2.77590732258326e-07, "loss": 0.4835, "step": 28830 }, { "epoch": 1.9349350692929768, "grad_norm": 0.9249545931816101, "learning_rate": 2.7644826323522943e-07, "loss": 0.433, "step": 28832 }, { "epoch": 1.935069292976746, "grad_norm": 1.1061577796936035, "learning_rate": 2.75308143552172e-07, "loss": 0.4839, "step": 28834 }, { "epoch": 1.9352035166605148, "grad_norm": 1.0632753372192383, "learning_rate": 2.741703732630274e-07, "loss": 0.4484, "step": 28836 }, { "epoch": 1.9353377403442837, "grad_norm": 1.08527672290802, "learning_rate": 2.7303495242155254e-07, "loss": 0.4027, "step": 28838 }, { "epoch": 1.9354719640280527, "grad_norm": 1.001092791557312, "learning_rate": 2.7190188108138784e-07, "loss": 0.4591, "step": 28840 }, { "epoch": 1.9356061877118218, "grad_norm": 0.8313817977905273, "learning_rate": 2.707711592960793e-07, "loss": 0.3729, "step": 28842 }, { "epoch": 1.9357404113955907, "grad_norm": 1.118964672088623, "learning_rate": 2.696427871190399e-07, "loss": 0.4507, "step": 28844 }, { "epoch": 1.9358746350793599, "grad_norm": 1.0571320056915283, "learning_rate": 2.6851676460359355e-07, "loss": 0.5072, "step": 28846 }, { "epoch": 1.9360088587631288, "grad_norm": 0.9531912803649902, "learning_rate": 2.67393091802931e-07, "loss": 0.469, "step": 28848 }, { "epoch": 1.9361430824468977, "grad_norm": 1.1145519018173218, "learning_rate": 2.6627176877015435e-07, "loss": 0.5439, "step": 28850 }, { "epoch": 1.9362773061306666, "grad_norm": 1.0474146604537964, "learning_rate": 2.651527955582378e-07, "loss": 0.5029, "step": 28852 }, { "epoch": 1.9364115298144358, "grad_norm": 0.9402778744697571, "learning_rate": 2.6403617222005017e-07, "loss": 0.5068, "step": 28854 }, { "epoch": 1.936545753498205, "grad_norm": 1.0296565294265747, "learning_rate": 2.629218988083548e-07, "loss": 0.4601, "step": 28856 }, { "epoch": 1.9366799771819738, "grad_norm": 1.130409598350525, "learning_rate": 2.6180997537579856e-07, "loss": 0.4967, "step": 28858 }, { "epoch": 1.9368142008657427, "grad_norm": 1.2263824939727783, "learning_rate": 2.60700401974906e-07, "loss": 0.5095, "step": 28860 }, { "epoch": 1.9369484245495117, "grad_norm": 1.068340539932251, "learning_rate": 2.595931786581185e-07, "loss": 0.4784, "step": 28862 }, { "epoch": 1.9370826482332808, "grad_norm": 1.0405715703964233, "learning_rate": 2.584883054777443e-07, "loss": 0.4375, "step": 28864 }, { "epoch": 1.9372168719170497, "grad_norm": 1.063066005706787, "learning_rate": 2.5738578248598044e-07, "loss": 0.4243, "step": 28866 }, { "epoch": 1.9373510956008189, "grad_norm": 0.9734081029891968, "learning_rate": 2.562856097349242e-07, "loss": 0.4596, "step": 28868 }, { "epoch": 1.9374853192845878, "grad_norm": 1.0543913841247559, "learning_rate": 2.551877872765562e-07, "loss": 0.4521, "step": 28870 }, { "epoch": 1.9376195429683567, "grad_norm": 0.9148135185241699, "learning_rate": 2.540923151627461e-07, "loss": 0.3845, "step": 28872 }, { "epoch": 1.9377537666521256, "grad_norm": 0.9552580714225769, "learning_rate": 2.52999193445258e-07, "loss": 0.4932, "step": 28874 }, { "epoch": 1.9378879903358948, "grad_norm": 1.0944716930389404, "learning_rate": 2.5190842217573396e-07, "loss": 0.4781, "step": 28876 }, { "epoch": 1.938022214019664, "grad_norm": 0.974706768989563, "learning_rate": 2.5082000140570493e-07, "loss": 0.4606, "step": 28878 }, { "epoch": 1.9381564377034328, "grad_norm": 1.061537742614746, "learning_rate": 2.4973393118660757e-07, "loss": 0.4437, "step": 28880 }, { "epoch": 1.9382906613872017, "grad_norm": 1.0034071207046509, "learning_rate": 2.4865021156975085e-07, "loss": 0.4057, "step": 28882 }, { "epoch": 1.9384248850709707, "grad_norm": 1.1839606761932373, "learning_rate": 2.4756884260634384e-07, "loss": 0.498, "step": 28884 }, { "epoch": 1.9385591087547398, "grad_norm": 0.9004724621772766, "learning_rate": 2.464898243474734e-07, "loss": 0.4202, "step": 28886 }, { "epoch": 1.9386933324385087, "grad_norm": 1.0127958059310913, "learning_rate": 2.454131568441154e-07, "loss": 0.4522, "step": 28888 }, { "epoch": 1.9388275561222779, "grad_norm": 0.9200853705406189, "learning_rate": 2.443388401471569e-07, "loss": 0.4911, "step": 28890 }, { "epoch": 1.9389617798060468, "grad_norm": 0.9288998246192932, "learning_rate": 2.432668743073463e-07, "loss": 0.4558, "step": 28892 }, { "epoch": 1.9390960034898157, "grad_norm": 1.0898096561431885, "learning_rate": 2.421972593753319e-07, "loss": 0.4421, "step": 28894 }, { "epoch": 1.9392302271735846, "grad_norm": 0.9902854561805725, "learning_rate": 2.4112999540165103e-07, "loss": 0.4606, "step": 28896 }, { "epoch": 1.9393644508573538, "grad_norm": 0.9736539721488953, "learning_rate": 2.400650824367301e-07, "loss": 0.4236, "step": 28898 }, { "epoch": 1.939498674541123, "grad_norm": 1.0703511238098145, "learning_rate": 2.3900252053088435e-07, "loss": 0.5251, "step": 28900 }, { "epoch": 1.9396328982248918, "grad_norm": 1.0596925020217896, "learning_rate": 2.379423097343292e-07, "loss": 0.4527, "step": 28902 }, { "epoch": 1.9397671219086607, "grad_norm": 0.9808062314987183, "learning_rate": 2.3688445009713566e-07, "loss": 0.415, "step": 28904 }, { "epoch": 1.9399013455924297, "grad_norm": 1.0180929899215698, "learning_rate": 2.3582894166930268e-07, "loss": 0.4719, "step": 28906 }, { "epoch": 1.9400355692761988, "grad_norm": 0.9373030662536621, "learning_rate": 2.3477578450069038e-07, "loss": 0.4148, "step": 28908 }, { "epoch": 1.940169792959968, "grad_norm": 1.1346129179000854, "learning_rate": 2.3372497864106445e-07, "loss": 0.4429, "step": 28910 }, { "epoch": 1.9403040166437369, "grad_norm": 1.9116069078445435, "learning_rate": 2.3267652414007414e-07, "loss": 0.446, "step": 28912 }, { "epoch": 1.9404382403275058, "grad_norm": 1.026120901107788, "learning_rate": 2.316304210472575e-07, "loss": 0.4624, "step": 28914 }, { "epoch": 1.9405724640112747, "grad_norm": 1.02186918258667, "learning_rate": 2.3058666941203623e-07, "loss": 0.4748, "step": 28916 }, { "epoch": 1.9407066876950438, "grad_norm": 1.0325887203216553, "learning_rate": 2.2954526928372632e-07, "loss": 0.5107, "step": 28918 }, { "epoch": 1.9408409113788128, "grad_norm": 1.205985426902771, "learning_rate": 2.2850622071153293e-07, "loss": 0.4616, "step": 28920 }, { "epoch": 1.940975135062582, "grad_norm": 1.0128406286239624, "learning_rate": 2.2746952374455011e-07, "loss": 0.4771, "step": 28922 }, { "epoch": 1.9411093587463508, "grad_norm": 1.089816927909851, "learning_rate": 2.264351784317553e-07, "loss": 0.5144, "step": 28924 }, { "epoch": 1.9412435824301197, "grad_norm": 0.9186378717422485, "learning_rate": 2.2540318482202615e-07, "loss": 0.413, "step": 28926 }, { "epoch": 1.9413778061138887, "grad_norm": 1.0192416906356812, "learning_rate": 2.2437354296411805e-07, "loss": 0.5828, "step": 28928 }, { "epoch": 1.9415120297976578, "grad_norm": 1.1217014789581299, "learning_rate": 2.23346252906681e-07, "loss": 0.4275, "step": 28930 }, { "epoch": 1.941646253481427, "grad_norm": 1.148184895515442, "learning_rate": 2.223213146982539e-07, "loss": 0.4881, "step": 28932 }, { "epoch": 1.9417804771651959, "grad_norm": 1.0364254713058472, "learning_rate": 2.2129872838725364e-07, "loss": 0.4372, "step": 28934 }, { "epoch": 1.9419147008489648, "grad_norm": 0.9857342839241028, "learning_rate": 2.2027849402201373e-07, "loss": 0.5483, "step": 28936 }, { "epoch": 1.9420489245327337, "grad_norm": 1.150019645690918, "learning_rate": 2.1926061165071788e-07, "loss": 0.4502, "step": 28938 }, { "epoch": 1.9421831482165028, "grad_norm": 0.9097667932510376, "learning_rate": 2.1824508132147204e-07, "loss": 0.4826, "step": 28940 }, { "epoch": 1.9423173719002718, "grad_norm": 0.9855087995529175, "learning_rate": 2.1723190308225448e-07, "loss": 0.4767, "step": 28942 }, { "epoch": 1.942451595584041, "grad_norm": 1.1955652236938477, "learning_rate": 2.1622107698093808e-07, "loss": 0.5317, "step": 28944 }, { "epoch": 1.9425858192678098, "grad_norm": 0.9239872097969055, "learning_rate": 2.1521260306527903e-07, "loss": 0.4205, "step": 28946 }, { "epoch": 1.9427200429515787, "grad_norm": 1.0808923244476318, "learning_rate": 2.1420648138292253e-07, "loss": 0.4204, "step": 28948 }, { "epoch": 1.9428542666353477, "grad_norm": 1.0448721647262573, "learning_rate": 2.1320271198141395e-07, "loss": 0.4553, "step": 28950 }, { "epoch": 1.9429884903191168, "grad_norm": 1.039955735206604, "learning_rate": 2.122012949081764e-07, "loss": 0.4373, "step": 28952 }, { "epoch": 1.943122714002886, "grad_norm": 0.969295859336853, "learning_rate": 2.112022302105221e-07, "loss": 0.4894, "step": 28954 }, { "epoch": 1.9432569376866549, "grad_norm": 1.0688284635543823, "learning_rate": 2.1020551793565768e-07, "loss": 0.4254, "step": 28956 }, { "epoch": 1.9433911613704238, "grad_norm": 0.9468502998352051, "learning_rate": 2.092111581306788e-07, "loss": 0.449, "step": 28958 }, { "epoch": 1.9435253850541927, "grad_norm": 0.8899710774421692, "learning_rate": 2.0821915084255906e-07, "loss": 0.4741, "step": 28960 }, { "epoch": 1.9436596087379618, "grad_norm": 0.9876530766487122, "learning_rate": 2.0722949611817198e-07, "loss": 0.4862, "step": 28962 }, { "epoch": 1.9437938324217308, "grad_norm": 0.9413831830024719, "learning_rate": 2.0624219400428023e-07, "loss": 0.4808, "step": 28964 }, { "epoch": 1.9439280561055, "grad_norm": 1.0272241830825806, "learning_rate": 2.052572445475298e-07, "loss": 0.4382, "step": 28966 }, { "epoch": 1.9440622797892688, "grad_norm": 0.9331052899360657, "learning_rate": 2.0427464779445572e-07, "loss": 0.4401, "step": 28968 }, { "epoch": 1.9441965034730377, "grad_norm": 0.9220208525657654, "learning_rate": 2.0329440379148746e-07, "loss": 0.5323, "step": 28970 }, { "epoch": 1.9443307271568067, "grad_norm": 1.0369528532028198, "learning_rate": 2.02316512584938e-07, "loss": 0.4699, "step": 28972 }, { "epoch": 1.9444649508405758, "grad_norm": 1.038794755935669, "learning_rate": 2.0134097422100928e-07, "loss": 0.4573, "step": 28974 }, { "epoch": 1.944599174524345, "grad_norm": 1.0068680047988892, "learning_rate": 2.0036778874579775e-07, "loss": 0.4714, "step": 28976 }, { "epoch": 1.9447333982081139, "grad_norm": 0.9134820103645325, "learning_rate": 1.9939695620527777e-07, "loss": 0.4858, "step": 28978 }, { "epoch": 1.9448676218918828, "grad_norm": 1.061395525932312, "learning_rate": 1.984284766453237e-07, "loss": 0.4697, "step": 28980 }, { "epoch": 1.9450018455756517, "grad_norm": 1.0704432725906372, "learning_rate": 1.9746235011169344e-07, "loss": 0.4647, "step": 28982 }, { "epoch": 1.9451360692594208, "grad_norm": 1.048775553703308, "learning_rate": 1.964985766500338e-07, "loss": 0.4801, "step": 28984 }, { "epoch": 1.94527029294319, "grad_norm": 1.0661295652389526, "learning_rate": 1.9553715630588053e-07, "loss": 0.5136, "step": 28986 }, { "epoch": 1.945404516626959, "grad_norm": 1.0962748527526855, "learning_rate": 1.9457808912466402e-07, "loss": 0.4616, "step": 28988 }, { "epoch": 1.9455387403107278, "grad_norm": 0.9793387651443481, "learning_rate": 1.9362137515169242e-07, "loss": 0.421, "step": 28990 }, { "epoch": 1.9456729639944967, "grad_norm": 0.9833454489707947, "learning_rate": 1.9266701443217406e-07, "loss": 0.4683, "step": 28992 }, { "epoch": 1.9458071876782659, "grad_norm": 1.0015058517456055, "learning_rate": 1.9171500701119504e-07, "loss": 0.441, "step": 28994 }, { "epoch": 1.9459414113620348, "grad_norm": 1.0324184894561768, "learning_rate": 1.907653529337361e-07, "loss": 0.4315, "step": 28996 }, { "epoch": 1.946075635045804, "grad_norm": 1.0735946893692017, "learning_rate": 1.8981805224467242e-07, "loss": 0.4409, "step": 28998 }, { "epoch": 1.9462098587295729, "grad_norm": 1.0525434017181396, "learning_rate": 1.8887310498875155e-07, "loss": 0.4902, "step": 29000 }, { "epoch": 1.9463440824133418, "grad_norm": 1.0508967638015747, "learning_rate": 1.879305112106322e-07, "loss": 0.4963, "step": 29002 }, { "epoch": 1.9464783060971107, "grad_norm": 1.0489139556884766, "learning_rate": 1.8699027095484545e-07, "loss": 0.4933, "step": 29004 }, { "epoch": 1.9466125297808798, "grad_norm": 0.9729229211807251, "learning_rate": 1.8605238426581683e-07, "loss": 0.4763, "step": 29006 }, { "epoch": 1.946746753464649, "grad_norm": 1.1164216995239258, "learning_rate": 1.8511685118785538e-07, "loss": 0.5097, "step": 29008 }, { "epoch": 1.946880977148418, "grad_norm": 0.9576023817062378, "learning_rate": 1.841836717651646e-07, "loss": 0.4712, "step": 29010 }, { "epoch": 1.9470152008321868, "grad_norm": 1.105007290840149, "learning_rate": 1.83252846041837e-07, "loss": 0.4731, "step": 29012 }, { "epoch": 1.9471494245159557, "grad_norm": 1.041167974472046, "learning_rate": 1.823243740618541e-07, "loss": 0.4661, "step": 29014 }, { "epoch": 1.9472836481997249, "grad_norm": 1.1564899682998657, "learning_rate": 1.8139825586908076e-07, "loss": 0.4755, "step": 29016 }, { "epoch": 1.9474178718834938, "grad_norm": 1.0982869863510132, "learning_rate": 1.8047449150727648e-07, "loss": 0.4441, "step": 29018 }, { "epoch": 1.947552095567263, "grad_norm": 0.9591256976127625, "learning_rate": 1.795530810200896e-07, "loss": 0.4082, "step": 29020 }, { "epoch": 1.9476863192510319, "grad_norm": 1.0350632667541504, "learning_rate": 1.786340244510465e-07, "loss": 0.4866, "step": 29022 }, { "epoch": 1.9478205429348008, "grad_norm": 0.9792650938034058, "learning_rate": 1.7771732184357904e-07, "loss": 0.4856, "step": 29024 }, { "epoch": 1.9479547666185697, "grad_norm": 0.9688316583633423, "learning_rate": 1.7680297324099703e-07, "loss": 0.4627, "step": 29026 }, { "epoch": 1.9480889903023388, "grad_norm": 0.9843719601631165, "learning_rate": 1.758909786864993e-07, "loss": 0.4657, "step": 29028 }, { "epoch": 1.948223213986108, "grad_norm": 0.9596292972564697, "learning_rate": 1.7498133822317908e-07, "loss": 0.4706, "step": 29030 }, { "epoch": 1.948357437669877, "grad_norm": 0.9700624942779541, "learning_rate": 1.7407405189401315e-07, "loss": 0.4184, "step": 29032 }, { "epoch": 1.9484916613536458, "grad_norm": 1.0020383596420288, "learning_rate": 1.7316911974187276e-07, "loss": 0.4551, "step": 29034 }, { "epoch": 1.9486258850374147, "grad_norm": 1.0333776473999023, "learning_rate": 1.7226654180950708e-07, "loss": 0.4732, "step": 29036 }, { "epoch": 1.9487601087211839, "grad_norm": 1.095948338508606, "learning_rate": 1.7136631813957082e-07, "loss": 0.4572, "step": 29038 }, { "epoch": 1.9488943324049528, "grad_norm": 1.1414443254470825, "learning_rate": 1.7046844877458556e-07, "loss": 0.4756, "step": 29040 }, { "epoch": 1.949028556088722, "grad_norm": 1.041707158088684, "learning_rate": 1.6957293375698403e-07, "loss": 0.4828, "step": 29042 }, { "epoch": 1.9491627797724909, "grad_norm": 1.217383623123169, "learning_rate": 1.6867977312907678e-07, "loss": 0.5353, "step": 29044 }, { "epoch": 1.9492970034562598, "grad_norm": 1.0687098503112793, "learning_rate": 1.677889669330579e-07, "loss": 0.4779, "step": 29046 }, { "epoch": 1.9494312271400287, "grad_norm": 1.08873450756073, "learning_rate": 1.6690051521102146e-07, "loss": 0.5443, "step": 29048 }, { "epoch": 1.9495654508237978, "grad_norm": 1.04291570186615, "learning_rate": 1.6601441800493945e-07, "loss": 0.5015, "step": 29050 }, { "epoch": 1.949699674507567, "grad_norm": 1.0000983476638794, "learning_rate": 1.6513067535668392e-07, "loss": 0.5141, "step": 29052 }, { "epoch": 1.949833898191336, "grad_norm": 0.8735694885253906, "learning_rate": 1.6424928730801036e-07, "loss": 0.4608, "step": 29054 }, { "epoch": 1.9499681218751048, "grad_norm": 1.115757942199707, "learning_rate": 1.633702539005577e-07, "loss": 0.4715, "step": 29056 }, { "epoch": 1.9501023455588737, "grad_norm": 1.030385971069336, "learning_rate": 1.6249357517585938e-07, "loss": 0.4762, "step": 29058 }, { "epoch": 1.9502365692426429, "grad_norm": 1.125352144241333, "learning_rate": 1.6161925117533783e-07, "loss": 0.4587, "step": 29060 }, { "epoch": 1.950370792926412, "grad_norm": 1.1368461847305298, "learning_rate": 1.6074728194030442e-07, "loss": 0.4317, "step": 29062 }, { "epoch": 1.950505016610181, "grad_norm": 1.1113425493240356, "learning_rate": 1.5987766751195953e-07, "loss": 0.4119, "step": 29064 }, { "epoch": 1.9506392402939499, "grad_norm": 1.0342289209365845, "learning_rate": 1.5901040793138699e-07, "loss": 0.458, "step": 29066 }, { "epoch": 1.9507734639777188, "grad_norm": 1.0388826131820679, "learning_rate": 1.5814550323957066e-07, "loss": 0.5091, "step": 29068 }, { "epoch": 1.950907687661488, "grad_norm": 1.1173781156539917, "learning_rate": 1.5728295347736123e-07, "loss": 0.5012, "step": 29070 }, { "epoch": 1.9510419113452568, "grad_norm": 1.036182165145874, "learning_rate": 1.5642275868552602e-07, "loss": 0.4631, "step": 29072 }, { "epoch": 1.951176135029026, "grad_norm": 1.0020720958709717, "learning_rate": 1.5556491890469927e-07, "loss": 0.4646, "step": 29074 }, { "epoch": 1.951310358712795, "grad_norm": 0.8500409722328186, "learning_rate": 1.5470943417541518e-07, "loss": 0.4186, "step": 29076 }, { "epoch": 1.9514445823965638, "grad_norm": 0.9407816529273987, "learning_rate": 1.538563045380914e-07, "loss": 0.5105, "step": 29078 }, { "epoch": 1.9515788060803327, "grad_norm": 1.0555976629257202, "learning_rate": 1.5300553003304575e-07, "loss": 0.4626, "step": 29080 }, { "epoch": 1.9517130297641019, "grad_norm": 1.0615906715393066, "learning_rate": 1.521571107004627e-07, "loss": 0.4693, "step": 29082 }, { "epoch": 1.951847253447871, "grad_norm": 1.036577820777893, "learning_rate": 1.5131104658043794e-07, "loss": 0.4407, "step": 29084 }, { "epoch": 1.95198147713164, "grad_norm": 1.1040834188461304, "learning_rate": 1.5046733771293953e-07, "loss": 0.4067, "step": 29086 }, { "epoch": 1.9521157008154089, "grad_norm": 0.9711515307426453, "learning_rate": 1.4962598413784113e-07, "loss": 0.4554, "step": 29088 }, { "epoch": 1.9522499244991778, "grad_norm": 1.0760223865509033, "learning_rate": 1.4878698589488315e-07, "loss": 0.4432, "step": 29090 }, { "epoch": 1.952384148182947, "grad_norm": 1.018295168876648, "learning_rate": 1.4795034302371168e-07, "loss": 0.5183, "step": 29092 }, { "epoch": 1.9525183718667158, "grad_norm": 1.0077235698699951, "learning_rate": 1.4711605556385622e-07, "loss": 0.4489, "step": 29094 }, { "epoch": 1.952652595550485, "grad_norm": 0.9149937629699707, "learning_rate": 1.4628412355474076e-07, "loss": 0.4419, "step": 29096 }, { "epoch": 1.952786819234254, "grad_norm": 1.0449026823043823, "learning_rate": 1.4545454703566165e-07, "loss": 0.4531, "step": 29098 }, { "epoch": 1.9529210429180228, "grad_norm": 1.10755455493927, "learning_rate": 1.4462732604582086e-07, "loss": 0.4817, "step": 29100 }, { "epoch": 1.9530552666017917, "grad_norm": 1.0468051433563232, "learning_rate": 1.4380246062430935e-07, "loss": 0.4701, "step": 29102 }, { "epoch": 1.9531894902855609, "grad_norm": 0.9423577189445496, "learning_rate": 1.4297995081008487e-07, "loss": 0.4525, "step": 29104 }, { "epoch": 1.95332371396933, "grad_norm": 0.8814041614532471, "learning_rate": 1.4215979664202183e-07, "loss": 0.4456, "step": 29106 }, { "epoch": 1.953457937653099, "grad_norm": 1.0376014709472656, "learning_rate": 1.4134199815886705e-07, "loss": 0.5057, "step": 29108 }, { "epoch": 1.9535921613368679, "grad_norm": 1.0050175189971924, "learning_rate": 1.4052655539926184e-07, "loss": 0.4042, "step": 29110 }, { "epoch": 1.9537263850206368, "grad_norm": 1.1026122570037842, "learning_rate": 1.3971346840173095e-07, "loss": 0.4135, "step": 29112 }, { "epoch": 1.953860608704406, "grad_norm": 0.948855459690094, "learning_rate": 1.389027372046936e-07, "loss": 0.4329, "step": 29114 }, { "epoch": 1.9539948323881748, "grad_norm": 1.0344352722167969, "learning_rate": 1.3809436184645252e-07, "loss": 0.4875, "step": 29116 }, { "epoch": 1.954129056071944, "grad_norm": 1.8970121145248413, "learning_rate": 1.3728834236520493e-07, "loss": 0.4427, "step": 29118 }, { "epoch": 1.954263279755713, "grad_norm": 0.8916929960250854, "learning_rate": 1.3648467879902594e-07, "loss": 0.4658, "step": 29120 }, { "epoch": 1.9543975034394818, "grad_norm": 0.9808556437492371, "learning_rate": 1.356833711859018e-07, "loss": 0.5233, "step": 29122 }, { "epoch": 1.9545317271232507, "grad_norm": 0.9045658111572266, "learning_rate": 1.3488441956368003e-07, "loss": 0.4995, "step": 29124 }, { "epoch": 1.9546659508070199, "grad_norm": 1.0025702714920044, "learning_rate": 1.3408782397011932e-07, "loss": 0.5065, "step": 29126 }, { "epoch": 1.954800174490789, "grad_norm": 0.7962448596954346, "learning_rate": 1.3329358444284513e-07, "loss": 0.4051, "step": 29128 }, { "epoch": 1.954934398174558, "grad_norm": 1.0446126461029053, "learning_rate": 1.3250170101939407e-07, "loss": 0.4518, "step": 29130 }, { "epoch": 1.9550686218583269, "grad_norm": 1.0310949087142944, "learning_rate": 1.3171217373717516e-07, "loss": 0.4761, "step": 29132 }, { "epoch": 1.9552028455420958, "grad_norm": 1.033321499824524, "learning_rate": 1.309250026334974e-07, "loss": 0.4402, "step": 29134 }, { "epoch": 1.955337069225865, "grad_norm": 0.9540125131607056, "learning_rate": 1.301401877455477e-07, "loss": 0.4348, "step": 29136 }, { "epoch": 1.955471292909634, "grad_norm": 0.9811615347862244, "learning_rate": 1.293577291104131e-07, "loss": 0.5095, "step": 29138 }, { "epoch": 1.955605516593403, "grad_norm": 1.055335283279419, "learning_rate": 1.285776267650529e-07, "loss": 0.4345, "step": 29140 }, { "epoch": 1.955739740277172, "grad_norm": 1.139459252357483, "learning_rate": 1.2779988074633765e-07, "loss": 0.4297, "step": 29142 }, { "epoch": 1.9558739639609408, "grad_norm": 0.95600426197052, "learning_rate": 1.2702449109100455e-07, "loss": 0.4257, "step": 29144 }, { "epoch": 1.95600818764471, "grad_norm": 1.0222060680389404, "learning_rate": 1.2625145783569658e-07, "loss": 0.4449, "step": 29146 }, { "epoch": 1.9561424113284789, "grad_norm": 1.061171293258667, "learning_rate": 1.2548078101692894e-07, "loss": 0.4535, "step": 29148 }, { "epoch": 1.956276635012248, "grad_norm": 1.103115200996399, "learning_rate": 1.2471246067112807e-07, "loss": 0.4707, "step": 29150 }, { "epoch": 1.956410858696017, "grad_norm": 0.9497417211532593, "learning_rate": 1.239464968345816e-07, "loss": 0.462, "step": 29152 }, { "epoch": 1.9565450823797859, "grad_norm": 1.0168882608413696, "learning_rate": 1.2318288954348833e-07, "loss": 0.4242, "step": 29154 }, { "epoch": 1.9566793060635548, "grad_norm": 1.0620505809783936, "learning_rate": 1.22421638833925e-07, "loss": 0.4896, "step": 29156 }, { "epoch": 1.956813529747324, "grad_norm": 1.0867334604263306, "learning_rate": 1.216627447418628e-07, "loss": 0.4628, "step": 29158 }, { "epoch": 1.956947753431093, "grad_norm": 0.9913773536682129, "learning_rate": 1.2090620730315084e-07, "loss": 0.4797, "step": 29160 }, { "epoch": 1.957081977114862, "grad_norm": 0.995578944683075, "learning_rate": 1.201520265535383e-07, "loss": 0.4371, "step": 29162 }, { "epoch": 1.957216200798631, "grad_norm": 0.9965330958366394, "learning_rate": 1.1940020252865226e-07, "loss": 0.5014, "step": 29164 }, { "epoch": 1.9573504244823998, "grad_norm": 1.0482277870178223, "learning_rate": 1.1865073526402537e-07, "loss": 0.4753, "step": 29166 }, { "epoch": 1.957484648166169, "grad_norm": 1.0561718940734863, "learning_rate": 1.1790362479506822e-07, "loss": 0.3859, "step": 29168 }, { "epoch": 1.9576188718499379, "grad_norm": 1.0040631294250488, "learning_rate": 1.1715887115706926e-07, "loss": 0.5141, "step": 29170 }, { "epoch": 1.957753095533707, "grad_norm": 1.0110114812850952, "learning_rate": 1.1641647438522807e-07, "loss": 0.4536, "step": 29172 }, { "epoch": 1.957887319217476, "grad_norm": 0.9060198664665222, "learning_rate": 1.1567643451461108e-07, "loss": 0.4565, "step": 29174 }, { "epoch": 1.9580215429012449, "grad_norm": 1.0186243057250977, "learning_rate": 1.1493875158019584e-07, "loss": 0.4437, "step": 29176 }, { "epoch": 1.9581557665850138, "grad_norm": 1.1048253774642944, "learning_rate": 1.1420342561682673e-07, "loss": 0.474, "step": 29178 }, { "epoch": 1.958289990268783, "grad_norm": 1.0118054151535034, "learning_rate": 1.1347045665924816e-07, "loss": 0.4664, "step": 29180 }, { "epoch": 1.958424213952552, "grad_norm": 0.8971180319786072, "learning_rate": 1.1273984474209354e-07, "loss": 0.4557, "step": 29182 }, { "epoch": 1.958558437636321, "grad_norm": 1.033743977546692, "learning_rate": 1.1201158989988525e-07, "loss": 0.4718, "step": 29184 }, { "epoch": 1.95869266132009, "grad_norm": 0.9839541912078857, "learning_rate": 1.1128569216702356e-07, "loss": 0.4439, "step": 29186 }, { "epoch": 1.9588268850038588, "grad_norm": 0.991456151008606, "learning_rate": 1.1056215157781435e-07, "loss": 0.4654, "step": 29188 }, { "epoch": 1.958961108687628, "grad_norm": 0.9525729417800903, "learning_rate": 1.098409681664414e-07, "loss": 0.388, "step": 29190 }, { "epoch": 1.9590953323713969, "grad_norm": 1.019943356513977, "learning_rate": 1.091221419669719e-07, "loss": 0.4282, "step": 29192 }, { "epoch": 1.959229556055166, "grad_norm": 0.8993679881095886, "learning_rate": 1.0840567301337868e-07, "loss": 0.448, "step": 29194 }, { "epoch": 1.959363779738935, "grad_norm": 1.0546905994415283, "learning_rate": 1.0769156133951241e-07, "loss": 0.435, "step": 29196 }, { "epoch": 1.9594980034227039, "grad_norm": 1.1095595359802246, "learning_rate": 1.0697980697910725e-07, "loss": 0.4396, "step": 29198 }, { "epoch": 1.9596322271064728, "grad_norm": 1.655231237411499, "learning_rate": 1.0627040996579741e-07, "loss": 0.4249, "step": 29200 }, { "epoch": 1.959766450790242, "grad_norm": 0.9991711378097534, "learning_rate": 1.0556337033310048e-07, "loss": 0.4191, "step": 29202 }, { "epoch": 1.959900674474011, "grad_norm": 1.16497004032135, "learning_rate": 1.0485868811441757e-07, "loss": 0.4507, "step": 29204 }, { "epoch": 1.96003489815778, "grad_norm": 1.2120740413665771, "learning_rate": 1.0415636334304979e-07, "loss": 0.4976, "step": 29206 }, { "epoch": 1.960169121841549, "grad_norm": 1.0939778089523315, "learning_rate": 1.0345639605217616e-07, "loss": 0.4828, "step": 29208 }, { "epoch": 1.9603033455253178, "grad_norm": 0.9883015751838684, "learning_rate": 1.0275878627487023e-07, "loss": 0.4519, "step": 29210 }, { "epoch": 1.960437569209087, "grad_norm": 0.9145817160606384, "learning_rate": 1.0206353404409452e-07, "loss": 0.4446, "step": 29212 }, { "epoch": 1.960571792892856, "grad_norm": 1.0710937976837158, "learning_rate": 1.0137063939269497e-07, "loss": 0.4951, "step": 29214 }, { "epoch": 1.960706016576625, "grad_norm": 1.0684460401535034, "learning_rate": 1.0068010235341208e-07, "loss": 0.4343, "step": 29216 }, { "epoch": 1.960840240260394, "grad_norm": 0.9165230393409729, "learning_rate": 9.999192295886972e-08, "loss": 0.4662, "step": 29218 }, { "epoch": 1.9609744639441629, "grad_norm": 0.965201199054718, "learning_rate": 9.930610124158634e-08, "loss": 0.4806, "step": 29220 }, { "epoch": 1.961108687627932, "grad_norm": 1.0687464475631714, "learning_rate": 9.862263723396382e-08, "loss": 0.5493, "step": 29222 }, { "epoch": 1.961242911311701, "grad_norm": 1.1389684677124023, "learning_rate": 9.79415309682985e-08, "loss": 0.5155, "step": 29224 }, { "epoch": 1.96137713499547, "grad_norm": 0.9733030200004578, "learning_rate": 9.726278247676468e-08, "loss": 0.4105, "step": 29226 }, { "epoch": 1.961511358679239, "grad_norm": 1.0655992031097412, "learning_rate": 9.658639179143669e-08, "loss": 0.4438, "step": 29228 }, { "epoch": 1.961645582363008, "grad_norm": 1.075886845588684, "learning_rate": 9.591235894426675e-08, "loss": 0.4108, "step": 29230 }, { "epoch": 1.9617798060467768, "grad_norm": 0.9965173006057739, "learning_rate": 9.524068396710717e-08, "loss": 0.5006, "step": 29232 }, { "epoch": 1.961914029730546, "grad_norm": 0.8664084076881409, "learning_rate": 9.457136689169366e-08, "loss": 0.3999, "step": 29234 }, { "epoch": 1.962048253414315, "grad_norm": 1.0834602117538452, "learning_rate": 9.390440774965092e-08, "loss": 0.516, "step": 29236 }, { "epoch": 1.962182477098084, "grad_norm": 0.9731799960136414, "learning_rate": 9.323980657248154e-08, "loss": 0.4889, "step": 29238 }, { "epoch": 1.962316700781853, "grad_norm": 0.9713391065597534, "learning_rate": 9.257756339159929e-08, "loss": 0.4285, "step": 29240 }, { "epoch": 1.9624509244656219, "grad_norm": 1.1183282136917114, "learning_rate": 9.191767823828467e-08, "loss": 0.4732, "step": 29242 }, { "epoch": 1.962585148149391, "grad_norm": 1.0748634338378906, "learning_rate": 9.126015114372388e-08, "loss": 0.4712, "step": 29244 }, { "epoch": 1.96271937183316, "grad_norm": 0.9243441224098206, "learning_rate": 9.060498213897539e-08, "loss": 0.4211, "step": 29246 }, { "epoch": 1.962853595516929, "grad_norm": 0.9169735908508301, "learning_rate": 8.995217125500333e-08, "loss": 0.4266, "step": 29248 }, { "epoch": 1.962987819200698, "grad_norm": 0.9653085470199585, "learning_rate": 8.930171852264413e-08, "loss": 0.4021, "step": 29250 }, { "epoch": 1.963122042884467, "grad_norm": 0.9383867383003235, "learning_rate": 8.865362397263433e-08, "loss": 0.433, "step": 29252 }, { "epoch": 1.9632562665682358, "grad_norm": 1.1528648138046265, "learning_rate": 8.800788763559386e-08, "loss": 0.4796, "step": 29254 }, { "epoch": 1.963390490252005, "grad_norm": 0.9118273258209229, "learning_rate": 8.736450954203168e-08, "loss": 0.4181, "step": 29256 }, { "epoch": 1.963524713935774, "grad_norm": 0.8596152663230896, "learning_rate": 8.672348972235122e-08, "loss": 0.4154, "step": 29258 }, { "epoch": 1.963658937619543, "grad_norm": 1.0593302249908447, "learning_rate": 8.608482820682828e-08, "loss": 0.4761, "step": 29260 }, { "epoch": 1.963793161303312, "grad_norm": 1.0908609628677368, "learning_rate": 8.544852502565537e-08, "loss": 0.4457, "step": 29262 }, { "epoch": 1.9639273849870809, "grad_norm": 0.8625156879425049, "learning_rate": 8.481458020888066e-08, "loss": 0.4576, "step": 29264 }, { "epoch": 1.96406160867085, "grad_norm": 1.1236494779586792, "learning_rate": 8.418299378646355e-08, "loss": 0.5115, "step": 29266 }, { "epoch": 1.964195832354619, "grad_norm": 1.000876545906067, "learning_rate": 8.355376578824681e-08, "loss": 0.4208, "step": 29268 }, { "epoch": 1.964330056038388, "grad_norm": 1.1040709018707275, "learning_rate": 8.292689624395666e-08, "loss": 0.4387, "step": 29270 }, { "epoch": 1.964464279722157, "grad_norm": 0.9892479777336121, "learning_rate": 8.230238518321387e-08, "loss": 0.4468, "step": 29272 }, { "epoch": 1.964598503405926, "grad_norm": 1.0196044445037842, "learning_rate": 8.168023263552815e-08, "loss": 0.4451, "step": 29274 }, { "epoch": 1.9647327270896948, "grad_norm": 0.9883550405502319, "learning_rate": 8.106043863028157e-08, "loss": 0.4487, "step": 29276 }, { "epoch": 1.964866950773464, "grad_norm": 1.0349782705307007, "learning_rate": 8.044300319677844e-08, "loss": 0.4504, "step": 29278 }, { "epoch": 1.965001174457233, "grad_norm": 1.0604530572891235, "learning_rate": 7.982792636417324e-08, "loss": 0.4185, "step": 29280 }, { "epoch": 1.965135398141002, "grad_norm": 1.0895187854766846, "learning_rate": 7.921520816153716e-08, "loss": 0.4678, "step": 29282 }, { "epoch": 1.965269621824771, "grad_norm": 0.9991888999938965, "learning_rate": 7.860484861781925e-08, "loss": 0.4507, "step": 29284 }, { "epoch": 1.9654038455085399, "grad_norm": 0.9821450114250183, "learning_rate": 7.799684776185201e-08, "loss": 0.4439, "step": 29286 }, { "epoch": 1.965538069192309, "grad_norm": 1.1980267763137817, "learning_rate": 7.739120562236802e-08, "loss": 0.5169, "step": 29288 }, { "epoch": 1.9656722928760781, "grad_norm": 0.896367609500885, "learning_rate": 7.678792222798325e-08, "loss": 0.4158, "step": 29290 }, { "epoch": 1.965806516559847, "grad_norm": 1.0855690240859985, "learning_rate": 7.618699760719716e-08, "loss": 0.4198, "step": 29292 }, { "epoch": 1.965940740243616, "grad_norm": 1.1153912544250488, "learning_rate": 7.558843178840924e-08, "loss": 0.4523, "step": 29294 }, { "epoch": 1.966074963927385, "grad_norm": 0.9311449527740479, "learning_rate": 7.499222479989132e-08, "loss": 0.39, "step": 29296 }, { "epoch": 1.966209187611154, "grad_norm": 1.0853956937789917, "learning_rate": 7.43983766698153e-08, "loss": 0.474, "step": 29298 }, { "epoch": 1.966343411294923, "grad_norm": 0.9984995126724243, "learning_rate": 7.380688742624209e-08, "loss": 0.4446, "step": 29300 }, { "epoch": 1.966477634978692, "grad_norm": 1.0153001546859741, "learning_rate": 7.321775709712153e-08, "loss": 0.4443, "step": 29302 }, { "epoch": 1.966611858662461, "grad_norm": 1.0898973941802979, "learning_rate": 7.263098571028138e-08, "loss": 0.5207, "step": 29304 }, { "epoch": 1.96674608234623, "grad_norm": 1.0372871160507202, "learning_rate": 7.204657329345498e-08, "loss": 0.4754, "step": 29306 }, { "epoch": 1.9668803060299989, "grad_norm": 1.0598440170288086, "learning_rate": 7.146451987424252e-08, "loss": 0.4786, "step": 29308 }, { "epoch": 1.967014529713768, "grad_norm": 1.0670109987258911, "learning_rate": 7.088482548015529e-08, "loss": 0.5069, "step": 29310 }, { "epoch": 1.9671487533975371, "grad_norm": 0.9652385115623474, "learning_rate": 7.030749013857696e-08, "loss": 0.4041, "step": 29312 }, { "epoch": 1.967282977081306, "grad_norm": 0.8895744681358337, "learning_rate": 6.97325138767857e-08, "loss": 0.4055, "step": 29314 }, { "epoch": 1.967417200765075, "grad_norm": 0.9850452542304993, "learning_rate": 6.915989672195422e-08, "loss": 0.4344, "step": 29316 }, { "epoch": 1.967551424448844, "grad_norm": 1.1335660219192505, "learning_rate": 6.858963870112756e-08, "loss": 0.4607, "step": 29318 }, { "epoch": 1.967685648132613, "grad_norm": 1.2784786224365234, "learning_rate": 6.802173984125637e-08, "loss": 0.4969, "step": 29320 }, { "epoch": 1.967819871816382, "grad_norm": 0.9682573080062866, "learning_rate": 6.745620016917476e-08, "loss": 0.4367, "step": 29322 }, { "epoch": 1.967954095500151, "grad_norm": 0.9108989238739014, "learning_rate": 6.689301971159467e-08, "loss": 0.4489, "step": 29324 }, { "epoch": 1.96808831918392, "grad_norm": 1.087185025215149, "learning_rate": 6.633219849513372e-08, "loss": 0.4899, "step": 29326 }, { "epoch": 1.968222542867689, "grad_norm": 1.101282000541687, "learning_rate": 6.577373654628183e-08, "loss": 0.5098, "step": 29328 }, { "epoch": 1.9683567665514579, "grad_norm": 1.0934100151062012, "learning_rate": 6.521763389142899e-08, "loss": 0.4456, "step": 29330 }, { "epoch": 1.968490990235227, "grad_norm": 1.023734211921692, "learning_rate": 6.466389055685418e-08, "loss": 0.5027, "step": 29332 }, { "epoch": 1.9686252139189961, "grad_norm": 1.0267611742019653, "learning_rate": 6.411250656871426e-08, "loss": 0.4883, "step": 29334 }, { "epoch": 1.968759437602765, "grad_norm": 1.0175542831420898, "learning_rate": 6.356348195306616e-08, "loss": 0.4625, "step": 29336 }, { "epoch": 1.968893661286534, "grad_norm": 1.2178868055343628, "learning_rate": 6.301681673585025e-08, "loss": 0.5294, "step": 29338 }, { "epoch": 1.969027884970303, "grad_norm": 1.1020519733428955, "learning_rate": 6.24725109428903e-08, "loss": 0.4643, "step": 29340 }, { "epoch": 1.969162108654072, "grad_norm": 1.027024507522583, "learning_rate": 6.193056459990465e-08, "loss": 0.4884, "step": 29342 }, { "epoch": 1.969296332337841, "grad_norm": 0.9886162877082825, "learning_rate": 6.139097773250057e-08, "loss": 0.4458, "step": 29344 }, { "epoch": 1.96943055602161, "grad_norm": 1.0386093854904175, "learning_rate": 6.085375036617436e-08, "loss": 0.447, "step": 29346 }, { "epoch": 1.969564779705379, "grad_norm": 1.0407241582870483, "learning_rate": 6.031888252630569e-08, "loss": 0.435, "step": 29348 }, { "epoch": 1.969699003389148, "grad_norm": 1.0747627019882202, "learning_rate": 5.97863742381688e-08, "loss": 0.4715, "step": 29350 }, { "epoch": 1.9698332270729169, "grad_norm": 1.035626769065857, "learning_rate": 5.9256225526921336e-08, "loss": 0.5026, "step": 29352 }, { "epoch": 1.969967450756686, "grad_norm": 1.252779483795166, "learning_rate": 5.8728436417615494e-08, "loss": 0.4384, "step": 29354 }, { "epoch": 1.9701016744404551, "grad_norm": 0.9608078002929688, "learning_rate": 5.820300693518133e-08, "loss": 0.4549, "step": 29356 }, { "epoch": 1.970235898124224, "grad_norm": 1.155197024345398, "learning_rate": 5.7679937104454516e-08, "loss": 0.4433, "step": 29358 }, { "epoch": 1.970370121807993, "grad_norm": 1.0186712741851807, "learning_rate": 5.715922695013753e-08, "loss": 0.4711, "step": 29360 }, { "epoch": 1.970504345491762, "grad_norm": 1.0469697713851929, "learning_rate": 5.664087649684402e-08, "loss": 0.4341, "step": 29362 }, { "epoch": 1.970638569175531, "grad_norm": 1.1047803163528442, "learning_rate": 5.6124885769054394e-08, "loss": 0.4584, "step": 29364 }, { "epoch": 1.9707727928593002, "grad_norm": 1.0283458232879639, "learning_rate": 5.5611254791154696e-08, "loss": 0.4277, "step": 29366 }, { "epoch": 1.970907016543069, "grad_norm": 1.0908386707305908, "learning_rate": 5.509998358741442e-08, "loss": 0.4204, "step": 29368 }, { "epoch": 1.971041240226838, "grad_norm": 1.1800463199615479, "learning_rate": 5.4591072181986444e-08, "loss": 0.4752, "step": 29370 }, { "epoch": 1.971175463910607, "grad_norm": 0.9975547194480896, "learning_rate": 5.408452059891822e-08, "loss": 0.4199, "step": 29372 }, { "epoch": 1.971309687594376, "grad_norm": 1.0553650856018066, "learning_rate": 5.358032886214059e-08, "loss": 0.4354, "step": 29374 }, { "epoch": 1.971443911278145, "grad_norm": 0.9321396350860596, "learning_rate": 5.307849699547895e-08, "loss": 0.4533, "step": 29376 }, { "epoch": 1.9715781349619141, "grad_norm": 1.0798766613006592, "learning_rate": 5.257902502263656e-08, "loss": 0.4708, "step": 29378 }, { "epoch": 1.971712358645683, "grad_norm": 1.0483887195587158, "learning_rate": 5.208191296722231e-08, "loss": 0.4684, "step": 29380 }, { "epoch": 1.971846582329452, "grad_norm": 1.0603145360946655, "learning_rate": 5.158716085271742e-08, "loss": 0.4299, "step": 29382 }, { "epoch": 1.971980806013221, "grad_norm": 1.0174487829208374, "learning_rate": 5.109476870250318e-08, "loss": 0.4365, "step": 29384 }, { "epoch": 1.97211502969699, "grad_norm": 0.9578461647033691, "learning_rate": 5.060473653983877e-08, "loss": 0.4683, "step": 29386 }, { "epoch": 1.9722492533807592, "grad_norm": 1.0425300598144531, "learning_rate": 5.0117064387877885e-08, "loss": 0.4852, "step": 29388 }, { "epoch": 1.972383477064528, "grad_norm": 1.0657423734664917, "learning_rate": 4.9631752269663213e-08, "loss": 0.4743, "step": 29390 }, { "epoch": 1.972517700748297, "grad_norm": 1.0595128536224365, "learning_rate": 4.91488002081264e-08, "loss": 0.4258, "step": 29392 }, { "epoch": 1.972651924432066, "grad_norm": 0.9602314829826355, "learning_rate": 4.8668208226088085e-08, "loss": 0.3623, "step": 29394 }, { "epoch": 1.972786148115835, "grad_norm": 0.9397258758544922, "learning_rate": 4.818997634624678e-08, "loss": 0.497, "step": 29396 }, { "epoch": 1.972920371799604, "grad_norm": 1.1596510410308838, "learning_rate": 4.771410459120662e-08, "loss": 0.458, "step": 29398 }, { "epoch": 1.9730545954833731, "grad_norm": 1.1102371215820312, "learning_rate": 4.724059298344408e-08, "loss": 0.4556, "step": 29400 }, { "epoch": 1.973188819167142, "grad_norm": 1.0936411619186401, "learning_rate": 4.676944154533569e-08, "loss": 0.459, "step": 29402 }, { "epoch": 1.973323042850911, "grad_norm": 1.1787388324737549, "learning_rate": 4.630065029914698e-08, "loss": 0.4643, "step": 29404 }, { "epoch": 1.97345726653468, "grad_norm": 0.9891636371612549, "learning_rate": 4.583421926701581e-08, "loss": 0.4165, "step": 29406 }, { "epoch": 1.973591490218449, "grad_norm": 1.1265305280685425, "learning_rate": 4.537014847099119e-08, "loss": 0.4794, "step": 29408 }, { "epoch": 1.9737257139022182, "grad_norm": 1.053418517112732, "learning_rate": 4.490843793300003e-08, "loss": 0.4857, "step": 29410 }, { "epoch": 1.973859937585987, "grad_norm": 1.0158971548080444, "learning_rate": 4.4449087674847125e-08, "loss": 0.4282, "step": 29412 }, { "epoch": 1.973994161269756, "grad_norm": 0.9685587882995605, "learning_rate": 4.399209771824287e-08, "loss": 0.4591, "step": 29414 }, { "epoch": 1.974128384953525, "grad_norm": 0.9691962003707886, "learning_rate": 4.353746808477554e-08, "loss": 0.472, "step": 29416 }, { "epoch": 1.974262608637294, "grad_norm": 0.9528734683990479, "learning_rate": 4.3085198795933536e-08, "loss": 0.4626, "step": 29418 }, { "epoch": 1.974396832321063, "grad_norm": 1.0498555898666382, "learning_rate": 4.263528987307197e-08, "loss": 0.4523, "step": 29420 }, { "epoch": 1.9745310560048321, "grad_norm": 1.0218913555145264, "learning_rate": 4.2187741337462724e-08, "loss": 0.4251, "step": 29422 }, { "epoch": 1.974665279688601, "grad_norm": 1.146499752998352, "learning_rate": 4.1742553210238896e-08, "loss": 0.5049, "step": 29424 }, { "epoch": 1.97479950337237, "grad_norm": 1.0821541547775269, "learning_rate": 4.129972551244476e-08, "loss": 0.4064, "step": 29426 }, { "epoch": 1.974933727056139, "grad_norm": 1.0310195684432983, "learning_rate": 4.085925826499692e-08, "loss": 0.5104, "step": 29428 }, { "epoch": 1.975067950739908, "grad_norm": 1.0251456499099731, "learning_rate": 4.0421151488712064e-08, "loss": 0.4923, "step": 29430 }, { "epoch": 1.9752021744236772, "grad_norm": 1.0707616806030273, "learning_rate": 3.998540520428473e-08, "loss": 0.5032, "step": 29432 }, { "epoch": 1.975336398107446, "grad_norm": 1.109976887702942, "learning_rate": 3.955201943230402e-08, "loss": 0.4809, "step": 29434 }, { "epoch": 1.975470621791215, "grad_norm": 0.980389416217804, "learning_rate": 3.9120994193247994e-08, "loss": 0.4546, "step": 29436 }, { "epoch": 1.975604845474984, "grad_norm": 1.1692322492599487, "learning_rate": 3.869232950747814e-08, "loss": 0.4648, "step": 29438 }, { "epoch": 1.975739069158753, "grad_norm": 1.0092170238494873, "learning_rate": 3.826602539525603e-08, "loss": 0.3806, "step": 29440 }, { "epoch": 1.9758732928425222, "grad_norm": 0.8585953712463379, "learning_rate": 3.784208187671556e-08, "loss": 0.4346, "step": 29442 }, { "epoch": 1.9760075165262911, "grad_norm": 1.019691824913025, "learning_rate": 3.7420498971890706e-08, "loss": 0.423, "step": 29444 }, { "epoch": 1.97614174021006, "grad_norm": 1.096204400062561, "learning_rate": 3.700127670070441e-08, "loss": 0.4462, "step": 29446 }, { "epoch": 1.976275963893829, "grad_norm": 0.977575957775116, "learning_rate": 3.658441508295196e-08, "loss": 0.4838, "step": 29448 }, { "epoch": 1.9764101875775981, "grad_norm": 0.9653611779212952, "learning_rate": 3.616991413834536e-08, "loss": 0.5223, "step": 29450 }, { "epoch": 1.976544411261367, "grad_norm": 1.0307095050811768, "learning_rate": 3.57577738864523e-08, "loss": 0.5271, "step": 29452 }, { "epoch": 1.9766786349451362, "grad_norm": 1.3023886680603027, "learning_rate": 3.534799434676273e-08, "loss": 0.4955, "step": 29454 }, { "epoch": 1.976812858628905, "grad_norm": 0.9763998985290527, "learning_rate": 3.494057553862229e-08, "loss": 0.4732, "step": 29456 }, { "epoch": 1.976947082312674, "grad_norm": 1.1294244527816772, "learning_rate": 3.453551748128781e-08, "loss": 0.4562, "step": 29458 }, { "epoch": 1.977081305996443, "grad_norm": 0.9462315440177917, "learning_rate": 3.4132820193899514e-08, "loss": 0.4131, "step": 29460 }, { "epoch": 1.977215529680212, "grad_norm": 1.0207066535949707, "learning_rate": 3.3732483695481097e-08, "loss": 0.4648, "step": 29462 }, { "epoch": 1.9773497533639812, "grad_norm": 0.995894193649292, "learning_rate": 3.333450800495075e-08, "loss": 0.4475, "step": 29464 }, { "epoch": 1.9774839770477501, "grad_norm": 1.0989102125167847, "learning_rate": 3.2938893141110094e-08, "loss": 0.4478, "step": 29466 }, { "epoch": 1.977618200731519, "grad_norm": 1.0410295724868774, "learning_rate": 3.254563912264419e-08, "loss": 0.5155, "step": 29468 }, { "epoch": 1.977752424415288, "grad_norm": 0.9531170129776001, "learning_rate": 3.215474596814372e-08, "loss": 0.4988, "step": 29470 }, { "epoch": 1.9778866480990571, "grad_norm": 0.8869547247886658, "learning_rate": 3.176621369607724e-08, "loss": 0.452, "step": 29472 }, { "epoch": 1.978020871782826, "grad_norm": 1.021654725074768, "learning_rate": 3.138004232479674e-08, "loss": 0.4554, "step": 29474 }, { "epoch": 1.9781550954665952, "grad_norm": 1.0772607326507568, "learning_rate": 3.099623187254874e-08, "loss": 0.4541, "step": 29476 }, { "epoch": 1.978289319150364, "grad_norm": 1.1076545715332031, "learning_rate": 3.061478235746873e-08, "loss": 0.4521, "step": 29478 }, { "epoch": 1.978423542834133, "grad_norm": 0.977824866771698, "learning_rate": 3.023569379758118e-08, "loss": 0.4835, "step": 29480 }, { "epoch": 1.978557766517902, "grad_norm": 0.9521993398666382, "learning_rate": 2.985896621079398e-08, "loss": 0.4299, "step": 29482 }, { "epoch": 1.978691990201671, "grad_norm": 1.0588092803955078, "learning_rate": 2.948459961490957e-08, "loss": 0.4646, "step": 29484 }, { "epoch": 1.9788262138854402, "grad_norm": 1.0376704931259155, "learning_rate": 2.9112594027619346e-08, "loss": 0.4729, "step": 29486 }, { "epoch": 1.9789604375692091, "grad_norm": 1.0790448188781738, "learning_rate": 2.8742949466487036e-08, "loss": 0.4117, "step": 29488 }, { "epoch": 1.979094661252978, "grad_norm": 1.0049660205841064, "learning_rate": 2.8375665948993103e-08, "loss": 0.4509, "step": 29490 }, { "epoch": 1.979228884936747, "grad_norm": 1.0287073850631714, "learning_rate": 2.801074349247923e-08, "loss": 0.446, "step": 29492 }, { "epoch": 1.9793631086205161, "grad_norm": 0.9084295630455017, "learning_rate": 2.7648182114198285e-08, "loss": 0.4474, "step": 29494 }, { "epoch": 1.979497332304285, "grad_norm": 1.1138800382614136, "learning_rate": 2.7287981831269905e-08, "loss": 0.4446, "step": 29496 }, { "epoch": 1.9796315559880542, "grad_norm": 1.0034202337265015, "learning_rate": 2.693014266071381e-08, "loss": 0.3988, "step": 29498 }, { "epoch": 1.979765779671823, "grad_norm": 0.9081588387489319, "learning_rate": 2.6574664619444244e-08, "loss": 0.4977, "step": 29500 }, { "epoch": 1.979900003355592, "grad_norm": 1.0774911642074585, "learning_rate": 2.6221547724253337e-08, "loss": 0.4847, "step": 29502 }, { "epoch": 1.980034227039361, "grad_norm": 0.9915654063224792, "learning_rate": 2.5870791991827737e-08, "loss": 0.529, "step": 29504 }, { "epoch": 1.98016845072313, "grad_norm": 1.0009660720825195, "learning_rate": 2.552239743873197e-08, "loss": 0.4338, "step": 29506 }, { "epoch": 1.9803026744068992, "grad_norm": 1.005157709121704, "learning_rate": 2.51763640814362e-08, "loss": 0.4364, "step": 29508 }, { "epoch": 1.9804368980906681, "grad_norm": 1.1098984479904175, "learning_rate": 2.4832691936282902e-08, "loss": 0.4328, "step": 29510 }, { "epoch": 1.980571121774437, "grad_norm": 0.9147427082061768, "learning_rate": 2.4491381019520198e-08, "loss": 0.4075, "step": 29512 }, { "epoch": 1.980705345458206, "grad_norm": 1.0833896398544312, "learning_rate": 2.415243134725742e-08, "loss": 0.4841, "step": 29514 }, { "epoch": 1.9808395691419751, "grad_norm": 0.9254489541053772, "learning_rate": 2.381584293552619e-08, "loss": 0.4969, "step": 29516 }, { "epoch": 1.9809737928257443, "grad_norm": 1.0122753381729126, "learning_rate": 2.3481615800219347e-08, "loss": 0.4171, "step": 29518 }, { "epoch": 1.9811080165095132, "grad_norm": 1.0806469917297363, "learning_rate": 2.3149749957129818e-08, "loss": 0.4285, "step": 29520 }, { "epoch": 1.981242240193282, "grad_norm": 1.0207287073135376, "learning_rate": 2.28202454219395e-08, "loss": 0.4306, "step": 29522 }, { "epoch": 1.981376463877051, "grad_norm": 1.008135199546814, "learning_rate": 2.2493102210219275e-08, "loss": 0.4515, "step": 29524 }, { "epoch": 1.9815106875608202, "grad_norm": 1.120957612991333, "learning_rate": 2.2168320337423442e-08, "loss": 0.4642, "step": 29526 }, { "epoch": 1.981644911244589, "grad_norm": 1.0079890489578247, "learning_rate": 2.1845899818895287e-08, "loss": 0.4304, "step": 29528 }, { "epoch": 1.9817791349283582, "grad_norm": 1.0214899778366089, "learning_rate": 2.152584066987262e-08, "loss": 0.4601, "step": 29530 }, { "epoch": 1.9819133586121271, "grad_norm": 0.9742206335067749, "learning_rate": 2.120814290547668e-08, "loss": 0.4746, "step": 29532 }, { "epoch": 1.982047582295896, "grad_norm": 1.4604252576828003, "learning_rate": 2.089280654071213e-08, "loss": 0.4276, "step": 29534 }, { "epoch": 1.982181805979665, "grad_norm": 1.1026074886322021, "learning_rate": 2.057983159048926e-08, "loss": 0.4672, "step": 29536 }, { "epoch": 1.9823160296634341, "grad_norm": 1.0693613290786743, "learning_rate": 2.026921806958515e-08, "loss": 0.4698, "step": 29538 }, { "epoch": 1.9824502533472033, "grad_norm": 1.0676215887069702, "learning_rate": 1.996096599267694e-08, "loss": 0.4791, "step": 29540 }, { "epoch": 1.9825844770309722, "grad_norm": 0.9879465699195862, "learning_rate": 1.965507537433631e-08, "loss": 0.4965, "step": 29542 }, { "epoch": 1.982718700714741, "grad_norm": 0.9821873307228088, "learning_rate": 1.9351546229007256e-08, "loss": 0.4378, "step": 29544 }, { "epoch": 1.98285292439851, "grad_norm": 1.057045340538025, "learning_rate": 1.9050378571039418e-08, "loss": 0.4422, "step": 29546 }, { "epoch": 1.9829871480822792, "grad_norm": 1.1257404088974, "learning_rate": 1.875157241465475e-08, "loss": 0.4876, "step": 29548 }, { "epoch": 1.983121371766048, "grad_norm": 0.8850076794624329, "learning_rate": 1.845512777397529e-08, "loss": 0.4808, "step": 29550 }, { "epoch": 1.9832555954498172, "grad_norm": 1.1537803411483765, "learning_rate": 1.8161044663000948e-08, "loss": 0.4219, "step": 29552 }, { "epoch": 1.9833898191335861, "grad_norm": 1.0371828079223633, "learning_rate": 1.786932309564282e-08, "loss": 0.4715, "step": 29554 }, { "epoch": 1.983524042817355, "grad_norm": 0.9885537028312683, "learning_rate": 1.7579963085667672e-08, "loss": 0.471, "step": 29556 }, { "epoch": 1.983658266501124, "grad_norm": 0.9527798295021057, "learning_rate": 1.7292964646753453e-08, "loss": 0.4321, "step": 29558 }, { "epoch": 1.9837924901848931, "grad_norm": 1.0112323760986328, "learning_rate": 1.700832779245598e-08, "loss": 0.4197, "step": 29560 }, { "epoch": 1.9839267138686623, "grad_norm": 0.7867613434791565, "learning_rate": 1.672605253623116e-08, "loss": 0.4374, "step": 29562 }, { "epoch": 1.9840609375524312, "grad_norm": 0.9947951436042786, "learning_rate": 1.6446138891412777e-08, "loss": 0.4336, "step": 29564 }, { "epoch": 1.9841951612362, "grad_norm": 1.0333178043365479, "learning_rate": 1.616858687122913e-08, "loss": 0.4762, "step": 29566 }, { "epoch": 1.984329384919969, "grad_norm": 1.0098247528076172, "learning_rate": 1.5893396488786407e-08, "loss": 0.4625, "step": 29568 }, { "epoch": 1.9844636086037382, "grad_norm": 0.9545372128486633, "learning_rate": 1.5620567757090865e-08, "loss": 0.4488, "step": 29570 }, { "epoch": 1.984597832287507, "grad_norm": 0.9851608276367188, "learning_rate": 1.535010068903775e-08, "loss": 0.4743, "step": 29572 }, { "epoch": 1.9847320559712762, "grad_norm": 1.3361846208572388, "learning_rate": 1.5081995297400177e-08, "loss": 0.4902, "step": 29574 }, { "epoch": 1.9848662796550451, "grad_norm": 1.1021267175674438, "learning_rate": 1.4816251594845787e-08, "loss": 0.4729, "step": 29576 }, { "epoch": 1.985000503338814, "grad_norm": 1.0077482461929321, "learning_rate": 1.4552869593931207e-08, "loss": 0.4361, "step": 29578 }, { "epoch": 1.985134727022583, "grad_norm": 1.046972632408142, "learning_rate": 1.4291849307102034e-08, "loss": 0.4368, "step": 29580 }, { "epoch": 1.9852689507063521, "grad_norm": 0.9834826588630676, "learning_rate": 1.4033190746687297e-08, "loss": 0.5096, "step": 29582 }, { "epoch": 1.9854031743901213, "grad_norm": 1.05916428565979, "learning_rate": 1.37768939249161e-08, "loss": 0.4548, "step": 29584 }, { "epoch": 1.9855373980738902, "grad_norm": 0.9710681438446045, "learning_rate": 1.3522958853889877e-08, "loss": 0.5093, "step": 29586 }, { "epoch": 1.985671621757659, "grad_norm": 0.9356110692024231, "learning_rate": 1.3271385545610137e-08, "loss": 0.4597, "step": 29588 }, { "epoch": 1.985805845441428, "grad_norm": 1.1384472846984863, "learning_rate": 1.302217401196737e-08, "loss": 0.5089, "step": 29590 }, { "epoch": 1.9859400691251972, "grad_norm": 0.8650492429733276, "learning_rate": 1.2775324264724386e-08, "loss": 0.4664, "step": 29592 }, { "epoch": 1.9860742928089663, "grad_norm": 1.0428403615951538, "learning_rate": 1.2530836315555183e-08, "loss": 0.5201, "step": 29594 }, { "epoch": 1.9862085164927352, "grad_norm": 1.044148564338684, "learning_rate": 1.228871017601163e-08, "loss": 0.5132, "step": 29596 }, { "epoch": 1.9863427401765041, "grad_norm": 1.0618407726287842, "learning_rate": 1.2048945857523475e-08, "loss": 0.4978, "step": 29598 }, { "epoch": 1.986476963860273, "grad_norm": 1.0523346662521362, "learning_rate": 1.1811543371431644e-08, "loss": 0.437, "step": 29600 }, { "epoch": 1.9866111875440422, "grad_norm": 0.9107396006584167, "learning_rate": 1.1576502728938287e-08, "loss": 0.456, "step": 29602 }, { "epoch": 1.9867454112278111, "grad_norm": 1.0899020433425903, "learning_rate": 1.134382394116229e-08, "loss": 0.4223, "step": 29604 }, { "epoch": 1.9868796349115803, "grad_norm": 0.9853198528289795, "learning_rate": 1.111350701909486e-08, "loss": 0.4546, "step": 29606 }, { "epoch": 1.9870138585953492, "grad_norm": 1.0362210273742676, "learning_rate": 1.088555197361063e-08, "loss": 0.5492, "step": 29608 }, { "epoch": 1.987148082279118, "grad_norm": 0.9253741502761841, "learning_rate": 1.0659958815489868e-08, "loss": 0.4201, "step": 29610 }, { "epoch": 1.987282305962887, "grad_norm": 1.087234377861023, "learning_rate": 1.043672755537961e-08, "loss": 0.4905, "step": 29612 }, { "epoch": 1.9874165296466562, "grad_norm": 0.912765383720398, "learning_rate": 1.021585820383808e-08, "loss": 0.4179, "step": 29614 }, { "epoch": 1.9875507533304253, "grad_norm": 0.8833821415901184, "learning_rate": 9.997350771295821e-09, "loss": 0.4319, "step": 29616 }, { "epoch": 1.9876849770141942, "grad_norm": 1.0871175527572632, "learning_rate": 9.781205268077908e-09, "loss": 0.4756, "step": 29618 }, { "epoch": 1.9878192006979631, "grad_norm": 1.0188250541687012, "learning_rate": 9.567421704392843e-09, "loss": 0.4835, "step": 29620 }, { "epoch": 1.987953424381732, "grad_norm": 1.129906415939331, "learning_rate": 9.356000090349204e-09, "loss": 0.4576, "step": 29622 }, { "epoch": 1.9880876480655012, "grad_norm": 0.8076863288879395, "learning_rate": 9.146940435933449e-09, "loss": 0.406, "step": 29624 }, { "epoch": 1.9882218717492701, "grad_norm": 1.024215579032898, "learning_rate": 8.94024275102101e-09, "loss": 0.4292, "step": 29626 }, { "epoch": 1.9883560954330393, "grad_norm": 0.9850940704345703, "learning_rate": 8.735907045376301e-09, "loss": 0.441, "step": 29628 }, { "epoch": 1.9884903191168082, "grad_norm": 1.049185872077942, "learning_rate": 8.533933328658262e-09, "loss": 0.5439, "step": 29630 }, { "epoch": 1.988624542800577, "grad_norm": 1.0200152397155762, "learning_rate": 8.334321610403706e-09, "loss": 0.4219, "step": 29632 }, { "epoch": 1.988758766484346, "grad_norm": 1.0623215436935425, "learning_rate": 8.137071900055082e-09, "loss": 0.5007, "step": 29634 }, { "epoch": 1.9888929901681152, "grad_norm": 1.196020483970642, "learning_rate": 7.942184206921611e-09, "loss": 0.5173, "step": 29636 }, { "epoch": 1.9890272138518843, "grad_norm": 1.135179042816162, "learning_rate": 7.74965854021259e-09, "loss": 0.5123, "step": 29638 }, { "epoch": 1.9891614375356532, "grad_norm": 1.08401358127594, "learning_rate": 7.5594949090263e-09, "loss": 0.5092, "step": 29640 }, { "epoch": 1.9892956612194221, "grad_norm": 0.9641003012657166, "learning_rate": 7.371693322349993e-09, "loss": 0.4967, "step": 29642 }, { "epoch": 1.989429884903191, "grad_norm": 1.1795014142990112, "learning_rate": 7.186253789059905e-09, "loss": 0.4712, "step": 29644 }, { "epoch": 1.9895641085869602, "grad_norm": 1.0693312883377075, "learning_rate": 7.003176317904591e-09, "loss": 0.4727, "step": 29646 }, { "epoch": 1.9896983322707291, "grad_norm": 1.0542054176330566, "learning_rate": 6.822460917549345e-09, "loss": 0.488, "step": 29648 }, { "epoch": 1.9898325559544983, "grad_norm": 0.99620121717453, "learning_rate": 6.644107596520677e-09, "loss": 0.4666, "step": 29650 }, { "epoch": 1.9899667796382672, "grad_norm": 1.1178016662597656, "learning_rate": 6.4681163632507314e-09, "loss": 0.499, "step": 29652 }, { "epoch": 1.990101003322036, "grad_norm": 1.054388403892517, "learning_rate": 6.294487226055079e-09, "loss": 0.4872, "step": 29654 }, { "epoch": 1.990235227005805, "grad_norm": 1.0646460056304932, "learning_rate": 6.123220193132717e-09, "loss": 0.4861, "step": 29656 }, { "epoch": 1.9903694506895742, "grad_norm": 1.1594460010528564, "learning_rate": 5.9543152725827226e-09, "loss": 0.4703, "step": 29658 }, { "epoch": 1.9905036743733433, "grad_norm": 1.0529216527938843, "learning_rate": 5.787772472382047e-09, "loss": 0.4481, "step": 29660 }, { "epoch": 1.9906378980571122, "grad_norm": 0.9020050168037415, "learning_rate": 5.623591800402173e-09, "loss": 0.4546, "step": 29662 }, { "epoch": 1.9907721217408811, "grad_norm": 0.9900140166282654, "learning_rate": 5.461773264398007e-09, "loss": 0.4591, "step": 29664 }, { "epoch": 1.99090634542465, "grad_norm": 1.078150987625122, "learning_rate": 5.302316872013435e-09, "loss": 0.5246, "step": 29666 }, { "epoch": 1.9910405691084192, "grad_norm": 0.9780765771865845, "learning_rate": 5.145222630781321e-09, "loss": 0.4966, "step": 29668 }, { "epoch": 1.9911747927921883, "grad_norm": 1.0546802282333374, "learning_rate": 4.990490548129057e-09, "loss": 0.4696, "step": 29670 }, { "epoch": 1.9913090164759573, "grad_norm": 0.9650538563728333, "learning_rate": 4.838120631361909e-09, "loss": 0.5195, "step": 29672 }, { "epoch": 1.9914432401597262, "grad_norm": 1.0090038776397705, "learning_rate": 4.688112887685225e-09, "loss": 0.4585, "step": 29674 }, { "epoch": 1.991577463843495, "grad_norm": 1.0023903846740723, "learning_rate": 4.540467324187781e-09, "loss": 0.4605, "step": 29676 }, { "epoch": 1.9917116875272642, "grad_norm": 0.8548973202705383, "learning_rate": 4.3951839478362235e-09, "loss": 0.4265, "step": 29678 }, { "epoch": 1.9918459112110332, "grad_norm": 1.024346947669983, "learning_rate": 4.2522627655028346e-09, "loss": 0.3986, "step": 29680 }, { "epoch": 1.9919801348948023, "grad_norm": 1.0842660665512085, "learning_rate": 4.111703783932219e-09, "loss": 0.4498, "step": 29682 }, { "epoch": 1.9921143585785712, "grad_norm": 0.9837660193443298, "learning_rate": 3.973507009774613e-09, "loss": 0.477, "step": 29684 }, { "epoch": 1.9922485822623401, "grad_norm": 0.9357654452323914, "learning_rate": 3.8376724495581276e-09, "loss": 0.428, "step": 29686 }, { "epoch": 1.992382805946109, "grad_norm": 0.9697017669677734, "learning_rate": 3.7042001096943e-09, "loss": 0.3818, "step": 29688 }, { "epoch": 1.9925170296298782, "grad_norm": 1.0058552026748657, "learning_rate": 3.5730899964947495e-09, "loss": 0.453, "step": 29690 }, { "epoch": 1.9926512533136473, "grad_norm": 1.1216505765914917, "learning_rate": 3.4443421161545197e-09, "loss": 0.4774, "step": 29692 }, { "epoch": 1.9927854769974163, "grad_norm": 1.0027047395706177, "learning_rate": 3.317956474757633e-09, "loss": 0.4588, "step": 29694 }, { "epoch": 1.9929197006811852, "grad_norm": 1.1020907163619995, "learning_rate": 3.193933078265987e-09, "loss": 0.4751, "step": 29696 }, { "epoch": 1.993053924364954, "grad_norm": 1.0000810623168945, "learning_rate": 3.0722719325526615e-09, "loss": 0.4252, "step": 29698 }, { "epoch": 1.9931881480487232, "grad_norm": 1.25977623462677, "learning_rate": 2.95297304335751e-09, "loss": 0.41, "step": 29700 }, { "epoch": 1.9933223717324922, "grad_norm": 1.0130668878555298, "learning_rate": 2.8360364163149135e-09, "loss": 0.4406, "step": 29702 }, { "epoch": 1.9934565954162613, "grad_norm": 1.1364526748657227, "learning_rate": 2.721462056959334e-09, "loss": 0.4426, "step": 29704 }, { "epoch": 1.9935908191000302, "grad_norm": 0.9602410793304443, "learning_rate": 2.609249970697558e-09, "loss": 0.4955, "step": 29706 }, { "epoch": 1.9937250427837991, "grad_norm": 0.9666051268577576, "learning_rate": 2.4994001628364517e-09, "loss": 0.5171, "step": 29708 }, { "epoch": 1.993859266467568, "grad_norm": 1.0599029064178467, "learning_rate": 2.391912638560756e-09, "loss": 0.4649, "step": 29710 }, { "epoch": 1.9939934901513372, "grad_norm": 0.8977524042129517, "learning_rate": 2.2867874029497415e-09, "loss": 0.4292, "step": 29712 }, { "epoch": 1.9941277138351063, "grad_norm": 1.0639896392822266, "learning_rate": 2.1840244609716565e-09, "loss": 0.4996, "step": 29714 }, { "epoch": 1.9942619375188753, "grad_norm": 1.0090793371200562, "learning_rate": 2.0836238174837264e-09, "loss": 0.4571, "step": 29716 }, { "epoch": 1.9943961612026442, "grad_norm": 1.0224015712738037, "learning_rate": 1.9855854772266037e-09, "loss": 0.4612, "step": 29718 }, { "epoch": 1.994530384886413, "grad_norm": 1.0634790658950806, "learning_rate": 1.8899094448354693e-09, "loss": 0.4556, "step": 29720 }, { "epoch": 1.9946646085701822, "grad_norm": 1.0072276592254639, "learning_rate": 1.796595724828931e-09, "loss": 0.4535, "step": 29722 }, { "epoch": 1.9947988322539512, "grad_norm": 1.0097860097885132, "learning_rate": 1.7056443216145746e-09, "loss": 0.458, "step": 29724 }, { "epoch": 1.9949330559377203, "grad_norm": 1.0248351097106934, "learning_rate": 1.6170552394889627e-09, "loss": 0.4061, "step": 29726 }, { "epoch": 1.9950672796214892, "grad_norm": 0.9212617874145508, "learning_rate": 1.5308284826431874e-09, "loss": 0.4112, "step": 29728 }, { "epoch": 1.9952015033052581, "grad_norm": 1.416978120803833, "learning_rate": 1.446964055146216e-09, "loss": 0.491, "step": 29730 }, { "epoch": 1.995335726989027, "grad_norm": 1.0837624073028564, "learning_rate": 1.365461960961545e-09, "loss": 0.465, "step": 29732 }, { "epoch": 1.9954699506727962, "grad_norm": 1.1062414646148682, "learning_rate": 1.2863222039416478e-09, "loss": 0.4684, "step": 29734 }, { "epoch": 1.9956041743565653, "grad_norm": 1.0122276544570923, "learning_rate": 1.2095447878279765e-09, "loss": 0.4579, "step": 29736 }, { "epoch": 1.9957383980403343, "grad_norm": 1.0660096406936646, "learning_rate": 1.1351297162398578e-09, "loss": 0.4577, "step": 29738 }, { "epoch": 1.9958726217241032, "grad_norm": 0.970309317111969, "learning_rate": 1.06307699270225e-09, "loss": 0.4796, "step": 29740 }, { "epoch": 1.996006845407872, "grad_norm": 0.8794416785240173, "learning_rate": 9.933866206124353e-10, "loss": 0.4244, "step": 29742 }, { "epoch": 1.9961410690916412, "grad_norm": 0.9972506165504456, "learning_rate": 9.260586032677765e-10, "loss": 0.4375, "step": 29744 }, { "epoch": 1.9962752927754104, "grad_norm": 1.0356035232543945, "learning_rate": 8.610929438490623e-10, "loss": 0.4301, "step": 29746 }, { "epoch": 1.9964095164591793, "grad_norm": 1.0438393354415894, "learning_rate": 7.984896454260593e-10, "loss": 0.4381, "step": 29748 }, { "epoch": 1.9965437401429482, "grad_norm": 0.9698221683502197, "learning_rate": 7.382487109519609e-10, "loss": 0.4574, "step": 29750 }, { "epoch": 1.9966779638267171, "grad_norm": 0.9586147665977478, "learning_rate": 6.803701432744891e-10, "loss": 0.4187, "step": 29752 }, { "epoch": 1.9968121875104863, "grad_norm": 1.0892906188964844, "learning_rate": 6.248539451358948e-10, "loss": 0.5244, "step": 29754 }, { "epoch": 1.9969464111942552, "grad_norm": 0.9858735799789429, "learning_rate": 5.71700119145202e-10, "loss": 0.4165, "step": 29756 }, { "epoch": 1.9970806348780243, "grad_norm": 1.0821709632873535, "learning_rate": 5.209086678281683e-10, "loss": 0.4559, "step": 29758 }, { "epoch": 1.9972148585617933, "grad_norm": 1.021770715713501, "learning_rate": 4.724795935773241e-10, "loss": 0.5047, "step": 29760 }, { "epoch": 1.9973490822455622, "grad_norm": 1.1175012588500977, "learning_rate": 4.2641289868528e-10, "loss": 0.57, "step": 29762 }, { "epoch": 1.997483305929331, "grad_norm": 1.196708083152771, "learning_rate": 3.82708585316971e-10, "loss": 0.4568, "step": 29764 }, { "epoch": 1.9976175296131002, "grad_norm": 1.007369875907898, "learning_rate": 3.413666555540651e-10, "loss": 0.5123, "step": 29766 }, { "epoch": 1.9977517532968694, "grad_norm": 1.0676368474960327, "learning_rate": 3.023871113339016e-10, "loss": 0.4617, "step": 29768 }, { "epoch": 1.9978859769806383, "grad_norm": 1.0238417387008667, "learning_rate": 2.65769954510553e-10, "loss": 0.4486, "step": 29770 }, { "epoch": 1.9980202006644072, "grad_norm": 1.0677101612091064, "learning_rate": 2.3151518681041595e-10, "loss": 0.4439, "step": 29772 }, { "epoch": 1.9981544243481761, "grad_norm": 1.0551823377609253, "learning_rate": 1.9962280984886507e-10, "loss": 0.4729, "step": 29774 }, { "epoch": 1.9982886480319453, "grad_norm": 1.0276211500167847, "learning_rate": 1.700928251358036e-10, "loss": 0.4413, "step": 29776 }, { "epoch": 1.9984228717157142, "grad_norm": 1.2995710372924805, "learning_rate": 1.429252340645615e-10, "loss": 0.5142, "step": 29778 }, { "epoch": 1.9985570953994833, "grad_norm": 1.126254916191101, "learning_rate": 1.181200379174463e-10, "loss": 0.4627, "step": 29780 }, { "epoch": 1.9986913190832523, "grad_norm": 0.997976541519165, "learning_rate": 9.567723787129445e-11, "loss": 0.4312, "step": 29782 }, { "epoch": 1.9988255427670212, "grad_norm": 0.9253402352333069, "learning_rate": 7.559683498081782e-11, "loss": 0.4378, "step": 29784 }, { "epoch": 1.99895976645079, "grad_norm": 1.0526634454727173, "learning_rate": 5.7878830200808196e-11, "loss": 0.4286, "step": 29786 }, { "epoch": 1.9990939901345592, "grad_norm": 1.0962398052215576, "learning_rate": 4.2523224363932856e-11, "loss": 0.4924, "step": 29788 }, { "epoch": 1.9992282138183284, "grad_norm": 1.120545506477356, "learning_rate": 2.953001819738788e-11, "loss": 0.4837, "step": 29790 }, { "epoch": 1.9993624375020973, "grad_norm": 1.0252364873886108, "learning_rate": 1.8899212317347036e-11, "loss": 0.4587, "step": 29792 }, { "epoch": 1.9994966611858662, "grad_norm": 0.8036876320838928, "learning_rate": 1.06308072234107e-11, "loss": 0.4413, "step": 29794 }, { "epoch": 1.9996308848696351, "grad_norm": 0.9679793119430542, "learning_rate": 4.7248033041569216e-12, "loss": 0.5315, "step": 29796 }, { "epoch": 1.9997651085534043, "grad_norm": 1.013262391090393, "learning_rate": 1.1812008371414608e-12, "loss": 0.4365, "step": 29798 }, { "epoch": 1.9998993322371732, "grad_norm": 1.0657564401626587, "learning_rate": 0.0, "loss": 0.4019, "step": 29800 } ], "logging_steps": 2, "max_steps": 29800, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.9092202292301005e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }