{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0036429872495445, "eval_steps": 500, "global_step": 1100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 0.46431864432643544, "learning_rate": 1.2121212121212122e-06, "loss": 1.4151, "step": 1 }, { "epoch": 0.0, "grad_norm": 0.42969176658539837, "learning_rate": 2.4242424242424244e-06, "loss": 1.3729, "step": 2 }, { "epoch": 0.01, "grad_norm": 0.5004307936270223, "learning_rate": 3.636363636363636e-06, "loss": 1.3989, "step": 3 }, { "epoch": 0.01, "grad_norm": 0.43666634920041486, "learning_rate": 4.848484848484849e-06, "loss": 1.3363, "step": 4 }, { "epoch": 0.01, "grad_norm": 0.4691114419353825, "learning_rate": 6.060606060606061e-06, "loss": 1.4293, "step": 5 }, { "epoch": 0.01, "grad_norm": 0.4277596061377729, "learning_rate": 7.272727272727272e-06, "loss": 1.4343, "step": 6 }, { "epoch": 0.01, "grad_norm": 0.4238339229382504, "learning_rate": 8.484848484848486e-06, "loss": 1.4462, "step": 7 }, { "epoch": 0.01, "grad_norm": 0.40992048534183273, "learning_rate": 9.696969696969698e-06, "loss": 1.2756, "step": 8 }, { "epoch": 0.02, "grad_norm": 0.37885700313540693, "learning_rate": 1.0909090909090909e-05, "loss": 1.3464, "step": 9 }, { "epoch": 0.02, "grad_norm": 0.36066141170123023, "learning_rate": 1.2121212121212122e-05, "loss": 1.3419, "step": 10 }, { "epoch": 0.02, "grad_norm": 0.35617169386863406, "learning_rate": 1.3333333333333333e-05, "loss": 1.3533, "step": 11 }, { "epoch": 0.02, "grad_norm": 0.3040188564782602, "learning_rate": 1.4545454545454545e-05, "loss": 1.2395, "step": 12 }, { "epoch": 0.02, "grad_norm": 0.31038319439216566, "learning_rate": 1.5757575757575756e-05, "loss": 1.3082, "step": 13 }, { "epoch": 0.03, "grad_norm": 0.26683768372135835, "learning_rate": 1.6969696969696972e-05, "loss": 1.3063, "step": 14 }, { "epoch": 0.03, "grad_norm": 0.3652323682563078, "learning_rate": 1.8181818181818182e-05, "loss": 1.3045, "step": 15 }, { "epoch": 0.03, "grad_norm": 0.23559121485457843, "learning_rate": 1.9393939393939395e-05, "loss": 1.2366, "step": 16 }, { "epoch": 0.03, "grad_norm": 0.2342299313020104, "learning_rate": 2.0606060606060608e-05, "loss": 1.2831, "step": 17 }, { "epoch": 0.03, "grad_norm": 0.2202931700357255, "learning_rate": 2.1818181818181818e-05, "loss": 1.3064, "step": 18 }, { "epoch": 0.03, "grad_norm": 0.2097660599292375, "learning_rate": 2.3030303030303034e-05, "loss": 1.2376, "step": 19 }, { "epoch": 0.04, "grad_norm": 0.2356785314652122, "learning_rate": 2.4242424242424244e-05, "loss": 1.2802, "step": 20 }, { "epoch": 0.04, "grad_norm": 0.24639302530564244, "learning_rate": 2.5454545454545454e-05, "loss": 1.3016, "step": 21 }, { "epoch": 0.04, "grad_norm": 0.24373126133228787, "learning_rate": 2.6666666666666667e-05, "loss": 1.3407, "step": 22 }, { "epoch": 0.04, "grad_norm": 0.24488805144123432, "learning_rate": 2.7878787878787883e-05, "loss": 1.3325, "step": 23 }, { "epoch": 0.04, "grad_norm": 0.2653033507571198, "learning_rate": 2.909090909090909e-05, "loss": 1.2811, "step": 24 }, { "epoch": 0.05, "grad_norm": 1.2841724819336817, "learning_rate": 3.0303030303030306e-05, "loss": 1.2837, "step": 25 }, { "epoch": 0.05, "grad_norm": 0.2183883020111492, "learning_rate": 3.151515151515151e-05, "loss": 1.2472, "step": 26 }, { "epoch": 0.05, "grad_norm": 0.2137995163762026, "learning_rate": 3.272727272727273e-05, "loss": 1.2854, "step": 27 }, { "epoch": 0.05, "grad_norm": 0.19499006223503876, "learning_rate": 3.3939393939393945e-05, "loss": 1.3018, "step": 28 }, { "epoch": 0.05, "grad_norm": 0.17367919355340256, "learning_rate": 3.515151515151515e-05, "loss": 1.2824, "step": 29 }, { "epoch": 0.05, "grad_norm": 0.18326045693683557, "learning_rate": 3.6363636363636364e-05, "loss": 1.2192, "step": 30 }, { "epoch": 0.06, "grad_norm": 0.17474388188066411, "learning_rate": 3.757575757575758e-05, "loss": 1.2078, "step": 31 }, { "epoch": 0.06, "grad_norm": 0.17856970178098716, "learning_rate": 3.878787878787879e-05, "loss": 1.2683, "step": 32 }, { "epoch": 0.06, "grad_norm": 0.18617589704298348, "learning_rate": 4e-05, "loss": 1.2265, "step": 33 }, { "epoch": 0.06, "grad_norm": 0.17653209733215317, "learning_rate": 4.1212121212121216e-05, "loss": 1.319, "step": 34 }, { "epoch": 0.06, "grad_norm": 0.1722921367585233, "learning_rate": 4.242424242424243e-05, "loss": 1.2117, "step": 35 }, { "epoch": 0.07, "grad_norm": 0.176642606378719, "learning_rate": 4.3636363636363636e-05, "loss": 1.2512, "step": 36 }, { "epoch": 0.07, "grad_norm": 0.16696442324691066, "learning_rate": 4.484848484848485e-05, "loss": 1.2637, "step": 37 }, { "epoch": 0.07, "grad_norm": 0.17035384059517106, "learning_rate": 4.606060606060607e-05, "loss": 1.2699, "step": 38 }, { "epoch": 0.07, "grad_norm": 0.15545801881444482, "learning_rate": 4.7272727272727275e-05, "loss": 1.2939, "step": 39 }, { "epoch": 0.07, "grad_norm": 0.17111439344347512, "learning_rate": 4.848484848484849e-05, "loss": 1.3033, "step": 40 }, { "epoch": 0.07, "grad_norm": 0.16994151343455458, "learning_rate": 4.9696969696969694e-05, "loss": 1.2603, "step": 41 }, { "epoch": 0.08, "grad_norm": 0.15929214926453447, "learning_rate": 5.090909090909091e-05, "loss": 1.2626, "step": 42 }, { "epoch": 0.08, "grad_norm": 0.16761261516238699, "learning_rate": 5.212121212121213e-05, "loss": 1.296, "step": 43 }, { "epoch": 0.08, "grad_norm": 0.15754700542426123, "learning_rate": 5.333333333333333e-05, "loss": 1.278, "step": 44 }, { "epoch": 0.08, "grad_norm": 0.15522526683877644, "learning_rate": 5.4545454545454546e-05, "loss": 1.2355, "step": 45 }, { "epoch": 0.08, "grad_norm": 0.1577929926930023, "learning_rate": 5.5757575757575766e-05, "loss": 1.2879, "step": 46 }, { "epoch": 0.09, "grad_norm": 0.31075066632858317, "learning_rate": 5.696969696969697e-05, "loss": 1.2202, "step": 47 }, { "epoch": 0.09, "grad_norm": 0.1663780653395111, "learning_rate": 5.818181818181818e-05, "loss": 1.2319, "step": 48 }, { "epoch": 0.09, "grad_norm": 0.16049499655883026, "learning_rate": 5.93939393939394e-05, "loss": 1.2801, "step": 49 }, { "epoch": 0.09, "grad_norm": 0.14515773124436285, "learning_rate": 6.060606060606061e-05, "loss": 1.2588, "step": 50 }, { "epoch": 0.09, "grad_norm": 0.14653064850325623, "learning_rate": 6.181818181818182e-05, "loss": 1.2677, "step": 51 }, { "epoch": 0.09, "grad_norm": 0.17193239746689878, "learning_rate": 6.303030303030302e-05, "loss": 1.2742, "step": 52 }, { "epoch": 0.1, "grad_norm": 0.1967020450342533, "learning_rate": 6.424242424242424e-05, "loss": 1.1545, "step": 53 }, { "epoch": 0.1, "grad_norm": 0.16247531997247225, "learning_rate": 6.545454545454546e-05, "loss": 1.222, "step": 54 }, { "epoch": 0.1, "grad_norm": 0.14990706377244528, "learning_rate": 6.666666666666667e-05, "loss": 1.2103, "step": 55 }, { "epoch": 0.1, "grad_norm": 0.1412817445239095, "learning_rate": 6.787878787878789e-05, "loss": 1.2169, "step": 56 }, { "epoch": 0.1, "grad_norm": 0.14575971073482757, "learning_rate": 6.90909090909091e-05, "loss": 1.2751, "step": 57 }, { "epoch": 0.11, "grad_norm": 0.13714747569950891, "learning_rate": 7.03030303030303e-05, "loss": 1.2508, "step": 58 }, { "epoch": 0.11, "grad_norm": 0.14334695156859903, "learning_rate": 7.151515151515152e-05, "loss": 1.2721, "step": 59 }, { "epoch": 0.11, "grad_norm": 0.1456824177522916, "learning_rate": 7.272727272727273e-05, "loss": 1.2649, "step": 60 }, { "epoch": 0.11, "grad_norm": 0.15030318240210044, "learning_rate": 7.393939393939395e-05, "loss": 1.2167, "step": 61 }, { "epoch": 0.11, "grad_norm": 0.1651326066719482, "learning_rate": 7.515151515151515e-05, "loss": 1.3126, "step": 62 }, { "epoch": 0.11, "grad_norm": 0.1408250406479118, "learning_rate": 7.636363636363637e-05, "loss": 1.2891, "step": 63 }, { "epoch": 0.12, "grad_norm": 0.21501384376905694, "learning_rate": 7.757575757575758e-05, "loss": 1.3019, "step": 64 }, { "epoch": 0.12, "grad_norm": 0.1365168726167339, "learning_rate": 7.878787878787879e-05, "loss": 1.2498, "step": 65 }, { "epoch": 0.12, "grad_norm": 0.1431463689660936, "learning_rate": 8e-05, "loss": 1.2793, "step": 66 }, { "epoch": 0.12, "grad_norm": 0.13689045214286194, "learning_rate": 8.121212121212121e-05, "loss": 1.2295, "step": 67 }, { "epoch": 0.12, "grad_norm": 0.13483608710081227, "learning_rate": 8.242424242424243e-05, "loss": 1.2258, "step": 68 }, { "epoch": 0.13, "grad_norm": 0.13707618564415613, "learning_rate": 8.363636363636364e-05, "loss": 1.2252, "step": 69 }, { "epoch": 0.13, "grad_norm": 0.13780236215967515, "learning_rate": 8.484848484848486e-05, "loss": 1.2565, "step": 70 }, { "epoch": 0.13, "grad_norm": 0.14036805493494423, "learning_rate": 8.606060606060606e-05, "loss": 1.3023, "step": 71 }, { "epoch": 0.13, "grad_norm": 0.12776919439147982, "learning_rate": 8.727272727272727e-05, "loss": 1.2292, "step": 72 }, { "epoch": 0.13, "grad_norm": 0.1289941815481437, "learning_rate": 8.848484848484849e-05, "loss": 1.2191, "step": 73 }, { "epoch": 0.13, "grad_norm": 0.13943952294847306, "learning_rate": 8.96969696969697e-05, "loss": 1.2915, "step": 74 }, { "epoch": 0.14, "grad_norm": 0.1493528502117281, "learning_rate": 9.090909090909092e-05, "loss": 1.2797, "step": 75 }, { "epoch": 0.14, "grad_norm": 0.1252401242451818, "learning_rate": 9.212121212121214e-05, "loss": 1.2552, "step": 76 }, { "epoch": 0.14, "grad_norm": 0.13969800467546992, "learning_rate": 9.333333333333334e-05, "loss": 1.3147, "step": 77 }, { "epoch": 0.14, "grad_norm": 0.1277258491470434, "learning_rate": 9.454545454545455e-05, "loss": 1.2089, "step": 78 }, { "epoch": 0.14, "grad_norm": 0.133041369314817, "learning_rate": 9.575757575757576e-05, "loss": 1.2761, "step": 79 }, { "epoch": 0.15, "grad_norm": 0.14564572037181842, "learning_rate": 9.696969696969698e-05, "loss": 1.1901, "step": 80 }, { "epoch": 0.15, "grad_norm": 0.13666505656492195, "learning_rate": 9.818181818181818e-05, "loss": 1.2615, "step": 81 }, { "epoch": 0.15, "grad_norm": 0.135007805210003, "learning_rate": 9.939393939393939e-05, "loss": 1.2669, "step": 82 }, { "epoch": 0.15, "grad_norm": 0.17287563365884975, "learning_rate": 0.00010060606060606062, "loss": 1.2669, "step": 83 }, { "epoch": 0.15, "grad_norm": 0.12934306326048103, "learning_rate": 0.00010181818181818181, "loss": 1.1979, "step": 84 }, { "epoch": 0.15, "grad_norm": 0.13517436169178096, "learning_rate": 0.00010303030303030303, "loss": 1.2226, "step": 85 }, { "epoch": 0.16, "grad_norm": 0.12105351159271568, "learning_rate": 0.00010424242424242425, "loss": 1.1172, "step": 86 }, { "epoch": 0.16, "grad_norm": 0.1281676431775383, "learning_rate": 0.00010545454545454545, "loss": 1.2046, "step": 87 }, { "epoch": 0.16, "grad_norm": 0.11730963057933333, "learning_rate": 0.00010666666666666667, "loss": 1.1883, "step": 88 }, { "epoch": 0.16, "grad_norm": 0.12655235108503246, "learning_rate": 0.00010787878787878789, "loss": 1.1331, "step": 89 }, { "epoch": 0.16, "grad_norm": 0.13047560307970027, "learning_rate": 0.00010909090909090909, "loss": 1.2731, "step": 90 }, { "epoch": 0.17, "grad_norm": 0.12193522973752649, "learning_rate": 0.00011030303030303031, "loss": 1.2161, "step": 91 }, { "epoch": 0.17, "grad_norm": 0.12804360300116346, "learning_rate": 0.00011151515151515153, "loss": 1.3062, "step": 92 }, { "epoch": 0.17, "grad_norm": 0.15991741754516206, "learning_rate": 0.00011272727272727272, "loss": 1.239, "step": 93 }, { "epoch": 0.17, "grad_norm": 0.15140182244454561, "learning_rate": 0.00011393939393939394, "loss": 1.2349, "step": 94 }, { "epoch": 0.17, "grad_norm": 0.12320241076263434, "learning_rate": 0.00011515151515151516, "loss": 1.2875, "step": 95 }, { "epoch": 0.17, "grad_norm": 0.13235998458230466, "learning_rate": 0.00011636363636363636, "loss": 1.2218, "step": 96 }, { "epoch": 0.18, "grad_norm": 0.11783688734798668, "learning_rate": 0.00011757575757575758, "loss": 1.1864, "step": 97 }, { "epoch": 0.18, "grad_norm": 0.3151933420750235, "learning_rate": 0.0001187878787878788, "loss": 1.3023, "step": 98 }, { "epoch": 0.18, "grad_norm": 0.12665632567219295, "learning_rate": 0.00012, "loss": 1.2249, "step": 99 }, { "epoch": 0.18, "grad_norm": 0.1228886740460738, "learning_rate": 0.00012121212121212122, "loss": 1.2517, "step": 100 }, { "epoch": 0.18, "grad_norm": 0.11892005244989344, "learning_rate": 0.00012242424242424243, "loss": 1.2586, "step": 101 }, { "epoch": 0.19, "grad_norm": 0.1232340827222201, "learning_rate": 0.00012363636363636364, "loss": 1.3217, "step": 102 }, { "epoch": 0.19, "grad_norm": 0.13837226869323116, "learning_rate": 0.00012484848484848487, "loss": 1.2693, "step": 103 }, { "epoch": 0.19, "grad_norm": 0.12068217991774362, "learning_rate": 0.00012606060606060605, "loss": 1.2623, "step": 104 }, { "epoch": 0.19, "grad_norm": 0.16779277284606545, "learning_rate": 0.00012727272727272728, "loss": 1.2415, "step": 105 }, { "epoch": 0.19, "grad_norm": 0.13396891539963085, "learning_rate": 0.0001284848484848485, "loss": 1.2313, "step": 106 }, { "epoch": 0.19, "grad_norm": 0.12457104490772812, "learning_rate": 0.0001296969696969697, "loss": 1.1758, "step": 107 }, { "epoch": 0.2, "grad_norm": 0.12676816816563452, "learning_rate": 0.00013090909090909093, "loss": 1.2478, "step": 108 }, { "epoch": 0.2, "grad_norm": 0.11973639622066906, "learning_rate": 0.00013212121212121213, "loss": 1.2335, "step": 109 }, { "epoch": 0.2, "grad_norm": 0.1330159646034068, "learning_rate": 0.00013333333333333334, "loss": 1.26, "step": 110 }, { "epoch": 0.2, "grad_norm": 0.1298003025338099, "learning_rate": 0.00013454545454545455, "loss": 1.1907, "step": 111 }, { "epoch": 0.2, "grad_norm": 0.1226154813287666, "learning_rate": 0.00013575757575757578, "loss": 1.1807, "step": 112 }, { "epoch": 0.21, "grad_norm": 0.12533753244302145, "learning_rate": 0.00013696969696969696, "loss": 1.2098, "step": 113 }, { "epoch": 0.21, "grad_norm": 0.12673266503840944, "learning_rate": 0.0001381818181818182, "loss": 1.2265, "step": 114 }, { "epoch": 0.21, "grad_norm": 0.1299039569361384, "learning_rate": 0.0001393939393939394, "loss": 1.2534, "step": 115 }, { "epoch": 0.21, "grad_norm": 0.13023496663090803, "learning_rate": 0.0001406060606060606, "loss": 1.2453, "step": 116 }, { "epoch": 0.21, "grad_norm": 0.12001793500864573, "learning_rate": 0.00014181818181818184, "loss": 1.1608, "step": 117 }, { "epoch": 0.21, "grad_norm": 0.14561862193041028, "learning_rate": 0.00014303030303030304, "loss": 1.2233, "step": 118 }, { "epoch": 0.22, "grad_norm": 0.12636130876430832, "learning_rate": 0.00014424242424242425, "loss": 1.2833, "step": 119 }, { "epoch": 0.22, "grad_norm": 0.189556849271166, "learning_rate": 0.00014545454545454546, "loss": 1.3105, "step": 120 }, { "epoch": 0.22, "grad_norm": 0.12409073764495662, "learning_rate": 0.00014666666666666666, "loss": 1.1534, "step": 121 }, { "epoch": 0.22, "grad_norm": 0.12149212466969316, "learning_rate": 0.0001478787878787879, "loss": 1.3039, "step": 122 }, { "epoch": 0.22, "grad_norm": 0.12147336887953522, "learning_rate": 0.0001490909090909091, "loss": 1.326, "step": 123 }, { "epoch": 0.23, "grad_norm": 0.1176585016163167, "learning_rate": 0.0001503030303030303, "loss": 1.191, "step": 124 }, { "epoch": 0.23, "grad_norm": 0.2066428974234372, "learning_rate": 0.00015151515151515152, "loss": 1.3054, "step": 125 }, { "epoch": 0.23, "grad_norm": 0.29582724255710047, "learning_rate": 0.00015272727272727275, "loss": 1.2032, "step": 126 }, { "epoch": 0.23, "grad_norm": 0.13084381204119358, "learning_rate": 0.00015393939393939393, "loss": 1.2289, "step": 127 }, { "epoch": 0.23, "grad_norm": 0.1294157600411397, "learning_rate": 0.00015515151515151516, "loss": 1.2561, "step": 128 }, { "epoch": 0.23, "grad_norm": 0.14039614543447027, "learning_rate": 0.00015636363636363637, "loss": 1.243, "step": 129 }, { "epoch": 0.24, "grad_norm": 0.19939984917282128, "learning_rate": 0.00015757575757575757, "loss": 1.1286, "step": 130 }, { "epoch": 0.24, "grad_norm": 0.14402764349968203, "learning_rate": 0.0001587878787878788, "loss": 1.1959, "step": 131 }, { "epoch": 0.24, "grad_norm": 0.13970978861500938, "learning_rate": 0.00016, "loss": 1.1814, "step": 132 }, { "epoch": 0.24, "grad_norm": 0.14539538472563127, "learning_rate": 0.00016121212121212122, "loss": 1.2317, "step": 133 }, { "epoch": 0.24, "grad_norm": 0.13456425455391557, "learning_rate": 0.00016242424242424243, "loss": 1.2239, "step": 134 }, { "epoch": 0.25, "grad_norm": 0.1314997837157779, "learning_rate": 0.00016363636363636366, "loss": 1.1986, "step": 135 }, { "epoch": 0.25, "grad_norm": 0.14046946525591422, "learning_rate": 0.00016484848484848487, "loss": 1.2238, "step": 136 }, { "epoch": 0.25, "grad_norm": 0.6095538041505763, "learning_rate": 0.00016606060606060607, "loss": 1.2332, "step": 137 }, { "epoch": 0.25, "grad_norm": 0.17707289712054367, "learning_rate": 0.00016727272727272728, "loss": 1.2401, "step": 138 }, { "epoch": 0.25, "grad_norm": 0.19335172179099247, "learning_rate": 0.00016848484848484848, "loss": 1.2361, "step": 139 }, { "epoch": 0.26, "grad_norm": 0.13725591818701255, "learning_rate": 0.00016969696969696972, "loss": 1.193, "step": 140 }, { "epoch": 0.26, "grad_norm": 0.15535575462507384, "learning_rate": 0.0001709090909090909, "loss": 1.2769, "step": 141 }, { "epoch": 0.26, "grad_norm": 0.14909436560898923, "learning_rate": 0.00017212121212121213, "loss": 1.2602, "step": 142 }, { "epoch": 0.26, "grad_norm": 0.15054368082407957, "learning_rate": 0.00017333333333333334, "loss": 1.2607, "step": 143 }, { "epoch": 0.26, "grad_norm": 0.13386897838741724, "learning_rate": 0.00017454545454545454, "loss": 1.168, "step": 144 }, { "epoch": 0.26, "grad_norm": 0.13567889528730145, "learning_rate": 0.00017575757575757578, "loss": 1.1984, "step": 145 }, { "epoch": 0.27, "grad_norm": 0.13994382298003089, "learning_rate": 0.00017696969696969698, "loss": 1.2795, "step": 146 }, { "epoch": 0.27, "grad_norm": 0.13941573210713187, "learning_rate": 0.0001781818181818182, "loss": 1.2303, "step": 147 }, { "epoch": 0.27, "grad_norm": 0.18302605925485763, "learning_rate": 0.0001793939393939394, "loss": 1.2696, "step": 148 }, { "epoch": 0.27, "grad_norm": 0.1547402223275396, "learning_rate": 0.00018060606060606063, "loss": 1.1276, "step": 149 }, { "epoch": 0.27, "grad_norm": 0.19947594494850646, "learning_rate": 0.00018181818181818183, "loss": 1.271, "step": 150 }, { "epoch": 0.28, "grad_norm": 0.1517101450465788, "learning_rate": 0.00018303030303030304, "loss": 1.2193, "step": 151 }, { "epoch": 0.28, "grad_norm": 0.19251063116857103, "learning_rate": 0.00018424242424242427, "loss": 1.2703, "step": 152 }, { "epoch": 0.28, "grad_norm": 0.16789099560498666, "learning_rate": 0.00018545454545454545, "loss": 1.2244, "step": 153 }, { "epoch": 0.28, "grad_norm": 0.14907376557922342, "learning_rate": 0.0001866666666666667, "loss": 1.264, "step": 154 }, { "epoch": 0.28, "grad_norm": 0.14276598263036905, "learning_rate": 0.0001878787878787879, "loss": 1.2545, "step": 155 }, { "epoch": 0.28, "grad_norm": 0.14526753816999002, "learning_rate": 0.0001890909090909091, "loss": 1.2912, "step": 156 }, { "epoch": 0.29, "grad_norm": 0.1627048894660859, "learning_rate": 0.0001903030303030303, "loss": 1.2573, "step": 157 }, { "epoch": 0.29, "grad_norm": 0.16405036632332695, "learning_rate": 0.0001915151515151515, "loss": 1.2359, "step": 158 }, { "epoch": 0.29, "grad_norm": 0.14533427219788658, "learning_rate": 0.00019272727272727274, "loss": 1.1718, "step": 159 }, { "epoch": 0.29, "grad_norm": 0.13802382666732702, "learning_rate": 0.00019393939393939395, "loss": 1.2297, "step": 160 }, { "epoch": 0.29, "grad_norm": 0.15620193618511755, "learning_rate": 0.00019515151515151516, "loss": 1.2287, "step": 161 }, { "epoch": 0.3, "grad_norm": 0.1401696295700075, "learning_rate": 0.00019636363636363636, "loss": 1.2231, "step": 162 }, { "epoch": 0.3, "grad_norm": 0.15816133304035035, "learning_rate": 0.0001975757575757576, "loss": 1.2804, "step": 163 }, { "epoch": 0.3, "grad_norm": 0.14626275180535692, "learning_rate": 0.00019878787878787878, "loss": 1.2115, "step": 164 }, { "epoch": 0.3, "grad_norm": 0.13100680398305042, "learning_rate": 0.0002, "loss": 1.2524, "step": 165 }, { "epoch": 0.3, "grad_norm": 0.14849458896148926, "learning_rate": 0.00019999977531546566, "loss": 1.2161, "step": 166 }, { "epoch": 0.3, "grad_norm": 0.13628125499037252, "learning_rate": 0.0001999991012628722, "loss": 1.2452, "step": 167 }, { "epoch": 0.31, "grad_norm": 0.18617698759086793, "learning_rate": 0.00019999797784524866, "loss": 1.2197, "step": 168 }, { "epoch": 0.31, "grad_norm": 0.14416004826313944, "learning_rate": 0.00019999640506764336, "loss": 1.2796, "step": 169 }, { "epoch": 0.31, "grad_norm": 0.13807081386834757, "learning_rate": 0.0001999943829371238, "loss": 1.2732, "step": 170 }, { "epoch": 0.31, "grad_norm": 0.16526927436841996, "learning_rate": 0.0001999919114627769, "loss": 1.3016, "step": 171 }, { "epoch": 0.31, "grad_norm": 0.14479672734919855, "learning_rate": 0.0001999889906557086, "loss": 1.3106, "step": 172 }, { "epoch": 0.32, "grad_norm": 0.13829284006072087, "learning_rate": 0.00019998562052904418, "loss": 1.3355, "step": 173 }, { "epoch": 0.32, "grad_norm": 0.13484630104616105, "learning_rate": 0.0001999818010979279, "loss": 1.1928, "step": 174 }, { "epoch": 0.32, "grad_norm": 0.14972770674556948, "learning_rate": 0.00019997753237952317, "loss": 1.2559, "step": 175 }, { "epoch": 0.32, "grad_norm": 0.13378525020528342, "learning_rate": 0.00019997281439301218, "loss": 1.2673, "step": 176 }, { "epoch": 0.32, "grad_norm": 0.13242998699125438, "learning_rate": 0.00019996764715959618, "loss": 1.2272, "step": 177 }, { "epoch": 0.32, "grad_norm": 0.12938881004364342, "learning_rate": 0.00019996203070249516, "loss": 1.2035, "step": 178 }, { "epoch": 0.33, "grad_norm": 0.13388032350164566, "learning_rate": 0.00019995596504694763, "loss": 1.2642, "step": 179 }, { "epoch": 0.33, "grad_norm": 0.13893372222140873, "learning_rate": 0.00019994945022021082, "loss": 1.2235, "step": 180 }, { "epoch": 0.33, "grad_norm": 0.14131710715500717, "learning_rate": 0.00019994248625156038, "loss": 1.1095, "step": 181 }, { "epoch": 0.33, "grad_norm": 0.13448100369103572, "learning_rate": 0.0001999350731722902, "loss": 1.1879, "step": 182 }, { "epoch": 0.33, "grad_norm": 0.13862444003216381, "learning_rate": 0.00019992721101571236, "loss": 1.2227, "step": 183 }, { "epoch": 0.34, "grad_norm": 0.13506115547921224, "learning_rate": 0.00019991889981715698, "loss": 1.2833, "step": 184 }, { "epoch": 0.34, "grad_norm": 0.13174857502600473, "learning_rate": 0.00019991013961397197, "loss": 1.2394, "step": 185 }, { "epoch": 0.34, "grad_norm": 0.1290276308949748, "learning_rate": 0.00019990093044552304, "loss": 1.2659, "step": 186 }, { "epoch": 0.34, "grad_norm": 0.1388159912078538, "learning_rate": 0.0001998912723531933, "loss": 1.3052, "step": 187 }, { "epoch": 0.34, "grad_norm": 0.1256806205303357, "learning_rate": 0.00019988116538038325, "loss": 1.2031, "step": 188 }, { "epoch": 0.34, "grad_norm": 0.13256850855084143, "learning_rate": 0.00019987060957251047, "loss": 1.211, "step": 189 }, { "epoch": 0.35, "grad_norm": 0.13197363789890235, "learning_rate": 0.0001998596049770095, "loss": 1.2256, "step": 190 }, { "epoch": 0.35, "grad_norm": 0.13277364593883098, "learning_rate": 0.00019984815164333163, "loss": 1.2174, "step": 191 }, { "epoch": 0.35, "grad_norm": 0.13838072824574454, "learning_rate": 0.00019983624962294458, "loss": 1.3128, "step": 192 }, { "epoch": 0.35, "grad_norm": 0.13524759737199996, "learning_rate": 0.0001998238989693323, "loss": 1.1806, "step": 193 }, { "epoch": 0.35, "grad_norm": 0.12669987683723832, "learning_rate": 0.0001998110997379949, "loss": 1.2171, "step": 194 }, { "epoch": 0.36, "grad_norm": 0.1461834612451898, "learning_rate": 0.00019979785198644806, "loss": 1.2231, "step": 195 }, { "epoch": 0.36, "grad_norm": 0.13265793664862735, "learning_rate": 0.0001997841557742232, "loss": 1.1718, "step": 196 }, { "epoch": 0.36, "grad_norm": 0.12842971557690963, "learning_rate": 0.00019977001116286674, "loss": 1.2758, "step": 197 }, { "epoch": 0.36, "grad_norm": 0.12188365921206967, "learning_rate": 0.00019975541821594026, "loss": 1.2457, "step": 198 }, { "epoch": 0.36, "grad_norm": 0.12679949330022622, "learning_rate": 0.00019974037699901993, "loss": 1.1825, "step": 199 }, { "epoch": 0.36, "grad_norm": 0.12949746150357985, "learning_rate": 0.00019972488757969635, "loss": 1.2666, "step": 200 }, { "epoch": 0.37, "grad_norm": 0.1363496149379173, "learning_rate": 0.00019970895002757413, "loss": 1.2031, "step": 201 }, { "epoch": 0.37, "grad_norm": 0.14218340110669314, "learning_rate": 0.0001996925644142717, "loss": 1.3073, "step": 202 }, { "epoch": 0.37, "grad_norm": 0.14234535389443218, "learning_rate": 0.00019967573081342103, "loss": 1.2444, "step": 203 }, { "epoch": 0.37, "grad_norm": 0.12866113026310516, "learning_rate": 0.000199658449300667, "loss": 1.2257, "step": 204 }, { "epoch": 0.37, "grad_norm": 0.1324053366295965, "learning_rate": 0.00019964071995366744, "loss": 1.2374, "step": 205 }, { "epoch": 0.38, "grad_norm": 0.12906841330218152, "learning_rate": 0.00019962254285209254, "loss": 1.2334, "step": 206 }, { "epoch": 0.38, "grad_norm": 0.13620873131846425, "learning_rate": 0.00019960391807762463, "loss": 1.242, "step": 207 }, { "epoch": 0.38, "grad_norm": 0.14877366842835116, "learning_rate": 0.00019958484571395757, "loss": 1.1772, "step": 208 }, { "epoch": 0.38, "grad_norm": 0.13914108740445985, "learning_rate": 0.00019956532584679675, "loss": 1.2734, "step": 209 }, { "epoch": 0.38, "grad_norm": 0.13198394930310692, "learning_rate": 0.00019954535856385837, "loss": 1.1728, "step": 210 }, { "epoch": 0.38, "grad_norm": 0.3807736597404611, "learning_rate": 0.0001995249439548693, "loss": 1.2089, "step": 211 }, { "epoch": 0.39, "grad_norm": 0.1682550557564819, "learning_rate": 0.00019950408211156636, "loss": 1.2423, "step": 212 }, { "epoch": 0.39, "grad_norm": 0.2102196862007261, "learning_rate": 0.0001994827731276963, "loss": 1.2096, "step": 213 }, { "epoch": 0.39, "grad_norm": 0.154346739470422, "learning_rate": 0.00019946101709901514, "loss": 1.2847, "step": 214 }, { "epoch": 0.39, "grad_norm": 0.16416668358293746, "learning_rate": 0.0001994388141232876, "loss": 1.2503, "step": 215 }, { "epoch": 0.39, "grad_norm": 0.13134349458231093, "learning_rate": 0.0001994161643002871, "loss": 1.1231, "step": 216 }, { "epoch": 0.4, "grad_norm": 0.15083246389185287, "learning_rate": 0.00019939306773179497, "loss": 1.1614, "step": 217 }, { "epoch": 0.4, "grad_norm": 0.1742387260929692, "learning_rate": 0.00019936952452159995, "loss": 1.3568, "step": 218 }, { "epoch": 0.4, "grad_norm": 0.18146911432436974, "learning_rate": 0.00019934553477549794, "loss": 1.2686, "step": 219 }, { "epoch": 0.4, "grad_norm": 0.1393593447949332, "learning_rate": 0.00019932109860129154, "loss": 1.1141, "step": 220 }, { "epoch": 0.4, "grad_norm": 0.14856124153987935, "learning_rate": 0.00019929621610878927, "loss": 1.234, "step": 221 }, { "epoch": 0.4, "grad_norm": 0.14820851831477327, "learning_rate": 0.0001992708874098054, "loss": 1.2069, "step": 222 }, { "epoch": 0.41, "grad_norm": 0.17893142790958147, "learning_rate": 0.00019924511261815926, "loss": 1.1278, "step": 223 }, { "epoch": 0.41, "grad_norm": 0.14573658703265605, "learning_rate": 0.00019921889184967476, "loss": 1.2292, "step": 224 }, { "epoch": 0.41, "grad_norm": 0.15282321197574994, "learning_rate": 0.00019919222522217996, "loss": 1.2482, "step": 225 }, { "epoch": 0.41, "grad_norm": 0.16342112084119492, "learning_rate": 0.00019916511285550642, "loss": 1.2172, "step": 226 }, { "epoch": 0.41, "grad_norm": 0.1475889153814455, "learning_rate": 0.00019913755487148876, "loss": 1.1747, "step": 227 }, { "epoch": 0.42, "grad_norm": 0.163738064491857, "learning_rate": 0.00019910955139396396, "loss": 1.3007, "step": 228 }, { "epoch": 0.42, "grad_norm": 0.14427856196022704, "learning_rate": 0.00019908110254877106, "loss": 1.2464, "step": 229 }, { "epoch": 0.42, "grad_norm": 0.20204742660246344, "learning_rate": 0.00019905220846375032, "loss": 1.2515, "step": 230 }, { "epoch": 0.42, "grad_norm": 0.15134144918251685, "learning_rate": 0.0001990228692687429, "loss": 1.1786, "step": 231 }, { "epoch": 0.42, "grad_norm": 0.1636590177812163, "learning_rate": 0.00019899308509558998, "loss": 1.1974, "step": 232 }, { "epoch": 0.42, "grad_norm": 0.15552319776955892, "learning_rate": 0.00019896285607813244, "loss": 1.2308, "step": 233 }, { "epoch": 0.43, "grad_norm": 0.17104898009833774, "learning_rate": 0.00019893218235221015, "loss": 1.2828, "step": 234 }, { "epoch": 0.43, "grad_norm": 0.16387378763964267, "learning_rate": 0.00019890106405566138, "loss": 1.2779, "step": 235 }, { "epoch": 0.43, "grad_norm": 0.14622126798612248, "learning_rate": 0.00019886950132832207, "loss": 1.2894, "step": 236 }, { "epoch": 0.43, "grad_norm": 0.16619841547518147, "learning_rate": 0.0001988374943120254, "loss": 1.2133, "step": 237 }, { "epoch": 0.43, "grad_norm": 0.12664832399697545, "learning_rate": 0.00019880504315060096, "loss": 1.1807, "step": 238 }, { "epoch": 0.44, "grad_norm": 0.2015108381613456, "learning_rate": 0.00019877214798987426, "loss": 1.1876, "step": 239 }, { "epoch": 0.44, "grad_norm": 0.14468620723711506, "learning_rate": 0.00019873880897766598, "loss": 1.1883, "step": 240 }, { "epoch": 0.44, "grad_norm": 0.1549018650770757, "learning_rate": 0.00019870502626379127, "loss": 1.2896, "step": 241 }, { "epoch": 0.44, "grad_norm": 0.1492917963684983, "learning_rate": 0.0001986708000000593, "loss": 1.2102, "step": 242 }, { "epoch": 0.44, "grad_norm": 0.178606606459489, "learning_rate": 0.00019863613034027224, "loss": 1.2292, "step": 243 }, { "epoch": 0.44, "grad_norm": 0.206170239681528, "learning_rate": 0.00019860101744022485, "loss": 1.2666, "step": 244 }, { "epoch": 0.45, "grad_norm": 0.13741043007948167, "learning_rate": 0.0001985654614577036, "loss": 1.2022, "step": 245 }, { "epoch": 0.45, "grad_norm": 0.1595080658199459, "learning_rate": 0.0001985294625524861, "loss": 1.1203, "step": 246 }, { "epoch": 0.45, "grad_norm": 0.13929705183853777, "learning_rate": 0.00019849302088634034, "loss": 1.1505, "step": 247 }, { "epoch": 0.45, "grad_norm": 0.14045247607912964, "learning_rate": 0.00019845613662302383, "loss": 1.1897, "step": 248 }, { "epoch": 0.45, "grad_norm": 0.15002651347444407, "learning_rate": 0.00019841880992828306, "loss": 1.2133, "step": 249 }, { "epoch": 0.46, "grad_norm": 0.1567929487810952, "learning_rate": 0.00019838104096985267, "loss": 1.129, "step": 250 }, { "epoch": 0.46, "grad_norm": 0.15240634543877116, "learning_rate": 0.00019834282991745464, "loss": 1.1995, "step": 251 }, { "epoch": 0.46, "grad_norm": 0.151807679821367, "learning_rate": 0.00019830417694279766, "loss": 1.25, "step": 252 }, { "epoch": 0.46, "grad_norm": 0.1648599156208311, "learning_rate": 0.0001982650822195762, "loss": 1.2511, "step": 253 }, { "epoch": 0.46, "grad_norm": 0.15363401233808713, "learning_rate": 0.00019822554592346993, "loss": 1.1794, "step": 254 }, { "epoch": 0.46, "grad_norm": 0.1569644350778875, "learning_rate": 0.00019818556823214268, "loss": 1.2033, "step": 255 }, { "epoch": 0.47, "grad_norm": 0.15996552747294254, "learning_rate": 0.0001981451493252418, "loss": 1.2809, "step": 256 }, { "epoch": 0.47, "grad_norm": 0.15863104885072635, "learning_rate": 0.0001981042893843974, "loss": 1.1667, "step": 257 }, { "epoch": 0.47, "grad_norm": 0.2887466971861171, "learning_rate": 0.0001980629885932214, "loss": 1.1915, "step": 258 }, { "epoch": 0.47, "grad_norm": 0.15233015979193984, "learning_rate": 0.00019802124713730681, "loss": 1.1734, "step": 259 }, { "epoch": 0.47, "grad_norm": 0.18207884538436447, "learning_rate": 0.00019797906520422677, "loss": 1.2575, "step": 260 }, { "epoch": 0.48, "grad_norm": 0.17323546756038308, "learning_rate": 0.0001979364429835339, "loss": 1.1704, "step": 261 }, { "epoch": 0.48, "grad_norm": 0.14592153602263633, "learning_rate": 0.00019789338066675922, "loss": 1.192, "step": 262 }, { "epoch": 0.48, "grad_norm": 0.19250697792287097, "learning_rate": 0.0001978498784474115, "loss": 1.2779, "step": 263 }, { "epoch": 0.48, "grad_norm": 0.1429107680887097, "learning_rate": 0.0001978059365209762, "loss": 1.2529, "step": 264 }, { "epoch": 0.48, "grad_norm": 0.48514081074992116, "learning_rate": 0.00019776155508491482, "loss": 1.1917, "step": 265 }, { "epoch": 0.48, "grad_norm": 0.1534376167748161, "learning_rate": 0.0001977167343386638, "loss": 1.2384, "step": 266 }, { "epoch": 0.49, "grad_norm": 0.16744875760032166, "learning_rate": 0.00019767147448363366, "loss": 1.1744, "step": 267 }, { "epoch": 0.49, "grad_norm": 0.29195538170738244, "learning_rate": 0.00019762577572320824, "loss": 1.1418, "step": 268 }, { "epoch": 0.49, "grad_norm": 0.1820804717651353, "learning_rate": 0.00019757963826274357, "loss": 1.2815, "step": 269 }, { "epoch": 0.49, "grad_norm": 0.17522345110441973, "learning_rate": 0.00019753306230956718, "loss": 1.2363, "step": 270 }, { "epoch": 0.49, "grad_norm": 0.16354388270886613, "learning_rate": 0.000197486048072977, "loss": 1.2845, "step": 271 }, { "epoch": 0.5, "grad_norm": 0.17590082756401024, "learning_rate": 0.0001974385957642404, "loss": 1.192, "step": 272 }, { "epoch": 0.5, "grad_norm": 0.17345720403188775, "learning_rate": 0.00019739070559659347, "loss": 1.2068, "step": 273 }, { "epoch": 0.5, "grad_norm": 0.16070434867766506, "learning_rate": 0.00019734237778523976, "loss": 1.189, "step": 274 }, { "epoch": 0.5, "grad_norm": 0.18983443066710415, "learning_rate": 0.0001972936125473495, "loss": 1.2223, "step": 275 }, { "epoch": 0.5, "grad_norm": 0.15724400187981355, "learning_rate": 0.00019724441010205863, "loss": 1.2292, "step": 276 }, { "epoch": 0.5, "grad_norm": 0.14570729442956004, "learning_rate": 0.00019719477067046766, "loss": 1.1421, "step": 277 }, { "epoch": 0.51, "grad_norm": 0.1559242881177266, "learning_rate": 0.00019714469447564088, "loss": 1.2598, "step": 278 }, { "epoch": 0.51, "grad_norm": 0.16621830243096108, "learning_rate": 0.0001970941817426052, "loss": 1.3038, "step": 279 }, { "epoch": 0.51, "grad_norm": 4.675483994100576, "learning_rate": 0.00019704323269834927, "loss": 1.2298, "step": 280 }, { "epoch": 0.51, "grad_norm": 0.2769699381619058, "learning_rate": 0.00019699184757182225, "loss": 1.2566, "step": 281 }, { "epoch": 0.51, "grad_norm": 0.20189839889100783, "learning_rate": 0.00019694002659393305, "loss": 1.3181, "step": 282 }, { "epoch": 0.52, "grad_norm": 0.19497107359413876, "learning_rate": 0.00019688776999754912, "loss": 1.1502, "step": 283 }, { "epoch": 0.52, "grad_norm": 0.1982266815755412, "learning_rate": 0.00019683507801749545, "loss": 1.2053, "step": 284 }, { "epoch": 0.52, "grad_norm": 0.1924340950322314, "learning_rate": 0.00019678195089055346, "loss": 1.2149, "step": 285 }, { "epoch": 0.52, "grad_norm": 0.1725322346446431, "learning_rate": 0.00019672838885546008, "loss": 1.2553, "step": 286 }, { "epoch": 0.52, "grad_norm": 0.2535488743520272, "learning_rate": 0.00019667439215290648, "loss": 1.2576, "step": 287 }, { "epoch": 0.52, "grad_norm": 0.37837586860064026, "learning_rate": 0.00019661996102553718, "loss": 1.1815, "step": 288 }, { "epoch": 0.53, "grad_norm": 0.17520419597901843, "learning_rate": 0.00019656509571794878, "loss": 1.1932, "step": 289 }, { "epoch": 0.53, "grad_norm": 0.17056234784450633, "learning_rate": 0.00019650979647668906, "loss": 1.163, "step": 290 }, { "epoch": 0.53, "grad_norm": 0.18272246580207432, "learning_rate": 0.00019645406355025565, "loss": 1.1887, "step": 291 }, { "epoch": 0.53, "grad_norm": 0.17889037954429915, "learning_rate": 0.00019639789718909508, "loss": 1.2126, "step": 292 }, { "epoch": 0.53, "grad_norm": 0.23993734971101424, "learning_rate": 0.00019634129764560168, "loss": 1.2485, "step": 293 }, { "epoch": 0.54, "grad_norm": 0.1847578318208199, "learning_rate": 0.00019628426517411625, "loss": 1.2549, "step": 294 }, { "epoch": 0.54, "grad_norm": 0.23185098827091005, "learning_rate": 0.00019622680003092503, "loss": 1.1599, "step": 295 }, { "epoch": 0.54, "grad_norm": 0.220638044092583, "learning_rate": 0.00019616890247425866, "loss": 1.2281, "step": 296 }, { "epoch": 0.54, "grad_norm": 0.2303439219825616, "learning_rate": 0.00019611057276429085, "loss": 1.2208, "step": 297 }, { "epoch": 0.54, "grad_norm": 0.1744807302230573, "learning_rate": 0.00019605181116313724, "loss": 1.2303, "step": 298 }, { "epoch": 0.54, "grad_norm": 0.17510946821872422, "learning_rate": 0.0001959926179348543, "loss": 1.2385, "step": 299 }, { "epoch": 0.55, "grad_norm": 0.2218474349751746, "learning_rate": 0.00019593299334543808, "loss": 1.2153, "step": 300 }, { "epoch": 0.55, "grad_norm": 0.1742070481516402, "learning_rate": 0.00019587293766282308, "loss": 1.1628, "step": 301 }, { "epoch": 0.55, "grad_norm": 0.15250311715180823, "learning_rate": 0.00019581245115688094, "loss": 1.1632, "step": 302 }, { "epoch": 0.55, "grad_norm": 0.1744397677094501, "learning_rate": 0.0001957515340994193, "loss": 1.254, "step": 303 }, { "epoch": 0.55, "grad_norm": 0.1686772182789891, "learning_rate": 0.00019569018676418053, "loss": 1.2169, "step": 304 }, { "epoch": 0.56, "grad_norm": 0.16404966161017623, "learning_rate": 0.00019562840942684067, "loss": 1.2221, "step": 305 }, { "epoch": 0.56, "grad_norm": 0.16052011449463713, "learning_rate": 0.00019556620236500793, "loss": 1.2045, "step": 306 }, { "epoch": 0.56, "grad_norm": 0.16343251390831215, "learning_rate": 0.0001955035658582216, "loss": 1.2289, "step": 307 }, { "epoch": 0.56, "grad_norm": 0.14387162360389305, "learning_rate": 0.00019544050018795075, "loss": 1.1365, "step": 308 }, { "epoch": 0.56, "grad_norm": 0.15304461439740238, "learning_rate": 0.00019537700563759304, "loss": 1.1931, "step": 309 }, { "epoch": 0.56, "grad_norm": 0.17059958050065627, "learning_rate": 0.00019531308249247327, "loss": 1.2166, "step": 310 }, { "epoch": 0.57, "grad_norm": 0.17633385530926995, "learning_rate": 0.00019524873103984235, "loss": 1.2604, "step": 311 }, { "epoch": 0.57, "grad_norm": 0.17855814403303746, "learning_rate": 0.00019518395156887576, "loss": 1.1615, "step": 312 }, { "epoch": 0.57, "grad_norm": 0.19823982444256988, "learning_rate": 0.00019511874437067243, "loss": 1.2153, "step": 313 }, { "epoch": 0.57, "grad_norm": 0.1570784627362585, "learning_rate": 0.0001950531097382533, "loss": 1.2788, "step": 314 }, { "epoch": 0.57, "grad_norm": 0.2183125402112695, "learning_rate": 0.00019498704796656018, "loss": 1.2966, "step": 315 }, { "epoch": 0.58, "grad_norm": 0.18173933276147194, "learning_rate": 0.00019492055935245418, "loss": 1.2978, "step": 316 }, { "epoch": 0.58, "grad_norm": 0.17483116680914407, "learning_rate": 0.00019485364419471454, "loss": 1.258, "step": 317 }, { "epoch": 0.58, "grad_norm": 0.15490767356815494, "learning_rate": 0.0001947863027940374, "loss": 1.2088, "step": 318 }, { "epoch": 0.58, "grad_norm": 0.14703966491934156, "learning_rate": 0.00019471853545303405, "loss": 1.2355, "step": 319 }, { "epoch": 0.58, "grad_norm": 0.14386689086661608, "learning_rate": 0.00019465034247623003, "loss": 1.2583, "step": 320 }, { "epoch": 0.58, "grad_norm": 0.18818904376313625, "learning_rate": 0.00019458172417006347, "loss": 1.2181, "step": 321 }, { "epoch": 0.59, "grad_norm": 0.17393313719202513, "learning_rate": 0.00019451268084288385, "loss": 1.3453, "step": 322 }, { "epoch": 0.59, "grad_norm": 0.14706823379985753, "learning_rate": 0.00019444321280495043, "loss": 1.2234, "step": 323 }, { "epoch": 0.59, "grad_norm": 0.15282014755252687, "learning_rate": 0.00019437332036843118, "loss": 1.1262, "step": 324 }, { "epoch": 0.59, "grad_norm": 0.1618727884326225, "learning_rate": 0.00019430300384740105, "loss": 1.3136, "step": 325 }, { "epoch": 0.59, "grad_norm": 0.16090758705378874, "learning_rate": 0.00019423226355784077, "loss": 1.2055, "step": 326 }, { "epoch": 0.6, "grad_norm": 0.15241156801091013, "learning_rate": 0.00019416109981763526, "loss": 1.2678, "step": 327 }, { "epoch": 0.6, "grad_norm": 0.14216697909809062, "learning_rate": 0.0001940895129465724, "loss": 1.2841, "step": 328 }, { "epoch": 0.6, "grad_norm": 0.15790232415414485, "learning_rate": 0.00019401750326634144, "loss": 1.3119, "step": 329 }, { "epoch": 0.6, "grad_norm": 0.13322691961062616, "learning_rate": 0.0001939450711005316, "loss": 1.1293, "step": 330 }, { "epoch": 0.6, "grad_norm": 0.14075018938835404, "learning_rate": 0.00019387221677463062, "loss": 1.2176, "step": 331 }, { "epoch": 0.6, "grad_norm": 0.21565975459393052, "learning_rate": 0.00019379894061602335, "loss": 1.1723, "step": 332 }, { "epoch": 0.61, "grad_norm": 0.17967631394222838, "learning_rate": 0.00019372524295399013, "loss": 1.239, "step": 333 }, { "epoch": 0.61, "grad_norm": 0.21187969201978435, "learning_rate": 0.0001936511241197055, "loss": 1.2207, "step": 334 }, { "epoch": 0.61, "grad_norm": 0.16967789022974608, "learning_rate": 0.00019357658444623654, "loss": 1.2478, "step": 335 }, { "epoch": 0.61, "grad_norm": 0.14810621660374448, "learning_rate": 0.0001935016242685415, "loss": 1.1223, "step": 336 }, { "epoch": 0.61, "grad_norm": 0.1489106421847434, "learning_rate": 0.00019342624392346824, "loss": 1.1592, "step": 337 }, { "epoch": 0.62, "grad_norm": 0.17625176068748855, "learning_rate": 0.0001933504437497527, "loss": 1.2145, "step": 338 }, { "epoch": 0.62, "grad_norm": 0.17250255512763446, "learning_rate": 0.00019327422408801744, "loss": 1.2504, "step": 339 }, { "epoch": 0.62, "grad_norm": 0.16079375745566896, "learning_rate": 0.00019319758528077, "loss": 1.1795, "step": 340 }, { "epoch": 0.62, "grad_norm": 0.15454466809245995, "learning_rate": 0.0001931205276724015, "loss": 1.2123, "step": 341 }, { "epoch": 0.62, "grad_norm": 0.7021323604447972, "learning_rate": 0.000193043051609185, "loss": 1.2239, "step": 342 }, { "epoch": 0.62, "grad_norm": 0.1572764339385847, "learning_rate": 0.00019296515743927399, "loss": 1.2516, "step": 343 }, { "epoch": 0.63, "grad_norm": 0.2136637778252246, "learning_rate": 0.00019288684551270073, "loss": 1.2321, "step": 344 }, { "epoch": 0.63, "grad_norm": 0.4546540454773654, "learning_rate": 0.00019280811618137484, "loss": 1.18, "step": 345 }, { "epoch": 0.63, "grad_norm": 0.9809832576786297, "learning_rate": 0.00019272896979908154, "loss": 1.2081, "step": 346 }, { "epoch": 0.63, "grad_norm": 0.5246256133291822, "learning_rate": 0.00019264940672148018, "loss": 1.2722, "step": 347 }, { "epoch": 0.63, "grad_norm": 0.24941717134878091, "learning_rate": 0.00019256942730610268, "loss": 1.2352, "step": 348 }, { "epoch": 0.64, "grad_norm": 0.3356068462072784, "learning_rate": 0.00019248903191235176, "loss": 1.2225, "step": 349 }, { "epoch": 0.64, "grad_norm": 0.19535845221880543, "learning_rate": 0.00019240822090149944, "loss": 1.1669, "step": 350 }, { "epoch": 0.64, "grad_norm": 0.22306941566416597, "learning_rate": 0.00019232699463668542, "loss": 1.2281, "step": 351 }, { "epoch": 0.64, "grad_norm": 0.2700134013989352, "learning_rate": 0.00019224535348291542, "loss": 1.1939, "step": 352 }, { "epoch": 0.64, "grad_norm": 0.24406908935562743, "learning_rate": 0.00019216329780705953, "loss": 1.1839, "step": 353 }, { "epoch": 0.64, "grad_norm": 0.20465183000217488, "learning_rate": 0.00019208082797785055, "loss": 1.2277, "step": 354 }, { "epoch": 0.65, "grad_norm": 0.21324820828129784, "learning_rate": 0.00019199794436588243, "loss": 1.2072, "step": 355 }, { "epoch": 0.65, "grad_norm": 0.1780562512431263, "learning_rate": 0.00019191464734360844, "loss": 1.2082, "step": 356 }, { "epoch": 0.65, "grad_norm": 0.16547971467615655, "learning_rate": 0.00019183093728533966, "loss": 1.1978, "step": 357 }, { "epoch": 0.65, "grad_norm": 0.22904664933247196, "learning_rate": 0.00019174681456724318, "loss": 1.1562, "step": 358 }, { "epoch": 0.65, "grad_norm": 0.1737397860007602, "learning_rate": 0.00019166227956734052, "loss": 1.2383, "step": 359 }, { "epoch": 0.66, "grad_norm": 0.1589465455917568, "learning_rate": 0.00019157733266550575, "loss": 1.2158, "step": 360 }, { "epoch": 0.66, "grad_norm": 0.16253126221999709, "learning_rate": 0.00019149197424346405, "loss": 1.1952, "step": 361 }, { "epoch": 0.66, "grad_norm": 0.22436676243032663, "learning_rate": 0.00019140620468478968, "loss": 1.2315, "step": 362 }, { "epoch": 0.66, "grad_norm": 0.19291682612950423, "learning_rate": 0.00019132002437490458, "loss": 1.2283, "step": 363 }, { "epoch": 0.66, "grad_norm": 0.1519191258459668, "learning_rate": 0.00019123343370107637, "loss": 1.1151, "step": 364 }, { "epoch": 0.66, "grad_norm": 0.17179909633547025, "learning_rate": 0.00019114643305241676, "loss": 1.1576, "step": 365 }, { "epoch": 0.67, "grad_norm": 0.17992599023321432, "learning_rate": 0.00019105902281987976, "loss": 1.2592, "step": 366 }, { "epoch": 0.67, "grad_norm": 0.17714099390314453, "learning_rate": 0.00019097120339625994, "loss": 1.2578, "step": 367 }, { "epoch": 0.67, "grad_norm": 0.2455577642687935, "learning_rate": 0.00019088297517619055, "loss": 1.2361, "step": 368 }, { "epoch": 0.67, "grad_norm": 0.18398518628783986, "learning_rate": 0.00019079433855614201, "loss": 1.1906, "step": 369 }, { "epoch": 0.67, "grad_norm": 0.18944067022821645, "learning_rate": 0.00019070529393441985, "loss": 1.237, "step": 370 }, { "epoch": 0.68, "grad_norm": 0.17639967519781063, "learning_rate": 0.00019061584171116303, "loss": 1.1841, "step": 371 }, { "epoch": 0.68, "grad_norm": 0.15947129998283005, "learning_rate": 0.00019052598228834217, "loss": 1.1722, "step": 372 }, { "epoch": 0.68, "grad_norm": 0.1693354353719105, "learning_rate": 0.00019043571606975777, "loss": 1.2204, "step": 373 }, { "epoch": 0.68, "grad_norm": 0.16236190451963983, "learning_rate": 0.00019034504346103823, "loss": 1.1778, "step": 374 }, { "epoch": 0.68, "grad_norm": 0.17702370729269964, "learning_rate": 0.00019025396486963827, "loss": 1.2065, "step": 375 }, { "epoch": 0.68, "grad_norm": 0.19388150596154238, "learning_rate": 0.00019016248070483687, "loss": 1.2942, "step": 376 }, { "epoch": 0.69, "grad_norm": 0.16152000400319103, "learning_rate": 0.0001900705913777356, "loss": 1.1784, "step": 377 }, { "epoch": 0.69, "grad_norm": 0.1545267913996029, "learning_rate": 0.00018997829730125663, "loss": 1.1829, "step": 378 }, { "epoch": 0.69, "grad_norm": 0.15421727704318197, "learning_rate": 0.000189885598890141, "loss": 1.177, "step": 379 }, { "epoch": 0.69, "grad_norm": 0.1624966073814206, "learning_rate": 0.00018979249656094673, "loss": 1.2439, "step": 380 }, { "epoch": 0.69, "grad_norm": 0.9490737312904575, "learning_rate": 0.00018969899073204686, "loss": 1.2085, "step": 381 }, { "epoch": 0.7, "grad_norm": 0.8982903208613089, "learning_rate": 0.00018960508182362768, "loss": 1.2347, "step": 382 }, { "epoch": 0.7, "grad_norm": 0.3771428474797688, "learning_rate": 0.00018951077025768678, "loss": 1.2546, "step": 383 }, { "epoch": 0.7, "grad_norm": 0.4776152950069111, "learning_rate": 0.00018941605645803115, "loss": 1.2904, "step": 384 }, { "epoch": 0.7, "grad_norm": 0.18786943849618057, "learning_rate": 0.00018932094085027533, "loss": 1.2122, "step": 385 }, { "epoch": 0.7, "grad_norm": 1.6297025984167128, "learning_rate": 0.0001892254238618394, "loss": 1.171, "step": 386 }, { "epoch": 0.7, "grad_norm": 0.20382660707264952, "learning_rate": 0.0001891295059219472, "loss": 1.1874, "step": 387 }, { "epoch": 0.71, "grad_norm": 1.3580819775908755, "learning_rate": 0.00018903318746162429, "loss": 1.1531, "step": 388 }, { "epoch": 0.71, "grad_norm": 0.43619056173016185, "learning_rate": 0.00018893646891369602, "loss": 1.2289, "step": 389 }, { "epoch": 0.71, "grad_norm": 0.29385240705823723, "learning_rate": 0.0001888393507127856, "loss": 1.2073, "step": 390 }, { "epoch": 0.71, "grad_norm": 0.3136086850525623, "learning_rate": 0.00018874183329531223, "loss": 1.1898, "step": 391 }, { "epoch": 0.71, "grad_norm": 0.2307767217662562, "learning_rate": 0.000188643917099489, "loss": 1.207, "step": 392 }, { "epoch": 0.72, "grad_norm": 0.18703654518135468, "learning_rate": 0.000188545602565321, "loss": 1.1688, "step": 393 }, { "epoch": 0.72, "grad_norm": 0.4809351333934126, "learning_rate": 0.00018844689013460336, "loss": 1.2519, "step": 394 }, { "epoch": 0.72, "grad_norm": 0.40370101428544464, "learning_rate": 0.0001883477802509192, "loss": 1.2411, "step": 395 }, { "epoch": 0.72, "grad_norm": 0.2858848636432859, "learning_rate": 0.00018824827335963765, "loss": 1.194, "step": 396 }, { "epoch": 0.72, "grad_norm": 0.32195602638999565, "learning_rate": 0.000188148369907912, "loss": 1.0988, "step": 397 }, { "epoch": 0.72, "grad_norm": 0.23790306908901832, "learning_rate": 0.00018804807034467733, "loss": 1.2237, "step": 398 }, { "epoch": 0.73, "grad_norm": 0.20126988767112128, "learning_rate": 0.0001879473751206489, "loss": 1.2731, "step": 399 }, { "epoch": 0.73, "grad_norm": 0.3336380339194037, "learning_rate": 0.00018784628468831996, "loss": 1.2369, "step": 400 }, { "epoch": 0.73, "grad_norm": 0.5054330893305989, "learning_rate": 0.0001877447995019596, "loss": 1.2443, "step": 401 }, { "epoch": 0.73, "grad_norm": 0.2297866279715136, "learning_rate": 0.0001876429200176108, "loss": 1.2376, "step": 402 }, { "epoch": 0.73, "grad_norm": 0.39350567174184636, "learning_rate": 0.00018754064669308858, "loss": 1.2126, "step": 403 }, { "epoch": 0.74, "grad_norm": 0.2025361091435325, "learning_rate": 0.00018743797998797753, "loss": 1.2224, "step": 404 }, { "epoch": 0.74, "grad_norm": 0.31824903419753814, "learning_rate": 0.00018733492036363005, "loss": 1.2942, "step": 405 }, { "epoch": 0.74, "grad_norm": 2.4642066748643017, "learning_rate": 0.00018723146828316428, "loss": 1.2515, "step": 406 }, { "epoch": 0.74, "grad_norm": 0.7833055646295342, "learning_rate": 0.00018712762421146183, "loss": 1.2207, "step": 407 }, { "epoch": 0.74, "grad_norm": 0.2810249021786599, "learning_rate": 0.00018702338861516587, "loss": 1.2755, "step": 408 }, { "epoch": 0.74, "grad_norm": 0.460995724241333, "learning_rate": 0.0001869187619626789, "loss": 1.2856, "step": 409 }, { "epoch": 0.75, "grad_norm": 0.49139203044984286, "learning_rate": 0.00018681374472416073, "loss": 1.2392, "step": 410 }, { "epoch": 0.75, "grad_norm": 2.691604613969173, "learning_rate": 0.0001867083373715264, "loss": 1.2992, "step": 411 }, { "epoch": 0.75, "grad_norm": 0.8014112047318501, "learning_rate": 0.00018660254037844388, "loss": 1.2683, "step": 412 }, { "epoch": 0.75, "grad_norm": 0.31614342841331383, "learning_rate": 0.00018649635422033215, "loss": 1.2356, "step": 413 }, { "epoch": 0.75, "grad_norm": 0.2559855196513244, "learning_rate": 0.000186389779374359, "loss": 1.2053, "step": 414 }, { "epoch": 0.76, "grad_norm": 1.6613999986014714, "learning_rate": 0.0001862828163194388, "loss": 1.2568, "step": 415 }, { "epoch": 0.76, "grad_norm": 0.27190082167109786, "learning_rate": 0.0001861754655362304, "loss": 1.1288, "step": 416 }, { "epoch": 0.76, "grad_norm": 0.43819582203066043, "learning_rate": 0.00018606772750713504, "loss": 1.1758, "step": 417 }, { "epoch": 0.76, "grad_norm": 0.41738497400383384, "learning_rate": 0.0001859596027162941, "loss": 1.2993, "step": 418 }, { "epoch": 0.76, "grad_norm": 0.2595142634740817, "learning_rate": 0.000185851091649587, "loss": 1.269, "step": 419 }, { "epoch": 0.77, "grad_norm": 0.2795314201020271, "learning_rate": 0.00018574219479462878, "loss": 1.1915, "step": 420 }, { "epoch": 0.77, "grad_norm": 0.2502992494749938, "learning_rate": 0.00018563291264076835, "loss": 1.2157, "step": 421 }, { "epoch": 0.77, "grad_norm": 0.27422512335538374, "learning_rate": 0.00018552324567908585, "loss": 1.2541, "step": 422 }, { "epoch": 0.77, "grad_norm": 1.3360989016060905, "learning_rate": 0.00018541319440239066, "loss": 1.2666, "step": 423 }, { "epoch": 0.77, "grad_norm": 1.9441131913572127, "learning_rate": 0.00018530275930521924, "loss": 1.2924, "step": 424 }, { "epoch": 0.77, "grad_norm": 0.3772373301771213, "learning_rate": 0.00018519194088383273, "loss": 1.1952, "step": 425 }, { "epoch": 0.78, "grad_norm": 0.22091753616251295, "learning_rate": 0.0001850807396362148, "loss": 1.1858, "step": 426 }, { "epoch": 0.78, "grad_norm": 0.21423504993321807, "learning_rate": 0.00018496915606206951, "loss": 1.2245, "step": 427 }, { "epoch": 0.78, "grad_norm": 0.5238946238105926, "learning_rate": 0.00018485719066281892, "loss": 1.2351, "step": 428 }, { "epoch": 0.78, "grad_norm": 0.3037858949309141, "learning_rate": 0.0001847448439416009, "loss": 1.1669, "step": 429 }, { "epoch": 0.78, "grad_norm": 0.21553286799952254, "learning_rate": 0.00018463211640326686, "loss": 1.1454, "step": 430 }, { "epoch": 0.79, "grad_norm": 0.202875547805464, "learning_rate": 0.0001845190085543795, "loss": 1.188, "step": 431 }, { "epoch": 0.79, "grad_norm": 0.24385408620619278, "learning_rate": 0.00018440552090321047, "loss": 1.2307, "step": 432 }, { "epoch": 0.79, "grad_norm": 0.23793944272430378, "learning_rate": 0.0001842916539597382, "loss": 1.2253, "step": 433 }, { "epoch": 0.79, "grad_norm": 0.17062488448810784, "learning_rate": 0.0001841774082356455, "loss": 1.2681, "step": 434 }, { "epoch": 0.79, "grad_norm": 0.20003742001916064, "learning_rate": 0.00018406278424431736, "loss": 1.2428, "step": 435 }, { "epoch": 0.79, "grad_norm": 0.2696052831337752, "learning_rate": 0.0001839477825008385, "loss": 1.2945, "step": 436 }, { "epoch": 0.8, "grad_norm": 0.23302960820538443, "learning_rate": 0.00018383240352199117, "loss": 1.1718, "step": 437 }, { "epoch": 0.8, "grad_norm": 0.38187833239777536, "learning_rate": 0.00018371664782625287, "loss": 1.2311, "step": 438 }, { "epoch": 0.8, "grad_norm": 0.4052561772533732, "learning_rate": 0.00018360051593379383, "loss": 1.1639, "step": 439 }, { "epoch": 0.8, "grad_norm": 0.23379763821020377, "learning_rate": 0.0001834840083664749, "loss": 1.1809, "step": 440 }, { "epoch": 0.8, "grad_norm": 0.2368414607613928, "learning_rate": 0.00018336712564784503, "loss": 1.2357, "step": 441 }, { "epoch": 0.81, "grad_norm": 0.20230633988510938, "learning_rate": 0.000183249868303139, "loss": 1.1851, "step": 442 }, { "epoch": 0.81, "grad_norm": 0.170513157244292, "learning_rate": 0.00018313223685927505, "loss": 1.205, "step": 443 }, { "epoch": 0.81, "grad_norm": 0.18082295035256266, "learning_rate": 0.0001830142318448525, "loss": 1.2305, "step": 444 }, { "epoch": 0.81, "grad_norm": 0.18286299264146286, "learning_rate": 0.00018289585379014942, "loss": 1.23, "step": 445 }, { "epoch": 0.81, "grad_norm": 0.17868104103482751, "learning_rate": 0.00018277710322712012, "loss": 1.2894, "step": 446 }, { "epoch": 0.81, "grad_norm": 0.1820411127336495, "learning_rate": 0.00018265798068939294, "loss": 1.2395, "step": 447 }, { "epoch": 0.82, "grad_norm": 0.1738237541783663, "learning_rate": 0.0001825384867122677, "loss": 1.1576, "step": 448 }, { "epoch": 0.82, "grad_norm": 0.15693445967795147, "learning_rate": 0.0001824186218327134, "loss": 1.0809, "step": 449 }, { "epoch": 0.82, "grad_norm": 0.18509145652208978, "learning_rate": 0.00018229838658936564, "loss": 1.2717, "step": 450 }, { "epoch": 0.82, "grad_norm": 0.14702488366564262, "learning_rate": 0.0001821777815225245, "loss": 1.2236, "step": 451 }, { "epoch": 0.82, "grad_norm": 0.1828399354418095, "learning_rate": 0.00018205680717415187, "loss": 1.2565, "step": 452 }, { "epoch": 0.83, "grad_norm": 0.17460984182013486, "learning_rate": 0.00018193546408786898, "loss": 1.2474, "step": 453 }, { "epoch": 0.83, "grad_norm": 0.2001623109673152, "learning_rate": 0.00018181375280895416, "loss": 1.2544, "step": 454 }, { "epoch": 0.83, "grad_norm": 0.17228631742863837, "learning_rate": 0.00018169167388434025, "loss": 1.1851, "step": 455 }, { "epoch": 0.83, "grad_norm": 0.1644862232819482, "learning_rate": 0.00018156922786261216, "loss": 1.1817, "step": 456 }, { "epoch": 0.83, "grad_norm": 0.19775186397477057, "learning_rate": 0.00018144641529400446, "loss": 1.257, "step": 457 }, { "epoch": 0.83, "grad_norm": 0.1626281991220394, "learning_rate": 0.00018132323673039885, "loss": 1.2277, "step": 458 }, { "epoch": 0.84, "grad_norm": 0.16158256707311264, "learning_rate": 0.00018119969272532166, "loss": 1.1624, "step": 459 }, { "epoch": 0.84, "grad_norm": 0.17705809207051687, "learning_rate": 0.00018107578383394146, "loss": 1.2421, "step": 460 }, { "epoch": 0.84, "grad_norm": 0.17639060401882287, "learning_rate": 0.00018095151061306645, "loss": 1.285, "step": 461 }, { "epoch": 0.84, "grad_norm": 0.16918796486576196, "learning_rate": 0.00018082687362114212, "loss": 1.2606, "step": 462 }, { "epoch": 0.84, "grad_norm": 0.15968377185965665, "learning_rate": 0.0001807018734182485, "loss": 1.194, "step": 463 }, { "epoch": 0.85, "grad_norm": 0.17537027967397978, "learning_rate": 0.00018057651056609784, "loss": 1.1594, "step": 464 }, { "epoch": 0.85, "grad_norm": 0.15753665403127565, "learning_rate": 0.00018045078562803203, "loss": 1.1382, "step": 465 }, { "epoch": 0.85, "grad_norm": 0.17121200763916436, "learning_rate": 0.00018032469916902003, "loss": 1.2286, "step": 466 }, { "epoch": 0.85, "grad_norm": 0.19120510133331003, "learning_rate": 0.00018019825175565542, "loss": 1.2835, "step": 467 }, { "epoch": 0.85, "grad_norm": 0.1671735980123817, "learning_rate": 0.0001800714439561538, "loss": 1.2201, "step": 468 }, { "epoch": 0.85, "grad_norm": 0.1579098534969056, "learning_rate": 0.00017994427634035015, "loss": 1.2156, "step": 469 }, { "epoch": 0.86, "grad_norm": 0.1746075421158512, "learning_rate": 0.00017981674947969636, "loss": 1.2049, "step": 470 }, { "epoch": 0.86, "grad_norm": 0.16878182886737042, "learning_rate": 0.00017968886394725874, "loss": 1.2204, "step": 471 }, { "epoch": 0.86, "grad_norm": 0.16725956538286493, "learning_rate": 0.00017956062031771535, "loss": 1.2091, "step": 472 }, { "epoch": 0.86, "grad_norm": 0.18877845951705005, "learning_rate": 0.00017943201916735335, "loss": 1.241, "step": 473 }, { "epoch": 0.86, "grad_norm": 0.180337447476004, "learning_rate": 0.00017930306107406653, "loss": 1.2253, "step": 474 }, { "epoch": 0.87, "grad_norm": 0.16688572366717752, "learning_rate": 0.0001791737466173527, "loss": 1.239, "step": 475 }, { "epoch": 0.87, "grad_norm": 0.15385917621135983, "learning_rate": 0.00017904407637831099, "loss": 1.2476, "step": 476 }, { "epoch": 0.87, "grad_norm": 0.17725645269055587, "learning_rate": 0.00017891405093963938, "loss": 1.2599, "step": 477 }, { "epoch": 0.87, "grad_norm": 0.14758551718901028, "learning_rate": 0.00017878367088563195, "loss": 1.2249, "step": 478 }, { "epoch": 0.87, "grad_norm": 0.15216962408661316, "learning_rate": 0.00017865293680217637, "loss": 1.2346, "step": 479 }, { "epoch": 0.87, "grad_norm": 0.16679282848599514, "learning_rate": 0.00017852184927675112, "loss": 1.2443, "step": 480 }, { "epoch": 0.88, "grad_norm": 0.16723562739069214, "learning_rate": 0.00017839040889842305, "loss": 1.224, "step": 481 }, { "epoch": 0.88, "grad_norm": 0.15922276239929914, "learning_rate": 0.00017825861625784455, "loss": 1.2739, "step": 482 }, { "epoch": 0.88, "grad_norm": 0.1510107938469514, "learning_rate": 0.00017812647194725094, "loss": 1.1764, "step": 483 }, { "epoch": 0.88, "grad_norm": 0.16446999054333494, "learning_rate": 0.00017799397656045792, "loss": 1.2498, "step": 484 }, { "epoch": 0.88, "grad_norm": 0.18566301651865832, "learning_rate": 0.00017786113069285874, "loss": 1.232, "step": 485 }, { "epoch": 0.89, "grad_norm": 0.20592971655306183, "learning_rate": 0.00017772793494142167, "loss": 1.1586, "step": 486 }, { "epoch": 0.89, "grad_norm": 0.1581947714375729, "learning_rate": 0.00017759438990468725, "loss": 1.2502, "step": 487 }, { "epoch": 0.89, "grad_norm": 0.15466760695169174, "learning_rate": 0.00017746049618276545, "loss": 1.1605, "step": 488 }, { "epoch": 0.89, "grad_norm": 0.16041506222444918, "learning_rate": 0.00017732625437733335, "loss": 1.2778, "step": 489 }, { "epoch": 0.89, "grad_norm": 0.17168109661676773, "learning_rate": 0.0001771916650916321, "loss": 1.262, "step": 490 }, { "epoch": 0.89, "grad_norm": 0.1788973186498254, "learning_rate": 0.00017705672893046425, "loss": 1.2111, "step": 491 }, { "epoch": 0.9, "grad_norm": 0.1759644359346382, "learning_rate": 0.00017692144650019125, "loss": 1.2546, "step": 492 }, { "epoch": 0.9, "grad_norm": 0.15710749736088767, "learning_rate": 0.0001767858184087304, "loss": 1.2487, "step": 493 }, { "epoch": 0.9, "grad_norm": 0.1648235522911144, "learning_rate": 0.00017664984526555248, "loss": 1.2469, "step": 494 }, { "epoch": 0.9, "grad_norm": 0.15452607969890703, "learning_rate": 0.0001765135276816787, "loss": 1.1855, "step": 495 }, { "epoch": 0.9, "grad_norm": 0.1837695597880219, "learning_rate": 0.00017637686626967812, "loss": 1.2185, "step": 496 }, { "epoch": 0.91, "grad_norm": 0.15861390725762364, "learning_rate": 0.00017623986164366486, "loss": 1.2056, "step": 497 }, { "epoch": 0.91, "grad_norm": 0.1663260460966887, "learning_rate": 0.00017610251441929533, "loss": 1.1242, "step": 498 }, { "epoch": 0.91, "grad_norm": 0.1803309720529981, "learning_rate": 0.00017596482521376546, "loss": 1.2938, "step": 499 }, { "epoch": 0.91, "grad_norm": 0.14909085011764342, "learning_rate": 0.00017582679464580797, "loss": 1.1953, "step": 500 }, { "epoch": 0.91, "grad_norm": 0.15779022242482527, "learning_rate": 0.00017568842333568952, "loss": 1.2792, "step": 501 }, { "epoch": 0.91, "grad_norm": 0.1553327313967345, "learning_rate": 0.00017554971190520798, "loss": 1.2286, "step": 502 }, { "epoch": 0.92, "grad_norm": 0.16363964666273684, "learning_rate": 0.00017541066097768963, "loss": 1.2753, "step": 503 }, { "epoch": 0.92, "grad_norm": 0.16668099163659675, "learning_rate": 0.00017527127117798635, "loss": 1.185, "step": 504 }, { "epoch": 0.92, "grad_norm": 0.13957568397594883, "learning_rate": 0.0001751315431324727, "loss": 1.143, "step": 505 }, { "epoch": 0.92, "grad_norm": 0.1553111736740035, "learning_rate": 0.00017499147746904335, "loss": 1.2492, "step": 506 }, { "epoch": 0.92, "grad_norm": 0.1691517335818193, "learning_rate": 0.00017485107481711012, "loss": 1.2619, "step": 507 }, { "epoch": 0.93, "grad_norm": 0.15480883994395986, "learning_rate": 0.00017471033580759903, "loss": 1.2396, "step": 508 }, { "epoch": 0.93, "grad_norm": 0.1451690143792058, "learning_rate": 0.00017456926107294765, "loss": 1.1732, "step": 509 }, { "epoch": 0.93, "grad_norm": 0.1524398957482947, "learning_rate": 0.00017442785124710227, "loss": 1.2083, "step": 510 }, { "epoch": 0.93, "grad_norm": 0.16790264977550012, "learning_rate": 0.0001742861069655148, "loss": 1.2201, "step": 511 }, { "epoch": 0.93, "grad_norm": 0.1529847047636337, "learning_rate": 0.0001741440288651403, "loss": 1.243, "step": 512 }, { "epoch": 0.93, "grad_norm": 0.1485875402374676, "learning_rate": 0.00017400161758443375, "loss": 1.2053, "step": 513 }, { "epoch": 0.94, "grad_norm": 0.16950094279079617, "learning_rate": 0.00017385887376334742, "loss": 1.1944, "step": 514 }, { "epoch": 0.94, "grad_norm": 0.15289337084330445, "learning_rate": 0.00017371579804332789, "loss": 1.2503, "step": 515 }, { "epoch": 0.94, "grad_norm": 0.15337063655317973, "learning_rate": 0.00017357239106731317, "loss": 1.3092, "step": 516 }, { "epoch": 0.94, "grad_norm": 0.1458937961897621, "learning_rate": 0.00017342865347972988, "loss": 1.2244, "step": 517 }, { "epoch": 0.94, "grad_norm": 0.19897118610161338, "learning_rate": 0.00017328458592649027, "loss": 1.2238, "step": 518 }, { "epoch": 0.95, "grad_norm": 0.15850805264911003, "learning_rate": 0.00017314018905498931, "loss": 1.195, "step": 519 }, { "epoch": 0.95, "grad_norm": 0.14445183074519347, "learning_rate": 0.00017299546351410197, "loss": 1.1974, "step": 520 }, { "epoch": 0.95, "grad_norm": 0.18180731722745677, "learning_rate": 0.00017285040995418, "loss": 1.2107, "step": 521 }, { "epoch": 0.95, "grad_norm": 0.14943874953193587, "learning_rate": 0.00017270502902704926, "loss": 1.1843, "step": 522 }, { "epoch": 0.95, "grad_norm": 0.15767466790910512, "learning_rate": 0.00017255932138600665, "loss": 1.1409, "step": 523 }, { "epoch": 0.95, "grad_norm": 0.16402921378654775, "learning_rate": 0.00017241328768581726, "loss": 1.2135, "step": 524 }, { "epoch": 0.96, "grad_norm": 0.15526246786505485, "learning_rate": 0.00017226692858271134, "loss": 1.2255, "step": 525 }, { "epoch": 0.96, "grad_norm": 0.16608155892622348, "learning_rate": 0.00017212024473438147, "loss": 1.2691, "step": 526 }, { "epoch": 0.96, "grad_norm": 0.14913271520144072, "learning_rate": 0.00017197323679997943, "loss": 1.1574, "step": 527 }, { "epoch": 0.96, "grad_norm": 0.1471910610421707, "learning_rate": 0.00017182590544011347, "loss": 1.2774, "step": 528 }, { "epoch": 0.96, "grad_norm": 0.1417464185073962, "learning_rate": 0.00017167825131684513, "loss": 1.2446, "step": 529 }, { "epoch": 0.97, "grad_norm": 0.1610488125634495, "learning_rate": 0.0001715302750936864, "loss": 1.2862, "step": 530 }, { "epoch": 0.97, "grad_norm": 0.20227974555123074, "learning_rate": 0.00017138197743559654, "loss": 1.207, "step": 531 }, { "epoch": 0.97, "grad_norm": 0.1355502559749413, "learning_rate": 0.00017123335900897946, "loss": 1.1019, "step": 532 }, { "epoch": 0.97, "grad_norm": 0.1559423167028215, "learning_rate": 0.00017108442048168038, "loss": 1.2549, "step": 533 }, { "epoch": 0.97, "grad_norm": 0.15898973818185586, "learning_rate": 0.00017093516252298296, "loss": 1.2705, "step": 534 }, { "epoch": 0.97, "grad_norm": 0.15169569998999652, "learning_rate": 0.00017078558580360632, "loss": 1.2454, "step": 535 }, { "epoch": 0.98, "grad_norm": 0.15976111665597925, "learning_rate": 0.00017063569099570196, "loss": 1.2585, "step": 536 }, { "epoch": 0.98, "grad_norm": 0.14488877221999352, "learning_rate": 0.00017048547877285077, "loss": 1.2169, "step": 537 }, { "epoch": 0.98, "grad_norm": 0.14919533098974924, "learning_rate": 0.00017033494981006002, "loss": 1.2358, "step": 538 }, { "epoch": 0.98, "grad_norm": 0.15251746717084805, "learning_rate": 0.00017018410478376032, "loss": 1.2241, "step": 539 }, { "epoch": 0.98, "grad_norm": 0.1456060482002663, "learning_rate": 0.00017003294437180255, "loss": 1.2298, "step": 540 }, { "epoch": 0.99, "grad_norm": 0.17048886778787248, "learning_rate": 0.00016988146925345484, "loss": 1.2707, "step": 541 }, { "epoch": 0.99, "grad_norm": 0.15304381059310815, "learning_rate": 0.00016972968010939954, "loss": 1.1498, "step": 542 }, { "epoch": 0.99, "grad_norm": 0.16590055969071696, "learning_rate": 0.0001695775776217301, "loss": 1.2481, "step": 543 }, { "epoch": 0.99, "grad_norm": 0.14299575837437278, "learning_rate": 0.00016942516247394807, "loss": 1.2058, "step": 544 }, { "epoch": 0.99, "grad_norm": 0.14275107775859475, "learning_rate": 0.00016927243535095997, "loss": 1.2178, "step": 545 }, { "epoch": 0.99, "grad_norm": 0.1554250137491414, "learning_rate": 0.0001691193969390742, "loss": 1.1197, "step": 546 }, { "epoch": 1.0, "grad_norm": 0.16958418467021688, "learning_rate": 0.0001689660479259981, "loss": 1.1768, "step": 547 }, { "epoch": 1.0, "grad_norm": 0.1546216583314497, "learning_rate": 0.00016881238900083473, "loss": 1.1741, "step": 548 }, { "epoch": 1.0, "grad_norm": 0.15287056494787424, "learning_rate": 0.0001686584208540797, "loss": 1.2328, "step": 549 }, { "epoch": 1.0, "grad_norm": 0.1419329373337611, "learning_rate": 0.0001685041441776183, "loss": 1.1743, "step": 550 }, { "epoch": 1.0, "grad_norm": 0.15662192296485464, "learning_rate": 0.00016834955966472213, "loss": 1.1861, "step": 551 }, { "epoch": 1.01, "grad_norm": 0.16304778894798697, "learning_rate": 0.00016819466801004621, "loss": 1.2045, "step": 552 }, { "epoch": 1.01, "grad_norm": 0.15690090424895087, "learning_rate": 0.00016803946990962576, "loss": 1.1553, "step": 553 }, { "epoch": 1.01, "grad_norm": 0.15227523196027068, "learning_rate": 0.000167883966060873, "loss": 1.126, "step": 554 }, { "epoch": 1.01, "grad_norm": 0.15442134859978873, "learning_rate": 0.00016772815716257412, "loss": 1.169, "step": 555 }, { "epoch": 1.01, "grad_norm": 0.16235578615890994, "learning_rate": 0.00016757204391488613, "loss": 1.198, "step": 556 }, { "epoch": 1.01, "grad_norm": 0.16631365293435893, "learning_rate": 0.00016741562701933367, "loss": 1.2191, "step": 557 }, { "epoch": 1.02, "grad_norm": 0.15668252106453312, "learning_rate": 0.0001672589071788059, "loss": 1.1094, "step": 558 }, { "epoch": 1.02, "grad_norm": 0.14883987789397163, "learning_rate": 0.00016710188509755329, "loss": 1.0653, "step": 559 }, { "epoch": 1.02, "grad_norm": 0.14723412204861966, "learning_rate": 0.00016694456148118452, "loss": 1.0632, "step": 560 }, { "epoch": 1.02, "grad_norm": 0.2694273773199372, "learning_rate": 0.00016678693703666325, "loss": 1.1577, "step": 561 }, { "epoch": 1.02, "grad_norm": 0.3010203762394184, "learning_rate": 0.00016662901247230502, "loss": 1.1049, "step": 562 }, { "epoch": 1.03, "grad_norm": 0.16960159731526514, "learning_rate": 0.0001664707884977739, "loss": 1.2185, "step": 563 }, { "epoch": 1.03, "grad_norm": 0.1593257551859938, "learning_rate": 0.00016631226582407952, "loss": 1.1085, "step": 564 }, { "epoch": 1.03, "grad_norm": 0.1695114549114674, "learning_rate": 0.00016615344516357378, "loss": 1.1863, "step": 565 }, { "epoch": 1.03, "grad_norm": 0.17116727204552032, "learning_rate": 0.00016599432722994755, "loss": 1.0921, "step": 566 }, { "epoch": 1.03, "grad_norm": 0.16265948588203852, "learning_rate": 0.00016583491273822765, "loss": 1.1905, "step": 567 }, { "epoch": 1.03, "grad_norm": 0.16500105151611763, "learning_rate": 0.00016567520240477344, "loss": 1.1477, "step": 568 }, { "epoch": 1.04, "grad_norm": 0.15445707774433456, "learning_rate": 0.00016551519694727381, "loss": 1.1247, "step": 569 }, { "epoch": 1.04, "grad_norm": 0.17740657850033337, "learning_rate": 0.0001653548970847438, "loss": 1.0164, "step": 570 }, { "epoch": 1.04, "grad_norm": 0.16152198390194278, "learning_rate": 0.0001651943035375214, "loss": 1.1779, "step": 571 }, { "epoch": 1.04, "grad_norm": 0.16299638453443582, "learning_rate": 0.00016503341702726426, "loss": 1.1847, "step": 572 }, { "epoch": 1.04, "grad_norm": 0.1486770667934901, "learning_rate": 0.00016487223827694672, "loss": 1.1002, "step": 573 }, { "epoch": 1.05, "grad_norm": 0.15907088594322072, "learning_rate": 0.00016471076801085615, "loss": 1.1127, "step": 574 }, { "epoch": 1.05, "grad_norm": 0.15491780431797444, "learning_rate": 0.00016454900695458998, "loss": 1.1196, "step": 575 }, { "epoch": 1.05, "grad_norm": 0.16584586610927665, "learning_rate": 0.00016438695583505242, "loss": 1.1441, "step": 576 }, { "epoch": 1.05, "grad_norm": 0.15197831968756778, "learning_rate": 0.00016422461538045103, "loss": 1.0532, "step": 577 }, { "epoch": 1.05, "grad_norm": 0.15669296934233234, "learning_rate": 0.00016406198632029357, "loss": 1.1753, "step": 578 }, { "epoch": 1.05, "grad_norm": 0.1523894998655591, "learning_rate": 0.0001638990693853848, "loss": 1.1339, "step": 579 }, { "epoch": 1.06, "grad_norm": 0.1570057583413597, "learning_rate": 0.000163735865307823, "loss": 1.1117, "step": 580 }, { "epoch": 1.06, "grad_norm": 0.1637341350288717, "learning_rate": 0.00016357237482099684, "loss": 1.1369, "step": 581 }, { "epoch": 1.06, "grad_norm": 0.15703259344496112, "learning_rate": 0.0001634085986595819, "loss": 1.1448, "step": 582 }, { "epoch": 1.06, "grad_norm": 0.1672273979359662, "learning_rate": 0.00016324453755953773, "loss": 1.1972, "step": 583 }, { "epoch": 1.06, "grad_norm": 0.16032943590057747, "learning_rate": 0.0001630801922581041, "loss": 1.1811, "step": 584 }, { "epoch": 1.07, "grad_norm": 0.15414179865764022, "learning_rate": 0.00016291556349379795, "loss": 1.1126, "step": 585 }, { "epoch": 1.07, "grad_norm": 0.16133591336152153, "learning_rate": 0.00016275065200641004, "loss": 1.0664, "step": 586 }, { "epoch": 1.07, "grad_norm": 0.17496271909044792, "learning_rate": 0.00016258545853700158, "loss": 1.2042, "step": 587 }, { "epoch": 1.07, "grad_norm": 0.16189829224809252, "learning_rate": 0.00016241998382790095, "loss": 1.1896, "step": 588 }, { "epoch": 1.07, "grad_norm": 0.1508460775909727, "learning_rate": 0.00016225422862270027, "loss": 1.1205, "step": 589 }, { "epoch": 1.07, "grad_norm": 0.16848110855261456, "learning_rate": 0.00016208819366625218, "loss": 1.1132, "step": 590 }, { "epoch": 1.08, "grad_norm": 0.15996856299765072, "learning_rate": 0.00016192187970466644, "loss": 1.1022, "step": 591 }, { "epoch": 1.08, "grad_norm": 0.15697088450293906, "learning_rate": 0.0001617552874853065, "loss": 1.0483, "step": 592 }, { "epoch": 1.08, "grad_norm": 0.16338270251123885, "learning_rate": 0.0001615884177567863, "loss": 1.1049, "step": 593 }, { "epoch": 1.08, "grad_norm": 0.1605597226615145, "learning_rate": 0.0001614212712689668, "loss": 1.1319, "step": 594 }, { "epoch": 1.08, "grad_norm": 0.1694169270615921, "learning_rate": 0.00016125384877295257, "loss": 1.2143, "step": 595 }, { "epoch": 1.09, "grad_norm": 0.16306866904399633, "learning_rate": 0.00016108615102108855, "loss": 1.164, "step": 596 }, { "epoch": 1.09, "grad_norm": 0.1646758516051161, "learning_rate": 0.00016091817876695655, "loss": 1.2176, "step": 597 }, { "epoch": 1.09, "grad_norm": 0.16334394129553162, "learning_rate": 0.00016074993276537198, "loss": 1.1526, "step": 598 }, { "epoch": 1.09, "grad_norm": 0.16103491072240317, "learning_rate": 0.00016058141377238026, "loss": 1.1022, "step": 599 }, { "epoch": 1.09, "grad_norm": 0.16029258404100036, "learning_rate": 0.00016041262254525362, "loss": 1.1508, "step": 600 }, { "epoch": 1.09, "grad_norm": 0.15686217802450533, "learning_rate": 0.00016024355984248768, "loss": 1.1471, "step": 601 }, { "epoch": 1.1, "grad_norm": 0.1520812354157866, "learning_rate": 0.0001600742264237979, "loss": 1.0994, "step": 602 }, { "epoch": 1.1, "grad_norm": 0.161954690104593, "learning_rate": 0.0001599046230501163, "loss": 1.1475, "step": 603 }, { "epoch": 1.1, "grad_norm": 0.16470046818858164, "learning_rate": 0.00015973475048358795, "loss": 1.1425, "step": 604 }, { "epoch": 1.1, "grad_norm": 0.17783426542152342, "learning_rate": 0.00015956460948756765, "loss": 1.1933, "step": 605 }, { "epoch": 1.1, "grad_norm": 0.20834095672163827, "learning_rate": 0.0001593942008266164, "loss": 1.1545, "step": 606 }, { "epoch": 1.11, "grad_norm": 0.1601922371369916, "learning_rate": 0.00015922352526649803, "loss": 1.1626, "step": 607 }, { "epoch": 1.11, "grad_norm": 0.17590795577463478, "learning_rate": 0.00015905258357417569, "loss": 1.1428, "step": 608 }, { "epoch": 1.11, "grad_norm": 0.16438364538914646, "learning_rate": 0.00015888137651780845, "loss": 1.1237, "step": 609 }, { "epoch": 1.11, "grad_norm": 0.1532404311365829, "learning_rate": 0.00015870990486674792, "loss": 1.1292, "step": 610 }, { "epoch": 1.11, "grad_norm": 0.1658092976448754, "learning_rate": 0.0001585381693915346, "loss": 1.2051, "step": 611 }, { "epoch": 1.11, "grad_norm": 0.17009266166893572, "learning_rate": 0.00015836617086389468, "loss": 1.1574, "step": 612 }, { "epoch": 1.12, "grad_norm": 0.17117032466097273, "learning_rate": 0.00015819391005673626, "loss": 1.1715, "step": 613 }, { "epoch": 1.12, "grad_norm": 0.16766945870038483, "learning_rate": 0.00015802138774414622, "loss": 1.1322, "step": 614 }, { "epoch": 1.12, "grad_norm": 0.16603674100502688, "learning_rate": 0.00015784860470138633, "loss": 1.1682, "step": 615 }, { "epoch": 1.12, "grad_norm": 0.1538296997751752, "learning_rate": 0.00015767556170489025, "loss": 1.0979, "step": 616 }, { "epoch": 1.12, "grad_norm": 0.16355862351540293, "learning_rate": 0.00015750225953225968, "loss": 1.0859, "step": 617 }, { "epoch": 1.13, "grad_norm": 0.17100029950170442, "learning_rate": 0.00015732869896226094, "loss": 1.1513, "step": 618 }, { "epoch": 1.13, "grad_norm": 0.16981502045992727, "learning_rate": 0.0001571548807748215, "loss": 1.1665, "step": 619 }, { "epoch": 1.13, "grad_norm": 0.15981257682597005, "learning_rate": 0.00015698080575102661, "loss": 1.0862, "step": 620 }, { "epoch": 1.13, "grad_norm": 0.17114664647857297, "learning_rate": 0.00015680647467311557, "loss": 1.1883, "step": 621 }, { "epoch": 1.13, "grad_norm": 0.16818704328900685, "learning_rate": 0.00015663188832447833, "loss": 1.1345, "step": 622 }, { "epoch": 1.13, "grad_norm": 0.16716958487866437, "learning_rate": 0.0001564570474896519, "loss": 1.1613, "step": 623 }, { "epoch": 1.14, "grad_norm": 0.16425076739154273, "learning_rate": 0.00015628195295431697, "loss": 1.1153, "step": 624 }, { "epoch": 1.14, "grad_norm": 0.1668267227341853, "learning_rate": 0.0001561066055052941, "loss": 1.1796, "step": 625 }, { "epoch": 1.14, "grad_norm": 0.15748311738845247, "learning_rate": 0.00015593100593054064, "loss": 1.125, "step": 626 }, { "epoch": 1.14, "grad_norm": 0.1578229320002536, "learning_rate": 0.00015575515501914668, "loss": 1.1751, "step": 627 }, { "epoch": 1.14, "grad_norm": 0.17990453419427754, "learning_rate": 0.0001555790535613318, "loss": 1.2107, "step": 628 }, { "epoch": 1.15, "grad_norm": 0.17170087096955436, "learning_rate": 0.0001554027023484416, "loss": 1.1704, "step": 629 }, { "epoch": 1.15, "grad_norm": 0.1641233552033525, "learning_rate": 0.00015522610217294375, "loss": 1.1399, "step": 630 }, { "epoch": 1.15, "grad_norm": 0.22816800313793714, "learning_rate": 0.00015504925382842487, "loss": 1.1374, "step": 631 }, { "epoch": 1.15, "grad_norm": 0.1699514304776241, "learning_rate": 0.00015487215810958675, "loss": 1.1668, "step": 632 }, { "epoch": 1.15, "grad_norm": 0.16580705156733896, "learning_rate": 0.00015469481581224272, "loss": 1.2467, "step": 633 }, { "epoch": 1.15, "grad_norm": 0.1621295978537637, "learning_rate": 0.0001545172277333142, "loss": 1.095, "step": 634 }, { "epoch": 1.16, "grad_norm": 0.15661235975760687, "learning_rate": 0.00015433939467082713, "loss": 1.1437, "step": 635 }, { "epoch": 1.16, "grad_norm": 0.15913894864881573, "learning_rate": 0.00015416131742390827, "loss": 1.1682, "step": 636 }, { "epoch": 1.16, "grad_norm": 0.17078546916822446, "learning_rate": 0.00015398299679278172, "loss": 1.1673, "step": 637 }, { "epoch": 1.16, "grad_norm": 0.17593093163481005, "learning_rate": 0.00015380443357876518, "loss": 1.1588, "step": 638 }, { "epoch": 1.16, "grad_norm": 0.16563306408223222, "learning_rate": 0.00015362562858426654, "loss": 1.1542, "step": 639 }, { "epoch": 1.17, "grad_norm": 0.15824017459526013, "learning_rate": 0.0001534465826127801, "loss": 1.136, "step": 640 }, { "epoch": 1.17, "grad_norm": 0.1555748015065671, "learning_rate": 0.00015326729646888314, "loss": 1.1744, "step": 641 }, { "epoch": 1.17, "grad_norm": 0.15054650982383877, "learning_rate": 0.0001530877709582321, "loss": 1.0996, "step": 642 }, { "epoch": 1.17, "grad_norm": 0.1732483344367302, "learning_rate": 0.00015290800688755907, "loss": 1.1375, "step": 643 }, { "epoch": 1.17, "grad_norm": 0.17086276533942882, "learning_rate": 0.0001527280050646682, "loss": 1.1399, "step": 644 }, { "epoch": 1.17, "grad_norm": 0.14791498583408913, "learning_rate": 0.00015254776629843205, "loss": 1.0112, "step": 645 }, { "epoch": 1.18, "grad_norm": 0.16492159023612152, "learning_rate": 0.00015236729139878782, "loss": 1.1179, "step": 646 }, { "epoch": 1.18, "grad_norm": 0.15550107018855622, "learning_rate": 0.0001521865811767339, "loss": 1.0872, "step": 647 }, { "epoch": 1.18, "grad_norm": 0.17086993477527834, "learning_rate": 0.00015200563644432612, "loss": 1.1747, "step": 648 }, { "epoch": 1.18, "grad_norm": 0.16832629109129904, "learning_rate": 0.0001518244580146742, "loss": 1.2623, "step": 649 }, { "epoch": 1.18, "grad_norm": 0.15555098188988337, "learning_rate": 0.00015164304670193792, "loss": 1.1172, "step": 650 }, { "epoch": 1.19, "grad_norm": 0.16504655864531245, "learning_rate": 0.00015146140332132358, "loss": 1.1615, "step": 651 }, { "epoch": 1.19, "grad_norm": 0.15904615747067738, "learning_rate": 0.00015127952868908043, "loss": 1.1841, "step": 652 }, { "epoch": 1.19, "grad_norm": 0.1664190253819244, "learning_rate": 0.00015109742362249672, "loss": 1.2258, "step": 653 }, { "epoch": 1.19, "grad_norm": 0.17622098883404838, "learning_rate": 0.00015091508893989633, "loss": 1.2141, "step": 654 }, { "epoch": 1.19, "grad_norm": 0.17253106779139057, "learning_rate": 0.00015073252546063493, "loss": 1.1289, "step": 655 }, { "epoch": 1.19, "grad_norm": 0.15924181512457228, "learning_rate": 0.0001505497340050963, "loss": 1.1836, "step": 656 }, { "epoch": 1.2, "grad_norm": 0.1755114666296529, "learning_rate": 0.00015036671539468878, "loss": 1.2109, "step": 657 }, { "epoch": 1.2, "grad_norm": 0.1623211151951732, "learning_rate": 0.00015018347045184132, "loss": 1.1922, "step": 658 }, { "epoch": 1.2, "grad_norm": 0.16895842955233992, "learning_rate": 0.00015000000000000001, "loss": 1.2402, "step": 659 }, { "epoch": 1.2, "grad_norm": 0.17126264186496687, "learning_rate": 0.00014981630486362435, "loss": 1.2327, "step": 660 }, { "epoch": 1.2, "grad_norm": 0.19632534974860108, "learning_rate": 0.00014963238586818345, "loss": 1.1586, "step": 661 }, { "epoch": 1.21, "grad_norm": 0.16708414744148634, "learning_rate": 0.00014944824384015236, "loss": 1.1766, "step": 662 }, { "epoch": 1.21, "grad_norm": 0.1800283830073218, "learning_rate": 0.00014926387960700842, "loss": 1.0902, "step": 663 }, { "epoch": 1.21, "grad_norm": 0.17570100482291343, "learning_rate": 0.0001490792939972275, "loss": 1.1875, "step": 664 }, { "epoch": 1.21, "grad_norm": 0.15904497800793038, "learning_rate": 0.0001488944878402802, "loss": 1.083, "step": 665 }, { "epoch": 1.21, "grad_norm": 0.1629549994998532, "learning_rate": 0.00014870946196662822, "loss": 1.1505, "step": 666 }, { "epoch": 1.21, "grad_norm": 0.17495490666041913, "learning_rate": 0.00014852421720772062, "loss": 1.1107, "step": 667 }, { "epoch": 1.22, "grad_norm": 0.1814722009088628, "learning_rate": 0.00014833875439599004, "loss": 1.2089, "step": 668 }, { "epoch": 1.22, "grad_norm": 0.16312442272141373, "learning_rate": 0.00014815307436484898, "loss": 1.1518, "step": 669 }, { "epoch": 1.22, "grad_norm": 0.18878089936272002, "learning_rate": 0.00014796717794868607, "loss": 1.1806, "step": 670 }, { "epoch": 1.22, "grad_norm": 0.171347321491223, "learning_rate": 0.00014778106598286234, "loss": 1.3189, "step": 671 }, { "epoch": 1.22, "grad_norm": 0.15561777538677532, "learning_rate": 0.00014759473930370736, "loss": 1.1071, "step": 672 }, { "epoch": 1.23, "grad_norm": 0.16095434527629326, "learning_rate": 0.0001474081987485156, "loss": 1.173, "step": 673 }, { "epoch": 1.23, "grad_norm": 0.1594519070011647, "learning_rate": 0.00014722144515554264, "loss": 1.1099, "step": 674 }, { "epoch": 1.23, "grad_norm": 0.15853021638823037, "learning_rate": 0.00014703447936400134, "loss": 1.1563, "step": 675 }, { "epoch": 1.23, "grad_norm": 0.21942200368949344, "learning_rate": 0.00014684730221405814, "loss": 1.1226, "step": 676 }, { "epoch": 1.23, "grad_norm": 0.16328115550493413, "learning_rate": 0.00014665991454682924, "loss": 1.1878, "step": 677 }, { "epoch": 1.23, "grad_norm": 0.16551970389762746, "learning_rate": 0.00014647231720437686, "loss": 1.055, "step": 678 }, { "epoch": 1.24, "grad_norm": 0.17275962400330097, "learning_rate": 0.00014628451102970547, "loss": 1.1302, "step": 679 }, { "epoch": 1.24, "grad_norm": 0.17269943862734122, "learning_rate": 0.00014609649686675785, "loss": 1.1635, "step": 680 }, { "epoch": 1.24, "grad_norm": 0.18932844828433326, "learning_rate": 0.00014590827556041158, "loss": 1.0954, "step": 681 }, { "epoch": 1.24, "grad_norm": 0.1670913427771278, "learning_rate": 0.00014571984795647494, "loss": 1.1336, "step": 682 }, { "epoch": 1.24, "grad_norm": 0.1727053222121231, "learning_rate": 0.00014553121490168332, "loss": 1.1477, "step": 683 }, { "epoch": 1.25, "grad_norm": 0.20281162440108022, "learning_rate": 0.00014534237724369534, "loss": 1.1364, "step": 684 }, { "epoch": 1.25, "grad_norm": 0.17595782723751183, "learning_rate": 0.00014515333583108896, "loss": 1.1977, "step": 685 }, { "epoch": 1.25, "grad_norm": 0.16755958800968615, "learning_rate": 0.00014496409151335785, "loss": 1.1315, "step": 686 }, { "epoch": 1.25, "grad_norm": 0.17908885139779734, "learning_rate": 0.00014477464514090743, "loss": 1.1851, "step": 687 }, { "epoch": 1.25, "grad_norm": 0.16047769884625865, "learning_rate": 0.00014458499756505116, "loss": 1.1137, "step": 688 }, { "epoch": 1.26, "grad_norm": 0.18864957014640948, "learning_rate": 0.00014439514963800648, "loss": 1.1962, "step": 689 }, { "epoch": 1.26, "grad_norm": 0.16074020271372416, "learning_rate": 0.00014420510221289137, "loss": 1.1448, "step": 690 }, { "epoch": 1.26, "grad_norm": 0.17166433855606958, "learning_rate": 0.00014401485614372008, "loss": 1.0487, "step": 691 }, { "epoch": 1.26, "grad_norm": 0.1773682331611409, "learning_rate": 0.0001438244122853996, "loss": 1.16, "step": 692 }, { "epoch": 1.26, "grad_norm": 0.16596601643110168, "learning_rate": 0.00014363377149372584, "loss": 1.0843, "step": 693 }, { "epoch": 1.26, "grad_norm": 0.16748936526973612, "learning_rate": 0.0001434429346253794, "loss": 1.1619, "step": 694 }, { "epoch": 1.27, "grad_norm": 0.17600274748860967, "learning_rate": 0.00014325190253792222, "loss": 1.151, "step": 695 }, { "epoch": 1.27, "grad_norm": 0.1721019830304793, "learning_rate": 0.0001430606760897934, "loss": 1.0609, "step": 696 }, { "epoch": 1.27, "grad_norm": 0.172361054646693, "learning_rate": 0.00014286925614030542, "loss": 1.1777, "step": 697 }, { "epoch": 1.27, "grad_norm": 0.1938123204770535, "learning_rate": 0.00014267764354964038, "loss": 1.1078, "step": 698 }, { "epoch": 1.27, "grad_norm": 0.171152731950173, "learning_rate": 0.00014248583917884594, "loss": 1.2344, "step": 699 }, { "epoch": 1.28, "grad_norm": 0.176915280903416, "learning_rate": 0.00014229384388983167, "loss": 1.1443, "step": 700 }, { "epoch": 1.28, "grad_norm": 0.16620426221482948, "learning_rate": 0.00014210165854536494, "loss": 1.1635, "step": 701 }, { "epoch": 1.28, "grad_norm": 0.17368168932295722, "learning_rate": 0.0001419092840090673, "loss": 1.2076, "step": 702 }, { "epoch": 1.28, "grad_norm": 0.16396003568599715, "learning_rate": 0.0001417167211454104, "loss": 1.1798, "step": 703 }, { "epoch": 1.28, "grad_norm": 0.17474861300401356, "learning_rate": 0.0001415239708197122, "loss": 1.1951, "step": 704 }, { "epoch": 1.28, "grad_norm": 0.17687469132536782, "learning_rate": 0.00014133103389813302, "loss": 1.1863, "step": 705 }, { "epoch": 1.29, "grad_norm": 0.1729350163866909, "learning_rate": 0.0001411379112476717, "loss": 1.1791, "step": 706 }, { "epoch": 1.29, "grad_norm": 0.16183372577765218, "learning_rate": 0.0001409446037361617, "loss": 1.11, "step": 707 }, { "epoch": 1.29, "grad_norm": 0.17468439327379603, "learning_rate": 0.0001407511122322672, "loss": 1.1459, "step": 708 }, { "epoch": 1.29, "grad_norm": 0.1746393388604993, "learning_rate": 0.00014055743760547917, "loss": 1.1782, "step": 709 }, { "epoch": 1.29, "grad_norm": 0.1709730768079813, "learning_rate": 0.00014036358072611147, "loss": 1.1729, "step": 710 }, { "epoch": 1.3, "grad_norm": 0.16338226575160056, "learning_rate": 0.00014016954246529696, "loss": 1.1637, "step": 711 }, { "epoch": 1.3, "grad_norm": 0.16522796461612693, "learning_rate": 0.00013997532369498355, "loss": 1.1786, "step": 712 }, { "epoch": 1.3, "grad_norm": 0.16443017398263532, "learning_rate": 0.0001397809252879303, "loss": 1.1885, "step": 713 }, { "epoch": 1.3, "grad_norm": 0.16678395716368374, "learning_rate": 0.0001395863481177036, "loss": 1.1576, "step": 714 }, { "epoch": 1.3, "grad_norm": 0.15796662849679724, "learning_rate": 0.000139391593058673, "loss": 1.1719, "step": 715 }, { "epoch": 1.3, "grad_norm": 0.1605406960709786, "learning_rate": 0.00013919666098600753, "loss": 1.1442, "step": 716 }, { "epoch": 1.31, "grad_norm": 0.16720350808668913, "learning_rate": 0.00013900155277567157, "loss": 1.1231, "step": 717 }, { "epoch": 1.31, "grad_norm": 0.16025549611436293, "learning_rate": 0.00013880626930442113, "loss": 1.0959, "step": 718 }, { "epoch": 1.31, "grad_norm": 0.1682327263659866, "learning_rate": 0.00013861081144979974, "loss": 1.1816, "step": 719 }, { "epoch": 1.31, "grad_norm": 0.16843691606816277, "learning_rate": 0.00013841518009013445, "loss": 1.1607, "step": 720 }, { "epoch": 1.31, "grad_norm": 0.17753480625595375, "learning_rate": 0.0001382193761045322, "loss": 1.2085, "step": 721 }, { "epoch": 1.32, "grad_norm": 0.175444570487372, "learning_rate": 0.0001380234003728754, "loss": 1.1883, "step": 722 }, { "epoch": 1.32, "grad_norm": 0.17844909537485765, "learning_rate": 0.00013782725377581848, "loss": 1.2548, "step": 723 }, { "epoch": 1.32, "grad_norm": 0.17827531614400435, "learning_rate": 0.00013763093719478358, "loss": 1.1753, "step": 724 }, { "epoch": 1.32, "grad_norm": 0.16568782268956592, "learning_rate": 0.00013743445151195657, "loss": 1.1269, "step": 725 }, { "epoch": 1.32, "grad_norm": 0.16229148230338714, "learning_rate": 0.00013723779761028347, "loss": 1.1233, "step": 726 }, { "epoch": 1.32, "grad_norm": 0.18071555500117806, "learning_rate": 0.000137040976373466, "loss": 1.1691, "step": 727 }, { "epoch": 1.33, "grad_norm": 0.16528568438217509, "learning_rate": 0.000136843988685958, "loss": 1.1044, "step": 728 }, { "epoch": 1.33, "grad_norm": 0.15621961073605914, "learning_rate": 0.00013664683543296112, "loss": 1.0585, "step": 729 }, { "epoch": 1.33, "grad_norm": 0.1821012365890871, "learning_rate": 0.00013644951750042114, "loss": 1.1061, "step": 730 }, { "epoch": 1.33, "grad_norm": 0.18832728793079637, "learning_rate": 0.00013625203577502382, "loss": 1.2088, "step": 731 }, { "epoch": 1.33, "grad_norm": 0.17821269011741103, "learning_rate": 0.00013605439114419094, "loss": 1.1076, "step": 732 }, { "epoch": 1.34, "grad_norm": 0.15816928920099374, "learning_rate": 0.00013585658449607633, "loss": 1.057, "step": 733 }, { "epoch": 1.34, "grad_norm": 0.1634582673036166, "learning_rate": 0.00013565861671956187, "loss": 1.2206, "step": 734 }, { "epoch": 1.34, "grad_norm": 0.16402056158577089, "learning_rate": 0.00013546048870425356, "loss": 1.0809, "step": 735 }, { "epoch": 1.34, "grad_norm": 0.18674575425072043, "learning_rate": 0.0001352622013404774, "loss": 1.0979, "step": 736 }, { "epoch": 1.34, "grad_norm": 0.16768128529630819, "learning_rate": 0.00013506375551927547, "loss": 1.1813, "step": 737 }, { "epoch": 1.34, "grad_norm": 0.17842009039615817, "learning_rate": 0.00013486515213240188, "loss": 1.1767, "step": 738 }, { "epoch": 1.35, "grad_norm": 0.16460609654310757, "learning_rate": 0.0001346663920723188, "loss": 1.1711, "step": 739 }, { "epoch": 1.35, "grad_norm": 0.16130836204900895, "learning_rate": 0.00013446747623219255, "loss": 1.1506, "step": 740 }, { "epoch": 1.35, "grad_norm": 0.19357518099915463, "learning_rate": 0.00013426840550588933, "loss": 1.1454, "step": 741 }, { "epoch": 1.35, "grad_norm": 0.16818483070815926, "learning_rate": 0.0001340691807879714, "loss": 1.105, "step": 742 }, { "epoch": 1.35, "grad_norm": 0.17634247767232533, "learning_rate": 0.00013386980297369307, "loss": 1.1173, "step": 743 }, { "epoch": 1.36, "grad_norm": 0.17275588243641285, "learning_rate": 0.0001336702729589965, "loss": 1.1614, "step": 744 }, { "epoch": 1.36, "grad_norm": 0.18630211892037335, "learning_rate": 0.00013347059164050794, "loss": 1.0822, "step": 745 }, { "epoch": 1.36, "grad_norm": 0.1610612417401003, "learning_rate": 0.0001332707599155334, "loss": 1.11, "step": 746 }, { "epoch": 1.36, "grad_norm": 0.1710900937487699, "learning_rate": 0.00013307077868205487, "loss": 1.1556, "step": 747 }, { "epoch": 1.36, "grad_norm": 0.16789088046962233, "learning_rate": 0.00013287064883872612, "loss": 1.166, "step": 748 }, { "epoch": 1.36, "grad_norm": 0.17152792861267624, "learning_rate": 0.00013267037128486883, "loss": 1.2076, "step": 749 }, { "epoch": 1.37, "grad_norm": 0.16987939988427406, "learning_rate": 0.00013246994692046836, "loss": 1.1892, "step": 750 }, { "epoch": 1.37, "grad_norm": 0.17050807620104472, "learning_rate": 0.00013226937664616976, "loss": 1.1405, "step": 751 }, { "epoch": 1.37, "grad_norm": 0.1619859173580362, "learning_rate": 0.00013206866136327388, "loss": 1.15, "step": 752 }, { "epoch": 1.37, "grad_norm": 0.1693102604231561, "learning_rate": 0.00013186780197373306, "loss": 1.1108, "step": 753 }, { "epoch": 1.37, "grad_norm": 0.15751817301555188, "learning_rate": 0.00013166679938014726, "loss": 1.1075, "step": 754 }, { "epoch": 1.38, "grad_norm": 0.17344055609711165, "learning_rate": 0.00013146565448576004, "loss": 1.2239, "step": 755 }, { "epoch": 1.38, "grad_norm": 0.16419033762795351, "learning_rate": 0.00013126436819445422, "loss": 1.2146, "step": 756 }, { "epoch": 1.38, "grad_norm": 0.17241446484871606, "learning_rate": 0.00013106294141074825, "loss": 1.1808, "step": 757 }, { "epoch": 1.38, "grad_norm": 0.16299230805525922, "learning_rate": 0.0001308613750397917, "loss": 1.1639, "step": 758 }, { "epoch": 1.38, "grad_norm": 0.15993384139183944, "learning_rate": 0.00013065966998736155, "loss": 1.0664, "step": 759 }, { "epoch": 1.38, "grad_norm": 0.17951699730636844, "learning_rate": 0.00013045782715985792, "loss": 1.1565, "step": 760 }, { "epoch": 1.39, "grad_norm": 0.16420615622643212, "learning_rate": 0.0001302558474643, "loss": 1.1675, "step": 761 }, { "epoch": 1.39, "grad_norm": 0.1718558627273243, "learning_rate": 0.0001300537318083221, "loss": 1.1758, "step": 762 }, { "epoch": 1.39, "grad_norm": 0.16040004880583997, "learning_rate": 0.00012985148110016947, "loss": 1.1286, "step": 763 }, { "epoch": 1.39, "grad_norm": 0.1644924005506398, "learning_rate": 0.0001296490962486942, "loss": 1.0901, "step": 764 }, { "epoch": 1.39, "grad_norm": 0.17508769995003778, "learning_rate": 0.00012944657816335123, "loss": 1.158, "step": 765 }, { "epoch": 1.4, "grad_norm": 0.16636235063611218, "learning_rate": 0.0001292439277541942, "loss": 1.1506, "step": 766 }, { "epoch": 1.4, "grad_norm": 0.1651752039403403, "learning_rate": 0.00012904114593187136, "loss": 1.1167, "step": 767 }, { "epoch": 1.4, "grad_norm": 0.1727449182027404, "learning_rate": 0.0001288382336076215, "loss": 1.1208, "step": 768 }, { "epoch": 1.4, "grad_norm": 0.1678888505155555, "learning_rate": 0.00012863519169326984, "loss": 1.1164, "step": 769 }, { "epoch": 1.4, "grad_norm": 0.1717914946350261, "learning_rate": 0.0001284320211012239, "loss": 1.174, "step": 770 }, { "epoch": 1.4, "grad_norm": 0.15833038102003086, "learning_rate": 0.00012822872274446958, "loss": 1.0557, "step": 771 }, { "epoch": 1.41, "grad_norm": 0.16693919668597118, "learning_rate": 0.00012802529753656668, "loss": 1.2202, "step": 772 }, { "epoch": 1.41, "grad_norm": 0.1608034670310254, "learning_rate": 0.0001278217463916453, "loss": 1.0761, "step": 773 }, { "epoch": 1.41, "grad_norm": 0.16522372158185394, "learning_rate": 0.0001276180702244012, "loss": 1.114, "step": 774 }, { "epoch": 1.41, "grad_norm": 0.15699403016449, "learning_rate": 0.00012741426995009213, "loss": 1.0487, "step": 775 }, { "epoch": 1.41, "grad_norm": 0.1724966872571711, "learning_rate": 0.00012721034648453353, "loss": 1.2198, "step": 776 }, { "epoch": 1.42, "grad_norm": 0.16612867386952304, "learning_rate": 0.00012700630074409427, "loss": 1.1481, "step": 777 }, { "epoch": 1.42, "grad_norm": 0.1755115998846381, "learning_rate": 0.0001268021336456929, "loss": 1.1295, "step": 778 }, { "epoch": 1.42, "grad_norm": 0.16610612624318807, "learning_rate": 0.00012659784610679318, "loss": 1.0735, "step": 779 }, { "epoch": 1.42, "grad_norm": 0.17210665028457767, "learning_rate": 0.0001263934390454001, "loss": 1.128, "step": 780 }, { "epoch": 1.42, "grad_norm": 0.17115180879996908, "learning_rate": 0.00012618891338005573, "loss": 1.2114, "step": 781 }, { "epoch": 1.42, "grad_norm": 0.1708121456807184, "learning_rate": 0.0001259842700298352, "loss": 1.2037, "step": 782 }, { "epoch": 1.43, "grad_norm": 0.16413369503599784, "learning_rate": 0.00012577950991434248, "loss": 1.115, "step": 783 }, { "epoch": 1.43, "grad_norm": 0.16575779512688402, "learning_rate": 0.0001255746339537061, "loss": 1.13, "step": 784 }, { "epoch": 1.43, "grad_norm": 0.16787768159350958, "learning_rate": 0.00012536964306857526, "loss": 1.1187, "step": 785 }, { "epoch": 1.43, "grad_norm": 0.161257935533138, "learning_rate": 0.00012516453818011566, "loss": 1.1272, "step": 786 }, { "epoch": 1.43, "grad_norm": 0.17163817923820932, "learning_rate": 0.00012495932021000517, "loss": 1.1371, "step": 787 }, { "epoch": 1.44, "grad_norm": 0.16451801686420228, "learning_rate": 0.0001247539900804299, "loss": 1.1234, "step": 788 }, { "epoch": 1.44, "grad_norm": 0.4897545674498291, "learning_rate": 0.00012454854871407994, "loss": 1.1276, "step": 789 }, { "epoch": 1.44, "grad_norm": 0.1674813107615878, "learning_rate": 0.00012434299703414524, "loss": 1.0717, "step": 790 }, { "epoch": 1.44, "grad_norm": 0.16889265391095676, "learning_rate": 0.0001241373359643114, "loss": 1.1094, "step": 791 }, { "epoch": 1.44, "grad_norm": 0.17807200243744, "learning_rate": 0.0001239315664287558, "loss": 1.2072, "step": 792 }, { "epoch": 1.44, "grad_norm": 0.1821416845887793, "learning_rate": 0.00012372568935214298, "loss": 1.1574, "step": 793 }, { "epoch": 1.45, "grad_norm": 0.16787645348154986, "learning_rate": 0.00012351970565962085, "loss": 1.0884, "step": 794 }, { "epoch": 1.45, "grad_norm": 0.17767787321450104, "learning_rate": 0.00012331361627681645, "loss": 1.1561, "step": 795 }, { "epoch": 1.45, "grad_norm": 0.16990542040230855, "learning_rate": 0.00012310742212983167, "loss": 1.1454, "step": 796 }, { "epoch": 1.45, "grad_norm": 0.17040623149046621, "learning_rate": 0.00012290112414523928, "loss": 1.1143, "step": 797 }, { "epoch": 1.45, "grad_norm": 0.17383642491755713, "learning_rate": 0.00012269472325007858, "loss": 1.1502, "step": 798 }, { "epoch": 1.46, "grad_norm": 0.1670271756098212, "learning_rate": 0.00012248822037185138, "loss": 1.1872, "step": 799 }, { "epoch": 1.46, "grad_norm": 0.17338597290651167, "learning_rate": 0.0001222816164385177, "loss": 1.201, "step": 800 }, { "epoch": 1.46, "grad_norm": 0.16342647338148947, "learning_rate": 0.00012207491237849172, "loss": 1.0994, "step": 801 }, { "epoch": 1.46, "grad_norm": 0.15847559756714308, "learning_rate": 0.0001218681091206376, "loss": 1.1272, "step": 802 }, { "epoch": 1.46, "grad_norm": 0.17295023549456498, "learning_rate": 0.00012166120759426514, "loss": 1.1026, "step": 803 }, { "epoch": 1.46, "grad_norm": 0.17574482838436642, "learning_rate": 0.00012145420872912585, "loss": 1.1785, "step": 804 }, { "epoch": 1.47, "grad_norm": 0.1636907763577917, "learning_rate": 0.0001212471134554086, "loss": 1.125, "step": 805 }, { "epoch": 1.47, "grad_norm": 0.1683476239550862, "learning_rate": 0.00012103992270373547, "loss": 1.1196, "step": 806 }, { "epoch": 1.47, "grad_norm": 0.1634846923426851, "learning_rate": 0.00012083263740515765, "loss": 1.1666, "step": 807 }, { "epoch": 1.47, "grad_norm": 0.20362360607016913, "learning_rate": 0.00012062525849115107, "loss": 1.2166, "step": 808 }, { "epoch": 1.47, "grad_norm": 0.15520795046759722, "learning_rate": 0.00012041778689361254, "loss": 1.1046, "step": 809 }, { "epoch": 1.48, "grad_norm": 0.17498352958051774, "learning_rate": 0.00012021022354485514, "loss": 1.1331, "step": 810 }, { "epoch": 1.48, "grad_norm": 0.17105423893572264, "learning_rate": 0.00012000256937760445, "loss": 1.0799, "step": 811 }, { "epoch": 1.48, "grad_norm": 0.16837214432904796, "learning_rate": 0.00011979482532499401, "loss": 1.1856, "step": 812 }, { "epoch": 1.48, "grad_norm": 0.1717737899447442, "learning_rate": 0.00011958699232056134, "loss": 1.1738, "step": 813 }, { "epoch": 1.48, "grad_norm": 0.16901679352588425, "learning_rate": 0.0001193790712982437, "loss": 1.1361, "step": 814 }, { "epoch": 1.48, "grad_norm": 0.39457461047161113, "learning_rate": 0.00011917106319237386, "loss": 1.1095, "step": 815 }, { "epoch": 1.49, "grad_norm": 0.1689082831860728, "learning_rate": 0.00011896296893767587, "loss": 1.1518, "step": 816 }, { "epoch": 1.49, "grad_norm": 0.17381008421938537, "learning_rate": 0.00011875478946926093, "loss": 1.1251, "step": 817 }, { "epoch": 1.49, "grad_norm": 0.1723002103848165, "learning_rate": 0.00011854652572262323, "loss": 1.1128, "step": 818 }, { "epoch": 1.49, "grad_norm": 0.1799457776202214, "learning_rate": 0.00011833817863363564, "loss": 1.1702, "step": 819 }, { "epoch": 1.49, "grad_norm": 0.18319688210909812, "learning_rate": 0.00011812974913854545, "loss": 1.1537, "step": 820 }, { "epoch": 1.5, "grad_norm": 0.1716604871697964, "learning_rate": 0.0001179212381739704, "loss": 1.1567, "step": 821 }, { "epoch": 1.5, "grad_norm": 0.17098997576829758, "learning_rate": 0.00011771264667689427, "loss": 1.1309, "step": 822 }, { "epoch": 1.5, "grad_norm": 0.2607304915983766, "learning_rate": 0.00011750397558466273, "loss": 1.1151, "step": 823 }, { "epoch": 1.5, "grad_norm": 0.1869030292628108, "learning_rate": 0.00011729522583497912, "loss": 1.0764, "step": 824 }, { "epoch": 1.5, "grad_norm": 0.1969233515350324, "learning_rate": 0.00011708639836590023, "loss": 1.0986, "step": 825 }, { "epoch": 1.5, "grad_norm": 0.17794486028770468, "learning_rate": 0.00011687749411583213, "loss": 1.1327, "step": 826 }, { "epoch": 1.51, "grad_norm": 0.19408346665780069, "learning_rate": 0.00011666851402352588, "loss": 1.0768, "step": 827 }, { "epoch": 1.51, "grad_norm": 0.5122187922492092, "learning_rate": 0.00011645945902807341, "loss": 1.1223, "step": 828 }, { "epoch": 1.51, "grad_norm": 0.18948615786386766, "learning_rate": 0.00011625033006890315, "loss": 1.1452, "step": 829 }, { "epoch": 1.51, "grad_norm": 0.16917117240636045, "learning_rate": 0.00011604112808577603, "loss": 1.131, "step": 830 }, { "epoch": 1.51, "grad_norm": 0.177460873056157, "learning_rate": 0.00011583185401878101, "loss": 1.1691, "step": 831 }, { "epoch": 1.52, "grad_norm": 0.2727518607538819, "learning_rate": 0.00011562250880833104, "loss": 1.185, "step": 832 }, { "epoch": 1.52, "grad_norm": 0.1722336921590558, "learning_rate": 0.00011541309339515869, "loss": 1.2202, "step": 833 }, { "epoch": 1.52, "grad_norm": 0.16731429410912582, "learning_rate": 0.00011520360872031209, "loss": 1.1114, "step": 834 }, { "epoch": 1.52, "grad_norm": 0.1747423744661194, "learning_rate": 0.00011499405572515059, "loss": 1.0929, "step": 835 }, { "epoch": 1.52, "grad_norm": 0.18705677818301408, "learning_rate": 0.00011478443535134049, "loss": 1.101, "step": 836 }, { "epoch": 1.52, "grad_norm": 0.18090585152937122, "learning_rate": 0.00011457474854085096, "loss": 1.2376, "step": 837 }, { "epoch": 1.53, "grad_norm": 0.1572625515956593, "learning_rate": 0.00011436499623594963, "loss": 1.1039, "step": 838 }, { "epoch": 1.53, "grad_norm": 0.18868561701551964, "learning_rate": 0.00011415517937919846, "loss": 1.1091, "step": 839 }, { "epoch": 1.53, "grad_norm": 0.3765201515989829, "learning_rate": 0.00011394529891344958, "loss": 1.1904, "step": 840 }, { "epoch": 1.53, "grad_norm": 0.17659930390592027, "learning_rate": 0.00011373535578184082, "loss": 1.1767, "step": 841 }, { "epoch": 1.53, "grad_norm": 0.16250915193456436, "learning_rate": 0.00011352535092779173, "loss": 1.0396, "step": 842 }, { "epoch": 1.54, "grad_norm": 0.17679862621844297, "learning_rate": 0.00011331528529499909, "loss": 1.0899, "step": 843 }, { "epoch": 1.54, "grad_norm": 0.17303502214887417, "learning_rate": 0.00011310515982743293, "loss": 1.1623, "step": 844 }, { "epoch": 1.54, "grad_norm": 0.17241019588300063, "learning_rate": 0.00011289497546933212, "loss": 1.1761, "step": 845 }, { "epoch": 1.54, "grad_norm": 0.19940689264510733, "learning_rate": 0.00011268473316520007, "loss": 1.186, "step": 846 }, { "epoch": 1.54, "grad_norm": 0.17205111928537736, "learning_rate": 0.00011247443385980078, "loss": 1.1669, "step": 847 }, { "epoch": 1.54, "grad_norm": 0.168610859884427, "learning_rate": 0.0001122640784981542, "loss": 1.1223, "step": 848 }, { "epoch": 1.55, "grad_norm": 0.1571679218925349, "learning_rate": 0.0001120536680255323, "loss": 1.1017, "step": 849 }, { "epoch": 1.55, "grad_norm": 0.1731635445936976, "learning_rate": 0.00011184320338745467, "loss": 1.1549, "step": 850 }, { "epoch": 1.55, "grad_norm": 0.16986390394426032, "learning_rate": 0.00011163268552968423, "loss": 1.1561, "step": 851 }, { "epoch": 1.55, "grad_norm": 0.17807860518694296, "learning_rate": 0.00011142211539822318, "loss": 1.1191, "step": 852 }, { "epoch": 1.55, "grad_norm": 0.1848822566355218, "learning_rate": 0.0001112114939393085, "loss": 1.1265, "step": 853 }, { "epoch": 1.56, "grad_norm": 0.16154933469362054, "learning_rate": 0.00011100082209940795, "loss": 1.1696, "step": 854 }, { "epoch": 1.56, "grad_norm": 0.17539593545795423, "learning_rate": 0.00011079010082521557, "loss": 1.2098, "step": 855 }, { "epoch": 1.56, "grad_norm": 0.16955097983529596, "learning_rate": 0.00011057933106364758, "loss": 1.134, "step": 856 }, { "epoch": 1.56, "grad_norm": 0.1627286025891699, "learning_rate": 0.00011036851376183812, "loss": 1.0968, "step": 857 }, { "epoch": 1.56, "grad_norm": 0.18241009216885773, "learning_rate": 0.0001101576498671349, "loss": 1.0784, "step": 858 }, { "epoch": 1.56, "grad_norm": 0.1602213810951013, "learning_rate": 0.00010994674032709513, "loss": 1.0272, "step": 859 }, { "epoch": 1.57, "grad_norm": 0.17710273355839293, "learning_rate": 0.00010973578608948094, "loss": 1.1807, "step": 860 }, { "epoch": 1.57, "grad_norm": 0.16846043025211901, "learning_rate": 0.00010952478810225548, "loss": 1.1237, "step": 861 }, { "epoch": 1.57, "grad_norm": 0.17578602240771765, "learning_rate": 0.00010931374731357841, "loss": 1.1305, "step": 862 }, { "epoch": 1.57, "grad_norm": 0.1741114211076835, "learning_rate": 0.0001091026646718018, "loss": 1.0663, "step": 863 }, { "epoch": 1.57, "grad_norm": 0.17419627851243186, "learning_rate": 0.0001088915411254657, "loss": 1.0864, "step": 864 }, { "epoch": 1.58, "grad_norm": 0.1941495171655647, "learning_rate": 0.00010868037762329404, "loss": 1.1471, "step": 865 }, { "epoch": 1.58, "grad_norm": 0.16828482820770493, "learning_rate": 0.0001084691751141903, "loss": 1.1386, "step": 866 }, { "epoch": 1.58, "grad_norm": 0.1812212801001379, "learning_rate": 0.00010825793454723325, "loss": 1.193, "step": 867 }, { "epoch": 1.58, "grad_norm": 0.31177412914918357, "learning_rate": 0.00010804665687167262, "loss": 1.1416, "step": 868 }, { "epoch": 1.58, "grad_norm": 0.16307908317649347, "learning_rate": 0.00010783534303692493, "loss": 1.0742, "step": 869 }, { "epoch": 1.58, "grad_norm": 0.17264040918528867, "learning_rate": 0.00010762399399256917, "loss": 1.1196, "step": 870 }, { "epoch": 1.59, "grad_norm": 0.16662882952717148, "learning_rate": 0.00010741261068834265, "loss": 1.1095, "step": 871 }, { "epoch": 1.59, "grad_norm": 0.16909860632274898, "learning_rate": 0.00010720119407413647, "loss": 1.0978, "step": 872 }, { "epoch": 1.59, "grad_norm": 0.1841227749300958, "learning_rate": 0.00010698974509999158, "loss": 1.1249, "step": 873 }, { "epoch": 1.59, "grad_norm": 0.163410587005958, "learning_rate": 0.00010677826471609422, "loss": 1.1027, "step": 874 }, { "epoch": 1.59, "grad_norm": 0.16473194468337188, "learning_rate": 0.00010656675387277182, "loss": 1.0452, "step": 875 }, { "epoch": 1.6, "grad_norm": 0.16465304101505895, "learning_rate": 0.00010635521352048872, "loss": 1.0975, "step": 876 }, { "epoch": 1.6, "grad_norm": 0.1648559678877552, "learning_rate": 0.00010614364460984176, "loss": 1.1865, "step": 877 }, { "epoch": 1.6, "grad_norm": 0.17182083260240222, "learning_rate": 0.00010593204809155628, "loss": 1.1666, "step": 878 }, { "epoch": 1.6, "grad_norm": 0.262043699845629, "learning_rate": 0.00010572042491648149, "loss": 1.1705, "step": 879 }, { "epoch": 1.6, "grad_norm": 0.1642146793634745, "learning_rate": 0.00010550877603558655, "loss": 1.0943, "step": 880 }, { "epoch": 1.6, "grad_norm": 0.1740460470258003, "learning_rate": 0.00010529710239995605, "loss": 1.0492, "step": 881 }, { "epoch": 1.61, "grad_norm": 0.16424504391420278, "learning_rate": 0.0001050854049607858, "loss": 1.0877, "step": 882 }, { "epoch": 1.61, "grad_norm": 0.1754464752434157, "learning_rate": 0.00010487368466937866, "loss": 1.2004, "step": 883 }, { "epoch": 1.61, "grad_norm": 0.17019599469065266, "learning_rate": 0.00010466194247714008, "loss": 1.1736, "step": 884 }, { "epoch": 1.61, "grad_norm": 0.16453531456339282, "learning_rate": 0.00010445017933557404, "loss": 1.0682, "step": 885 }, { "epoch": 1.61, "grad_norm": 0.16576841634828685, "learning_rate": 0.00010423839619627853, "loss": 1.1239, "step": 886 }, { "epoch": 1.62, "grad_norm": 0.17959933112877494, "learning_rate": 0.00010402659401094152, "loss": 1.1741, "step": 887 }, { "epoch": 1.62, "grad_norm": 0.1764446278189948, "learning_rate": 0.00010381477373133652, "loss": 1.1465, "step": 888 }, { "epoch": 1.62, "grad_norm": 0.1704634166756464, "learning_rate": 0.0001036029363093183, "loss": 1.2092, "step": 889 }, { "epoch": 1.62, "grad_norm": 0.16952721202785456, "learning_rate": 0.00010339108269681874, "loss": 1.1478, "step": 890 }, { "epoch": 1.62, "grad_norm": 0.1787903145977452, "learning_rate": 0.00010317921384584244, "loss": 1.1818, "step": 891 }, { "epoch": 1.62, "grad_norm": 0.1623533169508836, "learning_rate": 0.00010296733070846252, "loss": 1.1352, "step": 892 }, { "epoch": 1.63, "grad_norm": 0.17651784025914782, "learning_rate": 0.00010275543423681621, "loss": 1.175, "step": 893 }, { "epoch": 1.63, "grad_norm": 0.16473007072288753, "learning_rate": 0.00010254352538310075, "loss": 1.1837, "step": 894 }, { "epoch": 1.63, "grad_norm": 0.16909571695790346, "learning_rate": 0.00010233160509956894, "loss": 1.1435, "step": 895 }, { "epoch": 1.63, "grad_norm": 0.1904744004900901, "learning_rate": 0.000102119674338525, "loss": 1.1549, "step": 896 }, { "epoch": 1.63, "grad_norm": 0.17373949711988154, "learning_rate": 0.00010190773405232024, "loss": 1.202, "step": 897 }, { "epoch": 1.64, "grad_norm": 0.18192103377408994, "learning_rate": 0.00010169578519334873, "loss": 1.1851, "step": 898 }, { "epoch": 1.64, "grad_norm": 0.18229237608440976, "learning_rate": 0.0001014838287140431, "loss": 1.1233, "step": 899 }, { "epoch": 1.64, "grad_norm": 0.16940278596266165, "learning_rate": 0.00010127186556687019, "loss": 1.169, "step": 900 }, { "epoch": 1.64, "grad_norm": 0.18262078712459354, "learning_rate": 0.00010105989670432681, "loss": 1.199, "step": 901 }, { "epoch": 1.64, "grad_norm": 0.16747513400418446, "learning_rate": 0.00010084792307893552, "loss": 1.0666, "step": 902 }, { "epoch": 1.64, "grad_norm": 0.1746037115339502, "learning_rate": 0.00010063594564324012, "loss": 1.1885, "step": 903 }, { "epoch": 1.65, "grad_norm": 0.1732153359208303, "learning_rate": 0.00010042396534980176, "loss": 1.1155, "step": 904 }, { "epoch": 1.65, "grad_norm": 0.18911672921606115, "learning_rate": 0.00010021198315119424, "loss": 1.1495, "step": 905 }, { "epoch": 1.65, "grad_norm": 0.1810967486335703, "learning_rate": 0.0001, "loss": 1.2386, "step": 906 }, { "epoch": 1.65, "grad_norm": 0.16208600336701465, "learning_rate": 9.978801684880578e-05, "loss": 1.1016, "step": 907 }, { "epoch": 1.65, "grad_norm": 0.17108242412955507, "learning_rate": 9.957603465019826e-05, "loss": 1.1659, "step": 908 }, { "epoch": 1.66, "grad_norm": 0.17777524975299183, "learning_rate": 9.93640543567599e-05, "loss": 1.1877, "step": 909 }, { "epoch": 1.66, "grad_norm": 0.32271912031116656, "learning_rate": 9.91520769210645e-05, "loss": 1.1254, "step": 910 }, { "epoch": 1.66, "grad_norm": 0.16598806172835462, "learning_rate": 9.894010329567323e-05, "loss": 1.187, "step": 911 }, { "epoch": 1.66, "grad_norm": 0.17032736087667158, "learning_rate": 9.872813443312984e-05, "loss": 1.2073, "step": 912 }, { "epoch": 1.66, "grad_norm": 0.16171590357915658, "learning_rate": 9.851617128595694e-05, "loss": 1.0839, "step": 913 }, { "epoch": 1.66, "grad_norm": 0.22200097454760384, "learning_rate": 9.830421480665128e-05, "loss": 1.1299, "step": 914 }, { "epoch": 1.67, "grad_norm": 0.18159512381331983, "learning_rate": 9.809226594767978e-05, "loss": 1.1991, "step": 915 }, { "epoch": 1.67, "grad_norm": 0.1729262170321024, "learning_rate": 9.788032566147505e-05, "loss": 1.0962, "step": 916 }, { "epoch": 1.67, "grad_norm": 0.17432638214872634, "learning_rate": 9.766839490043108e-05, "loss": 1.1669, "step": 917 }, { "epoch": 1.67, "grad_norm": 0.16088749813770933, "learning_rate": 9.745647461689931e-05, "loss": 1.0548, "step": 918 }, { "epoch": 1.67, "grad_norm": 0.1741820101835776, "learning_rate": 9.724456576318381e-05, "loss": 1.1986, "step": 919 }, { "epoch": 1.68, "grad_norm": 0.1749247436598961, "learning_rate": 9.70326692915375e-05, "loss": 1.1474, "step": 920 }, { "epoch": 1.68, "grad_norm": 0.1637038092120215, "learning_rate": 9.682078615415754e-05, "loss": 1.1231, "step": 921 }, { "epoch": 1.68, "grad_norm": 0.18807058316399383, "learning_rate": 9.660891730318128e-05, "loss": 1.2045, "step": 922 }, { "epoch": 1.68, "grad_norm": 0.1748352594360658, "learning_rate": 9.639706369068171e-05, "loss": 1.1689, "step": 923 }, { "epoch": 1.68, "grad_norm": 0.1716288593635444, "learning_rate": 9.61852262686635e-05, "loss": 1.1572, "step": 924 }, { "epoch": 1.68, "grad_norm": 0.18271775085545605, "learning_rate": 9.597340598905852e-05, "loss": 1.21, "step": 925 }, { "epoch": 1.69, "grad_norm": 0.1717878271479738, "learning_rate": 9.576160380372149e-05, "loss": 1.0988, "step": 926 }, { "epoch": 1.69, "grad_norm": 0.1680182822271049, "learning_rate": 9.5549820664426e-05, "loss": 1.0968, "step": 927 }, { "epoch": 1.69, "grad_norm": 0.1742620341628087, "learning_rate": 9.533805752285993e-05, "loss": 1.1635, "step": 928 }, { "epoch": 1.69, "grad_norm": 0.16338762068272702, "learning_rate": 9.512631533062138e-05, "loss": 0.9968, "step": 929 }, { "epoch": 1.69, "grad_norm": 0.17102172627425277, "learning_rate": 9.491459503921421e-05, "loss": 1.1481, "step": 930 }, { "epoch": 1.7, "grad_norm": 0.166342175891061, "learning_rate": 9.470289760004398e-05, "loss": 1.0721, "step": 931 }, { "epoch": 1.7, "grad_norm": 0.2215993724024064, "learning_rate": 9.449122396441345e-05, "loss": 1.1877, "step": 932 }, { "epoch": 1.7, "grad_norm": 0.1659659894828161, "learning_rate": 9.427957508351852e-05, "loss": 1.1691, "step": 933 }, { "epoch": 1.7, "grad_norm": 0.1656871284817402, "learning_rate": 9.406795190844376e-05, "loss": 1.06, "step": 934 }, { "epoch": 1.7, "grad_norm": 0.1586845482565193, "learning_rate": 9.385635539015825e-05, "loss": 1.1, "step": 935 }, { "epoch": 1.7, "grad_norm": 0.16465420669616035, "learning_rate": 9.364478647951133e-05, "loss": 1.1403, "step": 936 }, { "epoch": 1.71, "grad_norm": 0.15668641868741373, "learning_rate": 9.343324612722819e-05, "loss": 1.1075, "step": 937 }, { "epoch": 1.71, "grad_norm": 0.16611169088286057, "learning_rate": 9.32217352839058e-05, "loss": 1.1256, "step": 938 }, { "epoch": 1.71, "grad_norm": 0.17301783500328655, "learning_rate": 9.301025490000841e-05, "loss": 1.1797, "step": 939 }, { "epoch": 1.71, "grad_norm": 0.17015053210523748, "learning_rate": 9.279880592586354e-05, "loss": 1.1087, "step": 940 }, { "epoch": 1.71, "grad_norm": 0.17913490315482883, "learning_rate": 9.25873893116574e-05, "loss": 1.1673, "step": 941 }, { "epoch": 1.72, "grad_norm": 0.18716341523653843, "learning_rate": 9.237600600743085e-05, "loss": 1.1448, "step": 942 }, { "epoch": 1.72, "grad_norm": 0.16020067287870843, "learning_rate": 9.216465696307512e-05, "loss": 1.1152, "step": 943 }, { "epoch": 1.72, "grad_norm": 0.18082661292174337, "learning_rate": 9.195334312832742e-05, "loss": 1.1923, "step": 944 }, { "epoch": 1.72, "grad_norm": 0.16932315705059348, "learning_rate": 9.174206545276677e-05, "loss": 1.2025, "step": 945 }, { "epoch": 1.72, "grad_norm": 0.1773723182281052, "learning_rate": 9.153082488580968e-05, "loss": 1.1583, "step": 946 }, { "epoch": 1.72, "grad_norm": 0.1650084811263765, "learning_rate": 9.131962237670598e-05, "loss": 1.1456, "step": 947 }, { "epoch": 1.73, "grad_norm": 0.17296637806848497, "learning_rate": 9.11084588745343e-05, "loss": 1.1188, "step": 948 }, { "epoch": 1.73, "grad_norm": 0.18441757049220042, "learning_rate": 9.089733532819824e-05, "loss": 1.1057, "step": 949 }, { "epoch": 1.73, "grad_norm": 0.16479278966433117, "learning_rate": 9.068625268642161e-05, "loss": 0.9686, "step": 950 }, { "epoch": 1.73, "grad_norm": 0.1721522611977686, "learning_rate": 9.047521189774455e-05, "loss": 1.1593, "step": 951 }, { "epoch": 1.73, "grad_norm": 0.17693809369137542, "learning_rate": 9.026421391051907e-05, "loss": 1.141, "step": 952 }, { "epoch": 1.74, "grad_norm": 0.16748786371543328, "learning_rate": 9.005325967290488e-05, "loss": 1.0954, "step": 953 }, { "epoch": 1.74, "grad_norm": 0.16738864271495277, "learning_rate": 8.984235013286511e-05, "loss": 1.1954, "step": 954 }, { "epoch": 1.74, "grad_norm": 0.17608800377093817, "learning_rate": 8.963148623816191e-05, "loss": 1.2054, "step": 955 }, { "epoch": 1.74, "grad_norm": 0.1572056318124515, "learning_rate": 8.942066893635246e-05, "loss": 1.062, "step": 956 }, { "epoch": 1.74, "grad_norm": 0.17321331852374994, "learning_rate": 8.920989917478447e-05, "loss": 1.1774, "step": 957 }, { "epoch": 1.74, "grad_norm": 0.24022668199221645, "learning_rate": 8.899917790059208e-05, "loss": 1.1422, "step": 958 }, { "epoch": 1.75, "grad_norm": 0.17078130002322278, "learning_rate": 8.878850606069153e-05, "loss": 1.102, "step": 959 }, { "epoch": 1.75, "grad_norm": 0.18816676453808523, "learning_rate": 8.857788460177686e-05, "loss": 1.1228, "step": 960 }, { "epoch": 1.75, "grad_norm": 0.17061330738089536, "learning_rate": 8.836731447031581e-05, "loss": 1.1664, "step": 961 }, { "epoch": 1.75, "grad_norm": 0.15744686305416977, "learning_rate": 8.815679661254537e-05, "loss": 1.0526, "step": 962 }, { "epoch": 1.75, "grad_norm": 0.17140964720406338, "learning_rate": 8.79463319744677e-05, "loss": 1.1015, "step": 963 }, { "epoch": 1.76, "grad_norm": 0.177839632569439, "learning_rate": 8.77359215018458e-05, "loss": 1.179, "step": 964 }, { "epoch": 1.76, "grad_norm": 0.18976111884283367, "learning_rate": 8.752556614019923e-05, "loss": 1.1901, "step": 965 }, { "epoch": 1.76, "grad_norm": 0.17757617341847579, "learning_rate": 8.731526683479992e-05, "loss": 1.1195, "step": 966 }, { "epoch": 1.76, "grad_norm": 0.18025608157061349, "learning_rate": 8.710502453066792e-05, "loss": 1.1681, "step": 967 }, { "epoch": 1.76, "grad_norm": 0.17991548690786335, "learning_rate": 8.689484017256711e-05, "loss": 1.0382, "step": 968 }, { "epoch": 1.77, "grad_norm": 0.18377930665665893, "learning_rate": 8.668471470500095e-05, "loss": 1.2159, "step": 969 }, { "epoch": 1.77, "grad_norm": 0.16785777204268718, "learning_rate": 8.647464907220832e-05, "loss": 1.1001, "step": 970 }, { "epoch": 1.77, "grad_norm": 0.1718542715910871, "learning_rate": 8.626464421815919e-05, "loss": 1.1362, "step": 971 }, { "epoch": 1.77, "grad_norm": 0.17567452985919224, "learning_rate": 8.605470108655045e-05, "loss": 1.0701, "step": 972 }, { "epoch": 1.77, "grad_norm": 0.16645014392512258, "learning_rate": 8.584482062080154e-05, "loss": 1.1224, "step": 973 }, { "epoch": 1.77, "grad_norm": 0.1683034297078507, "learning_rate": 8.563500376405041e-05, "loss": 1.1833, "step": 974 }, { "epoch": 1.78, "grad_norm": 0.17027006999646288, "learning_rate": 8.542525145914905e-05, "loss": 1.2105, "step": 975 }, { "epoch": 1.78, "grad_norm": 0.16661484142693264, "learning_rate": 8.521556464865954e-05, "loss": 1.1345, "step": 976 }, { "epoch": 1.78, "grad_norm": 0.16535108955142738, "learning_rate": 8.500594427484946e-05, "loss": 1.0801, "step": 977 }, { "epoch": 1.78, "grad_norm": 0.1619767851731314, "learning_rate": 8.479639127968792e-05, "loss": 1.1398, "step": 978 }, { "epoch": 1.78, "grad_norm": 0.16148830590153632, "learning_rate": 8.458690660484134e-05, "loss": 1.1714, "step": 979 }, { "epoch": 1.79, "grad_norm": 0.1718526387843433, "learning_rate": 8.4377491191669e-05, "loss": 1.1196, "step": 980 }, { "epoch": 1.79, "grad_norm": 0.1589791532312397, "learning_rate": 8.4168145981219e-05, "loss": 1.0975, "step": 981 }, { "epoch": 1.79, "grad_norm": 0.18228960673193118, "learning_rate": 8.395887191422397e-05, "loss": 1.0911, "step": 982 }, { "epoch": 1.79, "grad_norm": 0.17051635436251017, "learning_rate": 8.374966993109687e-05, "loss": 1.1156, "step": 983 }, { "epoch": 1.79, "grad_norm": 0.1734092864786433, "learning_rate": 8.35405409719266e-05, "loss": 1.1387, "step": 984 }, { "epoch": 1.79, "grad_norm": 0.17960160313198717, "learning_rate": 8.333148597647414e-05, "loss": 1.1255, "step": 985 }, { "epoch": 1.8, "grad_norm": 0.1667460148360989, "learning_rate": 8.312250588416791e-05, "loss": 1.0877, "step": 986 }, { "epoch": 1.8, "grad_norm": 0.17227304205526361, "learning_rate": 8.291360163409978e-05, "loss": 1.1643, "step": 987 }, { "epoch": 1.8, "grad_norm": 0.16172787345643957, "learning_rate": 8.270477416502091e-05, "loss": 1.0813, "step": 988 }, { "epoch": 1.8, "grad_norm": 0.17717380607042416, "learning_rate": 8.249602441533726e-05, "loss": 1.1653, "step": 989 }, { "epoch": 1.8, "grad_norm": 0.17646325594155327, "learning_rate": 8.228735332310575e-05, "loss": 1.1671, "step": 990 }, { "epoch": 1.81, "grad_norm": 0.16889550666185796, "learning_rate": 8.207876182602958e-05, "loss": 1.1068, "step": 991 }, { "epoch": 1.81, "grad_norm": 0.1668114849825541, "learning_rate": 8.187025086145458e-05, "loss": 1.1467, "step": 992 }, { "epoch": 1.81, "grad_norm": 0.1721245962364379, "learning_rate": 8.16618213663644e-05, "loss": 1.1377, "step": 993 }, { "epoch": 1.81, "grad_norm": 0.1929335873892253, "learning_rate": 8.145347427737678e-05, "loss": 1.119, "step": 994 }, { "epoch": 1.81, "grad_norm": 0.16623130316399012, "learning_rate": 8.12452105307391e-05, "loss": 1.0687, "step": 995 }, { "epoch": 1.81, "grad_norm": 0.18085333131496864, "learning_rate": 8.103703106232416e-05, "loss": 1.1517, "step": 996 }, { "epoch": 1.82, "grad_norm": 0.17906886829733978, "learning_rate": 8.082893680762619e-05, "loss": 1.1325, "step": 997 }, { "epoch": 1.82, "grad_norm": 0.16359826556853824, "learning_rate": 8.062092870175628e-05, "loss": 1.0353, "step": 998 }, { "epoch": 1.82, "grad_norm": 0.1626009671933035, "learning_rate": 8.041300767943867e-05, "loss": 1.1598, "step": 999 }, { "epoch": 1.82, "grad_norm": 0.16643269286545453, "learning_rate": 8.0205174675006e-05, "loss": 1.1226, "step": 1000 }, { "epoch": 1.82, "grad_norm": 0.1699654548245084, "learning_rate": 7.999743062239557e-05, "loss": 1.0999, "step": 1001 }, { "epoch": 1.83, "grad_norm": 0.1827266429905637, "learning_rate": 7.978977645514487e-05, "loss": 1.2791, "step": 1002 }, { "epoch": 1.83, "grad_norm": 0.44434388351574605, "learning_rate": 7.958221310638749e-05, "loss": 1.1693, "step": 1003 }, { "epoch": 1.83, "grad_norm": 0.17634755268047692, "learning_rate": 7.937474150884897e-05, "loss": 1.0845, "step": 1004 }, { "epoch": 1.83, "grad_norm": 0.177726814790075, "learning_rate": 7.916736259484239e-05, "loss": 1.1849, "step": 1005 }, { "epoch": 1.83, "grad_norm": 0.16479013332848777, "learning_rate": 7.896007729626457e-05, "loss": 1.1296, "step": 1006 }, { "epoch": 1.83, "grad_norm": 0.1607507116543152, "learning_rate": 7.875288654459144e-05, "loss": 1.0829, "step": 1007 }, { "epoch": 1.84, "grad_norm": 0.16387244898588724, "learning_rate": 7.854579127087417e-05, "loss": 1.1604, "step": 1008 }, { "epoch": 1.84, "grad_norm": 0.163742117997135, "learning_rate": 7.833879240573487e-05, "loss": 1.0757, "step": 1009 }, { "epoch": 1.84, "grad_norm": 0.16638334168591146, "learning_rate": 7.813189087936243e-05, "loss": 1.1637, "step": 1010 }, { "epoch": 1.84, "grad_norm": 0.16818864513814694, "learning_rate": 7.792508762150833e-05, "loss": 1.1564, "step": 1011 }, { "epoch": 1.84, "grad_norm": 0.17901829622442805, "learning_rate": 7.771838356148232e-05, "loss": 1.1841, "step": 1012 }, { "epoch": 1.85, "grad_norm": 0.16614583867770816, "learning_rate": 7.751177962814866e-05, "loss": 1.1255, "step": 1013 }, { "epoch": 1.85, "grad_norm": 0.19013974114118293, "learning_rate": 7.730527674992143e-05, "loss": 1.1488, "step": 1014 }, { "epoch": 1.85, "grad_norm": 0.17442438513245137, "learning_rate": 7.709887585476075e-05, "loss": 1.1066, "step": 1015 }, { "epoch": 1.85, "grad_norm": 0.17701832779209928, "learning_rate": 7.689257787016834e-05, "loss": 1.1755, "step": 1016 }, { "epoch": 1.85, "grad_norm": 0.16936912919421254, "learning_rate": 7.668638372318359e-05, "loss": 1.1057, "step": 1017 }, { "epoch": 1.85, "grad_norm": 0.19112857435989614, "learning_rate": 7.648029434037915e-05, "loss": 1.1412, "step": 1018 }, { "epoch": 1.86, "grad_norm": 0.16103962948152212, "learning_rate": 7.627431064785706e-05, "loss": 1.0929, "step": 1019 }, { "epoch": 1.86, "grad_norm": 0.16677548612612422, "learning_rate": 7.606843357124426e-05, "loss": 1.2046, "step": 1020 }, { "epoch": 1.86, "grad_norm": 0.16233738083605964, "learning_rate": 7.58626640356886e-05, "loss": 1.0678, "step": 1021 }, { "epoch": 1.86, "grad_norm": 0.16602043351095722, "learning_rate": 7.565700296585483e-05, "loss": 1.1216, "step": 1022 }, { "epoch": 1.86, "grad_norm": 0.16627151579102536, "learning_rate": 7.54514512859201e-05, "loss": 1.0911, "step": 1023 }, { "epoch": 1.87, "grad_norm": 0.20281510390834787, "learning_rate": 7.524600991957012e-05, "loss": 1.1722, "step": 1024 }, { "epoch": 1.87, "grad_norm": 0.16069933741616732, "learning_rate": 7.504067978999484e-05, "loss": 1.0839, "step": 1025 }, { "epoch": 1.87, "grad_norm": 0.18107514785611215, "learning_rate": 7.483546181988436e-05, "loss": 1.0717, "step": 1026 }, { "epoch": 1.87, "grad_norm": 0.1721242495617235, "learning_rate": 7.463035693142473e-05, "loss": 1.1599, "step": 1027 }, { "epoch": 1.87, "grad_norm": 0.17918262771658913, "learning_rate": 7.442536604629395e-05, "loss": 1.089, "step": 1028 }, { "epoch": 1.87, "grad_norm": 0.17784491718923554, "learning_rate": 7.422049008565757e-05, "loss": 1.176, "step": 1029 }, { "epoch": 1.88, "grad_norm": 0.15630505382249996, "learning_rate": 7.401572997016479e-05, "loss": 1.039, "step": 1030 }, { "epoch": 1.88, "grad_norm": 0.17070239450442767, "learning_rate": 7.381108661994429e-05, "loss": 1.1059, "step": 1031 }, { "epoch": 1.88, "grad_norm": 0.16222624124278331, "learning_rate": 7.360656095459995e-05, "loss": 1.024, "step": 1032 }, { "epoch": 1.88, "grad_norm": 0.17125506848213806, "learning_rate": 7.340215389320687e-05, "loss": 1.1567, "step": 1033 }, { "epoch": 1.88, "grad_norm": 0.17959953833242476, "learning_rate": 7.31978663543071e-05, "loss": 1.141, "step": 1034 }, { "epoch": 1.89, "grad_norm": 0.17010870239393844, "learning_rate": 7.299369925590574e-05, "loss": 1.1046, "step": 1035 }, { "epoch": 1.89, "grad_norm": 0.15998424148316903, "learning_rate": 7.278965351546648e-05, "loss": 1.0627, "step": 1036 }, { "epoch": 1.89, "grad_norm": 0.1710612967988063, "learning_rate": 7.258573004990788e-05, "loss": 1.1486, "step": 1037 }, { "epoch": 1.89, "grad_norm": 0.1607397255850716, "learning_rate": 7.238192977559884e-05, "loss": 1.0751, "step": 1038 }, { "epoch": 1.89, "grad_norm": 0.17530360487610264, "learning_rate": 7.217825360835473e-05, "loss": 1.1281, "step": 1039 }, { "epoch": 1.89, "grad_norm": 0.17374375355320706, "learning_rate": 7.197470246343333e-05, "loss": 1.1493, "step": 1040 }, { "epoch": 1.9, "grad_norm": 0.15987413285182245, "learning_rate": 7.177127725553045e-05, "loss": 1.1446, "step": 1041 }, { "epoch": 1.9, "grad_norm": 0.1789355664165954, "learning_rate": 7.156797889877613e-05, "loss": 1.2304, "step": 1042 }, { "epoch": 1.9, "grad_norm": 0.17282586483463735, "learning_rate": 7.136480830673019e-05, "loss": 1.1909, "step": 1043 }, { "epoch": 1.9, "grad_norm": 0.159213004708279, "learning_rate": 7.116176639237852e-05, "loss": 1.0624, "step": 1044 }, { "epoch": 1.9, "grad_norm": 0.16197929506166975, "learning_rate": 7.095885406812866e-05, "loss": 1.1283, "step": 1045 }, { "epoch": 1.91, "grad_norm": 0.16495179889101508, "learning_rate": 7.075607224580581e-05, "loss": 1.1378, "step": 1046 }, { "epoch": 1.91, "grad_norm": 0.1662731898113741, "learning_rate": 7.05534218366488e-05, "loss": 1.0542, "step": 1047 }, { "epoch": 1.91, "grad_norm": 0.1650210560996418, "learning_rate": 7.035090375130581e-05, "loss": 1.1326, "step": 1048 }, { "epoch": 1.91, "grad_norm": 0.1708049460231556, "learning_rate": 7.014851889983057e-05, "loss": 1.137, "step": 1049 }, { "epoch": 1.91, "grad_norm": 0.17558230008855352, "learning_rate": 6.994626819167789e-05, "loss": 1.1359, "step": 1050 }, { "epoch": 1.91, "grad_norm": 0.16645826974884173, "learning_rate": 6.974415253570003e-05, "loss": 1.1742, "step": 1051 }, { "epoch": 1.92, "grad_norm": 0.17184188151368943, "learning_rate": 6.954217284014211e-05, "loss": 1.1953, "step": 1052 }, { "epoch": 1.92, "grad_norm": 0.17068592296841717, "learning_rate": 6.934033001263847e-05, "loss": 1.1301, "step": 1053 }, { "epoch": 1.92, "grad_norm": 0.16302700472895656, "learning_rate": 6.913862496020831e-05, "loss": 1.1232, "step": 1054 }, { "epoch": 1.92, "grad_norm": 0.17057720351205383, "learning_rate": 6.893705858925178e-05, "loss": 1.1247, "step": 1055 }, { "epoch": 1.92, "grad_norm": 0.1582652133644645, "learning_rate": 6.873563180554583e-05, "loss": 1.1203, "step": 1056 }, { "epoch": 1.93, "grad_norm": 0.16895339487465857, "learning_rate": 6.853434551424e-05, "loss": 1.1846, "step": 1057 }, { "epoch": 1.93, "grad_norm": 0.1726993947504367, "learning_rate": 6.833320061985277e-05, "loss": 1.1963, "step": 1058 }, { "epoch": 1.93, "grad_norm": 0.16055788377798388, "learning_rate": 6.813219802626698e-05, "loss": 1.1439, "step": 1059 }, { "epoch": 1.93, "grad_norm": 0.17265920805658908, "learning_rate": 6.793133863672616e-05, "loss": 1.192, "step": 1060 }, { "epoch": 1.93, "grad_norm": 0.1833601588530699, "learning_rate": 6.773062335383024e-05, "loss": 1.1128, "step": 1061 }, { "epoch": 1.93, "grad_norm": 0.1717674054389487, "learning_rate": 6.753005307953167e-05, "loss": 1.134, "step": 1062 }, { "epoch": 1.94, "grad_norm": 0.18207071739270747, "learning_rate": 6.73296287151312e-05, "loss": 1.1723, "step": 1063 }, { "epoch": 1.94, "grad_norm": 0.1624856029472637, "learning_rate": 6.712935116127389e-05, "loss": 1.1144, "step": 1064 }, { "epoch": 1.94, "grad_norm": 0.19915390833072513, "learning_rate": 6.692922131794517e-05, "loss": 1.1072, "step": 1065 }, { "epoch": 1.94, "grad_norm": 0.1722567762020263, "learning_rate": 6.672924008446662e-05, "loss": 1.1425, "step": 1066 }, { "epoch": 1.94, "grad_norm": 0.15963511981578907, "learning_rate": 6.652940835949208e-05, "loss": 1.1211, "step": 1067 }, { "epoch": 1.95, "grad_norm": 0.16578458195408932, "learning_rate": 6.632972704100349e-05, "loss": 1.1244, "step": 1068 }, { "epoch": 1.95, "grad_norm": 0.17110384656595054, "learning_rate": 6.613019702630694e-05, "loss": 1.1795, "step": 1069 }, { "epoch": 1.95, "grad_norm": 0.16990207952122507, "learning_rate": 6.593081921202859e-05, "loss": 1.0817, "step": 1070 }, { "epoch": 1.95, "grad_norm": 0.16467375370533938, "learning_rate": 6.57315944941107e-05, "loss": 1.0977, "step": 1071 }, { "epoch": 1.95, "grad_norm": 0.17152460282733728, "learning_rate": 6.553252376780748e-05, "loss": 1.1868, "step": 1072 }, { "epoch": 1.95, "grad_norm": 0.18132610186641904, "learning_rate": 6.533360792768122e-05, "loss": 1.2005, "step": 1073 }, { "epoch": 1.96, "grad_norm": 0.16698085910075294, "learning_rate": 6.513484786759818e-05, "loss": 1.1362, "step": 1074 }, { "epoch": 1.96, "grad_norm": 0.16504199012773973, "learning_rate": 6.493624448072457e-05, "loss": 1.1303, "step": 1075 }, { "epoch": 1.96, "grad_norm": 0.1779596734054924, "learning_rate": 6.473779865952263e-05, "loss": 1.1894, "step": 1076 }, { "epoch": 1.96, "grad_norm": 0.16364715074236819, "learning_rate": 6.453951129574644e-05, "loss": 1.1443, "step": 1077 }, { "epoch": 1.96, "grad_norm": 0.16251550362999542, "learning_rate": 6.434138328043815e-05, "loss": 1.1238, "step": 1078 }, { "epoch": 1.97, "grad_norm": 0.1666990603079008, "learning_rate": 6.414341550392368e-05, "loss": 1.0909, "step": 1079 }, { "epoch": 1.97, "grad_norm": 0.15779118156939148, "learning_rate": 6.39456088558091e-05, "loss": 1.1109, "step": 1080 }, { "epoch": 1.97, "grad_norm": 0.1677767995314692, "learning_rate": 6.374796422497621e-05, "loss": 1.2103, "step": 1081 }, { "epoch": 1.97, "grad_norm": 0.17165919802575422, "learning_rate": 6.355048249957886e-05, "loss": 1.1879, "step": 1082 }, { "epoch": 1.97, "grad_norm": 0.1918026223221136, "learning_rate": 6.33531645670389e-05, "loss": 1.1733, "step": 1083 }, { "epoch": 1.97, "grad_norm": 0.1787527260843118, "learning_rate": 6.3156011314042e-05, "loss": 1.1338, "step": 1084 }, { "epoch": 1.98, "grad_norm": 0.1694260433762424, "learning_rate": 6.2959023626534e-05, "loss": 1.1537, "step": 1085 }, { "epoch": 1.98, "grad_norm": 0.18311039386845188, "learning_rate": 6.276220238971652e-05, "loss": 1.1976, "step": 1086 }, { "epoch": 1.98, "grad_norm": 0.17113962947413622, "learning_rate": 6.256554848804343e-05, "loss": 1.0813, "step": 1087 }, { "epoch": 1.98, "grad_norm": 0.1664054538940142, "learning_rate": 6.236906280521646e-05, "loss": 1.1514, "step": 1088 }, { "epoch": 1.98, "grad_norm": 0.16227842164425618, "learning_rate": 6.217274622418153e-05, "loss": 1.0597, "step": 1089 }, { "epoch": 1.99, "grad_norm": 0.17755371081391708, "learning_rate": 6.197659962712461e-05, "loss": 1.19, "step": 1090 }, { "epoch": 1.99, "grad_norm": 0.15844523213403552, "learning_rate": 6.178062389546784e-05, "loss": 1.0587, "step": 1091 }, { "epoch": 1.99, "grad_norm": 0.16903107223771263, "learning_rate": 6.158481990986557e-05, "loss": 1.1339, "step": 1092 }, { "epoch": 1.99, "grad_norm": 0.17002303254909312, "learning_rate": 6.138918855020028e-05, "loss": 1.1158, "step": 1093 }, { "epoch": 1.99, "grad_norm": 0.16613843815618992, "learning_rate": 6.11937306955789e-05, "loss": 1.1894, "step": 1094 }, { "epoch": 1.99, "grad_norm": 0.17111783959067672, "learning_rate": 6.099844722432843e-05, "loss": 1.1974, "step": 1095 }, { "epoch": 2.0, "grad_norm": 0.1670020953276936, "learning_rate": 6.080333901399251e-05, "loss": 1.1018, "step": 1096 }, { "epoch": 2.0, "grad_norm": 0.18080514034863332, "learning_rate": 6.060840694132701e-05, "loss": 1.1845, "step": 1097 }, { "epoch": 2.0, "grad_norm": 0.18525347317947505, "learning_rate": 6.0413651882296406e-05, "loss": 1.0796, "step": 1098 }, { "epoch": 2.0, "grad_norm": 0.20400403778138837, "learning_rate": 6.021907471206971e-05, "loss": 1.0117, "step": 1099 }, { "epoch": 2.0, "grad_norm": 0.1641784925803925, "learning_rate": 6.002467630501646e-05, "loss": 1.097, "step": 1100 } ], "logging_steps": 1.0, "max_steps": 1647, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 50, "total_flos": 1.0209608098381824e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }