{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9858705560619874, "eval_steps": 137, "global_step": 1096, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 0.07610916346311569, "learning_rate": 2e-05, "loss": 1.795, "step": 1 }, { "epoch": 0.0, "eval_loss": 1.8087825775146484, "eval_runtime": 75.9539, "eval_samples_per_second": 65.829, "eval_steps_per_second": 16.457, "step": 1 }, { "epoch": 0.0, "grad_norm": 0.0771929994225502, "learning_rate": 4e-05, "loss": 1.7825, "step": 2 }, { "epoch": 0.01, "grad_norm": 0.08941341191530228, "learning_rate": 6e-05, "loss": 1.7737, "step": 3 }, { "epoch": 0.01, "grad_norm": 0.08335491269826889, "learning_rate": 8e-05, "loss": 1.8004, "step": 4 }, { "epoch": 0.01, "grad_norm": 0.08835520595312119, "learning_rate": 0.0001, "loss": 1.8495, "step": 5 }, { "epoch": 0.01, "grad_norm": 0.08816578984260559, "learning_rate": 0.00012, "loss": 1.7758, "step": 6 }, { "epoch": 0.01, "grad_norm": 0.09536299854516983, "learning_rate": 0.00014, "loss": 1.8001, "step": 7 }, { "epoch": 0.01, "grad_norm": 0.07634323835372925, "learning_rate": 0.00016, "loss": 1.7022, "step": 8 }, { "epoch": 0.02, "grad_norm": 0.06886536628007889, "learning_rate": 0.00018, "loss": 1.8428, "step": 9 }, { "epoch": 0.02, "grad_norm": 0.07389801740646362, "learning_rate": 0.0002, "loss": 1.7598, "step": 10 }, { "epoch": 0.02, "grad_norm": 0.06829163432121277, "learning_rate": 0.00019999981517295864, "loss": 1.7479, "step": 11 }, { "epoch": 0.02, "grad_norm": 0.060045819729566574, "learning_rate": 0.0001999992606925178, "loss": 1.7454, "step": 12 }, { "epoch": 0.02, "grad_norm": 0.08187604695558548, "learning_rate": 0.0001999983365607271, "loss": 1.7679, "step": 13 }, { "epoch": 0.03, "grad_norm": 0.05995490401983261, "learning_rate": 0.00019999704278100263, "loss": 1.7599, "step": 14 }, { "epoch": 0.03, "grad_norm": 0.055336710065603256, "learning_rate": 0.00019999537935812698, "loss": 1.8244, "step": 15 }, { "epoch": 0.03, "grad_norm": 0.0541992112994194, "learning_rate": 0.00019999334629824895, "loss": 1.7756, "step": 16 }, { "epoch": 0.03, "grad_norm": 0.05088195204734802, "learning_rate": 0.00019999094360888392, "loss": 1.7352, "step": 17 }, { "epoch": 0.03, "grad_norm": 0.05157861113548279, "learning_rate": 0.00019998817129891346, "loss": 1.7634, "step": 18 }, { "epoch": 0.03, "grad_norm": 0.055710840970277786, "learning_rate": 0.00019998502937858557, "loss": 1.7802, "step": 19 }, { "epoch": 0.04, "grad_norm": 0.055150121450424194, "learning_rate": 0.00019998151785951448, "loss": 1.7445, "step": 20 }, { "epoch": 0.04, "grad_norm": 0.0526655912399292, "learning_rate": 0.0001999776367546806, "loss": 1.6634, "step": 21 }, { "epoch": 0.04, "grad_norm": 0.04809674620628357, "learning_rate": 0.00019997338607843075, "loss": 1.7277, "step": 22 }, { "epoch": 0.04, "grad_norm": 0.049412671476602554, "learning_rate": 0.00019996876584647754, "loss": 1.7357, "step": 23 }, { "epoch": 0.04, "grad_norm": 0.04948608949780464, "learning_rate": 0.00019996377607589997, "loss": 1.7323, "step": 24 }, { "epoch": 0.05, "grad_norm": 0.050225820392370224, "learning_rate": 0.00019995841678514294, "loss": 1.7273, "step": 25 }, { "epoch": 0.05, "grad_norm": 0.05085042864084244, "learning_rate": 0.00019995268799401718, "loss": 1.7564, "step": 26 }, { "epoch": 0.05, "grad_norm": 0.04916631057858467, "learning_rate": 0.00019994658972369948, "loss": 1.7439, "step": 27 }, { "epoch": 0.05, "grad_norm": 0.04791415110230446, "learning_rate": 0.00019994012199673234, "loss": 1.6813, "step": 28 }, { "epoch": 0.05, "grad_norm": 0.04975065216422081, "learning_rate": 0.00019993328483702393, "loss": 1.691, "step": 29 }, { "epoch": 0.05, "grad_norm": 0.055913638323545456, "learning_rate": 0.00019992607826984816, "loss": 1.7242, "step": 30 }, { "epoch": 0.06, "grad_norm": 0.045829374343156815, "learning_rate": 0.00019991850232184435, "loss": 1.7334, "step": 31 }, { "epoch": 0.06, "grad_norm": 0.053105831146240234, "learning_rate": 0.00019991055702101734, "loss": 1.7214, "step": 32 }, { "epoch": 0.06, "grad_norm": 0.04539350047707558, "learning_rate": 0.00019990224239673722, "loss": 1.7698, "step": 33 }, { "epoch": 0.06, "grad_norm": 0.046983517706394196, "learning_rate": 0.00019989355847973932, "loss": 1.6887, "step": 34 }, { "epoch": 0.06, "grad_norm": 0.0471692830324173, "learning_rate": 0.00019988450530212414, "loss": 1.7571, "step": 35 }, { "epoch": 0.07, "grad_norm": 0.046874694526195526, "learning_rate": 0.00019987508289735716, "loss": 1.7558, "step": 36 }, { "epoch": 0.07, "grad_norm": 0.04474163055419922, "learning_rate": 0.00019986529130026857, "loss": 1.7465, "step": 37 }, { "epoch": 0.07, "grad_norm": 0.044651810079813004, "learning_rate": 0.00019985513054705348, "loss": 1.6983, "step": 38 }, { "epoch": 0.07, "grad_norm": 0.04951983690261841, "learning_rate": 0.00019984460067527153, "loss": 1.761, "step": 39 }, { "epoch": 0.07, "grad_norm": 0.04424133151769638, "learning_rate": 0.00019983370172384682, "loss": 1.6383, "step": 40 }, { "epoch": 0.07, "grad_norm": 0.052418872714042664, "learning_rate": 0.00019982243373306772, "loss": 1.779, "step": 41 }, { "epoch": 0.08, "grad_norm": 0.04530750587582588, "learning_rate": 0.0001998107967445869, "loss": 1.6942, "step": 42 }, { "epoch": 0.08, "grad_norm": 0.04790988191962242, "learning_rate": 0.0001997987908014209, "loss": 1.7053, "step": 43 }, { "epoch": 0.08, "grad_norm": 0.04889607056975365, "learning_rate": 0.0001997864159479502, "loss": 1.7275, "step": 44 }, { "epoch": 0.08, "grad_norm": 0.04314807429909706, "learning_rate": 0.00019977367222991893, "loss": 1.7393, "step": 45 }, { "epoch": 0.08, "grad_norm": 0.04405505582690239, "learning_rate": 0.00019976055969443479, "loss": 1.7306, "step": 46 }, { "epoch": 0.09, "grad_norm": 0.04656574875116348, "learning_rate": 0.00019974707838996882, "loss": 1.7686, "step": 47 }, { "epoch": 0.09, "grad_norm": 0.04246290400624275, "learning_rate": 0.00019973322836635518, "loss": 1.7209, "step": 48 }, { "epoch": 0.09, "grad_norm": 0.05493748560547829, "learning_rate": 0.00019971900967479106, "loss": 1.7155, "step": 49 }, { "epoch": 0.09, "grad_norm": 0.0450466088950634, "learning_rate": 0.0001997044223678364, "loss": 1.6604, "step": 50 }, { "epoch": 0.09, "grad_norm": 0.08634985238313675, "learning_rate": 0.00019968946649941382, "loss": 1.7321, "step": 51 }, { "epoch": 0.09, "grad_norm": 0.04310084879398346, "learning_rate": 0.00019967414212480831, "loss": 1.7281, "step": 52 }, { "epoch": 0.1, "grad_norm": 0.04666193947196007, "learning_rate": 0.000199658449300667, "loss": 1.6787, "step": 53 }, { "epoch": 0.1, "grad_norm": 0.04957772046327591, "learning_rate": 0.00019964238808499907, "loss": 1.6919, "step": 54 }, { "epoch": 0.1, "grad_norm": 0.0421697273850441, "learning_rate": 0.00019962595853717548, "loss": 1.7245, "step": 55 }, { "epoch": 0.1, "grad_norm": 0.04654068127274513, "learning_rate": 0.0001996091607179287, "loss": 1.7123, "step": 56 }, { "epoch": 0.1, "grad_norm": 0.04076274484395981, "learning_rate": 0.00019959199468935258, "loss": 1.7066, "step": 57 }, { "epoch": 0.11, "grad_norm": 0.04215634986758232, "learning_rate": 0.00019957446051490198, "loss": 1.7748, "step": 58 }, { "epoch": 0.11, "grad_norm": 0.04252045601606369, "learning_rate": 0.0001995565582593928, "loss": 1.7396, "step": 59 }, { "epoch": 0.11, "grad_norm": 0.04455842077732086, "learning_rate": 0.00019953828798900135, "loss": 1.7236, "step": 60 }, { "epoch": 0.11, "grad_norm": 0.044083647429943085, "learning_rate": 0.0001995196497712645, "loss": 1.7416, "step": 61 }, { "epoch": 0.11, "grad_norm": 0.04511955380439758, "learning_rate": 0.00019950064367507916, "loss": 1.7481, "step": 62 }, { "epoch": 0.11, "grad_norm": 0.0424315445125103, "learning_rate": 0.00019948126977070217, "loss": 1.7712, "step": 63 }, { "epoch": 0.12, "grad_norm": 0.04309271275997162, "learning_rate": 0.00019946152812974993, "loss": 1.6927, "step": 64 }, { "epoch": 0.12, "grad_norm": 0.042915165424346924, "learning_rate": 0.00019944141882519817, "loss": 1.7465, "step": 65 }, { "epoch": 0.12, "grad_norm": 0.05950941890478134, "learning_rate": 0.00019942094193138186, "loss": 1.7035, "step": 66 }, { "epoch": 0.12, "grad_norm": 0.042048510164022446, "learning_rate": 0.0001994000975239946, "loss": 1.7521, "step": 67 }, { "epoch": 0.12, "grad_norm": 0.041577938944101334, "learning_rate": 0.00019937888568008862, "loss": 1.7439, "step": 68 }, { "epoch": 0.13, "grad_norm": 0.04538682475686073, "learning_rate": 0.00019935730647807436, "loss": 1.7528, "step": 69 }, { "epoch": 0.13, "grad_norm": 0.04102981090545654, "learning_rate": 0.00019933535999772025, "loss": 1.6828, "step": 70 }, { "epoch": 0.13, "grad_norm": 0.04318905994296074, "learning_rate": 0.00019931304632015228, "loss": 1.7532, "step": 71 }, { "epoch": 0.13, "grad_norm": 0.043007493019104004, "learning_rate": 0.00019929036552785397, "loss": 1.7353, "step": 72 }, { "epoch": 0.13, "grad_norm": 0.04308176040649414, "learning_rate": 0.00019926731770466568, "loss": 1.6882, "step": 73 }, { "epoch": 0.13, "grad_norm": 0.04227353632450104, "learning_rate": 0.00019924390293578472, "loss": 1.7302, "step": 74 }, { "epoch": 0.14, "grad_norm": 0.0429629310965538, "learning_rate": 0.0001992201213077647, "loss": 1.6822, "step": 75 }, { "epoch": 0.14, "grad_norm": 0.042203355580568314, "learning_rate": 0.00019919597290851538, "loss": 1.7601, "step": 76 }, { "epoch": 0.14, "grad_norm": 0.04265713319182396, "learning_rate": 0.00019917145782730232, "loss": 1.7725, "step": 77 }, { "epoch": 0.14, "grad_norm": 0.04848012328147888, "learning_rate": 0.00019914657615474653, "loss": 1.7587, "step": 78 }, { "epoch": 0.14, "grad_norm": 0.042650256305933, "learning_rate": 0.00019912132798282408, "loss": 1.7422, "step": 79 }, { "epoch": 0.15, "grad_norm": 0.04107372462749481, "learning_rate": 0.00019909571340486593, "loss": 1.7059, "step": 80 }, { "epoch": 0.15, "grad_norm": 0.04788720980286598, "learning_rate": 0.00019906973251555734, "loss": 1.7205, "step": 81 }, { "epoch": 0.15, "grad_norm": 0.041231803596019745, "learning_rate": 0.0001990433854109378, "loss": 1.7277, "step": 82 }, { "epoch": 0.15, "grad_norm": 0.04246293380856514, "learning_rate": 0.0001990166721884004, "loss": 1.7739, "step": 83 }, { "epoch": 0.15, "grad_norm": 0.04331424832344055, "learning_rate": 0.00019898959294669167, "loss": 1.6913, "step": 84 }, { "epoch": 0.15, "grad_norm": 0.04720227047801018, "learning_rate": 0.00019896214778591115, "loss": 1.7079, "step": 85 }, { "epoch": 0.16, "grad_norm": 0.05255519971251488, "learning_rate": 0.00019893433680751103, "loss": 1.7182, "step": 86 }, { "epoch": 0.16, "grad_norm": 0.042392294853925705, "learning_rate": 0.00019890616011429568, "loss": 1.778, "step": 87 }, { "epoch": 0.16, "grad_norm": 0.043008286505937576, "learning_rate": 0.0001988776178104214, "loss": 1.7518, "step": 88 }, { "epoch": 0.16, "grad_norm": 0.044135116040706635, "learning_rate": 0.00019884871000139595, "loss": 1.7534, "step": 89 }, { "epoch": 0.16, "grad_norm": 0.041827455163002014, "learning_rate": 0.00019881943679407832, "loss": 1.7291, "step": 90 }, { "epoch": 0.17, "grad_norm": 0.05515114963054657, "learning_rate": 0.00019878979829667803, "loss": 1.7471, "step": 91 }, { "epoch": 0.17, "grad_norm": 0.040826503187417984, "learning_rate": 0.00019875979461875503, "loss": 1.6408, "step": 92 }, { "epoch": 0.17, "grad_norm": 0.04585504159331322, "learning_rate": 0.00019872942587121915, "loss": 1.6874, "step": 93 }, { "epoch": 0.17, "grad_norm": 0.04665527120232582, "learning_rate": 0.00019869869216632968, "loss": 1.6968, "step": 94 }, { "epoch": 0.17, "grad_norm": 0.046703219413757324, "learning_rate": 0.000198667593617695, "loss": 1.7401, "step": 95 }, { "epoch": 0.18, "grad_norm": 0.04115475341677666, "learning_rate": 0.00019863613034027224, "loss": 1.7227, "step": 96 }, { "epoch": 0.18, "grad_norm": 0.04217168688774109, "learning_rate": 0.00019860430245036663, "loss": 1.7268, "step": 97 }, { "epoch": 0.18, "grad_norm": 0.044889383018016815, "learning_rate": 0.00019857211006563125, "loss": 1.7006, "step": 98 }, { "epoch": 0.18, "grad_norm": 0.04161443933844566, "learning_rate": 0.00019853955330506663, "loss": 1.7266, "step": 99 }, { "epoch": 0.18, "grad_norm": 0.042708829045295715, "learning_rate": 0.00019850663228902012, "loss": 1.7314, "step": 100 }, { "epoch": 0.18, "grad_norm": 0.046648308634757996, "learning_rate": 0.00019847334713918557, "loss": 1.7362, "step": 101 }, { "epoch": 0.19, "grad_norm": 0.04414999857544899, "learning_rate": 0.00019843969797860294, "loss": 1.7065, "step": 102 }, { "epoch": 0.19, "grad_norm": 0.04574083164334297, "learning_rate": 0.00019840568493165772, "loss": 1.7333, "step": 103 }, { "epoch": 0.19, "grad_norm": 0.041924796998500824, "learning_rate": 0.0001983713081240805, "loss": 1.6517, "step": 104 }, { "epoch": 0.19, "grad_norm": 0.04238827899098396, "learning_rate": 0.00019833656768294662, "loss": 1.776, "step": 105 }, { "epoch": 0.19, "grad_norm": 0.04292167350649834, "learning_rate": 0.00019830146373667548, "loss": 1.6601, "step": 106 }, { "epoch": 0.2, "grad_norm": 0.0433412566781044, "learning_rate": 0.00019826599641503025, "loss": 1.6841, "step": 107 }, { "epoch": 0.2, "grad_norm": 0.04201202839612961, "learning_rate": 0.00019823016584911735, "loss": 1.764, "step": 108 }, { "epoch": 0.2, "grad_norm": 0.04234587028622627, "learning_rate": 0.00019819397217138595, "loss": 1.7243, "step": 109 }, { "epoch": 0.2, "grad_norm": 0.04268571734428406, "learning_rate": 0.0001981574155156274, "loss": 1.7656, "step": 110 }, { "epoch": 0.2, "grad_norm": 0.041506245732307434, "learning_rate": 0.00019812049601697492, "loss": 1.6636, "step": 111 }, { "epoch": 0.2, "grad_norm": 0.04152766987681389, "learning_rate": 0.00019808321381190294, "loss": 1.7478, "step": 112 }, { "epoch": 0.21, "grad_norm": 0.041750356554985046, "learning_rate": 0.00019804556903822663, "loss": 1.7518, "step": 113 }, { "epoch": 0.21, "grad_norm": 0.04935223609209061, "learning_rate": 0.00019800756183510144, "loss": 1.7673, "step": 114 }, { "epoch": 0.21, "grad_norm": 0.042300984263420105, "learning_rate": 0.00019796919234302255, "loss": 1.7753, "step": 115 }, { "epoch": 0.21, "grad_norm": 0.04224342852830887, "learning_rate": 0.00019793046070382437, "loss": 1.7226, "step": 116 }, { "epoch": 0.21, "grad_norm": 0.044274065643548965, "learning_rate": 0.00019789136706067998, "loss": 1.7065, "step": 117 }, { "epoch": 0.22, "grad_norm": 0.04910755529999733, "learning_rate": 0.00019785191155810062, "loss": 1.6387, "step": 118 }, { "epoch": 0.22, "grad_norm": 0.04774147644639015, "learning_rate": 0.00019781209434193515, "loss": 1.7297, "step": 119 }, { "epoch": 0.22, "grad_norm": 0.04416586086153984, "learning_rate": 0.00019777191555936957, "loss": 1.8096, "step": 120 }, { "epoch": 0.22, "grad_norm": 0.04406105354428291, "learning_rate": 0.00019773137535892635, "loss": 1.7629, "step": 121 }, { "epoch": 0.22, "grad_norm": 0.043473679572343826, "learning_rate": 0.00019769047389046402, "loss": 1.6979, "step": 122 }, { "epoch": 0.22, "grad_norm": 0.04570621997117996, "learning_rate": 0.00019764921130517653, "loss": 1.7123, "step": 123 }, { "epoch": 0.23, "grad_norm": 0.04326749965548515, "learning_rate": 0.00019760758775559274, "loss": 1.716, "step": 124 }, { "epoch": 0.23, "grad_norm": 0.04397182539105415, "learning_rate": 0.00019756560339557572, "loss": 1.73, "step": 125 }, { "epoch": 0.23, "grad_norm": 0.04468885809183121, "learning_rate": 0.00019752325838032244, "loss": 1.7136, "step": 126 }, { "epoch": 0.23, "grad_norm": 0.04554520919919014, "learning_rate": 0.00019748055286636295, "loss": 1.7448, "step": 127 }, { "epoch": 0.23, "grad_norm": 0.04646708443760872, "learning_rate": 0.00019743748701155995, "loss": 1.6956, "step": 128 }, { "epoch": 0.24, "grad_norm": 0.042717937380075455, "learning_rate": 0.00019739406097510812, "loss": 1.7245, "step": 129 }, { "epoch": 0.24, "grad_norm": 0.04367038235068321, "learning_rate": 0.00019735027491753353, "loss": 1.7102, "step": 130 }, { "epoch": 0.24, "grad_norm": 0.04296841099858284, "learning_rate": 0.0001973061290006932, "loss": 1.7163, "step": 131 }, { "epoch": 0.24, "grad_norm": 0.043665811419487, "learning_rate": 0.00019726162338777424, "loss": 1.7172, "step": 132 }, { "epoch": 0.24, "grad_norm": 0.046134624630212784, "learning_rate": 0.00019721675824329354, "loss": 1.7327, "step": 133 }, { "epoch": 0.24, "grad_norm": 0.04857848584651947, "learning_rate": 0.00019717153373309692, "loss": 1.6647, "step": 134 }, { "epoch": 0.25, "grad_norm": 0.047723885625600815, "learning_rate": 0.00019712595002435861, "loss": 1.7422, "step": 135 }, { "epoch": 0.25, "grad_norm": 0.04413154348731041, "learning_rate": 0.00019708000728558064, "loss": 1.6943, "step": 136 }, { "epoch": 0.25, "grad_norm": 0.043105412274599075, "learning_rate": 0.00019703370568659225, "loss": 1.7519, "step": 137 }, { "epoch": 0.25, "eval_loss": 1.7284438610076904, "eval_runtime": 76.3963, "eval_samples_per_second": 65.448, "eval_steps_per_second": 16.362, "step": 137 }, { "epoch": 0.25, "grad_norm": 0.04300757125020027, "learning_rate": 0.00019698704539854918, "loss": 1.7341, "step": 138 }, { "epoch": 0.25, "grad_norm": 0.043961744755506516, "learning_rate": 0.00019694002659393305, "loss": 1.777, "step": 139 }, { "epoch": 0.26, "grad_norm": 0.04376057907938957, "learning_rate": 0.00019689264944655084, "loss": 1.7403, "step": 140 }, { "epoch": 0.26, "grad_norm": 0.04482461139559746, "learning_rate": 0.00019684491413153411, "loss": 1.6852, "step": 141 }, { "epoch": 0.26, "grad_norm": 0.045192863792181015, "learning_rate": 0.0001967968208253384, "loss": 1.7494, "step": 142 }, { "epoch": 0.26, "grad_norm": 0.04361759498715401, "learning_rate": 0.00019674836970574254, "loss": 1.7331, "step": 143 }, { "epoch": 0.26, "grad_norm": 0.04294734448194504, "learning_rate": 0.0001966995609518481, "loss": 1.6375, "step": 144 }, { "epoch": 0.26, "grad_norm": 0.04528161138296127, "learning_rate": 0.00019665039474407863, "loss": 1.746, "step": 145 }, { "epoch": 0.27, "grad_norm": 0.04510699212551117, "learning_rate": 0.00019660087126417906, "loss": 1.7053, "step": 146 }, { "epoch": 0.27, "grad_norm": 0.042807720601558685, "learning_rate": 0.00019655099069521486, "loss": 1.6748, "step": 147 }, { "epoch": 0.27, "grad_norm": 0.04657953232526779, "learning_rate": 0.00019650075322157168, "loss": 1.684, "step": 148 }, { "epoch": 0.27, "grad_norm": 0.04593012481927872, "learning_rate": 0.00019645015902895437, "loss": 1.7076, "step": 149 }, { "epoch": 0.27, "grad_norm": 0.04362139105796814, "learning_rate": 0.0001963992083043864, "loss": 1.6773, "step": 150 }, { "epoch": 0.28, "grad_norm": 0.04773354157805443, "learning_rate": 0.00019634790123620926, "loss": 1.7107, "step": 151 }, { "epoch": 0.28, "grad_norm": 0.05423569679260254, "learning_rate": 0.00019629623801408155, "loss": 1.7052, "step": 152 }, { "epoch": 0.28, "grad_norm": 0.043550509959459305, "learning_rate": 0.00019624421882897855, "loss": 1.7151, "step": 153 }, { "epoch": 0.28, "grad_norm": 0.04896851256489754, "learning_rate": 0.00019619184387319123, "loss": 1.6611, "step": 154 }, { "epoch": 0.28, "grad_norm": 0.04392845928668976, "learning_rate": 0.00019613911334032583, "loss": 1.738, "step": 155 }, { "epoch": 0.28, "grad_norm": 0.04582325741648674, "learning_rate": 0.00019608602742530283, "loss": 1.6885, "step": 156 }, { "epoch": 0.29, "grad_norm": 0.045696284621953964, "learning_rate": 0.00019603258632435656, "loss": 1.7365, "step": 157 }, { "epoch": 0.29, "grad_norm": 0.043873440474271774, "learning_rate": 0.00019597879023503417, "loss": 1.8094, "step": 158 }, { "epoch": 0.29, "grad_norm": 0.05078018456697464, "learning_rate": 0.00019592463935619517, "loss": 1.7341, "step": 159 }, { "epoch": 0.29, "grad_norm": 0.042483873665332794, "learning_rate": 0.00019587013388801047, "loss": 1.7351, "step": 160 }, { "epoch": 0.29, "grad_norm": 0.045154914259910583, "learning_rate": 0.00019581527403196168, "loss": 1.6645, "step": 161 }, { "epoch": 0.3, "grad_norm": 0.04563280567526817, "learning_rate": 0.0001957600599908406, "loss": 1.7069, "step": 162 }, { "epoch": 0.3, "grad_norm": 0.0451313816010952, "learning_rate": 0.00019570449196874815, "loss": 1.7392, "step": 163 }, { "epoch": 0.3, "grad_norm": 0.04682654142379761, "learning_rate": 0.0001956485701710938, "loss": 1.6987, "step": 164 }, { "epoch": 0.3, "grad_norm": 0.04211273416876793, "learning_rate": 0.00019559229480459474, "loss": 1.6973, "step": 165 }, { "epoch": 0.3, "grad_norm": 0.04460490494966507, "learning_rate": 0.00019553566607727517, "loss": 1.7233, "step": 166 }, { "epoch": 0.3, "grad_norm": 0.044608812779188156, "learning_rate": 0.00019547868419846548, "loss": 1.7371, "step": 167 }, { "epoch": 0.31, "grad_norm": 0.04518236592411995, "learning_rate": 0.00019542134937880154, "loss": 1.7257, "step": 168 }, { "epoch": 0.31, "grad_norm": 0.04374237731099129, "learning_rate": 0.00019536366183022384, "loss": 1.7136, "step": 169 }, { "epoch": 0.31, "grad_norm": 0.04429790750145912, "learning_rate": 0.00019530562176597673, "loss": 1.7216, "step": 170 }, { "epoch": 0.31, "grad_norm": 0.04807354509830475, "learning_rate": 0.0001952472294006077, "loss": 1.6568, "step": 171 }, { "epoch": 0.31, "grad_norm": 0.04785493016242981, "learning_rate": 0.00019518848494996655, "loss": 1.7272, "step": 172 }, { "epoch": 0.32, "grad_norm": 0.04472104460000992, "learning_rate": 0.0001951293886312045, "loss": 1.7283, "step": 173 }, { "epoch": 0.32, "grad_norm": 0.04852326214313507, "learning_rate": 0.00019506994066277348, "loss": 1.6968, "step": 174 }, { "epoch": 0.32, "grad_norm": 0.04624422639608383, "learning_rate": 0.0001950101412644254, "loss": 1.758, "step": 175 }, { "epoch": 0.32, "grad_norm": 0.044666189700365067, "learning_rate": 0.00019494999065721108, "loss": 1.6933, "step": 176 }, { "epoch": 0.32, "grad_norm": 0.05367857217788696, "learning_rate": 0.0001948894890634798, "loss": 1.7328, "step": 177 }, { "epoch": 0.32, "grad_norm": 0.046923939138650894, "learning_rate": 0.0001948286367068781, "loss": 1.7367, "step": 178 }, { "epoch": 0.33, "grad_norm": 0.04480034112930298, "learning_rate": 0.00019476743381234926, "loss": 1.7677, "step": 179 }, { "epoch": 0.33, "grad_norm": 0.045380428433418274, "learning_rate": 0.00019470588060613222, "loss": 1.7439, "step": 180 }, { "epoch": 0.33, "grad_norm": 0.04550057277083397, "learning_rate": 0.00019464397731576094, "loss": 1.6895, "step": 181 }, { "epoch": 0.33, "grad_norm": 0.049537234008312225, "learning_rate": 0.00019458172417006347, "loss": 1.7274, "step": 182 }, { "epoch": 0.33, "grad_norm": 0.04696514084935188, "learning_rate": 0.0001945191213991611, "loss": 1.7121, "step": 183 }, { "epoch": 0.34, "grad_norm": 0.04783783480525017, "learning_rate": 0.00019445616923446755, "loss": 1.6942, "step": 184 }, { "epoch": 0.34, "grad_norm": 0.04514686018228531, "learning_rate": 0.00019439286790868802, "loss": 1.7219, "step": 185 }, { "epoch": 0.34, "grad_norm": 0.045743513852357864, "learning_rate": 0.00019432921765581847, "loss": 1.76, "step": 186 }, { "epoch": 0.34, "grad_norm": 0.04406295716762543, "learning_rate": 0.00019426521871114468, "loss": 1.7531, "step": 187 }, { "epoch": 0.34, "grad_norm": 0.04445353150367737, "learning_rate": 0.00019420087131124131, "loss": 1.7742, "step": 188 }, { "epoch": 0.34, "grad_norm": 0.04396241530776024, "learning_rate": 0.0001941361756939712, "loss": 1.7701, "step": 189 }, { "epoch": 0.35, "grad_norm": 0.04415050894021988, "learning_rate": 0.0001940711320984843, "loss": 1.7062, "step": 190 }, { "epoch": 0.35, "grad_norm": 0.04672138765454292, "learning_rate": 0.00019400574076521693, "loss": 1.754, "step": 191 }, { "epoch": 0.35, "grad_norm": 0.04417939484119415, "learning_rate": 0.00019394000193589088, "loss": 1.7357, "step": 192 }, { "epoch": 0.35, "grad_norm": 0.04567494988441467, "learning_rate": 0.00019387391585351234, "loss": 1.752, "step": 193 }, { "epoch": 0.35, "grad_norm": 0.045080311596393585, "learning_rate": 0.00019380748276237123, "loss": 1.736, "step": 194 }, { "epoch": 0.36, "grad_norm": 0.04506627842783928, "learning_rate": 0.0001937407029080402, "loss": 1.6726, "step": 195 }, { "epoch": 0.36, "grad_norm": 0.04523961618542671, "learning_rate": 0.0001936735765373737, "loss": 1.7621, "step": 196 }, { "epoch": 0.36, "grad_norm": 0.04326867312192917, "learning_rate": 0.00019360610389850712, "loss": 1.7341, "step": 197 }, { "epoch": 0.36, "grad_norm": 0.05188523977994919, "learning_rate": 0.00019353828524085577, "loss": 1.7277, "step": 198 }, { "epoch": 0.36, "grad_norm": 0.04654062166810036, "learning_rate": 0.00019347012081511415, "loss": 1.6845, "step": 199 }, { "epoch": 0.36, "grad_norm": 0.044841405004262924, "learning_rate": 0.0001934016108732548, "loss": 1.6611, "step": 200 }, { "epoch": 0.37, "grad_norm": 0.0941338911652565, "learning_rate": 0.00019333275566852756, "loss": 1.6978, "step": 201 }, { "epoch": 0.37, "grad_norm": 0.05048836022615433, "learning_rate": 0.00019326355545545845, "loss": 1.7056, "step": 202 }, { "epoch": 0.37, "grad_norm": 0.046358656138181686, "learning_rate": 0.00019319401048984892, "loss": 1.649, "step": 203 }, { "epoch": 0.37, "grad_norm": 0.04557095095515251, "learning_rate": 0.00019312412102877473, "loss": 1.6793, "step": 204 }, { "epoch": 0.37, "grad_norm": 0.04551040008664131, "learning_rate": 0.0001930538873305852, "loss": 1.7339, "step": 205 }, { "epoch": 0.38, "grad_norm": 0.044258005917072296, "learning_rate": 0.000192983309654902, "loss": 1.6627, "step": 206 }, { "epoch": 0.38, "grad_norm": 0.0485963337123394, "learning_rate": 0.00019291238826261843, "loss": 1.715, "step": 207 }, { "epoch": 0.38, "grad_norm": 0.047103844583034515, "learning_rate": 0.00019284112341589832, "loss": 1.6855, "step": 208 }, { "epoch": 0.38, "grad_norm": 0.045252177864313126, "learning_rate": 0.000192769515378175, "loss": 1.7557, "step": 209 }, { "epoch": 0.38, "grad_norm": 0.049794841557741165, "learning_rate": 0.00019269756441415062, "loss": 1.7116, "step": 210 }, { "epoch": 0.38, "grad_norm": 0.04380947723984718, "learning_rate": 0.00019262527078979478, "loss": 1.7663, "step": 211 }, { "epoch": 0.39, "grad_norm": 0.046488065272569656, "learning_rate": 0.00019255263477234381, "loss": 1.6724, "step": 212 }, { "epoch": 0.39, "grad_norm": 0.0422043539583683, "learning_rate": 0.00019247965663029976, "loss": 1.7345, "step": 213 }, { "epoch": 0.39, "grad_norm": 0.05002991482615471, "learning_rate": 0.0001924063366334293, "loss": 1.7468, "step": 214 }, { "epoch": 0.39, "grad_norm": 0.04376322776079178, "learning_rate": 0.0001923326750527628, "loss": 1.7748, "step": 215 }, { "epoch": 0.39, "grad_norm": 0.04664807394146919, "learning_rate": 0.00019225867216059325, "loss": 1.7156, "step": 216 }, { "epoch": 0.4, "grad_norm": 0.047952812165021896, "learning_rate": 0.0001921843282304754, "loss": 1.7247, "step": 217 }, { "epoch": 0.4, "grad_norm": 0.045118216425180435, "learning_rate": 0.00019210964353722464, "loss": 1.7354, "step": 218 }, { "epoch": 0.4, "grad_norm": 0.054903436452150345, "learning_rate": 0.00019203461835691594, "loss": 1.7241, "step": 219 }, { "epoch": 0.4, "grad_norm": 0.04747498407959938, "learning_rate": 0.000191959252966883, "loss": 1.7498, "step": 220 }, { "epoch": 0.4, "grad_norm": 0.04605628177523613, "learning_rate": 0.000191883547645717, "loss": 1.6889, "step": 221 }, { "epoch": 0.4, "grad_norm": 0.04835960268974304, "learning_rate": 0.00019180750267326578, "loss": 1.715, "step": 222 }, { "epoch": 0.41, "grad_norm": 0.04828386381268501, "learning_rate": 0.00019173111833063273, "loss": 1.6931, "step": 223 }, { "epoch": 0.41, "grad_norm": 0.04604095220565796, "learning_rate": 0.0001916543949001756, "loss": 1.6717, "step": 224 }, { "epoch": 0.41, "grad_norm": 0.049674633890390396, "learning_rate": 0.00019157733266550575, "loss": 1.7746, "step": 225 }, { "epoch": 0.41, "grad_norm": 0.04439341649413109, "learning_rate": 0.00019149993191148687, "loss": 1.6925, "step": 226 }, { "epoch": 0.41, "grad_norm": 0.04741811007261276, "learning_rate": 0.00019142219292423395, "loss": 1.7219, "step": 227 }, { "epoch": 0.42, "grad_norm": 0.049409981817007065, "learning_rate": 0.00019134411599111242, "loss": 1.7306, "step": 228 }, { "epoch": 0.42, "grad_norm": 0.04618163779377937, "learning_rate": 0.00019126570140073676, "loss": 1.7271, "step": 229 }, { "epoch": 0.42, "grad_norm": 0.04557076469063759, "learning_rate": 0.0001911869494429698, "loss": 1.7188, "step": 230 }, { "epoch": 0.42, "grad_norm": 0.04645569249987602, "learning_rate": 0.0001911078604089213, "loss": 1.7191, "step": 231 }, { "epoch": 0.42, "grad_norm": 0.04584849998354912, "learning_rate": 0.0001910284345909471, "loss": 1.7592, "step": 232 }, { "epoch": 0.42, "grad_norm": 0.045582644641399384, "learning_rate": 0.000190948672282648, "loss": 1.6902, "step": 233 }, { "epoch": 0.43, "grad_norm": 0.04627401754260063, "learning_rate": 0.00019086857377886865, "loss": 1.6937, "step": 234 }, { "epoch": 0.43, "grad_norm": 0.04470285400748253, "learning_rate": 0.00019078813937569643, "loss": 1.6977, "step": 235 }, { "epoch": 0.43, "grad_norm": 0.05287547782063484, "learning_rate": 0.00019070736937046035, "loss": 1.7539, "step": 236 }, { "epoch": 0.43, "grad_norm": 0.04990493878722191, "learning_rate": 0.00019062626406173006, "loss": 1.7469, "step": 237 }, { "epoch": 0.43, "grad_norm": 0.048645589500665665, "learning_rate": 0.00019054482374931467, "loss": 1.7037, "step": 238 }, { "epoch": 0.44, "grad_norm": 0.04730357602238655, "learning_rate": 0.0001904630487342616, "loss": 1.7388, "step": 239 }, { "epoch": 0.44, "grad_norm": 0.04754168912768364, "learning_rate": 0.00019038093931885553, "loss": 1.7805, "step": 240 }, { "epoch": 0.44, "grad_norm": 0.04760801047086716, "learning_rate": 0.00019029849580661727, "loss": 1.7383, "step": 241 }, { "epoch": 0.44, "grad_norm": 0.048467203974723816, "learning_rate": 0.0001902157185023026, "loss": 1.7078, "step": 242 }, { "epoch": 0.44, "grad_norm": 0.0522041916847229, "learning_rate": 0.00019013260771190126, "loss": 1.7052, "step": 243 }, { "epoch": 0.44, "grad_norm": 0.0501788929104805, "learning_rate": 0.00019004916374263563, "loss": 1.7818, "step": 244 }, { "epoch": 0.45, "grad_norm": 0.04538620635867119, "learning_rate": 0.00018996538690295979, "loss": 1.6589, "step": 245 }, { "epoch": 0.45, "grad_norm": 0.04511679336428642, "learning_rate": 0.00018988127750255824, "loss": 1.7179, "step": 246 }, { "epoch": 0.45, "grad_norm": 0.04756203666329384, "learning_rate": 0.0001897968358523448, "loss": 1.7333, "step": 247 }, { "epoch": 0.45, "grad_norm": 0.05278336629271507, "learning_rate": 0.00018971206226446147, "loss": 1.7431, "step": 248 }, { "epoch": 0.45, "grad_norm": 0.05926801264286041, "learning_rate": 0.00018962695705227728, "loss": 1.7768, "step": 249 }, { "epoch": 0.46, "grad_norm": 0.049290940165519714, "learning_rate": 0.00018954152053038712, "loss": 1.7119, "step": 250 }, { "epoch": 0.46, "grad_norm": 0.04777907952666283, "learning_rate": 0.0001894557530146106, "loss": 1.7559, "step": 251 }, { "epoch": 0.46, "grad_norm": 0.04726920276880264, "learning_rate": 0.00018936965482199084, "loss": 1.7861, "step": 252 }, { "epoch": 0.46, "grad_norm": 0.04677857458591461, "learning_rate": 0.0001892832262707933, "loss": 1.7039, "step": 253 }, { "epoch": 0.46, "grad_norm": 0.04724700003862381, "learning_rate": 0.00018919646768050468, "loss": 1.6704, "step": 254 }, { "epoch": 0.46, "grad_norm": 0.04969072341918945, "learning_rate": 0.00018910937937183166, "loss": 1.7168, "step": 255 }, { "epoch": 0.47, "grad_norm": 0.04533353075385094, "learning_rate": 0.0001890219616666997, "loss": 1.6751, "step": 256 }, { "epoch": 0.47, "grad_norm": 0.04647386819124222, "learning_rate": 0.0001889342148882519, "loss": 1.7146, "step": 257 }, { "epoch": 0.47, "grad_norm": 0.047208696603775024, "learning_rate": 0.00018884613936084784, "loss": 1.7378, "step": 258 }, { "epoch": 0.47, "grad_norm": 0.04841624200344086, "learning_rate": 0.0001887577354100623, "loss": 1.7128, "step": 259 }, { "epoch": 0.47, "grad_norm": 0.05073019117116928, "learning_rate": 0.00018866900336268408, "loss": 1.7206, "step": 260 }, { "epoch": 0.48, "grad_norm": 0.051456011831760406, "learning_rate": 0.00018857994354671482, "loss": 1.755, "step": 261 }, { "epoch": 0.48, "grad_norm": 0.04637736827135086, "learning_rate": 0.0001884905562913678, "loss": 1.7395, "step": 262 }, { "epoch": 0.48, "grad_norm": 0.061346374452114105, "learning_rate": 0.00018840084192706658, "loss": 1.674, "step": 263 }, { "epoch": 0.48, "grad_norm": 0.04413258284330368, "learning_rate": 0.00018831080078544402, "loss": 1.7288, "step": 264 }, { "epoch": 0.48, "grad_norm": 0.0531301349401474, "learning_rate": 0.0001882204331993409, "loss": 1.7625, "step": 265 }, { "epoch": 0.48, "grad_norm": 0.05146196484565735, "learning_rate": 0.00018812973950280468, "loss": 1.6815, "step": 266 }, { "epoch": 0.49, "grad_norm": 0.047678787261247635, "learning_rate": 0.0001880387200310883, "loss": 1.7278, "step": 267 }, { "epoch": 0.49, "grad_norm": 0.0556582510471344, "learning_rate": 0.0001879473751206489, "loss": 1.74, "step": 268 }, { "epoch": 0.49, "grad_norm": 0.047515787184238434, "learning_rate": 0.00018785570510914678, "loss": 1.7207, "step": 269 }, { "epoch": 0.49, "grad_norm": 0.04592055827379227, "learning_rate": 0.0001877637103354438, "loss": 1.6589, "step": 270 }, { "epoch": 0.49, "grad_norm": 0.04531411454081535, "learning_rate": 0.0001876713911396024, "loss": 1.706, "step": 271 }, { "epoch": 0.5, "grad_norm": 0.04682420939207077, "learning_rate": 0.0001875787478628843, "loss": 1.7297, "step": 272 }, { "epoch": 0.5, "grad_norm": 0.04545978829264641, "learning_rate": 0.00018748578084774913, "loss": 1.6572, "step": 273 }, { "epoch": 0.5, "grad_norm": 0.04849430173635483, "learning_rate": 0.00018739249043785324, "loss": 1.7442, "step": 274 }, { "epoch": 0.5, "eval_loss": 1.726025938987732, "eval_runtime": 76.0967, "eval_samples_per_second": 65.706, "eval_steps_per_second": 16.426, "step": 274 }, { "epoch": 0.5, "grad_norm": 0.04745488613843918, "learning_rate": 0.00018729887697804847, "loss": 1.7398, "step": 275 }, { "epoch": 0.5, "grad_norm": 0.05489857494831085, "learning_rate": 0.00018720494081438078, "loss": 1.701, "step": 276 }, { "epoch": 0.51, "grad_norm": 0.04818108305335045, "learning_rate": 0.00018711068229408903, "loss": 1.7068, "step": 277 }, { "epoch": 0.51, "grad_norm": 0.04530555009841919, "learning_rate": 0.0001870161017656037, "loss": 1.6966, "step": 278 }, { "epoch": 0.51, "grad_norm": 0.045606572180986404, "learning_rate": 0.00018692119957854558, "loss": 1.7086, "step": 279 }, { "epoch": 0.51, "grad_norm": 0.04626869410276413, "learning_rate": 0.00018682597608372445, "loss": 1.6981, "step": 280 }, { "epoch": 0.51, "grad_norm": 0.04752146080136299, "learning_rate": 0.0001867304316331379, "loss": 1.692, "step": 281 }, { "epoch": 0.51, "grad_norm": 0.046230729669332504, "learning_rate": 0.0001866345665799698, "loss": 1.7338, "step": 282 }, { "epoch": 0.52, "grad_norm": 0.04928119108080864, "learning_rate": 0.00018653838127858933, "loss": 1.738, "step": 283 }, { "epoch": 0.52, "grad_norm": 0.04641352593898773, "learning_rate": 0.00018644187608454936, "loss": 1.6792, "step": 284 }, { "epoch": 0.52, "grad_norm": 0.04860611632466316, "learning_rate": 0.00018634505135458525, "loss": 1.663, "step": 285 }, { "epoch": 0.52, "grad_norm": 0.046515002846717834, "learning_rate": 0.00018624790744661355, "loss": 1.7327, "step": 286 }, { "epoch": 0.52, "grad_norm": 0.04668186604976654, "learning_rate": 0.00018615044471973074, "loss": 1.6987, "step": 287 }, { "epoch": 0.53, "grad_norm": 0.047913163900375366, "learning_rate": 0.00018605266353421176, "loss": 1.7953, "step": 288 }, { "epoch": 0.53, "grad_norm": 0.04924839362502098, "learning_rate": 0.00018595456425150872, "loss": 1.7891, "step": 289 }, { "epoch": 0.53, "grad_norm": 0.049241986125707626, "learning_rate": 0.00018585614723424962, "loss": 1.7451, "step": 290 }, { "epoch": 0.53, "grad_norm": 0.05132036283612251, "learning_rate": 0.00018575741284623703, "loss": 1.7598, "step": 291 }, { "epoch": 0.53, "grad_norm": 0.04659922048449516, "learning_rate": 0.00018565836145244662, "loss": 1.7331, "step": 292 }, { "epoch": 0.53, "grad_norm": 0.0466977022588253, "learning_rate": 0.0001855589934190259, "loss": 1.7171, "step": 293 }, { "epoch": 0.54, "grad_norm": 0.049368374049663544, "learning_rate": 0.00018545930911329287, "loss": 1.6929, "step": 294 }, { "epoch": 0.54, "grad_norm": 0.04552480950951576, "learning_rate": 0.00018535930890373466, "loss": 1.753, "step": 295 }, { "epoch": 0.54, "grad_norm": 0.04755065590143204, "learning_rate": 0.00018525899316000608, "loss": 1.7472, "step": 296 }, { "epoch": 0.54, "grad_norm": 0.050540413707494736, "learning_rate": 0.0001851583622529284, "loss": 1.7585, "step": 297 }, { "epoch": 0.54, "grad_norm": 0.04644971713423729, "learning_rate": 0.00018505741655448792, "loss": 1.7531, "step": 298 }, { "epoch": 0.55, "grad_norm": 0.05085503309965134, "learning_rate": 0.00018495615643783446, "loss": 1.6954, "step": 299 }, { "epoch": 0.55, "grad_norm": 0.0480993427336216, "learning_rate": 0.0001848545822772802, "loss": 1.6976, "step": 300 }, { "epoch": 0.55, "grad_norm": 0.0487300269305706, "learning_rate": 0.00018475269444829818, "loss": 1.7642, "step": 301 }, { "epoch": 0.55, "grad_norm": 0.04805615171790123, "learning_rate": 0.0001846504933275209, "loss": 1.6666, "step": 302 }, { "epoch": 0.55, "grad_norm": 0.045554857701063156, "learning_rate": 0.00018454797929273902, "loss": 1.7259, "step": 303 }, { "epoch": 0.55, "grad_norm": 0.04570743814110756, "learning_rate": 0.00018444515272289982, "loss": 1.7067, "step": 304 }, { "epoch": 0.56, "grad_norm": 0.047652073204517365, "learning_rate": 0.00018434201399810594, "loss": 1.8147, "step": 305 }, { "epoch": 0.56, "grad_norm": 0.046781569719314575, "learning_rate": 0.00018423856349961384, "loss": 1.7509, "step": 306 }, { "epoch": 0.56, "grad_norm": 0.04698612168431282, "learning_rate": 0.00018413480160983254, "loss": 1.7074, "step": 307 }, { "epoch": 0.56, "grad_norm": 0.04796341061592102, "learning_rate": 0.0001840307287123221, "loss": 1.7444, "step": 308 }, { "epoch": 0.56, "grad_norm": 0.047553375363349915, "learning_rate": 0.00018392634519179225, "loss": 1.7103, "step": 309 }, { "epoch": 0.57, "grad_norm": 0.046323925256729126, "learning_rate": 0.00018382165143410092, "loss": 1.716, "step": 310 }, { "epoch": 0.57, "grad_norm": 0.04571986570954323, "learning_rate": 0.00018371664782625287, "loss": 1.7035, "step": 311 }, { "epoch": 0.57, "grad_norm": 0.05170504003763199, "learning_rate": 0.0001836113347563982, "loss": 1.7151, "step": 312 }, { "epoch": 0.57, "grad_norm": 0.047869808971881866, "learning_rate": 0.000183505712613831, "loss": 1.7223, "step": 313 }, { "epoch": 0.57, "grad_norm": 0.0482964813709259, "learning_rate": 0.0001833997817889878, "loss": 1.6805, "step": 314 }, { "epoch": 0.57, "grad_norm": 0.0486602708697319, "learning_rate": 0.00018329354267344625, "loss": 1.7303, "step": 315 }, { "epoch": 0.58, "grad_norm": 0.046554964035749435, "learning_rate": 0.00018318699565992357, "loss": 1.7745, "step": 316 }, { "epoch": 0.58, "grad_norm": 0.047917045652866364, "learning_rate": 0.00018308014114227513, "loss": 1.718, "step": 317 }, { "epoch": 0.58, "grad_norm": 0.0479004867374897, "learning_rate": 0.00018297297951549304, "loss": 1.7707, "step": 318 }, { "epoch": 0.58, "grad_norm": 0.04681101068854332, "learning_rate": 0.0001828655111757046, "loss": 1.7646, "step": 319 }, { "epoch": 0.58, "grad_norm": 0.05201521888375282, "learning_rate": 0.00018275773652017097, "loss": 1.7479, "step": 320 }, { "epoch": 0.59, "grad_norm": 0.04852493852376938, "learning_rate": 0.00018264965594728548, "loss": 1.7463, "step": 321 }, { "epoch": 0.59, "grad_norm": 0.046121757477521896, "learning_rate": 0.00018254126985657246, "loss": 1.7444, "step": 322 }, { "epoch": 0.59, "grad_norm": 0.05163992941379547, "learning_rate": 0.00018243257864868548, "loss": 1.7134, "step": 323 }, { "epoch": 0.59, "grad_norm": 0.06267976760864258, "learning_rate": 0.00018232358272540604, "loss": 1.6712, "step": 324 }, { "epoch": 0.59, "grad_norm": 0.04854287579655647, "learning_rate": 0.00018221428248964202, "loss": 1.6932, "step": 325 }, { "epoch": 0.59, "grad_norm": 0.046650100499391556, "learning_rate": 0.00018210467834542615, "loss": 1.768, "step": 326 }, { "epoch": 0.6, "grad_norm": 0.04779491573572159, "learning_rate": 0.00018199477069791474, "loss": 1.7109, "step": 327 }, { "epoch": 0.6, "grad_norm": 0.05170130729675293, "learning_rate": 0.0001818845599533858, "loss": 1.6926, "step": 328 }, { "epoch": 0.6, "grad_norm": 0.04867775738239288, "learning_rate": 0.00018177404651923787, "loss": 1.6908, "step": 329 }, { "epoch": 0.6, "grad_norm": 0.04707460105419159, "learning_rate": 0.00018166323080398835, "loss": 1.7461, "step": 330 }, { "epoch": 0.6, "grad_norm": 0.048908475786447525, "learning_rate": 0.00018155211321727212, "loss": 1.7214, "step": 331 }, { "epoch": 0.61, "grad_norm": 0.04802173003554344, "learning_rate": 0.00018144069416983985, "loss": 1.7528, "step": 332 }, { "epoch": 0.61, "grad_norm": 0.04747573658823967, "learning_rate": 0.00018132897407355657, "loss": 1.6726, "step": 333 }, { "epoch": 0.61, "grad_norm": 0.049620069563388824, "learning_rate": 0.00018121695334140017, "loss": 1.7215, "step": 334 }, { "epoch": 0.61, "grad_norm": 0.047733817249536514, "learning_rate": 0.00018110463238745988, "loss": 1.7538, "step": 335 }, { "epoch": 0.61, "grad_norm": 0.04856455698609352, "learning_rate": 0.00018099201162693476, "loss": 1.6833, "step": 336 }, { "epoch": 0.61, "grad_norm": 0.04885758087038994, "learning_rate": 0.00018087909147613193, "loss": 1.7141, "step": 337 }, { "epoch": 0.62, "grad_norm": 0.047947369515895844, "learning_rate": 0.0001807658723524654, "loss": 1.733, "step": 338 }, { "epoch": 0.62, "grad_norm": 0.0499010868370533, "learning_rate": 0.0001806523546744543, "loss": 1.6825, "step": 339 }, { "epoch": 0.62, "grad_norm": 0.048193834722042084, "learning_rate": 0.0001805385388617213, "loss": 1.7282, "step": 340 }, { "epoch": 0.62, "grad_norm": 0.05272866412997246, "learning_rate": 0.00018042442533499123, "loss": 1.7599, "step": 341 }, { "epoch": 0.62, "grad_norm": 0.047657158225774765, "learning_rate": 0.00018031001451608943, "loss": 1.7292, "step": 342 }, { "epoch": 0.63, "grad_norm": 0.0498197004199028, "learning_rate": 0.00018019530682794014, "loss": 1.7417, "step": 343 }, { "epoch": 0.63, "grad_norm": 0.04958554729819298, "learning_rate": 0.00018008030269456505, "loss": 1.7274, "step": 344 }, { "epoch": 0.63, "grad_norm": 0.04730832576751709, "learning_rate": 0.00017996500254108152, "loss": 1.778, "step": 345 }, { "epoch": 0.63, "grad_norm": 0.050828639417886734, "learning_rate": 0.0001798494067937014, "loss": 1.7285, "step": 346 }, { "epoch": 0.63, "grad_norm": 0.046292368322610855, "learning_rate": 0.00017973351587972905, "loss": 1.7334, "step": 347 }, { "epoch": 0.63, "grad_norm": 0.04758565500378609, "learning_rate": 0.00017961733022755992, "loss": 1.6814, "step": 348 }, { "epoch": 0.64, "grad_norm": 0.050507742911577225, "learning_rate": 0.00017950085026667903, "loss": 1.6949, "step": 349 }, { "epoch": 0.64, "grad_norm": 0.04801836982369423, "learning_rate": 0.00017938407642765938, "loss": 1.6594, "step": 350 }, { "epoch": 0.64, "grad_norm": 0.04616666957736015, "learning_rate": 0.00017926700914216016, "loss": 1.6969, "step": 351 }, { "epoch": 0.64, "grad_norm": 0.048213839530944824, "learning_rate": 0.00017914964884292544, "loss": 1.6908, "step": 352 }, { "epoch": 0.64, "grad_norm": 0.04909725859761238, "learning_rate": 0.00017903199596378227, "loss": 1.7213, "step": 353 }, { "epoch": 0.65, "grad_norm": 0.050252340734004974, "learning_rate": 0.00017891405093963938, "loss": 1.7094, "step": 354 }, { "epoch": 0.65, "grad_norm": 0.05401075631380081, "learning_rate": 0.00017879581420648534, "loss": 1.7163, "step": 355 }, { "epoch": 0.65, "grad_norm": 0.05027545616030693, "learning_rate": 0.00017867728620138708, "loss": 1.7362, "step": 356 }, { "epoch": 0.65, "grad_norm": 0.047479428350925446, "learning_rate": 0.00017855846736248822, "loss": 1.6785, "step": 357 }, { "epoch": 0.65, "grad_norm": 0.05026884377002716, "learning_rate": 0.0001784393581290074, "loss": 1.7221, "step": 358 }, { "epoch": 0.65, "grad_norm": 0.04901432618498802, "learning_rate": 0.00017831995894123683, "loss": 1.6401, "step": 359 }, { "epoch": 0.66, "grad_norm": 0.04764765873551369, "learning_rate": 0.00017820027024054044, "loss": 1.7361, "step": 360 }, { "epoch": 0.66, "grad_norm": 0.046871528029441833, "learning_rate": 0.0001780802924693524, "loss": 1.7986, "step": 361 }, { "epoch": 0.66, "grad_norm": 0.05453401803970337, "learning_rate": 0.00017796002607117545, "loss": 1.7447, "step": 362 }, { "epoch": 0.66, "grad_norm": 0.04958674684166908, "learning_rate": 0.00017783947149057925, "loss": 1.7091, "step": 363 }, { "epoch": 0.66, "grad_norm": 0.053141675889492035, "learning_rate": 0.0001777186291731987, "loss": 1.6866, "step": 364 }, { "epoch": 0.67, "grad_norm": 0.047340743243694305, "learning_rate": 0.00017759749956573238, "loss": 1.7191, "step": 365 }, { "epoch": 0.67, "grad_norm": 0.051203418523073196, "learning_rate": 0.00017747608311594087, "loss": 1.7238, "step": 366 }, { "epoch": 0.67, "grad_norm": 0.047188933938741684, "learning_rate": 0.00017735438027264495, "loss": 1.762, "step": 367 }, { "epoch": 0.67, "grad_norm": 0.056479763239622116, "learning_rate": 0.00017723239148572422, "loss": 1.6587, "step": 368 }, { "epoch": 0.67, "grad_norm": 0.04922572523355484, "learning_rate": 0.00017711011720611514, "loss": 1.6988, "step": 369 }, { "epoch": 0.67, "grad_norm": 0.046839334070682526, "learning_rate": 0.00017698755788580963, "loss": 1.7092, "step": 370 }, { "epoch": 0.68, "grad_norm": 0.0491393506526947, "learning_rate": 0.0001768647139778532, "loss": 1.7313, "step": 371 }, { "epoch": 0.68, "grad_norm": 0.04811710864305496, "learning_rate": 0.0001767415859363434, "loss": 1.8071, "step": 372 }, { "epoch": 0.68, "grad_norm": 0.04601633548736572, "learning_rate": 0.00017661817421642804, "loss": 1.7594, "step": 373 }, { "epoch": 0.68, "grad_norm": 0.05098440870642662, "learning_rate": 0.00017649447927430362, "loss": 1.6524, "step": 374 }, { "epoch": 0.68, "grad_norm": 0.04978582262992859, "learning_rate": 0.00017637050156721346, "loss": 1.7448, "step": 375 }, { "epoch": 0.69, "grad_norm": 0.05097389221191406, "learning_rate": 0.00017624624155344626, "loss": 1.7362, "step": 376 }, { "epoch": 0.69, "grad_norm": 0.05258944630622864, "learning_rate": 0.00017612169969233424, "loss": 1.7033, "step": 377 }, { "epoch": 0.69, "grad_norm": 0.05384654179215431, "learning_rate": 0.0001759968764442515, "loss": 1.6349, "step": 378 }, { "epoch": 0.69, "grad_norm": 0.047803860157728195, "learning_rate": 0.00017587177227061226, "loss": 1.6655, "step": 379 }, { "epoch": 0.69, "grad_norm": 0.04812454432249069, "learning_rate": 0.00017574638763386916, "loss": 1.7064, "step": 380 }, { "epoch": 0.69, "grad_norm": 0.04860275238752365, "learning_rate": 0.00017562072299751163, "loss": 1.6648, "step": 381 }, { "epoch": 0.7, "grad_norm": 0.049836620688438416, "learning_rate": 0.00017549477882606418, "loss": 1.6957, "step": 382 }, { "epoch": 0.7, "grad_norm": 0.05114325135946274, "learning_rate": 0.00017536855558508458, "loss": 1.6257, "step": 383 }, { "epoch": 0.7, "grad_norm": 0.054609425365924835, "learning_rate": 0.00017524205374116214, "loss": 1.6854, "step": 384 }, { "epoch": 0.7, "grad_norm": 0.04757620766758919, "learning_rate": 0.00017511527376191618, "loss": 1.7425, "step": 385 }, { "epoch": 0.7, "grad_norm": 0.05384545028209686, "learning_rate": 0.00017498821611599397, "loss": 1.712, "step": 386 }, { "epoch": 0.71, "grad_norm": 0.04726232588291168, "learning_rate": 0.00017486088127306932, "loss": 1.701, "step": 387 }, { "epoch": 0.71, "grad_norm": 0.04885297268629074, "learning_rate": 0.0001747332697038407, "loss": 1.7227, "step": 388 }, { "epoch": 0.71, "grad_norm": 0.04793693870306015, "learning_rate": 0.00017460538188002946, "loss": 1.7058, "step": 389 }, { "epoch": 0.71, "grad_norm": 0.04942973330616951, "learning_rate": 0.0001744772182743782, "loss": 1.7443, "step": 390 }, { "epoch": 0.71, "grad_norm": 0.05246872082352638, "learning_rate": 0.00017434877936064886, "loss": 1.6807, "step": 391 }, { "epoch": 0.71, "grad_norm": 0.04894121363759041, "learning_rate": 0.0001742200656136212, "loss": 1.7963, "step": 392 }, { "epoch": 0.72, "grad_norm": 0.05082324892282486, "learning_rate": 0.00017409107750909078, "loss": 1.7024, "step": 393 }, { "epoch": 0.72, "grad_norm": 0.04718152433633804, "learning_rate": 0.00017396181552386741, "loss": 1.711, "step": 394 }, { "epoch": 0.72, "grad_norm": 0.05174902826547623, "learning_rate": 0.00017383228013577331, "loss": 1.7362, "step": 395 }, { "epoch": 0.72, "grad_norm": 0.048003047704696655, "learning_rate": 0.0001737024718236413, "loss": 1.6944, "step": 396 }, { "epoch": 0.72, "grad_norm": 0.0462164506316185, "learning_rate": 0.00017357239106731317, "loss": 1.7297, "step": 397 }, { "epoch": 0.73, "grad_norm": 0.04808316007256508, "learning_rate": 0.0001734420383476377, "loss": 1.6971, "step": 398 }, { "epoch": 0.73, "grad_norm": 0.05553476884961128, "learning_rate": 0.00017331141414646904, "loss": 1.7262, "step": 399 }, { "epoch": 0.73, "grad_norm": 0.046341411769390106, "learning_rate": 0.00017318051894666487, "loss": 1.7135, "step": 400 }, { "epoch": 0.73, "grad_norm": 0.048155754804611206, "learning_rate": 0.00017304935323208466, "loss": 1.7377, "step": 401 }, { "epoch": 0.73, "grad_norm": 0.05066389963030815, "learning_rate": 0.00017291791748758785, "loss": 1.6516, "step": 402 }, { "epoch": 0.73, "grad_norm": 0.05046610161662102, "learning_rate": 0.000172786212199032, "loss": 1.7536, "step": 403 }, { "epoch": 0.74, "grad_norm": 0.0542440302670002, "learning_rate": 0.00017265423785327107, "loss": 1.7857, "step": 404 }, { "epoch": 0.74, "grad_norm": 0.04833053797483444, "learning_rate": 0.0001725219949381537, "loss": 1.7594, "step": 405 }, { "epoch": 0.74, "grad_norm": 0.047335654497146606, "learning_rate": 0.00017238948394252115, "loss": 1.7495, "step": 406 }, { "epoch": 0.74, "grad_norm": 0.04961543157696724, "learning_rate": 0.00017225670535620576, "loss": 1.7201, "step": 407 }, { "epoch": 0.74, "grad_norm": 0.04761854186654091, "learning_rate": 0.00017212365967002893, "loss": 1.7522, "step": 408 }, { "epoch": 0.75, "grad_norm": 0.05010442063212395, "learning_rate": 0.0001719903473757996, "loss": 1.7535, "step": 409 }, { "epoch": 0.75, "grad_norm": 0.049323149025440216, "learning_rate": 0.000171856768966312, "loss": 1.6984, "step": 410 }, { "epoch": 0.75, "grad_norm": 0.08661342412233353, "learning_rate": 0.0001717229249353442, "loss": 1.7182, "step": 411 }, { "epoch": 0.75, "eval_loss": 1.724851131439209, "eval_runtime": 76.3068, "eval_samples_per_second": 65.525, "eval_steps_per_second": 16.381, "step": 411 }, { "epoch": 0.75, "grad_norm": 0.05118868127465248, "learning_rate": 0.00017158881577765612, "loss": 1.683, "step": 412 }, { "epoch": 0.75, "grad_norm": 0.053089968860149384, "learning_rate": 0.00017145444198898776, "loss": 1.7162, "step": 413 }, { "epoch": 0.75, "grad_norm": 0.05191902816295624, "learning_rate": 0.0001713198040660573, "loss": 1.7223, "step": 414 }, { "epoch": 0.76, "grad_norm": 0.05995416268706322, "learning_rate": 0.00017118490250655932, "loss": 1.7148, "step": 415 }, { "epoch": 0.76, "grad_norm": 0.04749016463756561, "learning_rate": 0.00017104973780916294, "loss": 1.7364, "step": 416 }, { "epoch": 0.76, "grad_norm": 0.047870930284261703, "learning_rate": 0.00017091431047351, "loss": 1.7607, "step": 417 }, { "epoch": 0.76, "grad_norm": 0.04802364483475685, "learning_rate": 0.00017077862100021318, "loss": 1.6957, "step": 418 }, { "epoch": 0.76, "grad_norm": 0.04796374961733818, "learning_rate": 0.00017064266989085412, "loss": 1.6972, "step": 419 }, { "epoch": 0.77, "grad_norm": 0.048874564468860626, "learning_rate": 0.00017050645764798164, "loss": 1.736, "step": 420 }, { "epoch": 0.77, "grad_norm": 0.052477337419986725, "learning_rate": 0.00017036998477510992, "loss": 1.7447, "step": 421 }, { "epoch": 0.77, "grad_norm": 0.049993280321359634, "learning_rate": 0.00017023325177671647, "loss": 1.7635, "step": 422 }, { "epoch": 0.77, "grad_norm": 0.09700744599103928, "learning_rate": 0.00017009625915824037, "loss": 1.7402, "step": 423 }, { "epoch": 0.77, "grad_norm": 0.048865802586078644, "learning_rate": 0.0001699590074260805, "loss": 1.7229, "step": 424 }, { "epoch": 0.77, "grad_norm": 0.04994821920990944, "learning_rate": 0.00016982149708759343, "loss": 1.672, "step": 425 }, { "epoch": 0.78, "grad_norm": 0.05008814111351967, "learning_rate": 0.00016968372865109176, "loss": 1.7338, "step": 426 }, { "epoch": 0.78, "grad_norm": 0.04830687865614891, "learning_rate": 0.00016954570262584214, "loss": 1.7177, "step": 427 }, { "epoch": 0.78, "grad_norm": 0.04781452193856239, "learning_rate": 0.0001694074195220634, "loss": 1.7628, "step": 428 }, { "epoch": 0.78, "grad_norm": 0.04739667847752571, "learning_rate": 0.00016926887985092468, "loss": 1.7107, "step": 429 }, { "epoch": 0.78, "grad_norm": 0.0481286458671093, "learning_rate": 0.00016913008412454357, "loss": 1.7646, "step": 430 }, { "epoch": 0.79, "grad_norm": 0.06283537298440933, "learning_rate": 0.0001689910328559841, "loss": 1.6896, "step": 431 }, { "epoch": 0.79, "grad_norm": 0.04944480583071709, "learning_rate": 0.00016885172655925495, "loss": 1.6931, "step": 432 }, { "epoch": 0.79, "grad_norm": 0.05051645264029503, "learning_rate": 0.00016871216574930754, "loss": 1.7752, "step": 433 }, { "epoch": 0.79, "grad_norm": 0.05406402051448822, "learning_rate": 0.0001685723509420341, "loss": 1.7203, "step": 434 }, { "epoch": 0.79, "grad_norm": 0.0995137020945549, "learning_rate": 0.00016843228265426584, "loss": 1.6454, "step": 435 }, { "epoch": 0.79, "grad_norm": 0.05356389284133911, "learning_rate": 0.00016829196140377085, "loss": 1.7327, "step": 436 }, { "epoch": 0.8, "grad_norm": 0.04902141913771629, "learning_rate": 0.0001681513877092523, "loss": 1.7262, "step": 437 }, { "epoch": 0.8, "grad_norm": 0.047820378094911575, "learning_rate": 0.00016801056209034672, "loss": 1.7294, "step": 438 }, { "epoch": 0.8, "grad_norm": 0.048359643667936325, "learning_rate": 0.00016786948506762164, "loss": 1.6959, "step": 439 }, { "epoch": 0.8, "grad_norm": 0.04830753803253174, "learning_rate": 0.00016772815716257412, "loss": 1.7714, "step": 440 }, { "epoch": 0.8, "grad_norm": 0.05318046733736992, "learning_rate": 0.0001675865788976285, "loss": 1.7325, "step": 441 }, { "epoch": 0.81, "grad_norm": 0.04992082715034485, "learning_rate": 0.0001674447507961346, "loss": 1.7866, "step": 442 }, { "epoch": 0.81, "grad_norm": 0.05253741890192032, "learning_rate": 0.0001673026733823658, "loss": 1.7273, "step": 443 }, { "epoch": 0.81, "grad_norm": 0.05121272802352905, "learning_rate": 0.00016716034718151706, "loss": 1.7063, "step": 444 }, { "epoch": 0.81, "grad_norm": 0.04715156927704811, "learning_rate": 0.000167017772719703, "loss": 1.7575, "step": 445 }, { "epoch": 0.81, "grad_norm": 0.05717930197715759, "learning_rate": 0.00016687495052395595, "loss": 1.7835, "step": 446 }, { "epoch": 0.81, "grad_norm": 0.04992460459470749, "learning_rate": 0.00016673188112222394, "loss": 1.7218, "step": 447 }, { "epoch": 0.82, "grad_norm": 0.0481155663728714, "learning_rate": 0.0001665885650433689, "loss": 1.7269, "step": 448 }, { "epoch": 0.82, "grad_norm": 0.0485762394964695, "learning_rate": 0.00016644500281716456, "loss": 1.6857, "step": 449 }, { "epoch": 0.82, "grad_norm": 0.04729575663805008, "learning_rate": 0.00016630119497429457, "loss": 1.7208, "step": 450 }, { "epoch": 0.82, "grad_norm": 0.051819782704114914, "learning_rate": 0.00016615714204635043, "loss": 1.7117, "step": 451 }, { "epoch": 0.82, "grad_norm": 0.052782051265239716, "learning_rate": 0.0001660128445658297, "loss": 1.7811, "step": 452 }, { "epoch": 0.83, "grad_norm": 0.05251288414001465, "learning_rate": 0.00016586830306613393, "loss": 1.7517, "step": 453 }, { "epoch": 0.83, "grad_norm": 0.047806352376937866, "learning_rate": 0.00016572351808156666, "loss": 1.7132, "step": 454 }, { "epoch": 0.83, "grad_norm": 0.05114049091935158, "learning_rate": 0.0001655784901473315, "loss": 1.7729, "step": 455 }, { "epoch": 0.83, "grad_norm": 0.04811178147792816, "learning_rate": 0.00016543321979953007, "loss": 1.7855, "step": 456 }, { "epoch": 0.83, "grad_norm": 0.05107167363166809, "learning_rate": 0.00016528770757516027, "loss": 1.7331, "step": 457 }, { "epoch": 0.84, "grad_norm": 0.04712466895580292, "learning_rate": 0.00016514195401211388, "loss": 1.7048, "step": 458 }, { "epoch": 0.84, "grad_norm": 0.05438878387212753, "learning_rate": 0.0001649959596491749, "loss": 1.753, "step": 459 }, { "epoch": 0.84, "grad_norm": 0.04884348064661026, "learning_rate": 0.00016484972502601753, "loss": 1.6734, "step": 460 }, { "epoch": 0.84, "grad_norm": 0.0536276139318943, "learning_rate": 0.00016470325068320392, "loss": 1.711, "step": 461 }, { "epoch": 0.84, "grad_norm": 0.05346493422985077, "learning_rate": 0.00016455653716218252, "loss": 1.7366, "step": 462 }, { "epoch": 0.84, "grad_norm": 0.05044522508978844, "learning_rate": 0.0001644095850052858, "loss": 1.7269, "step": 463 }, { "epoch": 0.85, "grad_norm": 0.05273488536477089, "learning_rate": 0.00016426239475572852, "loss": 1.7586, "step": 464 }, { "epoch": 0.85, "grad_norm": 0.053452517837285995, "learning_rate": 0.0001641149669576053, "loss": 1.7379, "step": 465 }, { "epoch": 0.85, "grad_norm": 0.047611016780138016, "learning_rate": 0.00016396730215588915, "loss": 1.7471, "step": 466 }, { "epoch": 0.85, "grad_norm": 0.05317235738039017, "learning_rate": 0.00016381940089642893, "loss": 1.6925, "step": 467 }, { "epoch": 0.85, "grad_norm": 0.049223560839891434, "learning_rate": 0.00016367126372594774, "loss": 1.7229, "step": 468 }, { "epoch": 0.86, "grad_norm": 0.047821756452322006, "learning_rate": 0.0001635228911920407, "loss": 1.7484, "step": 469 }, { "epoch": 0.86, "grad_norm": 0.05013042315840721, "learning_rate": 0.00016337428384317288, "loss": 1.7435, "step": 470 }, { "epoch": 0.86, "grad_norm": 0.04820725694298744, "learning_rate": 0.00016322544222867742, "loss": 1.7594, "step": 471 }, { "epoch": 0.86, "grad_norm": 0.04791193827986717, "learning_rate": 0.00016307636689875347, "loss": 1.644, "step": 472 }, { "epoch": 0.86, "grad_norm": 0.04905365779995918, "learning_rate": 0.00016292705840446404, "loss": 1.7144, "step": 473 }, { "epoch": 0.86, "grad_norm": 0.04875028133392334, "learning_rate": 0.00016277751729773407, "loss": 1.712, "step": 474 }, { "epoch": 0.87, "grad_norm": 0.05170164629817009, "learning_rate": 0.0001626277441313484, "loss": 1.7367, "step": 475 }, { "epoch": 0.87, "grad_norm": 0.05205371975898743, "learning_rate": 0.00016247773945894962, "loss": 1.689, "step": 476 }, { "epoch": 0.87, "grad_norm": 0.0485403798520565, "learning_rate": 0.00016232750383503617, "loss": 1.706, "step": 477 }, { "epoch": 0.87, "grad_norm": 0.0538201630115509, "learning_rate": 0.0001621770378149601, "loss": 1.7284, "step": 478 }, { "epoch": 0.87, "grad_norm": 0.04828377440571785, "learning_rate": 0.00016202634195492524, "loss": 1.661, "step": 479 }, { "epoch": 0.88, "grad_norm": 0.050310611724853516, "learning_rate": 0.000161875416811985, "loss": 1.6852, "step": 480 }, { "epoch": 0.88, "grad_norm": 0.050804853439331055, "learning_rate": 0.00016172426294404032, "loss": 1.7358, "step": 481 }, { "epoch": 0.88, "grad_norm": 0.051962971687316895, "learning_rate": 0.00016157288090983763, "loss": 1.6692, "step": 482 }, { "epoch": 0.88, "grad_norm": 0.05179814621806145, "learning_rate": 0.0001614212712689668, "loss": 1.6983, "step": 483 }, { "epoch": 0.88, "grad_norm": 0.05398216098546982, "learning_rate": 0.00016126943458185907, "loss": 1.7261, "step": 484 }, { "epoch": 0.88, "grad_norm": 0.049869704991579056, "learning_rate": 0.00016111737140978494, "loss": 1.6951, "step": 485 }, { "epoch": 0.89, "grad_norm": 0.048107776790857315, "learning_rate": 0.00016096508231485217, "loss": 1.6941, "step": 486 }, { "epoch": 0.89, "grad_norm": 0.05527656897902489, "learning_rate": 0.00016081256786000357, "loss": 1.7054, "step": 487 }, { "epoch": 0.89, "grad_norm": 0.05169270187616348, "learning_rate": 0.00016065982860901504, "loss": 1.7307, "step": 488 }, { "epoch": 0.89, "grad_norm": 0.04972197115421295, "learning_rate": 0.00016050686512649354, "loss": 1.6955, "step": 489 }, { "epoch": 0.89, "grad_norm": 0.05033208429813385, "learning_rate": 0.00016035367797787476, "loss": 1.7013, "step": 490 }, { "epoch": 0.9, "grad_norm": 0.05073223263025284, "learning_rate": 0.00016020026772942125, "loss": 1.6831, "step": 491 }, { "epoch": 0.9, "grad_norm": 0.056367356330156326, "learning_rate": 0.00016004663494822028, "loss": 1.6654, "step": 492 }, { "epoch": 0.9, "grad_norm": 0.049483008682727814, "learning_rate": 0.0001598927802021817, "loss": 1.7285, "step": 493 }, { "epoch": 0.9, "grad_norm": 0.052070703357458115, "learning_rate": 0.00015973870406003578, "loss": 1.7948, "step": 494 }, { "epoch": 0.9, "grad_norm": 0.05687413364648819, "learning_rate": 0.0001595844070913314, "loss": 1.7336, "step": 495 }, { "epoch": 0.9, "grad_norm": 0.048987727612257004, "learning_rate": 0.00015942988986643352, "loss": 1.6661, "step": 496 }, { "epoch": 0.91, "grad_norm": 0.05027730017900467, "learning_rate": 0.00015927515295652143, "loss": 1.7364, "step": 497 }, { "epoch": 0.91, "grad_norm": 0.048406291753053665, "learning_rate": 0.00015912019693358636, "loss": 1.6419, "step": 498 }, { "epoch": 0.91, "grad_norm": 0.05071192979812622, "learning_rate": 0.00015896502237042963, "loss": 1.6301, "step": 499 }, { "epoch": 0.91, "grad_norm": 0.05111885070800781, "learning_rate": 0.00015880962984066036, "loss": 1.7112, "step": 500 }, { "epoch": 0.91, "grad_norm": 0.06297910958528519, "learning_rate": 0.0001586540199186933, "loss": 1.7438, "step": 501 }, { "epoch": 0.92, "grad_norm": 0.04950469359755516, "learning_rate": 0.00015849819317974694, "loss": 1.6837, "step": 502 }, { "epoch": 0.92, "grad_norm": 0.04900701716542244, "learning_rate": 0.0001583421501998412, "loss": 1.7432, "step": 503 }, { "epoch": 0.92, "grad_norm": 0.04949019104242325, "learning_rate": 0.0001581858915557953, "loss": 1.688, "step": 504 }, { "epoch": 0.92, "grad_norm": 0.05047097057104111, "learning_rate": 0.00015802941782522569, "loss": 1.7256, "step": 505 }, { "epoch": 0.92, "grad_norm": 0.04921870306134224, "learning_rate": 0.0001578727295865439, "loss": 1.7723, "step": 506 }, { "epoch": 0.92, "grad_norm": 0.04841122031211853, "learning_rate": 0.0001577158274189544, "loss": 1.71, "step": 507 }, { "epoch": 0.93, "grad_norm": 0.04886234924197197, "learning_rate": 0.00015755871190245251, "loss": 1.6622, "step": 508 }, { "epoch": 0.93, "grad_norm": 0.04966573417186737, "learning_rate": 0.00015740138361782207, "loss": 1.7357, "step": 509 }, { "epoch": 0.93, "grad_norm": 0.050070296972990036, "learning_rate": 0.0001572438431466336, "loss": 1.6803, "step": 510 }, { "epoch": 0.93, "grad_norm": 0.054121073335409164, "learning_rate": 0.00015708609107124177, "loss": 1.7659, "step": 511 }, { "epoch": 0.93, "grad_norm": 0.05084529519081116, "learning_rate": 0.00015692812797478368, "loss": 1.6943, "step": 512 }, { "epoch": 0.94, "grad_norm": 0.056926507502794266, "learning_rate": 0.0001567699544411763, "loss": 1.6562, "step": 513 }, { "epoch": 0.94, "grad_norm": 0.05053721368312836, "learning_rate": 0.00015661157105511457, "loss": 1.7624, "step": 514 }, { "epoch": 0.94, "grad_norm": 0.048727016896009445, "learning_rate": 0.00015645297840206915, "loss": 1.7364, "step": 515 }, { "epoch": 0.94, "grad_norm": 0.051376283168792725, "learning_rate": 0.00015629417706828423, "loss": 1.699, "step": 516 }, { "epoch": 0.94, "grad_norm": 0.05029591917991638, "learning_rate": 0.00015613516764077548, "loss": 1.6972, "step": 517 }, { "epoch": 0.94, "grad_norm": 0.053968969732522964, "learning_rate": 0.00015597595070732765, "loss": 1.7128, "step": 518 }, { "epoch": 0.95, "grad_norm": 0.050694871693849564, "learning_rate": 0.00015581652685649276, "loss": 1.7681, "step": 519 }, { "epoch": 0.95, "grad_norm": 0.052369993180036545, "learning_rate": 0.00015565689667758746, "loss": 1.7321, "step": 520 }, { "epoch": 0.95, "grad_norm": 0.04850650206208229, "learning_rate": 0.00015549706076069128, "loss": 1.7162, "step": 521 }, { "epoch": 0.95, "grad_norm": 0.04979635775089264, "learning_rate": 0.00015533701969664424, "loss": 1.7429, "step": 522 }, { "epoch": 0.95, "grad_norm": 0.04920853301882744, "learning_rate": 0.0001551767740770446, "loss": 1.7103, "step": 523 }, { "epoch": 0.96, "grad_norm": 0.05081456899642944, "learning_rate": 0.0001550163244942469, "loss": 1.7781, "step": 524 }, { "epoch": 0.96, "grad_norm": 0.050754062831401825, "learning_rate": 0.00015485567154135952, "loss": 1.7496, "step": 525 }, { "epoch": 0.96, "grad_norm": 0.050315603613853455, "learning_rate": 0.00015469481581224272, "loss": 1.7303, "step": 526 }, { "epoch": 0.96, "grad_norm": 0.05050061643123627, "learning_rate": 0.00015453375790150617, "loss": 1.679, "step": 527 }, { "epoch": 0.96, "grad_norm": 0.06212810054421425, "learning_rate": 0.00015437249840450715, "loss": 1.713, "step": 528 }, { "epoch": 0.96, "grad_norm": 0.050966355949640274, "learning_rate": 0.00015421103791734786, "loss": 1.7551, "step": 529 }, { "epoch": 0.97, "grad_norm": 0.04892159253358841, "learning_rate": 0.00015404937703687363, "loss": 1.6758, "step": 530 }, { "epoch": 0.97, "grad_norm": 0.05551762133836746, "learning_rate": 0.00015388751636067052, "loss": 1.703, "step": 531 }, { "epoch": 0.97, "grad_norm": 0.0516047477722168, "learning_rate": 0.00015372545648706306, "loss": 1.7407, "step": 532 }, { "epoch": 0.97, "grad_norm": 0.05094458907842636, "learning_rate": 0.0001535631980151123, "loss": 1.6534, "step": 533 }, { "epoch": 0.97, "grad_norm": 0.05045678839087486, "learning_rate": 0.00015340074154461316, "loss": 1.7335, "step": 534 }, { "epoch": 0.98, "grad_norm": 0.05067756026983261, "learning_rate": 0.00015323808767609277, "loss": 1.7169, "step": 535 }, { "epoch": 0.98, "grad_norm": 0.05005278438329697, "learning_rate": 0.00015307523701080768, "loss": 1.7778, "step": 536 }, { "epoch": 0.98, "grad_norm": 0.04952746629714966, "learning_rate": 0.0001529121901507421, "loss": 1.7199, "step": 537 }, { "epoch": 0.98, "grad_norm": 0.04711218178272247, "learning_rate": 0.00015274894769860538, "loss": 1.734, "step": 538 }, { "epoch": 0.98, "grad_norm": 0.05313078686594963, "learning_rate": 0.0001525855102578299, "loss": 1.7733, "step": 539 }, { "epoch": 0.98, "grad_norm": 0.04977120831608772, "learning_rate": 0.0001524218784325688, "loss": 1.731, "step": 540 }, { "epoch": 0.99, "grad_norm": 0.05076899752020836, "learning_rate": 0.00015225805282769383, "loss": 1.7277, "step": 541 }, { "epoch": 0.99, "grad_norm": 0.049164701253175735, "learning_rate": 0.00015209403404879303, "loss": 1.7032, "step": 542 }, { "epoch": 0.99, "grad_norm": 0.0488349013030529, "learning_rate": 0.00015192982270216854, "loss": 1.765, "step": 543 }, { "epoch": 0.99, "grad_norm": 0.04831582307815552, "learning_rate": 0.0001517654193948343, "loss": 1.7548, "step": 544 }, { "epoch": 0.99, "grad_norm": 0.052940741181373596, "learning_rate": 0.00015160082473451378, "loss": 1.7209, "step": 545 }, { "epoch": 1.0, "grad_norm": 0.056908875703811646, "learning_rate": 0.00015143603932963795, "loss": 1.6537, "step": 546 }, { "epoch": 1.0, "grad_norm": 0.0509711354970932, "learning_rate": 0.00015127106378934273, "loss": 1.7151, "step": 547 }, { "epoch": 1.0, "grad_norm": 0.04795239865779877, "learning_rate": 0.000151105898723467, "loss": 1.743, "step": 548 }, { "epoch": 1.0, "eval_loss": 1.7236659526824951, "eval_runtime": 76.6784, "eval_samples_per_second": 65.207, "eval_steps_per_second": 16.302, "step": 548 }, { "epoch": 1.0, "grad_norm": 0.05828290060162544, "learning_rate": 0.00015094054474255007, "loss": 1.7014, "step": 549 }, { "epoch": 1.0, "grad_norm": 0.04827438294887543, "learning_rate": 0.00015077500245782978, "loss": 1.7124, "step": 550 }, { "epoch": 1.0, "grad_norm": 0.04962700232863426, "learning_rate": 0.0001506092724812399, "loss": 1.7496, "step": 551 }, { "epoch": 1.01, "grad_norm": 0.05015181377530098, "learning_rate": 0.00015044335542540804, "loss": 1.6653, "step": 552 }, { "epoch": 1.01, "grad_norm": 0.07125337421894073, "learning_rate": 0.0001502772519036534, "loss": 1.6938, "step": 553 }, { "epoch": 1.01, "grad_norm": 0.05031266435980797, "learning_rate": 0.0001501109625299844, "loss": 1.7782, "step": 554 }, { "epoch": 1.01, "grad_norm": 0.0487028993666172, "learning_rate": 0.00014994448791909656, "loss": 1.7202, "step": 555 }, { "epoch": 1.0, "grad_norm": 0.06726840883493423, "learning_rate": 0.00014977782868636999, "loss": 1.7504, "step": 556 }, { "epoch": 1.0, "grad_norm": 0.06244590878486633, "learning_rate": 0.00014961098544786743, "loss": 1.6834, "step": 557 }, { "epoch": 1.01, "grad_norm": 0.04934772849082947, "learning_rate": 0.00014944395882033167, "loss": 1.6822, "step": 558 }, { "epoch": 1.01, "grad_norm": 0.050311822444200516, "learning_rate": 0.00014927674942118345, "loss": 1.747, "step": 559 }, { "epoch": 1.01, "grad_norm": 0.051862068474292755, "learning_rate": 0.00014910935786851919, "loss": 1.7355, "step": 560 }, { "epoch": 1.01, "grad_norm": 0.049238841980695724, "learning_rate": 0.00014894178478110857, "loss": 1.6973, "step": 561 }, { "epoch": 1.01, "grad_norm": 0.05033009499311447, "learning_rate": 0.00014877403077839235, "loss": 1.6718, "step": 562 }, { "epoch": 1.01, "grad_norm": 0.04922296851873398, "learning_rate": 0.00014860609648048004, "loss": 1.7236, "step": 563 }, { "epoch": 1.02, "grad_norm": 0.05257139354944229, "learning_rate": 0.0001484379825081476, "loss": 1.6868, "step": 564 }, { "epoch": 1.02, "grad_norm": 0.05213212966918945, "learning_rate": 0.0001482696894828353, "loss": 1.726, "step": 565 }, { "epoch": 1.02, "grad_norm": 0.053737424314022064, "learning_rate": 0.00014810121802664512, "loss": 1.7046, "step": 566 }, { "epoch": 1.02, "grad_norm": 0.054125770926475525, "learning_rate": 0.0001479325687623386, "loss": 1.6106, "step": 567 }, { "epoch": 1.02, "grad_norm": 0.051876723766326904, "learning_rate": 0.00014776374231333477, "loss": 1.7354, "step": 568 }, { "epoch": 1.03, "grad_norm": 0.050595056265592575, "learning_rate": 0.00014759473930370736, "loss": 1.6947, "step": 569 }, { "epoch": 1.03, "grad_norm": 0.06360866129398346, "learning_rate": 0.00014742556035818297, "loss": 1.7379, "step": 570 }, { "epoch": 1.03, "grad_norm": 0.05476611480116844, "learning_rate": 0.0001472562061021385, "loss": 1.6392, "step": 571 }, { "epoch": 1.03, "grad_norm": 0.051338374614715576, "learning_rate": 0.0001470866771615988, "loss": 1.687, "step": 572 }, { "epoch": 1.03, "grad_norm": 0.05180288851261139, "learning_rate": 0.00014691697416323454, "loss": 1.6942, "step": 573 }, { "epoch": 1.03, "grad_norm": 0.05175211653113365, "learning_rate": 0.00014674709773435983, "loss": 1.6648, "step": 574 }, { "epoch": 1.04, "grad_norm": 0.055275119841098785, "learning_rate": 0.00014657704850292976, "loss": 1.7311, "step": 575 }, { "epoch": 1.04, "grad_norm": 0.053508460521698, "learning_rate": 0.00014640682709753832, "loss": 1.7118, "step": 576 }, { "epoch": 1.04, "grad_norm": 0.05283378064632416, "learning_rate": 0.00014623643414741585, "loss": 1.6675, "step": 577 }, { "epoch": 1.04, "grad_norm": 0.05684136226773262, "learning_rate": 0.00014606587028242682, "loss": 1.709, "step": 578 }, { "epoch": 1.04, "grad_norm": 0.0515415295958519, "learning_rate": 0.0001458951361330676, "loss": 1.653, "step": 579 }, { "epoch": 1.05, "grad_norm": 0.052131347358226776, "learning_rate": 0.00014572423233046386, "loss": 1.6497, "step": 580 }, { "epoch": 1.05, "grad_norm": 0.05229787901043892, "learning_rate": 0.00014555315950636854, "loss": 1.6209, "step": 581 }, { "epoch": 1.05, "grad_norm": 0.058796849101781845, "learning_rate": 0.00014538191829315927, "loss": 1.6907, "step": 582 }, { "epoch": 1.05, "grad_norm": 0.0535275973379612, "learning_rate": 0.00014521050932383625, "loss": 1.6765, "step": 583 }, { "epoch": 1.05, "grad_norm": 0.06131954491138458, "learning_rate": 0.00014503893323201966, "loss": 1.6963, "step": 584 }, { "epoch": 1.05, "grad_norm": 0.05318441987037659, "learning_rate": 0.00014486719065194757, "loss": 1.6693, "step": 585 }, { "epoch": 1.06, "grad_norm": 0.053547151386737823, "learning_rate": 0.00014469528221847344, "loss": 1.6265, "step": 586 }, { "epoch": 1.06, "grad_norm": 0.05694759264588356, "learning_rate": 0.00014452320856706382, "loss": 1.6998, "step": 587 }, { "epoch": 1.06, "grad_norm": 0.053848620504140854, "learning_rate": 0.00014435097033379596, "loss": 1.7248, "step": 588 }, { "epoch": 1.06, "grad_norm": 0.05272265151143074, "learning_rate": 0.00014417856815535554, "loss": 1.6973, "step": 589 }, { "epoch": 1.06, "grad_norm": 0.05548195540904999, "learning_rate": 0.00014400600266903423, "loss": 1.6912, "step": 590 }, { "epoch": 1.07, "grad_norm": 0.05391455814242363, "learning_rate": 0.00014383327451272744, "loss": 1.6507, "step": 591 }, { "epoch": 1.07, "grad_norm": 0.05697217211127281, "learning_rate": 0.00014366038432493181, "loss": 1.7277, "step": 592 }, { "epoch": 1.07, "grad_norm": 0.054713811725378036, "learning_rate": 0.000143487332744743, "loss": 1.7225, "step": 593 }, { "epoch": 1.07, "grad_norm": 0.05515265092253685, "learning_rate": 0.00014331412041185322, "loss": 1.6838, "step": 594 }, { "epoch": 1.07, "grad_norm": 0.054941218346357346, "learning_rate": 0.00014314074796654896, "loss": 1.6913, "step": 595 }, { "epoch": 1.07, "grad_norm": 0.05448353663086891, "learning_rate": 0.0001429672160497085, "loss": 1.6685, "step": 596 }, { "epoch": 1.08, "grad_norm": 0.058499112725257874, "learning_rate": 0.0001427935253027997, "loss": 1.6637, "step": 597 }, { "epoch": 1.08, "grad_norm": 0.0628763735294342, "learning_rate": 0.00014261967636787747, "loss": 1.7139, "step": 598 }, { "epoch": 1.08, "grad_norm": 0.05447819083929062, "learning_rate": 0.00014244566988758152, "loss": 1.6984, "step": 599 }, { "epoch": 1.08, "grad_norm": 0.05434316396713257, "learning_rate": 0.0001422715065051339, "loss": 1.6688, "step": 600 }, { "epoch": 1.08, "grad_norm": 0.052557747811079025, "learning_rate": 0.00014209718686433663, "loss": 1.7169, "step": 601 }, { "epoch": 1.09, "grad_norm": 0.054510824382305145, "learning_rate": 0.00014192271160956942, "loss": 1.6186, "step": 602 }, { "epoch": 1.09, "grad_norm": 0.0586363822221756, "learning_rate": 0.00014174808138578713, "loss": 1.7364, "step": 603 }, { "epoch": 1.09, "grad_norm": 0.05653434619307518, "learning_rate": 0.0001415732968385176, "loss": 1.77, "step": 604 }, { "epoch": 1.09, "grad_norm": 0.052821431308984756, "learning_rate": 0.00014139835861385892, "loss": 1.6599, "step": 605 }, { "epoch": 1.09, "grad_norm": 0.054437246173620224, "learning_rate": 0.00014122326735847748, "loss": 1.7026, "step": 606 }, { "epoch": 1.09, "grad_norm": 0.056837234646081924, "learning_rate": 0.00014104802371960523, "loss": 1.6475, "step": 607 }, { "epoch": 1.1, "grad_norm": 0.06032341718673706, "learning_rate": 0.0001408726283450374, "loss": 1.7482, "step": 608 }, { "epoch": 1.1, "grad_norm": 0.05582507699728012, "learning_rate": 0.00014069708188313017, "loss": 1.7046, "step": 609 }, { "epoch": 1.1, "grad_norm": 0.05785200744867325, "learning_rate": 0.00014052138498279828, "loss": 1.7234, "step": 610 }, { "epoch": 1.1, "grad_norm": 0.05540376156568527, "learning_rate": 0.00014034553829351236, "loss": 1.7157, "step": 611 }, { "epoch": 1.1, "grad_norm": 0.05743914842605591, "learning_rate": 0.00014016954246529696, "loss": 1.7548, "step": 612 }, { "epoch": 1.11, "grad_norm": 0.05496819317340851, "learning_rate": 0.00013999339814872784, "loss": 1.6913, "step": 613 }, { "epoch": 1.11, "grad_norm": 0.05739595368504524, "learning_rate": 0.00013981710599492964, "loss": 1.7232, "step": 614 }, { "epoch": 1.11, "grad_norm": 0.05653569847345352, "learning_rate": 0.00013964066665557348, "loss": 1.6953, "step": 615 }, { "epoch": 1.11, "grad_norm": 0.05570907145738602, "learning_rate": 0.00013946408078287462, "loss": 1.6858, "step": 616 }, { "epoch": 1.11, "grad_norm": 0.054925207048654556, "learning_rate": 0.00013928734902958996, "loss": 1.6248, "step": 617 }, { "epoch": 1.11, "grad_norm": 0.05743985250592232, "learning_rate": 0.0001391104720490156, "loss": 1.6627, "step": 618 }, { "epoch": 1.12, "grad_norm": 0.05516685172915459, "learning_rate": 0.00013893345049498457, "loss": 1.6714, "step": 619 }, { "epoch": 1.12, "grad_norm": 0.05717911571264267, "learning_rate": 0.0001387562850218642, "loss": 1.7124, "step": 620 }, { "epoch": 1.12, "grad_norm": 0.05529535561800003, "learning_rate": 0.00013857897628455397, "loss": 1.6451, "step": 621 }, { "epoch": 1.12, "grad_norm": 0.05724070221185684, "learning_rate": 0.00013840152493848284, "loss": 1.7274, "step": 622 }, { "epoch": 1.12, "grad_norm": 0.05622214823961258, "learning_rate": 0.0001382239316396069, "loss": 1.6506, "step": 623 }, { "epoch": 1.13, "grad_norm": 0.05893300846219063, "learning_rate": 0.00013804619704440714, "loss": 1.7037, "step": 624 }, { "epoch": 1.13, "grad_norm": 0.05549685284495354, "learning_rate": 0.00013786832180988666, "loss": 1.6894, "step": 625 }, { "epoch": 1.13, "grad_norm": 0.05931728705763817, "learning_rate": 0.00013769030659356853, "loss": 1.7189, "step": 626 }, { "epoch": 1.13, "grad_norm": 0.05465949699282646, "learning_rate": 0.0001375121520534933, "loss": 1.7016, "step": 627 }, { "epoch": 1.13, "grad_norm": 0.056453317403793335, "learning_rate": 0.00013733385884821648, "loss": 1.6711, "step": 628 }, { "epoch": 1.13, "grad_norm": 0.054540056735277176, "learning_rate": 0.00013715542763680623, "loss": 1.6638, "step": 629 }, { "epoch": 1.14, "grad_norm": 0.05919068679213524, "learning_rate": 0.00013697685907884072, "loss": 1.7241, "step": 630 }, { "epoch": 1.14, "grad_norm": 0.05730579420924187, "learning_rate": 0.00013679815383440603, "loss": 1.6946, "step": 631 }, { "epoch": 1.14, "grad_norm": 0.05658195540308952, "learning_rate": 0.00013661931256409325, "loss": 1.7038, "step": 632 }, { "epoch": 1.14, "grad_norm": 0.057528719305992126, "learning_rate": 0.00013644033592899658, "loss": 1.6853, "step": 633 }, { "epoch": 1.14, "grad_norm": 0.062490735203027725, "learning_rate": 0.00013626122459071033, "loss": 1.6733, "step": 634 }, { "epoch": 1.15, "grad_norm": 0.05776170268654823, "learning_rate": 0.00013608197921132696, "loss": 1.7351, "step": 635 }, { "epoch": 1.15, "grad_norm": 0.06134483963251114, "learning_rate": 0.00013590260045343432, "loss": 1.6203, "step": 636 }, { "epoch": 1.15, "grad_norm": 0.061270635575056076, "learning_rate": 0.0001357230889801133, "loss": 1.7268, "step": 637 }, { "epoch": 1.15, "grad_norm": 0.056105442345142365, "learning_rate": 0.00013554344545493535, "loss": 1.7171, "step": 638 }, { "epoch": 1.15, "grad_norm": 0.05647943168878555, "learning_rate": 0.0001353636705419602, "loss": 1.713, "step": 639 }, { "epoch": 1.15, "grad_norm": 0.05758386850357056, "learning_rate": 0.00013518376490573306, "loss": 1.6991, "step": 640 }, { "epoch": 1.16, "grad_norm": 0.05906842276453972, "learning_rate": 0.0001350037292112825, "loss": 1.6387, "step": 641 }, { "epoch": 1.16, "grad_norm": 0.06219753623008728, "learning_rate": 0.00013482356412411781, "loss": 1.7145, "step": 642 }, { "epoch": 1.16, "grad_norm": 0.05719519779086113, "learning_rate": 0.00013464327031022659, "loss": 1.7399, "step": 643 }, { "epoch": 1.16, "grad_norm": 0.08058752119541168, "learning_rate": 0.00013446284843607225, "loss": 1.6275, "step": 644 }, { "epoch": 1.16, "grad_norm": 0.06629724055528641, "learning_rate": 0.00013428229916859167, "loss": 1.6582, "step": 645 }, { "epoch": 1.17, "grad_norm": 0.05791241303086281, "learning_rate": 0.00013410162317519257, "loss": 1.6599, "step": 646 }, { "epoch": 1.17, "grad_norm": 0.06143872067332268, "learning_rate": 0.0001339208211237511, "loss": 1.6634, "step": 647 }, { "epoch": 1.17, "grad_norm": 0.06067274510860443, "learning_rate": 0.00013373989368260948, "loss": 1.6869, "step": 648 }, { "epoch": 1.17, "grad_norm": 0.06446303427219391, "learning_rate": 0.00013355884152057334, "loss": 1.6658, "step": 649 }, { "epoch": 1.17, "grad_norm": 0.05910011753439903, "learning_rate": 0.00013337766530690943, "loss": 1.683, "step": 650 }, { "epoch": 1.17, "grad_norm": 0.06423602253198624, "learning_rate": 0.00013319636571134297, "loss": 1.7058, "step": 651 }, { "epoch": 1.18, "grad_norm": 0.05743340775370598, "learning_rate": 0.00013301494340405535, "loss": 1.6491, "step": 652 }, { "epoch": 1.18, "grad_norm": 0.05755629763007164, "learning_rate": 0.00013283339905568157, "loss": 1.6606, "step": 653 }, { "epoch": 1.18, "grad_norm": 0.05766105651855469, "learning_rate": 0.00013265173333730764, "loss": 1.6855, "step": 654 }, { "epoch": 1.18, "grad_norm": 0.05892917141318321, "learning_rate": 0.00013246994692046836, "loss": 1.6398, "step": 655 }, { "epoch": 1.18, "grad_norm": 0.05860791355371475, "learning_rate": 0.00013228804047714463, "loss": 1.7089, "step": 656 }, { "epoch": 1.19, "grad_norm": 0.059190504252910614, "learning_rate": 0.00013210601467976104, "loss": 1.6703, "step": 657 }, { "epoch": 1.19, "grad_norm": 0.05735331028699875, "learning_rate": 0.0001319238702011834, "loss": 1.73, "step": 658 }, { "epoch": 1.19, "grad_norm": 0.05985163152217865, "learning_rate": 0.0001317416077147162, "loss": 1.6864, "step": 659 }, { "epoch": 1.19, "grad_norm": 0.05826161056756973, "learning_rate": 0.00013155922789410016, "loss": 1.6419, "step": 660 }, { "epoch": 1.19, "grad_norm": 0.059993255883455276, "learning_rate": 0.00013137673141350972, "loss": 1.7027, "step": 661 }, { "epoch": 1.19, "grad_norm": 0.06040223315358162, "learning_rate": 0.00013119411894755063, "loss": 1.7584, "step": 662 }, { "epoch": 1.2, "grad_norm": 0.056883446872234344, "learning_rate": 0.00013101139117125722, "loss": 1.6971, "step": 663 }, { "epoch": 1.2, "grad_norm": 0.05828433483839035, "learning_rate": 0.0001308285487600903, "loss": 1.6797, "step": 664 }, { "epoch": 1.2, "grad_norm": 0.0568573996424675, "learning_rate": 0.0001306455923899342, "loss": 1.6967, "step": 665 }, { "epoch": 1.2, "grad_norm": 0.05763811990618706, "learning_rate": 0.00013046252273709468, "loss": 1.7189, "step": 666 }, { "epoch": 1.2, "grad_norm": 0.05759183317422867, "learning_rate": 0.00013027934047829616, "loss": 1.7293, "step": 667 }, { "epoch": 1.21, "grad_norm": 0.06087080016732216, "learning_rate": 0.00013009604629067933, "loss": 1.7287, "step": 668 }, { "epoch": 1.21, "grad_norm": 0.05685460940003395, "learning_rate": 0.00012991264085179864, "loss": 1.6717, "step": 669 }, { "epoch": 1.21, "grad_norm": 0.06102333217859268, "learning_rate": 0.00012972912483961982, "loss": 1.7911, "step": 670 }, { "epoch": 1.21, "grad_norm": 0.05811255797743797, "learning_rate": 0.00012954549893251724, "loss": 1.7057, "step": 671 }, { "epoch": 1.21, "grad_norm": 0.05935278907418251, "learning_rate": 0.00012936176380927162, "loss": 1.6678, "step": 672 }, { "epoch": 1.21, "grad_norm": 0.06539764255285263, "learning_rate": 0.00012917792014906733, "loss": 1.6305, "step": 673 }, { "epoch": 1.22, "grad_norm": 0.059705205261707306, "learning_rate": 0.00012899396863148995, "loss": 1.7273, "step": 674 }, { "epoch": 1.22, "grad_norm": 0.05784007906913757, "learning_rate": 0.00012880990993652377, "loss": 1.6549, "step": 675 }, { "epoch": 1.22, "grad_norm": 0.07344791293144226, "learning_rate": 0.00012862574474454928, "loss": 1.6809, "step": 676 }, { "epoch": 1.22, "grad_norm": 0.06028100103139877, "learning_rate": 0.00012844147373634066, "loss": 1.6852, "step": 677 }, { "epoch": 1.22, "grad_norm": 0.06096576154232025, "learning_rate": 0.00012825709759306316, "loss": 1.7256, "step": 678 }, { "epoch": 1.23, "grad_norm": 0.060117993503808975, "learning_rate": 0.00012807261699627077, "loss": 1.7094, "step": 679 }, { "epoch": 1.23, "grad_norm": 0.06428851187229156, "learning_rate": 0.0001278880326279035, "loss": 1.6538, "step": 680 }, { "epoch": 1.23, "grad_norm": 0.060511935502290726, "learning_rate": 0.00012770334517028505, "loss": 1.6631, "step": 681 }, { "epoch": 1.23, "grad_norm": 0.05897079408168793, "learning_rate": 0.00012751855530612012, "loss": 1.6732, "step": 682 }, { "epoch": 1.23, "grad_norm": 0.05949567258358002, "learning_rate": 0.00012733366371849201, "loss": 1.6989, "step": 683 }, { "epoch": 1.23, "grad_norm": 0.05985894054174423, "learning_rate": 0.00012714867109086, "loss": 1.6983, "step": 684 }, { "epoch": 1.24, "grad_norm": 0.061160728335380554, "learning_rate": 0.0001269635781070569, "loss": 1.7075, "step": 685 }, { "epoch": 1.24, "eval_loss": 1.7264653444290161, "eval_runtime": 76.4445, "eval_samples_per_second": 65.407, "eval_steps_per_second": 16.352, "step": 685 }, { "epoch": 1.24, "grad_norm": 0.0652250349521637, "learning_rate": 0.00012677838545128647, "loss": 1.6851, "step": 686 }, { "epoch": 1.24, "grad_norm": 0.060404662042856216, "learning_rate": 0.00012659309380812092, "loss": 1.6539, "step": 687 }, { "epoch": 1.24, "grad_norm": 0.05635406821966171, "learning_rate": 0.0001264077038624984, "loss": 1.678, "step": 688 }, { "epoch": 1.24, "grad_norm": 0.06129194051027298, "learning_rate": 0.00012622221629972043, "loss": 1.6455, "step": 689 }, { "epoch": 1.25, "grad_norm": 0.06195101514458656, "learning_rate": 0.0001260366318054493, "loss": 1.7009, "step": 690 }, { "epoch": 1.25, "grad_norm": 0.06593389809131622, "learning_rate": 0.0001258509510657057, "loss": 1.6897, "step": 691 }, { "epoch": 1.25, "grad_norm": 0.0664474368095398, "learning_rate": 0.00012566517476686606, "loss": 1.6847, "step": 692 }, { "epoch": 1.25, "grad_norm": 0.06081750988960266, "learning_rate": 0.00012547930359566007, "loss": 1.6126, "step": 693 }, { "epoch": 1.25, "grad_norm": 0.06048804894089699, "learning_rate": 0.00012529333823916807, "loss": 1.7086, "step": 694 }, { "epoch": 1.25, "grad_norm": 0.06522712111473083, "learning_rate": 0.00012510727938481865, "loss": 1.6931, "step": 695 }, { "epoch": 1.26, "grad_norm": 0.0614117830991745, "learning_rate": 0.0001249211277203859, "loss": 1.7362, "step": 696 }, { "epoch": 1.26, "grad_norm": 0.05812584608793259, "learning_rate": 0.00012473488393398706, "loss": 1.7052, "step": 697 }, { "epoch": 1.26, "grad_norm": 0.059068553149700165, "learning_rate": 0.00012454854871407994, "loss": 1.6872, "step": 698 }, { "epoch": 1.26, "grad_norm": 0.06033525615930557, "learning_rate": 0.0001243621227494602, "loss": 1.6954, "step": 699 }, { "epoch": 1.26, "grad_norm": 0.06032804027199745, "learning_rate": 0.00012417560672925912, "loss": 1.6571, "step": 700 }, { "epoch": 1.27, "grad_norm": 0.06035863235592842, "learning_rate": 0.00012398900134294073, "loss": 1.6894, "step": 701 }, { "epoch": 1.27, "grad_norm": 0.059223804622888565, "learning_rate": 0.00012380230728029946, "loss": 1.711, "step": 702 }, { "epoch": 1.27, "grad_norm": 0.061139173805713654, "learning_rate": 0.00012361552523145757, "loss": 1.626, "step": 703 }, { "epoch": 1.27, "grad_norm": 0.06459489464759827, "learning_rate": 0.0001234286558868625, "loss": 1.7467, "step": 704 }, { "epoch": 1.27, "grad_norm": 0.06497075408697128, "learning_rate": 0.00012324169993728438, "loss": 1.7419, "step": 705 }, { "epoch": 1.27, "grad_norm": 0.06115833297371864, "learning_rate": 0.0001230546580738136, "loss": 1.6781, "step": 706 }, { "epoch": 1.28, "grad_norm": 0.06160350516438484, "learning_rate": 0.00012286753098785796, "loss": 1.6907, "step": 707 }, { "epoch": 1.28, "grad_norm": 0.06168088689446449, "learning_rate": 0.00012268031937114044, "loss": 1.7265, "step": 708 }, { "epoch": 1.28, "grad_norm": 0.06278149783611298, "learning_rate": 0.00012249302391569638, "loss": 1.7023, "step": 709 }, { "epoch": 1.28, "grad_norm": 0.06181812658905983, "learning_rate": 0.00012230564531387107, "loss": 1.6897, "step": 710 }, { "epoch": 1.28, "grad_norm": 0.05875727906823158, "learning_rate": 0.00012211818425831718, "loss": 1.644, "step": 711 }, { "epoch": 1.29, "grad_norm": 0.061242878437042236, "learning_rate": 0.00012193064144199218, "loss": 1.7256, "step": 712 }, { "epoch": 1.29, "grad_norm": 0.060726381838321686, "learning_rate": 0.00012174301755815571, "loss": 1.6871, "step": 713 }, { "epoch": 1.29, "grad_norm": 0.06219150498509407, "learning_rate": 0.00012155531330036712, "loss": 1.7048, "step": 714 }, { "epoch": 1.29, "grad_norm": 0.06084437295794487, "learning_rate": 0.0001213675293624829, "loss": 1.6888, "step": 715 }, { "epoch": 1.29, "grad_norm": 0.06178005784749985, "learning_rate": 0.00012117966643865398, "loss": 1.6791, "step": 716 }, { "epoch": 1.29, "grad_norm": 0.05991113558411598, "learning_rate": 0.00012099172522332338, "loss": 1.7318, "step": 717 }, { "epoch": 1.3, "grad_norm": 0.06223401054739952, "learning_rate": 0.00012080370641122345, "loss": 1.6417, "step": 718 }, { "epoch": 1.3, "grad_norm": 0.062392983585596085, "learning_rate": 0.00012061561069737343, "loss": 1.6411, "step": 719 }, { "epoch": 1.3, "grad_norm": 0.060492224991321564, "learning_rate": 0.00012042743877707678, "loss": 1.6717, "step": 720 }, { "epoch": 1.3, "grad_norm": 0.06418413668870926, "learning_rate": 0.0001202391913459187, "loss": 1.6751, "step": 721 }, { "epoch": 1.3, "grad_norm": 0.060530129820108414, "learning_rate": 0.0001200508690997635, "loss": 1.7175, "step": 722 }, { "epoch": 1.31, "grad_norm": 0.06409049779176712, "learning_rate": 0.00011986247273475206, "loss": 1.6953, "step": 723 }, { "epoch": 1.31, "grad_norm": 0.05866590142250061, "learning_rate": 0.0001196740029472992, "loss": 1.6935, "step": 724 }, { "epoch": 1.31, "grad_norm": 0.06476990133523941, "learning_rate": 0.00011948546043409123, "loss": 1.7017, "step": 725 }, { "epoch": 1.31, "grad_norm": 0.06523357331752777, "learning_rate": 0.00011929684589208326, "loss": 1.7183, "step": 726 }, { "epoch": 1.31, "grad_norm": 0.060969460755586624, "learning_rate": 0.00011910816001849654, "loss": 1.6887, "step": 727 }, { "epoch": 1.31, "grad_norm": 0.11310483515262604, "learning_rate": 0.00011891940351081625, "loss": 1.6816, "step": 728 }, { "epoch": 1.32, "grad_norm": 0.059255216270685196, "learning_rate": 0.00011873057706678843, "loss": 1.6554, "step": 729 }, { "epoch": 1.32, "grad_norm": 0.062034714967012405, "learning_rate": 0.00011854168138441775, "loss": 1.668, "step": 730 }, { "epoch": 1.32, "grad_norm": 0.06186864525079727, "learning_rate": 0.00011835271716196486, "loss": 1.6806, "step": 731 }, { "epoch": 1.32, "grad_norm": 0.06105494871735573, "learning_rate": 0.00011816368509794364, "loss": 1.6615, "step": 732 }, { "epoch": 1.32, "grad_norm": 0.06231169030070305, "learning_rate": 0.00011797458589111894, "loss": 1.6588, "step": 733 }, { "epoch": 1.33, "grad_norm": 0.06832422316074371, "learning_rate": 0.00011778542024050361, "loss": 1.6758, "step": 734 }, { "epoch": 1.33, "grad_norm": 0.06158846989274025, "learning_rate": 0.00011759618884535624, "loss": 1.7025, "step": 735 }, { "epoch": 1.33, "grad_norm": 0.07147394865751266, "learning_rate": 0.00011740689240517837, "loss": 1.6691, "step": 736 }, { "epoch": 1.33, "grad_norm": 0.06047786399722099, "learning_rate": 0.00011721753161971212, "loss": 1.6968, "step": 737 }, { "epoch": 1.33, "grad_norm": 0.0623675100505352, "learning_rate": 0.00011702810718893722, "loss": 1.7372, "step": 738 }, { "epoch": 1.34, "grad_norm": 0.06291418522596359, "learning_rate": 0.00011683861981306893, "loss": 1.7083, "step": 739 }, { "epoch": 1.34, "grad_norm": 0.059522755444049835, "learning_rate": 0.00011664907019255502, "loss": 1.6533, "step": 740 }, { "epoch": 1.34, "grad_norm": 0.060890860855579376, "learning_rate": 0.00011645945902807341, "loss": 1.6875, "step": 741 }, { "epoch": 1.34, "grad_norm": 0.060426972806453705, "learning_rate": 0.00011626978702052948, "loss": 1.6463, "step": 742 }, { "epoch": 1.34, "grad_norm": 0.062305621802806854, "learning_rate": 0.00011608005487105362, "loss": 1.6785, "step": 743 }, { "epoch": 1.34, "grad_norm": 0.06419097632169724, "learning_rate": 0.00011589026328099839, "loss": 1.6679, "step": 744 }, { "epoch": 1.35, "grad_norm": 0.06365741044282913, "learning_rate": 0.00011570041295193622, "loss": 1.6668, "step": 745 }, { "epoch": 1.35, "grad_norm": 0.0642697736620903, "learning_rate": 0.00011551050458565658, "loss": 1.7095, "step": 746 }, { "epoch": 1.35, "grad_norm": 0.06443499773740768, "learning_rate": 0.00011532053888416343, "loss": 1.6586, "step": 747 }, { "epoch": 1.35, "grad_norm": 0.06351306289434433, "learning_rate": 0.00011513051654967286, "loss": 1.6776, "step": 748 }, { "epoch": 1.35, "grad_norm": 0.06554794311523438, "learning_rate": 0.00011494043828461007, "loss": 1.7105, "step": 749 }, { "epoch": 1.36, "grad_norm": 0.10256826132535934, "learning_rate": 0.00011475030479160725, "loss": 1.7046, "step": 750 }, { "epoch": 1.36, "grad_norm": 0.06379935145378113, "learning_rate": 0.00011456011677350051, "loss": 1.711, "step": 751 }, { "epoch": 1.36, "grad_norm": 0.06044677272439003, "learning_rate": 0.00011436987493332767, "loss": 1.7186, "step": 752 }, { "epoch": 1.36, "grad_norm": 0.06297197192907333, "learning_rate": 0.00011417957997432546, "loss": 1.6453, "step": 753 }, { "epoch": 1.36, "grad_norm": 0.06677673757076263, "learning_rate": 0.00011398923259992697, "loss": 1.6443, "step": 754 }, { "epoch": 1.36, "grad_norm": 0.062335170805454254, "learning_rate": 0.00011379883351375901, "loss": 1.6738, "step": 755 }, { "epoch": 1.37, "grad_norm": 0.06286536902189255, "learning_rate": 0.00011360838341963964, "loss": 1.7081, "step": 756 }, { "epoch": 1.37, "grad_norm": 0.07303211838006973, "learning_rate": 0.00011341788302157536, "loss": 1.6906, "step": 757 }, { "epoch": 1.37, "grad_norm": 0.06304056942462921, "learning_rate": 0.00011322733302375863, "loss": 1.6783, "step": 758 }, { "epoch": 1.37, "grad_norm": 0.07291906327009201, "learning_rate": 0.00011303673413056541, "loss": 1.7162, "step": 759 }, { "epoch": 1.37, "grad_norm": 0.061802685260772705, "learning_rate": 0.00011284608704655215, "loss": 1.7375, "step": 760 }, { "epoch": 1.38, "grad_norm": 0.06205203756690025, "learning_rate": 0.00011265539247645373, "loss": 1.6617, "step": 761 }, { "epoch": 1.38, "grad_norm": 0.06457790732383728, "learning_rate": 0.0001124646511251803, "loss": 1.6395, "step": 762 }, { "epoch": 1.38, "grad_norm": 0.06102142482995987, "learning_rate": 0.00011227386369781508, "loss": 1.7031, "step": 763 }, { "epoch": 1.38, "grad_norm": 0.062267519533634186, "learning_rate": 0.00011208303089961161, "loss": 1.6889, "step": 764 }, { "epoch": 1.38, "grad_norm": 0.06354745477437973, "learning_rate": 0.00011189215343599109, "loss": 1.7099, "step": 765 }, { "epoch": 1.38, "grad_norm": 0.06255058199167252, "learning_rate": 0.00011170123201253986, "loss": 1.7092, "step": 766 }, { "epoch": 1.39, "grad_norm": 0.06354597955942154, "learning_rate": 0.00011151026733500677, "loss": 1.6462, "step": 767 }, { "epoch": 1.39, "grad_norm": 0.06314928829669952, "learning_rate": 0.00011131926010930058, "loss": 1.6377, "step": 768 }, { "epoch": 1.39, "grad_norm": 0.06911808252334595, "learning_rate": 0.00011112821104148723, "loss": 1.6787, "step": 769 }, { "epoch": 1.39, "grad_norm": 0.06356338411569595, "learning_rate": 0.00011093712083778746, "loss": 1.6657, "step": 770 }, { "epoch": 1.39, "grad_norm": 0.06266220659017563, "learning_rate": 0.00011074599020457395, "loss": 1.7108, "step": 771 }, { "epoch": 1.4, "grad_norm": 0.06397093832492828, "learning_rate": 0.00011055481984836893, "loss": 1.715, "step": 772 }, { "epoch": 1.4, "grad_norm": 0.06519615650177002, "learning_rate": 0.00011036361047584143, "loss": 1.6625, "step": 773 }, { "epoch": 1.4, "grad_norm": 0.06543872505426407, "learning_rate": 0.00011017236279380467, "loss": 1.6611, "step": 774 }, { "epoch": 1.4, "grad_norm": 0.06356982886791229, "learning_rate": 0.00010998107750921354, "loss": 1.6366, "step": 775 }, { "epoch": 1.4, "grad_norm": 0.06404688209295273, "learning_rate": 0.00010978975532916189, "loss": 1.689, "step": 776 }, { "epoch": 1.4, "grad_norm": 0.06206212565302849, "learning_rate": 0.00010959839696088001, "loss": 1.6853, "step": 777 }, { "epoch": 1.41, "grad_norm": 0.0640236884355545, "learning_rate": 0.00010940700311173184, "loss": 1.6874, "step": 778 }, { "epoch": 1.41, "grad_norm": 0.06289862096309662, "learning_rate": 0.00010921557448921267, "loss": 1.7187, "step": 779 }, { "epoch": 1.41, "grad_norm": 0.06534165889024734, "learning_rate": 0.00010902411180094607, "loss": 1.6285, "step": 780 }, { "epoch": 1.41, "grad_norm": 0.06411545723676682, "learning_rate": 0.00010883261575468184, "loss": 1.6932, "step": 781 }, { "epoch": 1.41, "grad_norm": 0.06283684074878693, "learning_rate": 0.00010864108705829282, "loss": 1.7544, "step": 782 }, { "epoch": 1.42, "grad_norm": 0.06294089555740356, "learning_rate": 0.00010844952641977273, "loss": 1.695, "step": 783 }, { "epoch": 1.42, "grad_norm": 0.06469050794839859, "learning_rate": 0.00010825793454723325, "loss": 1.654, "step": 784 }, { "epoch": 1.42, "grad_norm": 0.06504753977060318, "learning_rate": 0.00010806631214890155, "loss": 1.6641, "step": 785 }, { "epoch": 1.42, "grad_norm": 0.06289339065551758, "learning_rate": 0.00010787465993311768, "loss": 1.7246, "step": 786 }, { "epoch": 1.42, "grad_norm": 0.07044830918312073, "learning_rate": 0.00010768297860833185, "loss": 1.6784, "step": 787 }, { "epoch": 1.42, "grad_norm": 0.06241421401500702, "learning_rate": 0.00010749126888310197, "loss": 1.7413, "step": 788 }, { "epoch": 1.43, "grad_norm": 0.061875198036432266, "learning_rate": 0.00010729953146609076, "loss": 1.6837, "step": 789 }, { "epoch": 1.43, "grad_norm": 0.06335246562957764, "learning_rate": 0.00010710776706606349, "loss": 1.6713, "step": 790 }, { "epoch": 1.43, "grad_norm": 0.06218186393380165, "learning_rate": 0.00010691597639188507, "loss": 1.6563, "step": 791 }, { "epoch": 1.43, "grad_norm": 0.06283168494701385, "learning_rate": 0.00010672416015251757, "loss": 1.6672, "step": 792 }, { "epoch": 1.43, "grad_norm": 0.06283591687679291, "learning_rate": 0.00010653231905701748, "loss": 1.6719, "step": 793 }, { "epoch": 1.44, "grad_norm": 0.0629267543554306, "learning_rate": 0.00010634045381453337, "loss": 1.6764, "step": 794 }, { "epoch": 1.44, "grad_norm": 0.06264865398406982, "learning_rate": 0.00010614856513430284, "loss": 1.6874, "step": 795 }, { "epoch": 1.44, "grad_norm": 0.06411181390285492, "learning_rate": 0.00010595665372565027, "loss": 1.7095, "step": 796 }, { "epoch": 1.44, "grad_norm": 0.06262548267841339, "learning_rate": 0.00010576472029798399, "loss": 1.6898, "step": 797 }, { "epoch": 1.44, "grad_norm": 0.06278496235609055, "learning_rate": 0.00010557276556079378, "loss": 1.6055, "step": 798 }, { "epoch": 1.44, "grad_norm": 0.06674374639987946, "learning_rate": 0.00010538079022364819, "loss": 1.7226, "step": 799 }, { "epoch": 1.45, "grad_norm": 0.06753117591142654, "learning_rate": 0.00010518879499619181, "loss": 1.7008, "step": 800 }, { "epoch": 1.45, "grad_norm": 0.07137101143598557, "learning_rate": 0.0001049967805881429, "loss": 1.6945, "step": 801 }, { "epoch": 1.45, "grad_norm": 0.06417196989059448, "learning_rate": 0.00010480474770929054, "loss": 1.6662, "step": 802 }, { "epoch": 1.45, "grad_norm": 0.064505934715271, "learning_rate": 0.00010461269706949213, "loss": 1.6914, "step": 803 }, { "epoch": 1.45, "grad_norm": 0.06325452029705048, "learning_rate": 0.00010442062937867063, "loss": 1.6703, "step": 804 }, { "epoch": 1.46, "grad_norm": 0.0945320799946785, "learning_rate": 0.00010422854534681219, "loss": 1.6595, "step": 805 }, { "epoch": 1.46, "grad_norm": 0.07015063613653183, "learning_rate": 0.00010403644568396322, "loss": 1.7153, "step": 806 }, { "epoch": 1.46, "grad_norm": 0.06436234712600708, "learning_rate": 0.000103844331100228, "loss": 1.6767, "step": 807 }, { "epoch": 1.46, "grad_norm": 0.06437043845653534, "learning_rate": 0.0001036522023057659, "loss": 1.7026, "step": 808 }, { "epoch": 1.46, "grad_norm": 0.06160353124141693, "learning_rate": 0.00010346006001078885, "loss": 1.7112, "step": 809 }, { "epoch": 1.46, "grad_norm": 0.06519316881895065, "learning_rate": 0.00010326790492555876, "loss": 1.6611, "step": 810 }, { "epoch": 1.47, "grad_norm": 0.06452979147434235, "learning_rate": 0.00010307573776038462, "loss": 1.6291, "step": 811 }, { "epoch": 1.47, "grad_norm": 0.06813566386699677, "learning_rate": 0.00010288355922562034, "loss": 1.6432, "step": 812 }, { "epoch": 1.47, "grad_norm": 0.06800167262554169, "learning_rate": 0.0001026913700316616, "loss": 1.6739, "step": 813 }, { "epoch": 1.47, "grad_norm": 0.062173567712306976, "learning_rate": 0.0001024991708889437, "loss": 1.7207, "step": 814 }, { "epoch": 1.47, "grad_norm": 0.06301440298557281, "learning_rate": 0.00010230696250793856, "loss": 1.6348, "step": 815 }, { "epoch": 1.48, "grad_norm": 0.06262702494859695, "learning_rate": 0.00010211474559915233, "loss": 1.6982, "step": 816 }, { "epoch": 1.48, "grad_norm": 0.06448613107204437, "learning_rate": 0.00010192252087312265, "loss": 1.7004, "step": 817 }, { "epoch": 1.48, "grad_norm": 0.06269077211618423, "learning_rate": 0.00010173028904041606, "loss": 1.6981, "step": 818 }, { "epoch": 1.48, "grad_norm": 0.06326784938573837, "learning_rate": 0.00010153805081162539, "loss": 1.718, "step": 819 }, { "epoch": 1.48, "grad_norm": 0.06502313911914825, "learning_rate": 0.0001013458068973671, "loss": 1.6669, "step": 820 }, { "epoch": 1.48, "grad_norm": 0.06869412958621979, "learning_rate": 0.0001011535580082787, "loss": 1.6237, "step": 821 }, { "epoch": 1.49, "grad_norm": 0.0637192502617836, "learning_rate": 0.00010096130485501598, "loss": 1.7264, "step": 822 }, { "epoch": 1.49, "eval_loss": 1.7267118692398071, "eval_runtime": 76.2251, "eval_samples_per_second": 65.595, "eval_steps_per_second": 16.399, "step": 822 }, { "epoch": 1.49, "grad_norm": 0.06338479369878769, "learning_rate": 0.00010076904814825066, "loss": 1.66, "step": 823 }, { "epoch": 1.49, "grad_norm": 0.0718810185790062, "learning_rate": 0.0001005767885986674, "loss": 1.7044, "step": 824 }, { "epoch": 1.49, "grad_norm": 0.06428621709346771, "learning_rate": 0.00010038452691696161, "loss": 1.6375, "step": 825 }, { "epoch": 1.49, "grad_norm": 0.06198599189519882, "learning_rate": 0.00010019226381383633, "loss": 1.644, "step": 826 }, { "epoch": 1.5, "grad_norm": 0.0649799108505249, "learning_rate": 0.0001, "loss": 1.6751, "step": 827 }, { "epoch": 1.5, "grad_norm": 0.06546121090650558, "learning_rate": 9.980773618616371e-05, "loss": 1.6728, "step": 828 }, { "epoch": 1.5, "grad_norm": 0.0744151845574379, "learning_rate": 9.961547308303844e-05, "loss": 1.7465, "step": 829 }, { "epoch": 1.5, "grad_norm": 0.06264037638902664, "learning_rate": 9.942321140133261e-05, "loss": 1.6005, "step": 830 }, { "epoch": 1.5, "grad_norm": 0.06265675276517868, "learning_rate": 9.923095185174938e-05, "loss": 1.7181, "step": 831 }, { "epoch": 1.5, "grad_norm": 0.06809694319963455, "learning_rate": 9.903869514498402e-05, "loss": 1.6345, "step": 832 }, { "epoch": 1.51, "grad_norm": 0.06538775563240051, "learning_rate": 9.884644199172135e-05, "loss": 1.7251, "step": 833 }, { "epoch": 1.51, "grad_norm": 0.06529638916254044, "learning_rate": 9.865419310263292e-05, "loss": 1.6418, "step": 834 }, { "epoch": 1.51, "grad_norm": 0.08285729587078094, "learning_rate": 9.846194918837462e-05, "loss": 1.6837, "step": 835 }, { "epoch": 1.51, "grad_norm": 0.06490971148014069, "learning_rate": 9.826971095958395e-05, "loss": 1.6723, "step": 836 }, { "epoch": 1.51, "grad_norm": 0.06375712156295776, "learning_rate": 9.807747912687739e-05, "loss": 1.6838, "step": 837 }, { "epoch": 1.52, "grad_norm": 0.06696437299251556, "learning_rate": 9.788525440084771e-05, "loss": 1.6579, "step": 838 }, { "epoch": 1.52, "grad_norm": 0.06473565846681595, "learning_rate": 9.769303749206146e-05, "loss": 1.6489, "step": 839 }, { "epoch": 1.52, "grad_norm": 0.07211591303348541, "learning_rate": 9.750082911105634e-05, "loss": 1.7435, "step": 840 }, { "epoch": 1.52, "grad_norm": 0.06550677120685577, "learning_rate": 9.730862996833841e-05, "loss": 1.6935, "step": 841 }, { "epoch": 1.52, "grad_norm": 0.06820110231637955, "learning_rate": 9.711644077437968e-05, "loss": 1.6759, "step": 842 }, { "epoch": 1.52, "grad_norm": 0.06783100217580795, "learning_rate": 9.692426223961537e-05, "loss": 1.7081, "step": 843 }, { "epoch": 1.53, "grad_norm": 0.06615381687879562, "learning_rate": 9.67320950744413e-05, "loss": 1.7375, "step": 844 }, { "epoch": 1.53, "grad_norm": 0.0648663192987442, "learning_rate": 9.653993998921118e-05, "loss": 1.6836, "step": 845 }, { "epoch": 1.53, "grad_norm": 0.0639321506023407, "learning_rate": 9.63477976942341e-05, "loss": 1.6319, "step": 846 }, { "epoch": 1.53, "grad_norm": 0.06528212130069733, "learning_rate": 9.615566889977201e-05, "loss": 1.6675, "step": 847 }, { "epoch": 1.53, "grad_norm": 0.06574473530054092, "learning_rate": 9.59635543160368e-05, "loss": 1.6442, "step": 848 }, { "epoch": 1.54, "grad_norm": 0.06326039880514145, "learning_rate": 9.577145465318783e-05, "loss": 1.639, "step": 849 }, { "epoch": 1.54, "grad_norm": 0.06851720809936523, "learning_rate": 9.557937062132938e-05, "loss": 1.7044, "step": 850 }, { "epoch": 1.54, "grad_norm": 0.06546233594417572, "learning_rate": 9.538730293050792e-05, "loss": 1.7091, "step": 851 }, { "epoch": 1.54, "grad_norm": 0.0674884095788002, "learning_rate": 9.51952522907095e-05, "loss": 1.6572, "step": 852 }, { "epoch": 1.54, "grad_norm": 0.06366416811943054, "learning_rate": 9.50032194118571e-05, "loss": 1.6913, "step": 853 }, { "epoch": 1.54, "grad_norm": 0.065780408680439, "learning_rate": 9.481120500380818e-05, "loss": 1.7106, "step": 854 }, { "epoch": 1.55, "grad_norm": 0.06662867218255997, "learning_rate": 9.461920977635184e-05, "loss": 1.6486, "step": 855 }, { "epoch": 1.55, "grad_norm": 0.06339140236377716, "learning_rate": 9.442723443920623e-05, "loss": 1.6799, "step": 856 }, { "epoch": 1.55, "grad_norm": 0.06222783029079437, "learning_rate": 9.423527970201602e-05, "loss": 1.72, "step": 857 }, { "epoch": 1.55, "grad_norm": 0.06612752377986908, "learning_rate": 9.404334627434974e-05, "loss": 1.7294, "step": 858 }, { "epoch": 1.55, "grad_norm": 0.06335198134183884, "learning_rate": 9.385143486569718e-05, "loss": 1.6978, "step": 859 }, { "epoch": 1.56, "grad_norm": 0.0652630627155304, "learning_rate": 9.365954618546665e-05, "loss": 1.6808, "step": 860 }, { "epoch": 1.56, "grad_norm": 0.08252695202827454, "learning_rate": 9.346768094298252e-05, "loss": 1.7117, "step": 861 }, { "epoch": 1.56, "grad_norm": 0.0695163905620575, "learning_rate": 9.327583984748248e-05, "loss": 1.6948, "step": 862 }, { "epoch": 1.56, "grad_norm": 0.06612583249807358, "learning_rate": 9.308402360811497e-05, "loss": 1.705, "step": 863 }, { "epoch": 1.56, "grad_norm": 0.06415654718875885, "learning_rate": 9.289223293393652e-05, "loss": 1.6796, "step": 864 }, { "epoch": 1.56, "grad_norm": 0.06522924453020096, "learning_rate": 9.270046853390925e-05, "loss": 1.6783, "step": 865 }, { "epoch": 1.57, "grad_norm": 0.06422727555036545, "learning_rate": 9.250873111689808e-05, "loss": 1.709, "step": 866 }, { "epoch": 1.57, "grad_norm": 0.06485796719789505, "learning_rate": 9.231702139166816e-05, "loss": 1.6323, "step": 867 }, { "epoch": 1.57, "grad_norm": 0.06597612798213959, "learning_rate": 9.212534006688233e-05, "loss": 1.6578, "step": 868 }, { "epoch": 1.57, "grad_norm": 0.06861060112714767, "learning_rate": 9.193368785109844e-05, "loss": 1.6711, "step": 869 }, { "epoch": 1.57, "grad_norm": 0.07582002878189087, "learning_rate": 9.174206545276677e-05, "loss": 1.666, "step": 870 }, { "epoch": 1.58, "grad_norm": 0.06606924533843994, "learning_rate": 9.15504735802273e-05, "loss": 1.7304, "step": 871 }, { "epoch": 1.58, "grad_norm": 0.06642486900091171, "learning_rate": 9.135891294170718e-05, "loss": 1.7082, "step": 872 }, { "epoch": 1.58, "grad_norm": 0.072264164686203, "learning_rate": 9.11673842453182e-05, "loss": 1.6355, "step": 873 }, { "epoch": 1.58, "grad_norm": 0.06571400165557861, "learning_rate": 9.097588819905394e-05, "loss": 1.6383, "step": 874 }, { "epoch": 1.58, "grad_norm": 0.062258243560791016, "learning_rate": 9.078442551078736e-05, "loss": 1.6676, "step": 875 }, { "epoch": 1.58, "grad_norm": 0.06381349265575409, "learning_rate": 9.059299688826816e-05, "loss": 1.699, "step": 876 }, { "epoch": 1.59, "grad_norm": 0.06702978163957596, "learning_rate": 9.040160303912003e-05, "loss": 1.7245, "step": 877 }, { "epoch": 1.59, "grad_norm": 0.0637059286236763, "learning_rate": 9.021024467083812e-05, "loss": 1.6478, "step": 878 }, { "epoch": 1.59, "grad_norm": 0.0654047429561615, "learning_rate": 9.001892249078648e-05, "loss": 1.7275, "step": 879 }, { "epoch": 1.59, "grad_norm": 0.06602399051189423, "learning_rate": 8.982763720619533e-05, "loss": 1.7712, "step": 880 }, { "epoch": 1.59, "grad_norm": 0.06693969666957855, "learning_rate": 8.96363895241586e-05, "loss": 1.6684, "step": 881 }, { "epoch": 1.6, "grad_norm": 0.06519246846437454, "learning_rate": 8.944518015163108e-05, "loss": 1.6698, "step": 882 }, { "epoch": 1.6, "grad_norm": 0.06838595122098923, "learning_rate": 8.925400979542606e-05, "loss": 1.655, "step": 883 }, { "epoch": 1.6, "grad_norm": 0.06535571813583374, "learning_rate": 8.906287916221259e-05, "loss": 1.6858, "step": 884 }, { "epoch": 1.6, "grad_norm": 0.06805121898651123, "learning_rate": 8.887178895851279e-05, "loss": 1.6746, "step": 885 }, { "epoch": 1.6, "grad_norm": 0.0715852826833725, "learning_rate": 8.868073989069943e-05, "loss": 1.7676, "step": 886 }, { "epoch": 1.6, "grad_norm": 0.06408550590276718, "learning_rate": 8.848973266499322e-05, "loss": 1.6434, "step": 887 }, { "epoch": 1.61, "grad_norm": 0.0682334452867508, "learning_rate": 8.829876798746017e-05, "loss": 1.6663, "step": 888 }, { "epoch": 1.61, "grad_norm": 0.06532958894968033, "learning_rate": 8.810784656400895e-05, "loss": 1.6914, "step": 889 }, { "epoch": 1.61, "grad_norm": 0.06579031050205231, "learning_rate": 8.791696910038843e-05, "loss": 1.6359, "step": 890 }, { "epoch": 1.61, "grad_norm": 0.0659404769539833, "learning_rate": 8.772613630218492e-05, "loss": 1.7121, "step": 891 }, { "epoch": 1.61, "grad_norm": 0.06567792594432831, "learning_rate": 8.753534887481976e-05, "loss": 1.6565, "step": 892 }, { "epoch": 1.62, "grad_norm": 0.07625501602888107, "learning_rate": 8.734460752354629e-05, "loss": 1.6743, "step": 893 }, { "epoch": 1.62, "grad_norm": 0.06591348350048065, "learning_rate": 8.715391295344784e-05, "loss": 1.6265, "step": 894 }, { "epoch": 1.62, "grad_norm": 0.06538601964712143, "learning_rate": 8.696326586943464e-05, "loss": 1.7139, "step": 895 }, { "epoch": 1.62, "grad_norm": 0.06885919719934464, "learning_rate": 8.677266697624138e-05, "loss": 1.6884, "step": 896 }, { "epoch": 1.62, "grad_norm": 0.06452605873346329, "learning_rate": 8.658211697842466e-05, "loss": 1.6894, "step": 897 }, { "epoch": 1.62, "grad_norm": 0.06521788239479065, "learning_rate": 8.639161658036037e-05, "loss": 1.6943, "step": 898 }, { "epoch": 1.63, "grad_norm": 0.06771497428417206, "learning_rate": 8.6201166486241e-05, "loss": 1.6718, "step": 899 }, { "epoch": 1.63, "grad_norm": 0.0637250766158104, "learning_rate": 8.601076740007305e-05, "loss": 1.6842, "step": 900 }, { "epoch": 1.63, "grad_norm": 0.0656089335680008, "learning_rate": 8.582042002567456e-05, "loss": 1.6649, "step": 901 }, { "epoch": 1.63, "grad_norm": 0.06827680766582489, "learning_rate": 8.563012506667233e-05, "loss": 1.7095, "step": 902 }, { "epoch": 1.63, "grad_norm": 0.06502600759267807, "learning_rate": 8.543988322649954e-05, "loss": 1.6368, "step": 903 }, { "epoch": 1.64, "grad_norm": 0.06803898513317108, "learning_rate": 8.524969520839279e-05, "loss": 1.657, "step": 904 }, { "epoch": 1.64, "grad_norm": 0.06632059067487717, "learning_rate": 8.505956171538994e-05, "loss": 1.7279, "step": 905 }, { "epoch": 1.64, "grad_norm": 0.06838211417198181, "learning_rate": 8.486948345032719e-05, "loss": 1.6318, "step": 906 }, { "epoch": 1.64, "grad_norm": 0.0652574896812439, "learning_rate": 8.46794611158366e-05, "loss": 1.6307, "step": 907 }, { "epoch": 1.64, "grad_norm": 0.0648072361946106, "learning_rate": 8.448949541434346e-05, "loss": 1.6517, "step": 908 }, { "epoch": 1.64, "grad_norm": 0.06592056900262833, "learning_rate": 8.429958704806379e-05, "loss": 1.6958, "step": 909 }, { "epoch": 1.65, "grad_norm": 0.06285024434328079, "learning_rate": 8.410973671900162e-05, "loss": 1.666, "step": 910 }, { "epoch": 1.65, "grad_norm": 0.06529216468334198, "learning_rate": 8.391994512894641e-05, "loss": 1.6919, "step": 911 }, { "epoch": 1.65, "grad_norm": 0.06455468386411667, "learning_rate": 8.373021297947053e-05, "loss": 1.6217, "step": 912 }, { "epoch": 1.65, "grad_norm": 0.06522978842258453, "learning_rate": 8.35405409719266e-05, "loss": 1.6729, "step": 913 }, { "epoch": 1.65, "grad_norm": 0.06686036288738251, "learning_rate": 8.335092980744502e-05, "loss": 1.6324, "step": 914 }, { "epoch": 1.66, "grad_norm": 0.06648086756467819, "learning_rate": 8.316138018693108e-05, "loss": 1.6052, "step": 915 }, { "epoch": 1.66, "grad_norm": 0.06622032076120377, "learning_rate": 8.297189281106278e-05, "loss": 1.7219, "step": 916 }, { "epoch": 1.66, "grad_norm": 0.07183654606342316, "learning_rate": 8.278246838028793e-05, "loss": 1.7633, "step": 917 }, { "epoch": 1.66, "grad_norm": 0.06654607504606247, "learning_rate": 8.259310759482164e-05, "loss": 1.7602, "step": 918 }, { "epoch": 1.66, "grad_norm": 0.06768395006656647, "learning_rate": 8.240381115464377e-05, "loss": 1.678, "step": 919 }, { "epoch": 1.66, "grad_norm": 0.0649079754948616, "learning_rate": 8.22145797594964e-05, "loss": 1.7013, "step": 920 }, { "epoch": 1.67, "grad_norm": 0.06565246731042862, "learning_rate": 8.20254141088811e-05, "loss": 1.7064, "step": 921 }, { "epoch": 1.67, "grad_norm": 0.06477197259664536, "learning_rate": 8.183631490205637e-05, "loss": 1.7219, "step": 922 }, { "epoch": 1.67, "grad_norm": 0.06408128142356873, "learning_rate": 8.164728283803518e-05, "loss": 1.7337, "step": 923 }, { "epoch": 1.67, "grad_norm": 0.06464950740337372, "learning_rate": 8.145831861558225e-05, "loss": 1.6853, "step": 924 }, { "epoch": 1.67, "grad_norm": 0.06401928514242172, "learning_rate": 8.126942293321162e-05, "loss": 1.6587, "step": 925 }, { "epoch": 1.68, "grad_norm": 0.06978955864906311, "learning_rate": 8.108059648918377e-05, "loss": 1.7083, "step": 926 }, { "epoch": 1.68, "grad_norm": 0.06544001400470734, "learning_rate": 8.089183998150344e-05, "loss": 1.6318, "step": 927 }, { "epoch": 1.68, "grad_norm": 0.06558380275964737, "learning_rate": 8.070315410791679e-05, "loss": 1.6897, "step": 928 }, { "epoch": 1.68, "grad_norm": 0.06930231302976608, "learning_rate": 8.051453956590878e-05, "loss": 1.6266, "step": 929 }, { "epoch": 1.68, "grad_norm": 0.06593599915504456, "learning_rate": 8.03259970527008e-05, "loss": 1.7096, "step": 930 }, { "epoch": 1.69, "grad_norm": 0.06622833758592606, "learning_rate": 8.013752726524795e-05, "loss": 1.5817, "step": 931 }, { "epoch": 1.69, "grad_norm": 0.06626243144273758, "learning_rate": 7.994913090023651e-05, "loss": 1.6525, "step": 932 }, { "epoch": 1.69, "grad_norm": 0.0677393451333046, "learning_rate": 7.976080865408131e-05, "loss": 1.7158, "step": 933 }, { "epoch": 1.69, "grad_norm": 0.06529498845338821, "learning_rate": 7.957256122292323e-05, "loss": 1.7317, "step": 934 }, { "epoch": 1.69, "grad_norm": 0.07396451383829117, "learning_rate": 7.938438930262656e-05, "loss": 1.6791, "step": 935 }, { "epoch": 1.69, "grad_norm": 0.07032353430986404, "learning_rate": 7.919629358877657e-05, "loss": 1.7024, "step": 936 }, { "epoch": 1.7, "grad_norm": 0.06451990455389023, "learning_rate": 7.900827477667663e-05, "loss": 1.7266, "step": 937 }, { "epoch": 1.7, "grad_norm": 0.06694858521223068, "learning_rate": 7.882033356134603e-05, "loss": 1.6612, "step": 938 }, { "epoch": 1.7, "grad_norm": 0.06609500199556351, "learning_rate": 7.863247063751715e-05, "loss": 1.713, "step": 939 }, { "epoch": 1.7, "grad_norm": 0.06344272941350937, "learning_rate": 7.844468669963289e-05, "loss": 1.6219, "step": 940 }, { "epoch": 1.7, "grad_norm": 0.06307589262723923, "learning_rate": 7.825698244184431e-05, "loss": 1.7042, "step": 941 }, { "epoch": 1.71, "grad_norm": 0.06659837812185287, "learning_rate": 7.806935855800782e-05, "loss": 1.6993, "step": 942 }, { "epoch": 1.71, "grad_norm": 0.06524292379617691, "learning_rate": 7.788181574168283e-05, "loss": 1.6687, "step": 943 }, { "epoch": 1.71, "grad_norm": 0.06560816615819931, "learning_rate": 7.769435468612896e-05, "loss": 1.7081, "step": 944 }, { "epoch": 1.71, "grad_norm": 0.06725630909204483, "learning_rate": 7.750697608430365e-05, "loss": 1.7001, "step": 945 }, { "epoch": 1.71, "grad_norm": 0.06650066375732422, "learning_rate": 7.731968062885956e-05, "loss": 1.7225, "step": 946 }, { "epoch": 1.71, "grad_norm": 0.06517896801233292, "learning_rate": 7.713246901214206e-05, "loss": 1.6299, "step": 947 }, { "epoch": 1.72, "grad_norm": 0.06807747483253479, "learning_rate": 7.694534192618641e-05, "loss": 1.695, "step": 948 }, { "epoch": 1.72, "grad_norm": 0.06809186935424805, "learning_rate": 7.67583000627156e-05, "loss": 1.6611, "step": 949 }, { "epoch": 1.72, "grad_norm": 0.06693090498447418, "learning_rate": 7.657134411313753e-05, "loss": 1.6603, "step": 950 }, { "epoch": 1.72, "grad_norm": 0.06553305685520172, "learning_rate": 7.638447476854245e-05, "loss": 1.7036, "step": 951 }, { "epoch": 1.72, "grad_norm": 0.06823913007974625, "learning_rate": 7.619769271970056e-05, "loss": 1.6848, "step": 952 }, { "epoch": 1.73, "grad_norm": 0.0652228444814682, "learning_rate": 7.601099865705927e-05, "loss": 1.6893, "step": 953 }, { "epoch": 1.73, "grad_norm": 0.07233775407075882, "learning_rate": 7.58243932707409e-05, "loss": 1.6777, "step": 954 }, { "epoch": 1.73, "grad_norm": 0.07119675725698471, "learning_rate": 7.563787725053981e-05, "loss": 1.706, "step": 955 }, { "epoch": 1.73, "grad_norm": 0.06489936262369156, "learning_rate": 7.54514512859201e-05, "loss": 1.6538, "step": 956 }, { "epoch": 1.73, "grad_norm": 0.06696008145809174, "learning_rate": 7.526511606601293e-05, "loss": 1.6862, "step": 957 }, { "epoch": 1.73, "grad_norm": 0.06405473500490189, "learning_rate": 7.507887227961414e-05, "loss": 1.662, "step": 958 }, { "epoch": 1.74, "grad_norm": 0.06998445093631744, "learning_rate": 7.489272061518136e-05, "loss": 1.6604, "step": 959 }, { "epoch": 1.74, "eval_loss": 1.726022481918335, "eval_runtime": 76.3141, "eval_samples_per_second": 65.519, "eval_steps_per_second": 16.38, "step": 959 }, { "epoch": 1.74, "grad_norm": 0.06673965603113174, "learning_rate": 7.470666176083192e-05, "loss": 1.7049, "step": 960 }, { "epoch": 1.74, "grad_norm": 0.06746464222669601, "learning_rate": 7.452069640433997e-05, "loss": 1.6803, "step": 961 }, { "epoch": 1.74, "grad_norm": 0.06396359950304031, "learning_rate": 7.433482523313395e-05, "loss": 1.7104, "step": 962 }, { "epoch": 1.74, "grad_norm": 0.066098153591156, "learning_rate": 7.414904893429433e-05, "loss": 1.6527, "step": 963 }, { "epoch": 1.75, "grad_norm": 0.06473662704229355, "learning_rate": 7.39633681945507e-05, "loss": 1.6891, "step": 964 }, { "epoch": 1.75, "grad_norm": 0.07003339380025864, "learning_rate": 7.377778370027962e-05, "loss": 1.676, "step": 965 }, { "epoch": 1.75, "grad_norm": 0.06654497236013412, "learning_rate": 7.35922961375016e-05, "loss": 1.6601, "step": 966 }, { "epoch": 1.75, "grad_norm": 0.06775406002998352, "learning_rate": 7.340690619187908e-05, "loss": 1.6391, "step": 967 }, { "epoch": 1.75, "grad_norm": 0.06764483451843262, "learning_rate": 7.322161454871356e-05, "loss": 1.7057, "step": 968 }, { "epoch": 1.75, "grad_norm": 0.0728226825594902, "learning_rate": 7.303642189294316e-05, "loss": 1.6793, "step": 969 }, { "epoch": 1.76, "grad_norm": 0.06543935835361481, "learning_rate": 7.285132890914002e-05, "loss": 1.6962, "step": 970 }, { "epoch": 1.76, "grad_norm": 0.06830572336912155, "learning_rate": 7.266633628150801e-05, "loss": 1.6774, "step": 971 }, { "epoch": 1.76, "grad_norm": 0.07373080402612686, "learning_rate": 7.248144469387992e-05, "loss": 1.6815, "step": 972 }, { "epoch": 1.76, "grad_norm": 0.06465107947587967, "learning_rate": 7.229665482971499e-05, "loss": 1.6572, "step": 973 }, { "epoch": 1.76, "grad_norm": 0.06544660031795502, "learning_rate": 7.211196737209653e-05, "loss": 1.6841, "step": 974 }, { "epoch": 1.77, "grad_norm": 0.06559861451387405, "learning_rate": 7.192738300372925e-05, "loss": 1.6835, "step": 975 }, { "epoch": 1.77, "grad_norm": 0.06756362318992615, "learning_rate": 7.174290240693689e-05, "loss": 1.5912, "step": 976 }, { "epoch": 1.77, "grad_norm": 0.06515438854694366, "learning_rate": 7.155852626365938e-05, "loss": 1.6586, "step": 977 }, { "epoch": 1.77, "grad_norm": 0.06673271209001541, "learning_rate": 7.137425525545074e-05, "loss": 1.67, "step": 978 }, { "epoch": 1.77, "grad_norm": 0.06732840090990067, "learning_rate": 7.119009006347625e-05, "loss": 1.6262, "step": 979 }, { "epoch": 1.77, "grad_norm": 0.0666419267654419, "learning_rate": 7.100603136851009e-05, "loss": 1.6963, "step": 980 }, { "epoch": 1.78, "grad_norm": 0.07527624070644379, "learning_rate": 7.082207985093268e-05, "loss": 1.6903, "step": 981 }, { "epoch": 1.78, "grad_norm": 0.06989062577486038, "learning_rate": 7.063823619072838e-05, "loss": 1.6497, "step": 982 }, { "epoch": 1.78, "grad_norm": 0.0654689222574234, "learning_rate": 7.045450106748277e-05, "loss": 1.6782, "step": 983 }, { "epoch": 1.78, "grad_norm": 0.06511061638593674, "learning_rate": 7.027087516038022e-05, "loss": 1.6824, "step": 984 }, { "epoch": 1.78, "grad_norm": 0.06674464046955109, "learning_rate": 7.008735914820138e-05, "loss": 1.7367, "step": 985 }, { "epoch": 1.79, "grad_norm": 0.06592298299074173, "learning_rate": 6.990395370932068e-05, "loss": 1.6879, "step": 986 }, { "epoch": 1.79, "grad_norm": 0.06826543807983398, "learning_rate": 6.97206595217039e-05, "loss": 1.6682, "step": 987 }, { "epoch": 1.79, "grad_norm": 0.06695631891489029, "learning_rate": 6.953747726290535e-05, "loss": 1.7181, "step": 988 }, { "epoch": 1.79, "grad_norm": 0.06656961888074875, "learning_rate": 6.935440761006582e-05, "loss": 1.6778, "step": 989 }, { "epoch": 1.79, "grad_norm": 0.06611720472574234, "learning_rate": 6.917145123990973e-05, "loss": 1.6467, "step": 990 }, { "epoch": 1.79, "grad_norm": 0.06846632063388824, "learning_rate": 6.898860882874279e-05, "loss": 1.7165, "step": 991 }, { "epoch": 1.8, "grad_norm": 0.06631824374198914, "learning_rate": 6.88058810524494e-05, "loss": 1.7042, "step": 992 }, { "epoch": 1.8, "grad_norm": 0.06761027872562408, "learning_rate": 6.862326858649026e-05, "loss": 1.6822, "step": 993 }, { "epoch": 1.8, "grad_norm": 0.06898529827594757, "learning_rate": 6.844077210589986e-05, "loss": 1.6635, "step": 994 }, { "epoch": 1.8, "grad_norm": 0.06683610379695892, "learning_rate": 6.825839228528382e-05, "loss": 1.6949, "step": 995 }, { "epoch": 1.8, "grad_norm": 0.06670662760734558, "learning_rate": 6.807612979881661e-05, "loss": 1.6724, "step": 996 }, { "epoch": 1.81, "grad_norm": 0.19084873795509338, "learning_rate": 6.789398532023894e-05, "loss": 1.7499, "step": 997 }, { "epoch": 1.81, "grad_norm": 0.06561749428510666, "learning_rate": 6.77119595228554e-05, "loss": 1.6733, "step": 998 }, { "epoch": 1.81, "grad_norm": 0.07371030747890472, "learning_rate": 6.753005307953167e-05, "loss": 1.6607, "step": 999 }, { "epoch": 1.81, "grad_norm": 0.0679875835776329, "learning_rate": 6.734826666269238e-05, "loss": 1.6233, "step": 1000 }, { "epoch": 1.81, "grad_norm": 0.0667947381734848, "learning_rate": 6.716660094431846e-05, "loss": 1.6186, "step": 1001 }, { "epoch": 1.81, "grad_norm": 0.06578990817070007, "learning_rate": 6.698505659594466e-05, "loss": 1.6997, "step": 1002 }, { "epoch": 1.82, "grad_norm": 0.07320542633533478, "learning_rate": 6.680363428865704e-05, "loss": 1.6729, "step": 1003 }, { "epoch": 1.82, "grad_norm": 0.06879616528749466, "learning_rate": 6.662233469309058e-05, "loss": 1.6982, "step": 1004 }, { "epoch": 1.82, "grad_norm": 0.06353451311588287, "learning_rate": 6.644115847942667e-05, "loss": 1.6698, "step": 1005 }, { "epoch": 1.82, "grad_norm": 0.06664732843637466, "learning_rate": 6.626010631739054e-05, "loss": 1.6225, "step": 1006 }, { "epoch": 1.82, "grad_norm": 0.0662289708852768, "learning_rate": 6.60791788762489e-05, "loss": 1.713, "step": 1007 }, { "epoch": 1.83, "grad_norm": 0.06735072284936905, "learning_rate": 6.589837682480744e-05, "loss": 1.6431, "step": 1008 }, { "epoch": 1.83, "grad_norm": 0.06567612290382385, "learning_rate": 6.571770083140836e-05, "loss": 1.6972, "step": 1009 }, { "epoch": 1.83, "grad_norm": 0.06742958724498749, "learning_rate": 6.553715156392776e-05, "loss": 1.6439, "step": 1010 }, { "epoch": 1.83, "grad_norm": 0.06748675554990768, "learning_rate": 6.535672968977345e-05, "loss": 1.6711, "step": 1011 }, { "epoch": 1.83, "grad_norm": 0.07259120792150497, "learning_rate": 6.517643587588221e-05, "loss": 1.7223, "step": 1012 }, { "epoch": 1.83, "grad_norm": 0.07579007744789124, "learning_rate": 6.499627078871753e-05, "loss": 1.6614, "step": 1013 }, { "epoch": 1.84, "grad_norm": 0.07152054458856583, "learning_rate": 6.481623509426697e-05, "loss": 1.7038, "step": 1014 }, { "epoch": 1.84, "grad_norm": 0.06873390078544617, "learning_rate": 6.463632945803981e-05, "loss": 1.6602, "step": 1015 }, { "epoch": 1.84, "grad_norm": 0.0664227306842804, "learning_rate": 6.445655454506465e-05, "loss": 1.6916, "step": 1016 }, { "epoch": 1.84, "grad_norm": 0.06599757075309753, "learning_rate": 6.427691101988673e-05, "loss": 1.605, "step": 1017 }, { "epoch": 1.84, "grad_norm": 0.06476866453886032, "learning_rate": 6.40973995465657e-05, "loss": 1.6309, "step": 1018 }, { "epoch": 1.85, "grad_norm": 0.06668147444725037, "learning_rate": 6.391802078867304e-05, "loss": 1.684, "step": 1019 }, { "epoch": 1.85, "grad_norm": 0.06579145044088364, "learning_rate": 6.373877540928972e-05, "loss": 1.6277, "step": 1020 }, { "epoch": 1.85, "grad_norm": 0.06740958243608475, "learning_rate": 6.355966407100346e-05, "loss": 1.728, "step": 1021 }, { "epoch": 1.85, "grad_norm": 0.07092586159706116, "learning_rate": 6.338068743590676e-05, "loss": 1.7091, "step": 1022 }, { "epoch": 1.85, "grad_norm": 0.06797771900892258, "learning_rate": 6.320184616559402e-05, "loss": 1.6962, "step": 1023 }, { "epoch": 1.85, "grad_norm": 0.06833136081695557, "learning_rate": 6.30231409211593e-05, "loss": 1.6981, "step": 1024 }, { "epoch": 1.86, "grad_norm": 0.06703907996416092, "learning_rate": 6.284457236319381e-05, "loss": 1.7082, "step": 1025 }, { "epoch": 1.86, "grad_norm": 0.0666668489575386, "learning_rate": 6.266614115178351e-05, "loss": 1.6198, "step": 1026 }, { "epoch": 1.86, "grad_norm": 0.07242632657289505, "learning_rate": 6.248784794650672e-05, "loss": 1.705, "step": 1027 }, { "epoch": 1.86, "grad_norm": 0.06651555746793747, "learning_rate": 6.230969340643149e-05, "loss": 1.6417, "step": 1028 }, { "epoch": 1.86, "grad_norm": 0.06552428007125854, "learning_rate": 6.213167819011338e-05, "loss": 1.6917, "step": 1029 }, { "epoch": 1.87, "grad_norm": 0.06741311401128769, "learning_rate": 6.195380295559288e-05, "loss": 1.7241, "step": 1030 }, { "epoch": 1.87, "grad_norm": 0.06656550616025925, "learning_rate": 6.177606836039311e-05, "loss": 1.646, "step": 1031 }, { "epoch": 1.87, "grad_norm": 0.06896986067295074, "learning_rate": 6.159847506151719e-05, "loss": 1.6708, "step": 1032 }, { "epoch": 1.87, "grad_norm": 0.06811494380235672, "learning_rate": 6.142102371544604e-05, "loss": 1.6927, "step": 1033 }, { "epoch": 1.87, "grad_norm": 0.06616541743278503, "learning_rate": 6.124371497813582e-05, "loss": 1.6175, "step": 1034 }, { "epoch": 1.87, "grad_norm": 0.06697241216897964, "learning_rate": 6.106654950501547e-05, "loss": 1.6848, "step": 1035 }, { "epoch": 1.88, "grad_norm": 0.06779171526432037, "learning_rate": 6.0889527950984416e-05, "loss": 1.6566, "step": 1036 }, { "epoch": 1.88, "grad_norm": 0.0683891773223877, "learning_rate": 6.071265097041005e-05, "loss": 1.6258, "step": 1037 }, { "epoch": 1.88, "grad_norm": 0.06936081498861313, "learning_rate": 6.053591921712541e-05, "loss": 1.6115, "step": 1038 }, { "epoch": 1.88, "grad_norm": 0.0856877937912941, "learning_rate": 6.035933334442654e-05, "loss": 1.6742, "step": 1039 }, { "epoch": 1.88, "grad_norm": 0.07240041345357895, "learning_rate": 6.01828940050704e-05, "loss": 1.6901, "step": 1040 }, { "epoch": 1.89, "grad_norm": 0.0770583376288414, "learning_rate": 6.000660185127219e-05, "loss": 1.6803, "step": 1041 }, { "epoch": 1.89, "grad_norm": 0.06806863099336624, "learning_rate": 5.983045753470308e-05, "loss": 1.6561, "step": 1042 }, { "epoch": 1.89, "grad_norm": 0.06816756725311279, "learning_rate": 5.965446170648765e-05, "loss": 1.6635, "step": 1043 }, { "epoch": 1.89, "grad_norm": 0.06543378531932831, "learning_rate": 5.947861501720175e-05, "loss": 1.7153, "step": 1044 }, { "epoch": 1.89, "grad_norm": 0.06688012927770615, "learning_rate": 5.930291811686983e-05, "loss": 1.7142, "step": 1045 }, { "epoch": 1.89, "grad_norm": 0.071477010846138, "learning_rate": 5.9127371654962615e-05, "loss": 1.6804, "step": 1046 }, { "epoch": 1.9, "grad_norm": 0.06843505799770355, "learning_rate": 5.8951976280394795e-05, "loss": 1.7476, "step": 1047 }, { "epoch": 1.9, "grad_norm": 0.06697747856378555, "learning_rate": 5.8776732641522503e-05, "loss": 1.662, "step": 1048 }, { "epoch": 1.9, "grad_norm": 0.06771202385425568, "learning_rate": 5.86016413861411e-05, "loss": 1.655, "step": 1049 }, { "epoch": 1.9, "grad_norm": 0.07092612236738205, "learning_rate": 5.842670316148244e-05, "loss": 1.707, "step": 1050 }, { "epoch": 1.9, "grad_norm": 0.06740372627973557, "learning_rate": 5.825191861421285e-05, "loss": 1.673, "step": 1051 }, { "epoch": 1.91, "grad_norm": 0.06587556004524231, "learning_rate": 5.807728839043061e-05, "loss": 1.6879, "step": 1052 }, { "epoch": 1.91, "grad_norm": 0.06834732741117477, "learning_rate": 5.790281313566341e-05, "loss": 1.7233, "step": 1053 }, { "epoch": 1.91, "grad_norm": 0.06691209226846695, "learning_rate": 5.7728493494866134e-05, "loss": 1.6966, "step": 1054 }, { "epoch": 1.91, "grad_norm": 0.06715382635593414, "learning_rate": 5.755433011241851e-05, "loss": 1.7185, "step": 1055 }, { "epoch": 1.91, "grad_norm": 0.06831709295511246, "learning_rate": 5.738032363212258e-05, "loss": 1.6529, "step": 1056 }, { "epoch": 1.91, "grad_norm": 0.06592843681573868, "learning_rate": 5.720647469720033e-05, "loss": 1.6939, "step": 1057 }, { "epoch": 1.92, "grad_norm": 0.06575801223516464, "learning_rate": 5.70327839502915e-05, "loss": 1.6642, "step": 1058 }, { "epoch": 1.92, "grad_norm": 0.07193956524133682, "learning_rate": 5.685925203345108e-05, "loss": 1.6675, "step": 1059 }, { "epoch": 1.92, "grad_norm": 0.0670444443821907, "learning_rate": 5.6685879588146815e-05, "loss": 1.7136, "step": 1060 }, { "epoch": 1.92, "grad_norm": 0.07206844538450241, "learning_rate": 5.651266725525703e-05, "loss": 1.6999, "step": 1061 }, { "epoch": 1.92, "grad_norm": 0.0692375898361206, "learning_rate": 5.633961567506819e-05, "loss": 1.6782, "step": 1062 }, { "epoch": 1.93, "grad_norm": 0.06483175605535507, "learning_rate": 5.6166725487272576e-05, "loss": 1.6448, "step": 1063 }, { "epoch": 1.93, "grad_norm": 0.0667993351817131, "learning_rate": 5.5993997330965796e-05, "loss": 1.6683, "step": 1064 }, { "epoch": 1.93, "grad_norm": 0.0673048198223114, "learning_rate": 5.5821431844644476e-05, "loss": 1.6534, "step": 1065 }, { "epoch": 1.93, "grad_norm": 0.07212254405021667, "learning_rate": 5.564902966620408e-05, "loss": 1.7084, "step": 1066 }, { "epoch": 1.93, "grad_norm": 0.06697355955839157, "learning_rate": 5.547679143293624e-05, "loss": 1.7029, "step": 1067 }, { "epoch": 1.93, "grad_norm": 0.07669904828071594, "learning_rate": 5.530471778152658e-05, "loss": 1.7153, "step": 1068 }, { "epoch": 1.94, "grad_norm": 0.07381530106067657, "learning_rate": 5.513280934805243e-05, "loss": 1.6769, "step": 1069 }, { "epoch": 1.94, "grad_norm": 0.068946473300457, "learning_rate": 5.4961066767980363e-05, "loss": 1.6799, "step": 1070 }, { "epoch": 1.94, "grad_norm": 0.06763108819723129, "learning_rate": 5.478949067616381e-05, "loss": 1.7185, "step": 1071 }, { "epoch": 1.94, "grad_norm": 0.06624120473861694, "learning_rate": 5.4618081706840754e-05, "loss": 1.6972, "step": 1072 }, { "epoch": 1.94, "grad_norm": 0.06670323014259338, "learning_rate": 5.444684049363147e-05, "loss": 1.6826, "step": 1073 }, { "epoch": 1.95, "grad_norm": 0.06699904054403305, "learning_rate": 5.4275767669536146e-05, "loss": 1.643, "step": 1074 }, { "epoch": 1.95, "grad_norm": 0.07036450505256653, "learning_rate": 5.410486386693243e-05, "loss": 1.6719, "step": 1075 }, { "epoch": 1.95, "grad_norm": 0.06482276320457458, "learning_rate": 5.3934129717573165e-05, "loss": 1.6756, "step": 1076 }, { "epoch": 1.95, "grad_norm": 0.06716746836900711, "learning_rate": 5.3763565852584177e-05, "loss": 1.6995, "step": 1077 }, { "epoch": 1.95, "grad_norm": 0.06743574887514114, "learning_rate": 5.3593172902461717e-05, "loss": 1.7064, "step": 1078 }, { "epoch": 1.95, "grad_norm": 0.06770848482847214, "learning_rate": 5.342295149707025e-05, "loss": 1.6588, "step": 1079 }, { "epoch": 1.96, "grad_norm": 0.06666205823421478, "learning_rate": 5.325290226564017e-05, "loss": 1.6215, "step": 1080 }, { "epoch": 1.96, "grad_norm": 0.0728970617055893, "learning_rate": 5.308302583676548e-05, "loss": 1.6878, "step": 1081 }, { "epoch": 1.96, "grad_norm": 0.06758435070514679, "learning_rate": 5.291332283840125e-05, "loss": 1.6422, "step": 1082 }, { "epoch": 1.96, "grad_norm": 0.06901335716247559, "learning_rate": 5.274379389786154e-05, "loss": 1.7208, "step": 1083 }, { "epoch": 1.96, "grad_norm": 0.06578974425792694, "learning_rate": 5.2574439641817006e-05, "loss": 1.6822, "step": 1084 }, { "epoch": 1.97, "grad_norm": 0.08507327735424042, "learning_rate": 5.240526069629265e-05, "loss": 1.6697, "step": 1085 }, { "epoch": 1.97, "grad_norm": 0.06818517297506332, "learning_rate": 5.223625768666528e-05, "loss": 1.7514, "step": 1086 }, { "epoch": 1.97, "grad_norm": 0.06869194656610489, "learning_rate": 5.206743123766139e-05, "loss": 1.6667, "step": 1087 }, { "epoch": 1.97, "grad_norm": 0.06622481346130371, "learning_rate": 5.1898781973354914e-05, "loss": 1.6807, "step": 1088 }, { "epoch": 1.97, "grad_norm": 0.07047388702630997, "learning_rate": 5.173031051716472e-05, "loss": 1.7118, "step": 1089 }, { "epoch": 1.97, "grad_norm": 0.0671396255493164, "learning_rate": 5.1562017491852387e-05, "loss": 1.641, "step": 1090 }, { "epoch": 1.98, "grad_norm": 0.06699879467487335, "learning_rate": 5.139390351951997e-05, "loss": 1.689, "step": 1091 }, { "epoch": 1.98, "grad_norm": 0.06538563221693039, "learning_rate": 5.122596922160768e-05, "loss": 1.6552, "step": 1092 }, { "epoch": 1.98, "grad_norm": 0.06701681017875671, "learning_rate": 5.105821521889147e-05, "loss": 1.6229, "step": 1093 }, { "epoch": 1.98, "grad_norm": 0.06672403961420059, "learning_rate": 5.089064213148082e-05, "loss": 1.695, "step": 1094 }, { "epoch": 1.98, "grad_norm": 0.06800191104412079, "learning_rate": 5.0723250578816576e-05, "loss": 1.6773, "step": 1095 }, { "epoch": 1.99, "grad_norm": 0.066898874938488, "learning_rate": 5.0556041179668354e-05, "loss": 1.6562, "step": 1096 }, { "epoch": 1.99, "eval_loss": 1.7255171537399292, "eval_runtime": 76.5349, "eval_samples_per_second": 65.33, "eval_steps_per_second": 16.332, "step": 1096 } ], "logging_steps": 1, "max_steps": 1644, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 548, "total_flos": 3.2705098222896415e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }