|
{ |
|
"best_metric": 0.5804409980773926, |
|
"best_model_checkpoint": "accent_classification/checkpoint-396", |
|
"epoch": 1.0, |
|
"global_step": 396, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.000000000000001e-07, |
|
"loss": 0.6956, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.6979, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.8e-06, |
|
"loss": 0.7014, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 0.6984, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3e-06, |
|
"loss": 0.6978, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.6e-06, |
|
"loss": 0.6783, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.2000000000000004e-06, |
|
"loss": 0.6956, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 0.6994, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.4e-06, |
|
"loss": 0.6902, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6e-06, |
|
"loss": 0.6937, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.6e-06, |
|
"loss": 0.6497, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.2e-06, |
|
"loss": 0.6911, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.8e-06, |
|
"loss": 0.6904, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.400000000000001e-06, |
|
"loss": 0.6854, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9e-06, |
|
"loss": 0.6843, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.600000000000001e-06, |
|
"loss": 0.6851, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.02e-05, |
|
"loss": 0.6733, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.08e-05, |
|
"loss": 0.6819, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.1400000000000001e-05, |
|
"loss": 0.681, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.6998, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.26e-05, |
|
"loss": 0.6743, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.32e-05, |
|
"loss": 0.6818, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3800000000000002e-05, |
|
"loss": 0.5331, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.44e-05, |
|
"loss": 0.6681, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.6955, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.56e-05, |
|
"loss": 0.7, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.62e-05, |
|
"loss": 0.6975, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.6800000000000002e-05, |
|
"loss": 0.6249, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.74e-05, |
|
"loss": 0.6979, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8e-05, |
|
"loss": 0.6076, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.86e-05, |
|
"loss": 0.6091, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9200000000000003e-05, |
|
"loss": 0.6438, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.98e-05, |
|
"loss": 0.6624, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 2.04e-05, |
|
"loss": 0.6452, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 2.1e-05, |
|
"loss": 0.6963, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 2.16e-05, |
|
"loss": 0.758, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 2.22e-05, |
|
"loss": 0.6471, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 2.2800000000000002e-05, |
|
"loss": 0.6313, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 2.3400000000000003e-05, |
|
"loss": 0.7019, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.6336, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 2.4599999999999998e-05, |
|
"loss": 0.7091, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 2.52e-05, |
|
"loss": 0.701, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 2.58e-05, |
|
"loss": 0.7036, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 2.64e-05, |
|
"loss": 0.7073, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 0.614, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2.7600000000000003e-05, |
|
"loss": 0.7543, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2.8199999999999998e-05, |
|
"loss": 0.538, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2.88e-05, |
|
"loss": 0.5179, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2.94e-05, |
|
"loss": 0.6164, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 3e-05, |
|
"loss": 0.6093, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.99923273657289e-05, |
|
"loss": 0.6983, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.9984654731457804e-05, |
|
"loss": 0.6069, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.9976982097186702e-05, |
|
"loss": 0.5933, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.99693094629156e-05, |
|
"loss": 0.552, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.9961636828644502e-05, |
|
"loss": 0.5865, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.9953964194373404e-05, |
|
"loss": 0.7146, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.9946291560102302e-05, |
|
"loss": 0.5591, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.9938618925831203e-05, |
|
"loss": 0.7366, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.9930946291560102e-05, |
|
"loss": 0.7006, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.9923273657289003e-05, |
|
"loss": 0.4148, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.9915601023017905e-05, |
|
"loss": 0.9938, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.9907928388746803e-05, |
|
"loss": 0.911, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.99002557544757e-05, |
|
"loss": 0.5682, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.9892583120204606e-05, |
|
"loss": 0.5833, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.9884910485933505e-05, |
|
"loss": 0.7237, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.9877237851662403e-05, |
|
"loss": 0.4013, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.9869565217391304e-05, |
|
"loss": 0.9695, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.9861892583120206e-05, |
|
"loss": 0.568, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.9854219948849104e-05, |
|
"loss": 0.5697, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.9846547314578006e-05, |
|
"loss": 0.7675, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.9838874680306904e-05, |
|
"loss": 0.8831, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.9831202046035806e-05, |
|
"loss": 0.7008, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.9823529411764707e-05, |
|
"loss": 0.7085, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.9815856777493606e-05, |
|
"loss": 0.5374, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.9808184143222507e-05, |
|
"loss": 0.4131, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.980051150895141e-05, |
|
"loss": 0.7309, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.9792838874680307e-05, |
|
"loss": 0.4279, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.978516624040921e-05, |
|
"loss": 0.5813, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.977749360613811e-05, |
|
"loss": 0.6276, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.976982097186701e-05, |
|
"loss": 0.7892, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.976214833759591e-05, |
|
"loss": 0.7134, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.9754475703324808e-05, |
|
"loss": 0.3682, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.974680306905371e-05, |
|
"loss": 0.5562, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.973913043478261e-05, |
|
"loss": 0.7158, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.973145780051151e-05, |
|
"loss": 0.524, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.9723785166240408e-05, |
|
"loss": 0.3372, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.9716112531969313e-05, |
|
"loss": 0.4979, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.970843989769821e-05, |
|
"loss": 0.3177, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.970076726342711e-05, |
|
"loss": 0.3488, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.969309462915601e-05, |
|
"loss": 0.5158, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.9685421994884913e-05, |
|
"loss": 0.3208, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.967774936061381e-05, |
|
"loss": 0.6654, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.9670076726342712e-05, |
|
"loss": 0.5581, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.966240409207161e-05, |
|
"loss": 0.7456, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.9654731457800512e-05, |
|
"loss": 0.4437, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.9647058823529414e-05, |
|
"loss": 0.5525, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.9639386189258312e-05, |
|
"loss": 0.6673, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.963171355498721e-05, |
|
"loss": 0.435, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.9624040920716115e-05, |
|
"loss": 0.7415, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.9616368286445014e-05, |
|
"loss": 0.2981, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 2.9608695652173912e-05, |
|
"loss": 0.5179, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 2.9601023017902817e-05, |
|
"loss": 0.4005, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 2.9593350383631715e-05, |
|
"loss": 0.4598, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 2.9585677749360613e-05, |
|
"loss": 0.2553, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2.9578005115089515e-05, |
|
"loss": 0.4302, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2.9570332480818416e-05, |
|
"loss": 0.7894, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2.9562659846547315e-05, |
|
"loss": 0.213, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2.9554987212276216e-05, |
|
"loss": 0.5507, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 2.9547314578005115e-05, |
|
"loss": 0.7676, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 2.9539641943734016e-05, |
|
"loss": 0.3418, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 2.9531969309462918e-05, |
|
"loss": 0.8533, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 2.9524296675191816e-05, |
|
"loss": 0.3659, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.9516624040920714e-05, |
|
"loss": 0.6482, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.950895140664962e-05, |
|
"loss": 0.5484, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.9501278772378517e-05, |
|
"loss": 0.4508, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.9493606138107416e-05, |
|
"loss": 0.4401, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 2.9485933503836317e-05, |
|
"loss": 0.5455, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 2.947826086956522e-05, |
|
"loss": 0.722, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 2.9470588235294117e-05, |
|
"loss": 0.2874, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 2.946291560102302e-05, |
|
"loss": 0.5973, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 2.9455242966751917e-05, |
|
"loss": 0.5898, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 2.944757033248082e-05, |
|
"loss": 1.0125, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 2.943989769820972e-05, |
|
"loss": 0.2511, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 2.943222506393862e-05, |
|
"loss": 0.545, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.942455242966752e-05, |
|
"loss": 0.593, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.941687979539642e-05, |
|
"loss": 0.2319, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.940920716112532e-05, |
|
"loss": 0.2215, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.9401534526854218e-05, |
|
"loss": 0.1964, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 2.9393861892583123e-05, |
|
"loss": 0.1471, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 2.938618925831202e-05, |
|
"loss": 1.5965, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 2.9378516624040923e-05, |
|
"loss": 1.0772, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 2.937084398976982e-05, |
|
"loss": 0.2005, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 2.9363171355498723e-05, |
|
"loss": 0.9628, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 2.9355498721227624e-05, |
|
"loss": 0.1503, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 2.9347826086956523e-05, |
|
"loss": 0.6201, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 2.934015345268542e-05, |
|
"loss": 0.6094, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 2.9332480818414326e-05, |
|
"loss": 1.385, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 2.9324808184143224e-05, |
|
"loss": 0.1552, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 2.9317135549872122e-05, |
|
"loss": 0.5828, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 2.9309462915601024e-05, |
|
"loss": 0.6018, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.9301790281329925e-05, |
|
"loss": 1.0614, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.9294117647058824e-05, |
|
"loss": 0.566, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.9286445012787725e-05, |
|
"loss": 0.5958, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.9278772378516624e-05, |
|
"loss": 0.5838, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 2.9271099744245525e-05, |
|
"loss": 0.5909, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 2.9263427109974427e-05, |
|
"loss": 0.5862, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 2.9255754475703325e-05, |
|
"loss": 0.2031, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 2.9248081841432223e-05, |
|
"loss": 0.1995, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 2.9240409207161128e-05, |
|
"loss": 0.24, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 2.9232736572890026e-05, |
|
"loss": 0.5845, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 2.9225063938618925e-05, |
|
"loss": 0.964, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 2.921739130434783e-05, |
|
"loss": 0.9595, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 2.9209718670076728e-05, |
|
"loss": 0.2033, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 2.9202046035805626e-05, |
|
"loss": 0.213, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 2.9194373401534528e-05, |
|
"loss": 0.2042, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 2.918670076726343e-05, |
|
"loss": 0.5783, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.9179028132992328e-05, |
|
"loss": 1.0822, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.917135549872123e-05, |
|
"loss": 1.2643, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.9163682864450127e-05, |
|
"loss": 0.2593, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.915601023017903e-05, |
|
"loss": 0.5783, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.914833759590793e-05, |
|
"loss": 0.5721, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.914066496163683e-05, |
|
"loss": 0.5808, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.9132992327365727e-05, |
|
"loss": 1.2578, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.9125319693094632e-05, |
|
"loss": 0.5709, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.911764705882353e-05, |
|
"loss": 0.2159, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.910997442455243e-05, |
|
"loss": 0.2329, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.910230179028133e-05, |
|
"loss": 0.926, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.9094629156010232e-05, |
|
"loss": 0.9095, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.908695652173913e-05, |
|
"loss": 0.5622, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.907928388746803e-05, |
|
"loss": 0.5703, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.907161125319693e-05, |
|
"loss": 0.5669, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.906393861892583e-05, |
|
"loss": 0.2637, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 2.9056265984654733e-05, |
|
"loss": 1.1077, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 2.904859335038363e-05, |
|
"loss": 0.5653, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 2.9040920716112533e-05, |
|
"loss": 0.5584, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 2.9033248081841434e-05, |
|
"loss": 0.5544, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.9025575447570333e-05, |
|
"loss": 0.5663, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.901790281329923e-05, |
|
"loss": 1.0808, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.9010230179028136e-05, |
|
"loss": 0.8736, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.9002557544757034e-05, |
|
"loss": 0.6435, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.8994884910485932e-05, |
|
"loss": 0.786, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.8987212276214834e-05, |
|
"loss": 0.5717, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.8979539641943736e-05, |
|
"loss": 0.5618, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.8971867007672637e-05, |
|
"loss": 0.5785, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.8964194373401535e-05, |
|
"loss": 0.7601, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.8956521739130434e-05, |
|
"loss": 0.8112, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.894884910485934e-05, |
|
"loss": 0.5663, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.8941176470588237e-05, |
|
"loss": 0.5631, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.8933503836317135e-05, |
|
"loss": 0.7363, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.8925831202046037e-05, |
|
"loss": 0.4552, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.8918158567774938e-05, |
|
"loss": 0.4599, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.8910485933503836e-05, |
|
"loss": 0.7848, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.8902813299232738e-05, |
|
"loss": 0.5907, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.8895140664961636e-05, |
|
"loss": 0.4264, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.8887468030690538e-05, |
|
"loss": 0.5697, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.887979539641944e-05, |
|
"loss": 0.7643, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.8872122762148338e-05, |
|
"loss": 0.3635, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.8864450127877236e-05, |
|
"loss": 0.3715, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.885677749360614e-05, |
|
"loss": 0.554, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.884910485933504e-05, |
|
"loss": 0.8044, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.8841432225063937e-05, |
|
"loss": 0.5673, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.8833759590792842e-05, |
|
"loss": 0.8042, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.882608695652174e-05, |
|
"loss": 0.5638, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.881841432225064e-05, |
|
"loss": 0.3584, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.881074168797954e-05, |
|
"loss": 0.8211, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.8803069053708442e-05, |
|
"loss": 0.5708, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.879539641943734e-05, |
|
"loss": 0.8301, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.8787723785166242e-05, |
|
"loss": 0.8229, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.878005115089514e-05, |
|
"loss": 0.5645, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.8772378516624042e-05, |
|
"loss": 0.3529, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.8764705882352943e-05, |
|
"loss": 0.5578, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.875703324808184e-05, |
|
"loss": 0.5659, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.874936061381074e-05, |
|
"loss": 0.7495, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.8741687979539645e-05, |
|
"loss": 0.3162, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.8734015345268543e-05, |
|
"loss": 0.5616, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.872634271099744e-05, |
|
"loss": 0.4034, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.8718670076726343e-05, |
|
"loss": 0.3163, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.8710997442455245e-05, |
|
"loss": 0.5514, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.8703324808184143e-05, |
|
"loss": 0.826, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.8695652173913044e-05, |
|
"loss": 0.8363, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.8687979539641943e-05, |
|
"loss": 0.8288, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.8680306905370844e-05, |
|
"loss": 0.5929, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.8672634271099746e-05, |
|
"loss": 0.565, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.8664961636828644e-05, |
|
"loss": 0.5642, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.8657289002557546e-05, |
|
"loss": 0.5609, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.8649616368286447e-05, |
|
"loss": 0.5641, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.8641943734015345e-05, |
|
"loss": 0.8535, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.8634271099744244e-05, |
|
"loss": 0.5652, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.862659846547315e-05, |
|
"loss": 0.4397, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.8618925831202047e-05, |
|
"loss": 0.567, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.8611253196930945e-05, |
|
"loss": 0.7517, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.8603580562659847e-05, |
|
"loss": 0.5839, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.859590792838875e-05, |
|
"loss": 0.5677, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.8588235294117647e-05, |
|
"loss": 0.5678, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.8580562659846548e-05, |
|
"loss": 1.3045, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.8572890025575446e-05, |
|
"loss": 0.804, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.856521739130435e-05, |
|
"loss": 1.0303, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.855754475703325e-05, |
|
"loss": 0.3303, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.8549872122762148e-05, |
|
"loss": 0.5677, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.854219948849105e-05, |
|
"loss": 0.7972, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.853452685421995e-05, |
|
"loss": 0.5644, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.852685421994885e-05, |
|
"loss": 0.9294, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.851918158567775e-05, |
|
"loss": 0.5665, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.851150895140665e-05, |
|
"loss": 0.7963, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.850383631713555e-05, |
|
"loss": 0.5659, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8496163682864452e-05, |
|
"loss": 0.5718, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.848849104859335e-05, |
|
"loss": 0.3526, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.8480818414322252e-05, |
|
"loss": 0.574, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.8473145780051154e-05, |
|
"loss": 0.9996, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.8465473145780052e-05, |
|
"loss": 0.7634, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.845780051150895e-05, |
|
"loss": 0.5639, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8450127877237855e-05, |
|
"loss": 0.5737, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8442455242966753e-05, |
|
"loss": 0.7793, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8434782608695652e-05, |
|
"loss": 0.6769, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8427109974424553e-05, |
|
"loss": 0.5748, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8419437340153455e-05, |
|
"loss": 0.3602, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8411764705882353e-05, |
|
"loss": 0.5714, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8404092071611255e-05, |
|
"loss": 0.5738, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8396419437340153e-05, |
|
"loss": 0.5719, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.8388746803069055e-05, |
|
"loss": 0.5753, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.8381074168797956e-05, |
|
"loss": 0.5717, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.8373401534526854e-05, |
|
"loss": 0.9794, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.8365728900255753e-05, |
|
"loss": 0.5737, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.8358056265984658e-05, |
|
"loss": 0.5723, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.8350383631713556e-05, |
|
"loss": 0.5656, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.8342710997442454e-05, |
|
"loss": 0.4259, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.8335038363171356e-05, |
|
"loss": 0.3076, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.8327365728900257e-05, |
|
"loss": 0.338, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.8319693094629156e-05, |
|
"loss": 0.5612, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.8312020460358057e-05, |
|
"loss": 0.5249, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.8304347826086955e-05, |
|
"loss": 0.5671, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.8296675191815857e-05, |
|
"loss": 0.5574, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.828900255754476e-05, |
|
"loss": 0.8152, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.8281329923273657e-05, |
|
"loss": 0.5647, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.827365728900256e-05, |
|
"loss": 0.5632, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.826598465473146e-05, |
|
"loss": 0.3391, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.8258312020460358e-05, |
|
"loss": 0.55, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.8250639386189257e-05, |
|
"loss": 0.5672, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.824296675191816e-05, |
|
"loss": 0.285, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.823529411764706e-05, |
|
"loss": 0.2809, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.8227621483375958e-05, |
|
"loss": 0.5552, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.821994884910486e-05, |
|
"loss": 0.563, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.821227621483376e-05, |
|
"loss": 0.5628, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.820460358056266e-05, |
|
"loss": 0.2548, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.819693094629156e-05, |
|
"loss": 0.871, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.818925831202046e-05, |
|
"loss": 0.8797, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.818158567774936e-05, |
|
"loss": 0.2453, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.8173913043478262e-05, |
|
"loss": 0.2464, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.816624040920716e-05, |
|
"loss": 0.3134, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.8158567774936062e-05, |
|
"loss": 0.2431, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.8150895140664964e-05, |
|
"loss": 0.2308, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.8143222506393862e-05, |
|
"loss": 0.2173, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.8135549872122764e-05, |
|
"loss": 0.5665, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.8127877237851662e-05, |
|
"loss": 0.5592, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.8120204603580564e-05, |
|
"loss": 0.473, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.8112531969309465e-05, |
|
"loss": 0.9458, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.8104859335038363e-05, |
|
"loss": 0.1879, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.8097186700767265e-05, |
|
"loss": 0.6797, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8089514066496167e-05, |
|
"loss": 0.7592, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8081841432225065e-05, |
|
"loss": 0.4109, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8074168797953963e-05, |
|
"loss": 0.5821, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8066496163682868e-05, |
|
"loss": 0.2057, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.8058823529411766e-05, |
|
"loss": 0.5322, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.8051150895140665e-05, |
|
"loss": 0.5208, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.8043478260869566e-05, |
|
"loss": 0.6799, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.8035805626598468e-05, |
|
"loss": 0.5894, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8028132992327366e-05, |
|
"loss": 0.4051, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8020460358056268e-05, |
|
"loss": 0.3695, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8012787723785166e-05, |
|
"loss": 0.4869, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8005115089514067e-05, |
|
"loss": 0.6075, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.799744245524297e-05, |
|
"loss": 0.4245, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.7989769820971867e-05, |
|
"loss": 0.5122, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.7982097186700765e-05, |
|
"loss": 0.452, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.797442455242967e-05, |
|
"loss": 0.9568, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.796675191815857e-05, |
|
"loss": 0.5846, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.7959079283887467e-05, |
|
"loss": 0.3743, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.795140664961637e-05, |
|
"loss": 0.1922, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.794373401534527e-05, |
|
"loss": 0.5303, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.793606138107417e-05, |
|
"loss": 0.7253, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.792838874680307e-05, |
|
"loss": 0.2794, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.7920716112531968e-05, |
|
"loss": 0.577, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.791304347826087e-05, |
|
"loss": 0.614, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.790537084398977e-05, |
|
"loss": 0.5882, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.789769820971867e-05, |
|
"loss": 0.2503, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.789002557544757e-05, |
|
"loss": 0.4195, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.7882352941176473e-05, |
|
"loss": 0.2959, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.787468030690537e-05, |
|
"loss": 0.3732, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.786700767263427e-05, |
|
"loss": 0.5729, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.7859335038363174e-05, |
|
"loss": 0.3179, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.7851662404092073e-05, |
|
"loss": 0.6823, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.784398976982097e-05, |
|
"loss": 0.2663, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.7836317135549872e-05, |
|
"loss": 0.4436, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.7828644501278774e-05, |
|
"loss": 0.3089, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.7820971867007672e-05, |
|
"loss": 0.5888, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.7813299232736574e-05, |
|
"loss": 0.2115, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.7805626598465472e-05, |
|
"loss": 0.4005, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.7797953964194374e-05, |
|
"loss": 0.3097, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.7790281329923275e-05, |
|
"loss": 0.5847, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7782608695652174e-05, |
|
"loss": 1.0164, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7774936061381072e-05, |
|
"loss": 1.1321, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7767263427109977e-05, |
|
"loss": 0.592, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7759590792838875e-05, |
|
"loss": 1.0191, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.7751918158567777e-05, |
|
"loss": 0.9199, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.7744245524296675e-05, |
|
"loss": 0.2874, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.7736572890025576e-05, |
|
"loss": 0.2469, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.7728900255754478e-05, |
|
"loss": 0.8364, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.7721227621483376e-05, |
|
"loss": 0.7365, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.7713554987212278e-05, |
|
"loss": 0.5595, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.770588235294118e-05, |
|
"loss": 0.2546, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.7698209718670078e-05, |
|
"loss": 0.8183, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.7690537084398976e-05, |
|
"loss": 0.5385, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.768286445012788e-05, |
|
"loss": 0.2763, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.767519181585678e-05, |
|
"loss": 0.3045, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.7667519181585677e-05, |
|
"loss": 0.3354, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.765984654731458e-05, |
|
"loss": 0.9366, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.765217391304348e-05, |
|
"loss": 0.6212, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.764450127877238e-05, |
|
"loss": 0.1161, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.763682864450128e-05, |
|
"loss": 0.4929, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.762915601023018e-05, |
|
"loss": 0.8167, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.762148337595908e-05, |
|
"loss": 0.2521, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.7613810741687982e-05, |
|
"loss": 0.4174, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.760613810741688e-05, |
|
"loss": 0.3822, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.759846547314578e-05, |
|
"loss": 0.5355, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.7590792838874683e-05, |
|
"loss": 0.561, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.758312020460358e-05, |
|
"loss": 0.5126, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.757544757033248e-05, |
|
"loss": 0.2301, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.756777493606138e-05, |
|
"loss": 0.1698, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7560102301790283e-05, |
|
"loss": 0.1634, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.755242966751918e-05, |
|
"loss": 0.5955, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7544757033248083e-05, |
|
"loss": 0.1568, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.753708439897698e-05, |
|
"loss": 0.5167, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7529411764705883e-05, |
|
"loss": 1.0028, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7521739130434784e-05, |
|
"loss": 1.2567, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7514066496163682e-05, |
|
"loss": 0.6094, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7506393861892584e-05, |
|
"loss": 0.6059, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7498721227621486e-05, |
|
"loss": 1.0593, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7491048593350384e-05, |
|
"loss": 0.1506, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7483375959079282e-05, |
|
"loss": 1.5106, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.7475703324808187e-05, |
|
"loss": 0.1827, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.7468030690537085e-05, |
|
"loss": 1.4961, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.7460358056265984e-05, |
|
"loss": 1.0456, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.7452685421994885e-05, |
|
"loss": 1.2063, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.7445012787723787e-05, |
|
"loss": 0.5837, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.7437340153452685e-05, |
|
"loss": 1.4536, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.7429667519181587e-05, |
|
"loss": 0.5803, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.7421994884910485e-05, |
|
"loss": 0.5896, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.7414322250639386e-05, |
|
"loss": 0.1822, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.7406649616368288e-05, |
|
"loss": 1.7893, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.7398976982097186e-05, |
|
"loss": 0.9723, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.7391304347826085e-05, |
|
"loss": 1.2815, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.738363171355499e-05, |
|
"loss": 0.5765, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.7375959079283888e-05, |
|
"loss": 1.2429, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.7368286445012786e-05, |
|
"loss": 0.1343, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.7360613810741688e-05, |
|
"loss": 0.2306, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.735294117647059e-05, |
|
"loss": 0.624, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.734526854219949e-05, |
|
"loss": 0.2371, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.7348484848484849, |
|
"eval_loss": 0.5804409980773926, |
|
"eval_runtime": 57.3336, |
|
"eval_samples_per_second": 6.907, |
|
"eval_steps_per_second": 1.727, |
|
"step": 396 |
|
} |
|
], |
|
"max_steps": 3960, |
|
"num_train_epochs": 10, |
|
"total_flos": 1.437147546912e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|