|
{ |
|
"best_metric": 0.9879931389365352, |
|
"best_model_checkpoint": "videomae-base-finetuned-numbers-augmented/checkpoint-2816", |
|
"epoch": 3.25, |
|
"eval_steps": 500, |
|
"global_step": 2816, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0035511363636363635, |
|
"grad_norm": 11.368436813354492, |
|
"learning_rate": 1.7730496453900712e-06, |
|
"loss": 2.4086, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.007102272727272727, |
|
"grad_norm": 13.567573547363281, |
|
"learning_rate": 3.5460992907801423e-06, |
|
"loss": 2.4404, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01065340909090909, |
|
"grad_norm": 10.261510848999023, |
|
"learning_rate": 5.319148936170213e-06, |
|
"loss": 2.4437, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.014204545454545454, |
|
"grad_norm": 10.864927291870117, |
|
"learning_rate": 7.092198581560285e-06, |
|
"loss": 2.4537, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01775568181818182, |
|
"grad_norm": 13.574213027954102, |
|
"learning_rate": 8.865248226950355e-06, |
|
"loss": 2.3604, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02130681818181818, |
|
"grad_norm": 12.087525367736816, |
|
"learning_rate": 1.0638297872340426e-05, |
|
"loss": 2.4028, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.024857954545454544, |
|
"grad_norm": 16.888160705566406, |
|
"learning_rate": 1.2411347517730498e-05, |
|
"loss": 2.3285, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.028409090909090908, |
|
"grad_norm": 11.21091365814209, |
|
"learning_rate": 1.418439716312057e-05, |
|
"loss": 2.4292, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03196022727272727, |
|
"grad_norm": 12.686086654663086, |
|
"learning_rate": 1.595744680851064e-05, |
|
"loss": 2.3094, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03551136363636364, |
|
"grad_norm": 13.735329627990723, |
|
"learning_rate": 1.773049645390071e-05, |
|
"loss": 2.2478, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0390625, |
|
"grad_norm": 15.364631652832031, |
|
"learning_rate": 1.950354609929078e-05, |
|
"loss": 2.2576, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04261363636363636, |
|
"grad_norm": 17.786842346191406, |
|
"learning_rate": 2.1276595744680852e-05, |
|
"loss": 2.2132, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.04616477272727273, |
|
"grad_norm": 15.442058563232422, |
|
"learning_rate": 2.3049645390070924e-05, |
|
"loss": 2.1968, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04971590909090909, |
|
"grad_norm": 18.52874755859375, |
|
"learning_rate": 2.4822695035460995e-05, |
|
"loss": 1.9613, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.053267045454545456, |
|
"grad_norm": 16.437816619873047, |
|
"learning_rate": 2.6595744680851064e-05, |
|
"loss": 2.055, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.056818181818181816, |
|
"grad_norm": 16.25989532470703, |
|
"learning_rate": 2.836879432624114e-05, |
|
"loss": 2.2668, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.060369318181818184, |
|
"grad_norm": 16.64414405822754, |
|
"learning_rate": 3.0141843971631207e-05, |
|
"loss": 2.0752, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.06392045454545454, |
|
"grad_norm": 22.03922462463379, |
|
"learning_rate": 3.191489361702128e-05, |
|
"loss": 1.9947, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06747159090909091, |
|
"grad_norm": 20.077898025512695, |
|
"learning_rate": 3.3687943262411347e-05, |
|
"loss": 1.8888, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.07102272727272728, |
|
"grad_norm": 17.019630432128906, |
|
"learning_rate": 3.546099290780142e-05, |
|
"loss": 1.8263, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07457386363636363, |
|
"grad_norm": 22.599655151367188, |
|
"learning_rate": 3.723404255319149e-05, |
|
"loss": 1.6843, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.078125, |
|
"grad_norm": 17.443727493286133, |
|
"learning_rate": 3.900709219858156e-05, |
|
"loss": 1.7081, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.08167613636363637, |
|
"grad_norm": 12.519112586975098, |
|
"learning_rate": 4.078014184397163e-05, |
|
"loss": 1.4262, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.08522727272727272, |
|
"grad_norm": 21.003129959106445, |
|
"learning_rate": 4.2553191489361704e-05, |
|
"loss": 1.6321, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.08877840909090909, |
|
"grad_norm": 9.726280212402344, |
|
"learning_rate": 4.432624113475177e-05, |
|
"loss": 1.6712, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.09232954545454546, |
|
"grad_norm": 24.422882080078125, |
|
"learning_rate": 4.609929078014185e-05, |
|
"loss": 2.0967, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.09588068181818182, |
|
"grad_norm": 11.92298698425293, |
|
"learning_rate": 4.787234042553192e-05, |
|
"loss": 1.7163, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.09943181818181818, |
|
"grad_norm": 14.245928764343262, |
|
"learning_rate": 4.964539007092199e-05, |
|
"loss": 1.9799, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.10298295454545454, |
|
"grad_norm": 14.423157691955566, |
|
"learning_rate": 4.9842146803472775e-05, |
|
"loss": 1.6602, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.10653409090909091, |
|
"grad_norm": 11.99425983428955, |
|
"learning_rate": 4.9644830307813736e-05, |
|
"loss": 1.6844, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11008522727272728, |
|
"grad_norm": 13.40038776397705, |
|
"learning_rate": 4.94475138121547e-05, |
|
"loss": 1.6082, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.11363636363636363, |
|
"grad_norm": 9.363014221191406, |
|
"learning_rate": 4.925019731649566e-05, |
|
"loss": 1.7191, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.1171875, |
|
"grad_norm": 16.138887405395508, |
|
"learning_rate": 4.905288082083662e-05, |
|
"loss": 1.621, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.12073863636363637, |
|
"grad_norm": 13.552618980407715, |
|
"learning_rate": 4.885556432517759e-05, |
|
"loss": 1.4569, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.12428977272727272, |
|
"grad_norm": 16.55213737487793, |
|
"learning_rate": 4.865824782951855e-05, |
|
"loss": 1.5886, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1278409090909091, |
|
"grad_norm": 4.656749248504639, |
|
"learning_rate": 4.8460931333859514e-05, |
|
"loss": 1.2231, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.13139204545454544, |
|
"grad_norm": 9.768326759338379, |
|
"learning_rate": 4.8263614838200475e-05, |
|
"loss": 1.65, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.13494318181818182, |
|
"grad_norm": 20.023557662963867, |
|
"learning_rate": 4.806629834254144e-05, |
|
"loss": 1.4166, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.13849431818181818, |
|
"grad_norm": 18.180370330810547, |
|
"learning_rate": 4.78689818468824e-05, |
|
"loss": 1.1376, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.14204545454545456, |
|
"grad_norm": 16.621734619140625, |
|
"learning_rate": 4.767166535122337e-05, |
|
"loss": 1.5926, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.1455965909090909, |
|
"grad_norm": 17.98001480102539, |
|
"learning_rate": 4.747434885556433e-05, |
|
"loss": 1.3632, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.14914772727272727, |
|
"grad_norm": 20.42987060546875, |
|
"learning_rate": 4.727703235990529e-05, |
|
"loss": 1.3245, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.15269886363636365, |
|
"grad_norm": 11.056243896484375, |
|
"learning_rate": 4.707971586424625e-05, |
|
"loss": 1.2668, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.15625, |
|
"grad_norm": 13.91126537322998, |
|
"learning_rate": 4.6882399368587215e-05, |
|
"loss": 1.0435, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.15980113636363635, |
|
"grad_norm": 19.263099670410156, |
|
"learning_rate": 4.6685082872928176e-05, |
|
"loss": 1.3369, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.16335227272727273, |
|
"grad_norm": 23.511869430541992, |
|
"learning_rate": 4.6487766377269145e-05, |
|
"loss": 1.1933, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.1669034090909091, |
|
"grad_norm": 14.228583335876465, |
|
"learning_rate": 4.629044988161011e-05, |
|
"loss": 1.591, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.17045454545454544, |
|
"grad_norm": 19.068967819213867, |
|
"learning_rate": 4.609313338595106e-05, |
|
"loss": 0.9233, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.17400568181818182, |
|
"grad_norm": 11.136780738830566, |
|
"learning_rate": 4.589581689029203e-05, |
|
"loss": 1.0833, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.17755681818181818, |
|
"grad_norm": 10.600316047668457, |
|
"learning_rate": 4.569850039463299e-05, |
|
"loss": 1.1883, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.18110795454545456, |
|
"grad_norm": 14.483977317810059, |
|
"learning_rate": 4.550118389897396e-05, |
|
"loss": 1.2224, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.1846590909090909, |
|
"grad_norm": 20.28068733215332, |
|
"learning_rate": 4.530386740331492e-05, |
|
"loss": 0.9271, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.18821022727272727, |
|
"grad_norm": 8.706443786621094, |
|
"learning_rate": 4.5106550907655884e-05, |
|
"loss": 0.876, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.19176136363636365, |
|
"grad_norm": 24.446735382080078, |
|
"learning_rate": 4.4909234411996846e-05, |
|
"loss": 0.9443, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.1953125, |
|
"grad_norm": 18.790678024291992, |
|
"learning_rate": 4.471191791633781e-05, |
|
"loss": 1.5365, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.19886363636363635, |
|
"grad_norm": 12.385162353515625, |
|
"learning_rate": 4.451460142067877e-05, |
|
"loss": 1.1072, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.20241477272727273, |
|
"grad_norm": 10.875085830688477, |
|
"learning_rate": 4.431728492501974e-05, |
|
"loss": 1.3267, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.2059659090909091, |
|
"grad_norm": 23.155399322509766, |
|
"learning_rate": 4.41199684293607e-05, |
|
"loss": 1.0875, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.20951704545454544, |
|
"grad_norm": 16.206789016723633, |
|
"learning_rate": 4.392265193370166e-05, |
|
"loss": 1.532, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.21306818181818182, |
|
"grad_norm": 10.199226379394531, |
|
"learning_rate": 4.372533543804262e-05, |
|
"loss": 1.3542, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.21661931818181818, |
|
"grad_norm": 24.252439498901367, |
|
"learning_rate": 4.3528018942383585e-05, |
|
"loss": 0.9932, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.22017045454545456, |
|
"grad_norm": 21.493459701538086, |
|
"learning_rate": 4.333070244672455e-05, |
|
"loss": 0.9927, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.2237215909090909, |
|
"grad_norm": 14.551782608032227, |
|
"learning_rate": 4.3133385951065515e-05, |
|
"loss": 0.86, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.22727272727272727, |
|
"grad_norm": 36.133609771728516, |
|
"learning_rate": 4.293606945540648e-05, |
|
"loss": 0.7792, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.23082386363636365, |
|
"grad_norm": 13.04199504852295, |
|
"learning_rate": 4.273875295974744e-05, |
|
"loss": 0.9251, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.234375, |
|
"grad_norm": 13.920279502868652, |
|
"learning_rate": 4.25414364640884e-05, |
|
"loss": 1.0642, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.23792613636363635, |
|
"grad_norm": 4.325495719909668, |
|
"learning_rate": 4.234411996842936e-05, |
|
"loss": 0.9275, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.24147727272727273, |
|
"grad_norm": 23.056415557861328, |
|
"learning_rate": 4.2146803472770324e-05, |
|
"loss": 1.1907, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.2450284090909091, |
|
"grad_norm": 11.812889099121094, |
|
"learning_rate": 4.194948697711129e-05, |
|
"loss": 1.0329, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.24857954545454544, |
|
"grad_norm": 11.360099792480469, |
|
"learning_rate": 4.1752170481452254e-05, |
|
"loss": 0.8968, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_accuracy": 0.6878216123499142, |
|
"eval_f1": 0.6885242103977541, |
|
"eval_loss": 0.8689462542533875, |
|
"eval_precision": 0.742302721464857, |
|
"eval_recall": 0.688613815159033, |
|
"eval_runtime": 653.0033, |
|
"eval_samples_per_second": 0.893, |
|
"eval_steps_per_second": 0.224, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.0021306818181819, |
|
"grad_norm": 5.516427040100098, |
|
"learning_rate": 4.1554853985793216e-05, |
|
"loss": 0.9278, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.0056818181818181, |
|
"grad_norm": 8.10287094116211, |
|
"learning_rate": 4.135753749013418e-05, |
|
"loss": 0.8833, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.0092329545454546, |
|
"grad_norm": 21.60080337524414, |
|
"learning_rate": 4.116022099447514e-05, |
|
"loss": 0.6148, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.0127840909090908, |
|
"grad_norm": 17.968292236328125, |
|
"learning_rate": 4.09629044988161e-05, |
|
"loss": 0.8788, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.0163352272727273, |
|
"grad_norm": 8.538472175598145, |
|
"learning_rate": 4.076558800315707e-05, |
|
"loss": 0.778, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.0198863636363635, |
|
"grad_norm": 4.343443870544434, |
|
"learning_rate": 4.056827150749803e-05, |
|
"loss": 0.5165, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.0234375, |
|
"grad_norm": 6.1510009765625, |
|
"learning_rate": 4.037095501183899e-05, |
|
"loss": 0.8024, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.0269886363636365, |
|
"grad_norm": 23.44464111328125, |
|
"learning_rate": 4.0173638516179955e-05, |
|
"loss": 1.0074, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.0305397727272727, |
|
"grad_norm": 16.371910095214844, |
|
"learning_rate": 3.997632202052092e-05, |
|
"loss": 0.9794, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.0340909090909092, |
|
"grad_norm": 8.532549858093262, |
|
"learning_rate": 3.977900552486188e-05, |
|
"loss": 0.8435, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.0376420454545454, |
|
"grad_norm": 15.13945484161377, |
|
"learning_rate": 3.958168902920285e-05, |
|
"loss": 0.9073, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.0411931818181819, |
|
"grad_norm": 12.980558395385742, |
|
"learning_rate": 3.938437253354381e-05, |
|
"loss": 0.8077, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.0447443181818181, |
|
"grad_norm": 7.141788005828857, |
|
"learning_rate": 3.9187056037884764e-05, |
|
"loss": 0.9853, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.0482954545454546, |
|
"grad_norm": 5.774048328399658, |
|
"learning_rate": 3.898973954222573e-05, |
|
"loss": 0.5617, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.0518465909090908, |
|
"grad_norm": 30.009641647338867, |
|
"learning_rate": 3.8792423046566694e-05, |
|
"loss": 0.8913, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.0553977272727273, |
|
"grad_norm": 11.07997989654541, |
|
"learning_rate": 3.8595106550907656e-05, |
|
"loss": 0.6674, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.0589488636363635, |
|
"grad_norm": 13.393827438354492, |
|
"learning_rate": 3.8397790055248625e-05, |
|
"loss": 0.5832, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.0625, |
|
"grad_norm": 10.827837944030762, |
|
"learning_rate": 3.8200473559589587e-05, |
|
"loss": 0.6509, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.0660511363636365, |
|
"grad_norm": 27.103227615356445, |
|
"learning_rate": 3.800315706393054e-05, |
|
"loss": 0.7135, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.0696022727272727, |
|
"grad_norm": 16.949966430664062, |
|
"learning_rate": 3.780584056827151e-05, |
|
"loss": 1.0985, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.0731534090909092, |
|
"grad_norm": 11.158037185668945, |
|
"learning_rate": 3.760852407261247e-05, |
|
"loss": 0.7335, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.0767045454545454, |
|
"grad_norm": 26.55095863342285, |
|
"learning_rate": 3.7411207576953434e-05, |
|
"loss": 0.7143, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.0802556818181819, |
|
"grad_norm": 7.920712947845459, |
|
"learning_rate": 3.72138910812944e-05, |
|
"loss": 0.8197, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.0838068181818181, |
|
"grad_norm": 2.0704779624938965, |
|
"learning_rate": 3.7016574585635364e-05, |
|
"loss": 0.7739, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.0873579545454546, |
|
"grad_norm": 29.551998138427734, |
|
"learning_rate": 3.681925808997632e-05, |
|
"loss": 0.9224, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.0909090909090908, |
|
"grad_norm": 9.859967231750488, |
|
"learning_rate": 3.662194159431729e-05, |
|
"loss": 0.8151, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.0944602272727273, |
|
"grad_norm": 10.744651794433594, |
|
"learning_rate": 3.642462509865825e-05, |
|
"loss": 0.4832, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.0980113636363635, |
|
"grad_norm": 3.085749864578247, |
|
"learning_rate": 3.622730860299921e-05, |
|
"loss": 0.5775, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.1015625, |
|
"grad_norm": 11.848700523376465, |
|
"learning_rate": 3.602999210734018e-05, |
|
"loss": 0.4641, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.1051136363636365, |
|
"grad_norm": 5.603400230407715, |
|
"learning_rate": 3.583267561168114e-05, |
|
"loss": 0.5069, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.1086647727272727, |
|
"grad_norm": 8.615111351013184, |
|
"learning_rate": 3.5635359116022096e-05, |
|
"loss": 0.5031, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.1122159090909092, |
|
"grad_norm": 23.804479598999023, |
|
"learning_rate": 3.5438042620363065e-05, |
|
"loss": 0.7348, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.1157670454545454, |
|
"grad_norm": 1.017351508140564, |
|
"learning_rate": 3.5240726124704027e-05, |
|
"loss": 0.7015, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.1193181818181819, |
|
"grad_norm": 14.092010498046875, |
|
"learning_rate": 3.504340962904499e-05, |
|
"loss": 0.5592, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.1228693181818181, |
|
"grad_norm": 20.209379196166992, |
|
"learning_rate": 3.484609313338596e-05, |
|
"loss": 0.7704, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.1264204545454546, |
|
"grad_norm": 0.6501548290252686, |
|
"learning_rate": 3.464877663772691e-05, |
|
"loss": 0.5026, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.1299715909090908, |
|
"grad_norm": 1.0749989748001099, |
|
"learning_rate": 3.4451460142067874e-05, |
|
"loss": 0.5371, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.1335227272727273, |
|
"grad_norm": 10.37779712677002, |
|
"learning_rate": 3.425414364640884e-05, |
|
"loss": 0.4032, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.1370738636363638, |
|
"grad_norm": 30.61161231994629, |
|
"learning_rate": 3.4056827150749804e-05, |
|
"loss": 0.6612, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.140625, |
|
"grad_norm": 28.0980167388916, |
|
"learning_rate": 3.3859510655090766e-05, |
|
"loss": 0.5713, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.1441761363636362, |
|
"grad_norm": 25.723461151123047, |
|
"learning_rate": 3.3662194159431734e-05, |
|
"loss": 0.6669, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.1477272727272727, |
|
"grad_norm": 7.138387203216553, |
|
"learning_rate": 3.346487766377269e-05, |
|
"loss": 0.5781, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.1512784090909092, |
|
"grad_norm": 27.993236541748047, |
|
"learning_rate": 3.326756116811365e-05, |
|
"loss": 0.6033, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.1548295454545454, |
|
"grad_norm": 3.16493558883667, |
|
"learning_rate": 3.307024467245462e-05, |
|
"loss": 0.787, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.1583806818181819, |
|
"grad_norm": 11.943394660949707, |
|
"learning_rate": 3.287292817679558e-05, |
|
"loss": 0.657, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.1619318181818181, |
|
"grad_norm": 3.7604823112487793, |
|
"learning_rate": 3.267561168113654e-05, |
|
"loss": 0.4096, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.1654829545454546, |
|
"grad_norm": 21.66828727722168, |
|
"learning_rate": 3.247829518547751e-05, |
|
"loss": 0.5554, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.1690340909090908, |
|
"grad_norm": 3.944326162338257, |
|
"learning_rate": 3.2280978689818467e-05, |
|
"loss": 0.8685, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.1725852272727273, |
|
"grad_norm": 4.541408061981201, |
|
"learning_rate": 3.208366219415943e-05, |
|
"loss": 0.651, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.1761363636363638, |
|
"grad_norm": 16.059410095214844, |
|
"learning_rate": 3.18863456985004e-05, |
|
"loss": 0.5313, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.1796875, |
|
"grad_norm": 29.15842056274414, |
|
"learning_rate": 3.168902920284136e-05, |
|
"loss": 0.6873, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.1832386363636362, |
|
"grad_norm": 27.368764877319336, |
|
"learning_rate": 3.149171270718232e-05, |
|
"loss": 0.7953, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.1867897727272727, |
|
"grad_norm": 12.318428039550781, |
|
"learning_rate": 3.129439621152329e-05, |
|
"loss": 0.316, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.1903409090909092, |
|
"grad_norm": 7.381816864013672, |
|
"learning_rate": 3.1097079715864244e-05, |
|
"loss": 0.5909, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.1938920454545454, |
|
"grad_norm": 56.74608612060547, |
|
"learning_rate": 3.0899763220205206e-05, |
|
"loss": 0.5496, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.1974431818181819, |
|
"grad_norm": 5.5123090744018555, |
|
"learning_rate": 3.0702446724546174e-05, |
|
"loss": 0.4415, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.2009943181818181, |
|
"grad_norm": 3.0995726585388184, |
|
"learning_rate": 3.0505130228887136e-05, |
|
"loss": 0.4572, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.2045454545454546, |
|
"grad_norm": 1.3388721942901611, |
|
"learning_rate": 3.03078137332281e-05, |
|
"loss": 0.4261, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.2080965909090908, |
|
"grad_norm": 3.432847738265991, |
|
"learning_rate": 3.0110497237569063e-05, |
|
"loss": 0.5804, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.2116477272727273, |
|
"grad_norm": 5.881500244140625, |
|
"learning_rate": 2.9913180741910025e-05, |
|
"loss": 0.3265, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.2151988636363638, |
|
"grad_norm": 27.702701568603516, |
|
"learning_rate": 2.971586424625099e-05, |
|
"loss": 0.8524, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.21875, |
|
"grad_norm": 3.5202479362487793, |
|
"learning_rate": 2.951854775059195e-05, |
|
"loss": 0.7208, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.2223011363636362, |
|
"grad_norm": 14.859701156616211, |
|
"learning_rate": 2.9321231254932913e-05, |
|
"loss": 0.6345, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.2258522727272727, |
|
"grad_norm": 2.571312427520752, |
|
"learning_rate": 2.912391475927388e-05, |
|
"loss": 0.3549, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.2294034090909092, |
|
"grad_norm": 8.97351360321045, |
|
"learning_rate": 2.892659826361484e-05, |
|
"loss": 0.507, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.2329545454545454, |
|
"grad_norm": 13.665780067443848, |
|
"learning_rate": 2.8729281767955802e-05, |
|
"loss": 0.4561, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.2365056818181819, |
|
"grad_norm": 13.264731407165527, |
|
"learning_rate": 2.8531965272296767e-05, |
|
"loss": 0.539, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.2400568181818181, |
|
"grad_norm": 29.683090209960938, |
|
"learning_rate": 2.833464877663773e-05, |
|
"loss": 0.5715, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.2436079545454546, |
|
"grad_norm": 8.478317260742188, |
|
"learning_rate": 2.813733228097869e-05, |
|
"loss": 0.183, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.2471590909090908, |
|
"grad_norm": 28.569978713989258, |
|
"learning_rate": 2.7940015785319656e-05, |
|
"loss": 0.5002, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_accuracy": 0.8542024013722127, |
|
"eval_f1": 0.8530975388761288, |
|
"eval_loss": 0.4373819828033447, |
|
"eval_precision": 0.8718298587182326, |
|
"eval_recall": 0.853515201868564, |
|
"eval_runtime": 695.9457, |
|
"eval_samples_per_second": 0.838, |
|
"eval_steps_per_second": 0.21, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 2.000710227272727, |
|
"grad_norm": 25.39379119873047, |
|
"learning_rate": 2.7742699289660618e-05, |
|
"loss": 0.6607, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.0042613636363638, |
|
"grad_norm": 3.422687530517578, |
|
"learning_rate": 2.754538279400158e-05, |
|
"loss": 0.5038, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.0078125, |
|
"grad_norm": 1.3007397651672363, |
|
"learning_rate": 2.7348066298342545e-05, |
|
"loss": 0.3052, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.0113636363636362, |
|
"grad_norm": 15.327472686767578, |
|
"learning_rate": 2.7150749802683506e-05, |
|
"loss": 0.2933, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.014914772727273, |
|
"grad_norm": 23.62551498413086, |
|
"learning_rate": 2.6953433307024468e-05, |
|
"loss": 0.4165, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.018465909090909, |
|
"grad_norm": 8.549898147583008, |
|
"learning_rate": 2.6756116811365433e-05, |
|
"loss": 0.3321, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.0220170454545454, |
|
"grad_norm": 25.098161697387695, |
|
"learning_rate": 2.6558800315706395e-05, |
|
"loss": 0.4109, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.0255681818181817, |
|
"grad_norm": 9.298599243164062, |
|
"learning_rate": 2.6361483820047357e-05, |
|
"loss": 0.4261, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.0291193181818183, |
|
"grad_norm": 66.2896957397461, |
|
"learning_rate": 2.6164167324388322e-05, |
|
"loss": 0.3508, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.0326704545454546, |
|
"grad_norm": 1.1833637952804565, |
|
"learning_rate": 2.5966850828729284e-05, |
|
"loss": 0.1169, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.036221590909091, |
|
"grad_norm": 32.40926742553711, |
|
"learning_rate": 2.5769534333070246e-05, |
|
"loss": 0.7922, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.039772727272727, |
|
"grad_norm": 31.9284610748291, |
|
"learning_rate": 2.557221783741121e-05, |
|
"loss": 0.3044, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.0433238636363638, |
|
"grad_norm": 0.32966673374176025, |
|
"learning_rate": 2.5374901341752172e-05, |
|
"loss": 0.5598, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.046875, |
|
"grad_norm": 3.2222678661346436, |
|
"learning_rate": 2.5177584846093134e-05, |
|
"loss": 0.5305, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.0504261363636362, |
|
"grad_norm": 5.157206058502197, |
|
"learning_rate": 2.4980268350434096e-05, |
|
"loss": 0.3806, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.053977272727273, |
|
"grad_norm": 29.749786376953125, |
|
"learning_rate": 2.478295185477506e-05, |
|
"loss": 0.6572, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.057528409090909, |
|
"grad_norm": 5.4849534034729, |
|
"learning_rate": 2.4585635359116023e-05, |
|
"loss": 0.2828, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.0610795454545454, |
|
"grad_norm": 1.8682303428649902, |
|
"learning_rate": 2.4388318863456985e-05, |
|
"loss": 0.275, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.0646306818181817, |
|
"grad_norm": 31.073657989501953, |
|
"learning_rate": 2.419100236779795e-05, |
|
"loss": 0.2635, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.0681818181818183, |
|
"grad_norm": 2.873807668685913, |
|
"learning_rate": 2.399368587213891e-05, |
|
"loss": 0.2189, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.0717329545454546, |
|
"grad_norm": 6.81038761138916, |
|
"learning_rate": 2.3796369376479873e-05, |
|
"loss": 0.2335, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.075284090909091, |
|
"grad_norm": 5.240432262420654, |
|
"learning_rate": 2.359905288082084e-05, |
|
"loss": 0.513, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.078835227272727, |
|
"grad_norm": 2.8783762454986572, |
|
"learning_rate": 2.34017363851618e-05, |
|
"loss": 0.2719, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.0823863636363638, |
|
"grad_norm": 4.599855422973633, |
|
"learning_rate": 2.3204419889502762e-05, |
|
"loss": 0.2012, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.0859375, |
|
"grad_norm": 1.6575523614883423, |
|
"learning_rate": 2.3007103393843727e-05, |
|
"loss": 0.1299, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.0894886363636362, |
|
"grad_norm": 0.3831326961517334, |
|
"learning_rate": 2.280978689818469e-05, |
|
"loss": 0.1766, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.093039772727273, |
|
"grad_norm": 1.1560420989990234, |
|
"learning_rate": 2.261247040252565e-05, |
|
"loss": 0.3844, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.096590909090909, |
|
"grad_norm": 1.8366397619247437, |
|
"learning_rate": 2.2415153906866616e-05, |
|
"loss": 0.3113, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.1001420454545454, |
|
"grad_norm": 10.303657531738281, |
|
"learning_rate": 2.2217837411207578e-05, |
|
"loss": 0.1612, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.1036931818181817, |
|
"grad_norm": 35.21874237060547, |
|
"learning_rate": 2.202052091554854e-05, |
|
"loss": 0.454, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.1072443181818183, |
|
"grad_norm": 4.324901580810547, |
|
"learning_rate": 2.1823204419889505e-05, |
|
"loss": 0.5713, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.1107954545454546, |
|
"grad_norm": 4.857918739318848, |
|
"learning_rate": 2.1625887924230466e-05, |
|
"loss": 0.4748, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.114346590909091, |
|
"grad_norm": 0.15146464109420776, |
|
"learning_rate": 2.1428571428571428e-05, |
|
"loss": 0.3261, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.117897727272727, |
|
"grad_norm": 19.752939224243164, |
|
"learning_rate": 2.1231254932912393e-05, |
|
"loss": 0.2406, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.1214488636363638, |
|
"grad_norm": 18.78827476501465, |
|
"learning_rate": 2.1033938437253355e-05, |
|
"loss": 0.4846, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.125, |
|
"grad_norm": 0.27601099014282227, |
|
"learning_rate": 2.0836621941594317e-05, |
|
"loss": 0.2026, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.1285511363636362, |
|
"grad_norm": 0.5743858218193054, |
|
"learning_rate": 2.0639305445935282e-05, |
|
"loss": 0.2579, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.132102272727273, |
|
"grad_norm": 0.10733279585838318, |
|
"learning_rate": 2.0441988950276244e-05, |
|
"loss": 0.2348, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.135653409090909, |
|
"grad_norm": 2.2115983963012695, |
|
"learning_rate": 2.0244672454617206e-05, |
|
"loss": 0.2632, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.1392045454545454, |
|
"grad_norm": 37.26658248901367, |
|
"learning_rate": 2.004735595895817e-05, |
|
"loss": 0.4384, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.1427556818181817, |
|
"grad_norm": 10.85006332397461, |
|
"learning_rate": 1.9850039463299132e-05, |
|
"loss": 0.2217, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.1463068181818183, |
|
"grad_norm": 4.717857360839844, |
|
"learning_rate": 1.9652722967640098e-05, |
|
"loss": 0.2748, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.1498579545454546, |
|
"grad_norm": 2.3380446434020996, |
|
"learning_rate": 1.945540647198106e-05, |
|
"loss": 0.1738, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.153409090909091, |
|
"grad_norm": 1.6607507467269897, |
|
"learning_rate": 1.925808997632202e-05, |
|
"loss": 0.2641, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.156960227272727, |
|
"grad_norm": 12.202836036682129, |
|
"learning_rate": 1.9060773480662986e-05, |
|
"loss": 0.2995, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.1605113636363638, |
|
"grad_norm": 20.316972732543945, |
|
"learning_rate": 1.8863456985003948e-05, |
|
"loss": 0.1573, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.1640625, |
|
"grad_norm": 34.030296325683594, |
|
"learning_rate": 1.866614048934491e-05, |
|
"loss": 0.266, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.1676136363636362, |
|
"grad_norm": 4.90464973449707, |
|
"learning_rate": 1.8468823993685875e-05, |
|
"loss": 0.1064, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.171164772727273, |
|
"grad_norm": 38.46381378173828, |
|
"learning_rate": 1.8271507498026837e-05, |
|
"loss": 0.2679, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.174715909090909, |
|
"grad_norm": 52.19645690917969, |
|
"learning_rate": 1.80741910023678e-05, |
|
"loss": 0.166, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.1782670454545454, |
|
"grad_norm": 15.714051246643066, |
|
"learning_rate": 1.7876874506708764e-05, |
|
"loss": 0.4399, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"grad_norm": 50.66694641113281, |
|
"learning_rate": 1.7679558011049725e-05, |
|
"loss": 0.1816, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.1853693181818183, |
|
"grad_norm": 4.8143205642700195, |
|
"learning_rate": 1.7482241515390687e-05, |
|
"loss": 0.5087, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.1889204545454546, |
|
"grad_norm": 1.34380042552948, |
|
"learning_rate": 1.7284925019731652e-05, |
|
"loss": 0.4257, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.192471590909091, |
|
"grad_norm": 6.62140417098999, |
|
"learning_rate": 1.7087608524072614e-05, |
|
"loss": 0.1767, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.196022727272727, |
|
"grad_norm": 34.63706588745117, |
|
"learning_rate": 1.6890292028413576e-05, |
|
"loss": 0.3301, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.1995738636363638, |
|
"grad_norm": 3.23504638671875, |
|
"learning_rate": 1.669297553275454e-05, |
|
"loss": 0.2174, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.203125, |
|
"grad_norm": 16.90635871887207, |
|
"learning_rate": 1.64956590370955e-05, |
|
"loss": 0.4572, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.2066761363636362, |
|
"grad_norm": 20.676467895507812, |
|
"learning_rate": 1.6298342541436465e-05, |
|
"loss": 0.3199, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.210227272727273, |
|
"grad_norm": 0.31672224402427673, |
|
"learning_rate": 1.610102604577743e-05, |
|
"loss": 0.2, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.213778409090909, |
|
"grad_norm": 8.79170036315918, |
|
"learning_rate": 1.5903709550118388e-05, |
|
"loss": 0.1675, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.2173295454545454, |
|
"grad_norm": 2.0761942863464355, |
|
"learning_rate": 1.5706393054459353e-05, |
|
"loss": 0.2614, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.2208806818181817, |
|
"grad_norm": 0.06360641866922379, |
|
"learning_rate": 1.550907655880032e-05, |
|
"loss": 0.2335, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 2.2244318181818183, |
|
"grad_norm": 18.36467170715332, |
|
"learning_rate": 1.5311760063141277e-05, |
|
"loss": 0.5066, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.2279829545454546, |
|
"grad_norm": 9.642960548400879, |
|
"learning_rate": 1.5114443567482242e-05, |
|
"loss": 0.2035, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 2.231534090909091, |
|
"grad_norm": 6.050065040588379, |
|
"learning_rate": 1.4917127071823205e-05, |
|
"loss": 0.1467, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.235085227272727, |
|
"grad_norm": 0.06556593626737595, |
|
"learning_rate": 1.4719810576164167e-05, |
|
"loss": 0.1243, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.2386363636363638, |
|
"grad_norm": 0.29696065187454224, |
|
"learning_rate": 1.452249408050513e-05, |
|
"loss": 0.2465, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.2421875, |
|
"grad_norm": 1.2069284915924072, |
|
"learning_rate": 1.4325177584846094e-05, |
|
"loss": 0.191, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 2.2457386363636362, |
|
"grad_norm": 2.228632688522339, |
|
"learning_rate": 1.4127861089187056e-05, |
|
"loss": 0.1718, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.249289772727273, |
|
"grad_norm": 1.17184317111969, |
|
"learning_rate": 1.393054459352802e-05, |
|
"loss": 0.3627, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"eval_accuracy": 0.9622641509433962, |
|
"eval_f1": 0.9617697305162668, |
|
"eval_loss": 0.1109052374958992, |
|
"eval_precision": 0.9647490495721144, |
|
"eval_recall": 0.9614389495077621, |
|
"eval_runtime": 696.0893, |
|
"eval_samples_per_second": 0.838, |
|
"eval_steps_per_second": 0.21, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 3.002840909090909, |
|
"grad_norm": 0.03763706982135773, |
|
"learning_rate": 1.3733228097868983e-05, |
|
"loss": 0.0171, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 3.0063920454545454, |
|
"grad_norm": 1.5585871934890747, |
|
"learning_rate": 1.3535911602209945e-05, |
|
"loss": 0.2023, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 3.0099431818181817, |
|
"grad_norm": 20.643203735351562, |
|
"learning_rate": 1.3338595106550908e-05, |
|
"loss": 0.1212, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 3.0134943181818183, |
|
"grad_norm": 0.19967862963676453, |
|
"learning_rate": 1.3141278610891871e-05, |
|
"loss": 0.0201, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 3.0170454545454546, |
|
"grad_norm": 3.6144466400146484, |
|
"learning_rate": 1.2943962115232833e-05, |
|
"loss": 0.1455, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 3.020596590909091, |
|
"grad_norm": 7.9842610359191895, |
|
"learning_rate": 1.2746645619573797e-05, |
|
"loss": 0.042, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 3.024147727272727, |
|
"grad_norm": 3.346548080444336, |
|
"learning_rate": 1.254932912391476e-05, |
|
"loss": 0.0365, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 3.0276988636363638, |
|
"grad_norm": 0.10210000723600388, |
|
"learning_rate": 1.2352012628255722e-05, |
|
"loss": 0.118, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 3.03125, |
|
"grad_norm": 2.003911018371582, |
|
"learning_rate": 1.2154696132596685e-05, |
|
"loss": 0.0461, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 3.0348011363636362, |
|
"grad_norm": 0.06448382884263992, |
|
"learning_rate": 1.1957379636937649e-05, |
|
"loss": 0.1276, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 3.038352272727273, |
|
"grad_norm": 59.72112274169922, |
|
"learning_rate": 1.176006314127861e-05, |
|
"loss": 0.1744, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 3.041903409090909, |
|
"grad_norm": 26.68014907836914, |
|
"learning_rate": 1.1562746645619574e-05, |
|
"loss": 0.0988, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 3.0454545454545454, |
|
"grad_norm": 1.2907896041870117, |
|
"learning_rate": 1.1365430149960538e-05, |
|
"loss": 0.0432, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 3.0490056818181817, |
|
"grad_norm": 2.807907819747925, |
|
"learning_rate": 1.11681136543015e-05, |
|
"loss": 0.1671, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 3.0525568181818183, |
|
"grad_norm": 0.22834616899490356, |
|
"learning_rate": 1.0970797158642463e-05, |
|
"loss": 0.0643, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 3.0561079545454546, |
|
"grad_norm": 0.23886850476264954, |
|
"learning_rate": 1.0773480662983426e-05, |
|
"loss": 0.1166, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 3.059659090909091, |
|
"grad_norm": 2.190918445587158, |
|
"learning_rate": 1.0576164167324388e-05, |
|
"loss": 0.0236, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 3.063210227272727, |
|
"grad_norm": 1.075537919998169, |
|
"learning_rate": 1.0378847671665353e-05, |
|
"loss": 0.1185, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 3.0667613636363638, |
|
"grad_norm": 34.861881256103516, |
|
"learning_rate": 1.0181531176006315e-05, |
|
"loss": 0.1225, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 3.0703125, |
|
"grad_norm": 0.2916108965873718, |
|
"learning_rate": 9.984214680347277e-06, |
|
"loss": 0.0877, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 3.0738636363636362, |
|
"grad_norm": 24.542476654052734, |
|
"learning_rate": 9.786898184688242e-06, |
|
"loss": 0.1007, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 3.077414772727273, |
|
"grad_norm": 1.0566986799240112, |
|
"learning_rate": 9.589581689029204e-06, |
|
"loss": 0.0563, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 3.080965909090909, |
|
"grad_norm": 20.17405128479004, |
|
"learning_rate": 9.392265193370165e-06, |
|
"loss": 0.0374, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 3.0845170454545454, |
|
"grad_norm": 0.07401897758245468, |
|
"learning_rate": 9.194948697711129e-06, |
|
"loss": 0.0495, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 3.0880681818181817, |
|
"grad_norm": 4.210412502288818, |
|
"learning_rate": 8.997632202052092e-06, |
|
"loss": 0.1265, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 3.0916193181818183, |
|
"grad_norm": 27.83361053466797, |
|
"learning_rate": 8.800315706393054e-06, |
|
"loss": 0.2213, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 3.0951704545454546, |
|
"grad_norm": 0.04297183081507683, |
|
"learning_rate": 8.602999210734018e-06, |
|
"loss": 0.1469, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 3.098721590909091, |
|
"grad_norm": 0.7070327401161194, |
|
"learning_rate": 8.405682715074981e-06, |
|
"loss": 0.0219, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 3.102272727272727, |
|
"grad_norm": 0.22129130363464355, |
|
"learning_rate": 8.208366219415943e-06, |
|
"loss": 0.0115, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.1058238636363638, |
|
"grad_norm": 0.8595547080039978, |
|
"learning_rate": 8.011049723756906e-06, |
|
"loss": 0.0124, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 3.109375, |
|
"grad_norm": 0.13839711248874664, |
|
"learning_rate": 7.81373322809787e-06, |
|
"loss": 0.0865, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 3.1129261363636362, |
|
"grad_norm": 0.04325387626886368, |
|
"learning_rate": 7.6164167324388314e-06, |
|
"loss": 0.0793, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 3.116477272727273, |
|
"grad_norm": 3.90751051902771, |
|
"learning_rate": 7.419100236779796e-06, |
|
"loss": 0.1025, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 3.120028409090909, |
|
"grad_norm": 0.03920091316103935, |
|
"learning_rate": 7.2217837411207575e-06, |
|
"loss": 0.2589, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 3.1235795454545454, |
|
"grad_norm": 0.2701418995857239, |
|
"learning_rate": 7.02446724546172e-06, |
|
"loss": 0.0066, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 3.1271306818181817, |
|
"grad_norm": 0.3685694634914398, |
|
"learning_rate": 6.827150749802684e-06, |
|
"loss": 0.0537, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 3.1306818181818183, |
|
"grad_norm": 0.03167586028575897, |
|
"learning_rate": 6.629834254143646e-06, |
|
"loss": 0.2329, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 3.1342329545454546, |
|
"grad_norm": 24.94156265258789, |
|
"learning_rate": 6.4325177584846105e-06, |
|
"loss": 0.113, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 3.137784090909091, |
|
"grad_norm": 2.3878488540649414, |
|
"learning_rate": 6.235201262825572e-06, |
|
"loss": 0.0289, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.141335227272727, |
|
"grad_norm": 0.1500893086194992, |
|
"learning_rate": 6.037884767166536e-06, |
|
"loss": 0.0164, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 3.1448863636363638, |
|
"grad_norm": 1.0166605710983276, |
|
"learning_rate": 5.840568271507498e-06, |
|
"loss": 0.0665, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 3.1484375, |
|
"grad_norm": 8.010841369628906, |
|
"learning_rate": 5.643251775848461e-06, |
|
"loss": 0.2883, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 3.1519886363636362, |
|
"grad_norm": 11.449795722961426, |
|
"learning_rate": 5.445935280189424e-06, |
|
"loss": 0.0892, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 3.155539772727273, |
|
"grad_norm": 41.279239654541016, |
|
"learning_rate": 5.248618784530387e-06, |
|
"loss": 0.0529, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 3.159090909090909, |
|
"grad_norm": 0.06051742658019066, |
|
"learning_rate": 5.05130228887135e-06, |
|
"loss": 0.1401, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 3.1626420454545454, |
|
"grad_norm": 0.08330044150352478, |
|
"learning_rate": 4.853985793212313e-06, |
|
"loss": 0.0411, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 3.1661931818181817, |
|
"grad_norm": 44.56931686401367, |
|
"learning_rate": 4.656669297553276e-06, |
|
"loss": 0.0657, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 3.1697443181818183, |
|
"grad_norm": 1.3760157823562622, |
|
"learning_rate": 4.459352801894238e-06, |
|
"loss": 0.1188, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 3.1732954545454546, |
|
"grad_norm": 0.1809810996055603, |
|
"learning_rate": 4.262036306235202e-06, |
|
"loss": 0.0305, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 3.176846590909091, |
|
"grad_norm": 1.571562647819519, |
|
"learning_rate": 4.064719810576164e-06, |
|
"loss": 0.0448, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 3.180397727272727, |
|
"grad_norm": 0.04679562523961067, |
|
"learning_rate": 3.867403314917127e-06, |
|
"loss": 0.0876, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 3.1839488636363638, |
|
"grad_norm": 0.40420079231262207, |
|
"learning_rate": 3.67008681925809e-06, |
|
"loss": 0.0399, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 3.1875, |
|
"grad_norm": 0.8966354727745056, |
|
"learning_rate": 3.472770323599053e-06, |
|
"loss": 0.0747, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 3.1910511363636362, |
|
"grad_norm": 0.499515563249588, |
|
"learning_rate": 3.2754538279400157e-06, |
|
"loss": 0.1019, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 3.194602272727273, |
|
"grad_norm": 0.09785713255405426, |
|
"learning_rate": 3.0781373322809787e-06, |
|
"loss": 0.032, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 3.198153409090909, |
|
"grad_norm": 0.037965867668390274, |
|
"learning_rate": 2.8808208366219414e-06, |
|
"loss": 0.0693, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 3.2017045454545454, |
|
"grad_norm": 0.5521709322929382, |
|
"learning_rate": 2.683504340962905e-06, |
|
"loss": 0.0761, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 3.2052556818181817, |
|
"grad_norm": 0.09314560890197754, |
|
"learning_rate": 2.4861878453038674e-06, |
|
"loss": 0.1744, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 3.2088068181818183, |
|
"grad_norm": 12.886544227600098, |
|
"learning_rate": 2.2888713496448305e-06, |
|
"loss": 0.1549, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 3.2123579545454546, |
|
"grad_norm": 0.18088915944099426, |
|
"learning_rate": 2.091554853985793e-06, |
|
"loss": 0.186, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 3.215909090909091, |
|
"grad_norm": 0.18304170668125153, |
|
"learning_rate": 1.8942383583267563e-06, |
|
"loss": 0.0137, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 3.219460227272727, |
|
"grad_norm": 5.272530555725098, |
|
"learning_rate": 1.6969218626677192e-06, |
|
"loss": 0.0369, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 3.2230113636363638, |
|
"grad_norm": 0.17552484571933746, |
|
"learning_rate": 1.499605367008682e-06, |
|
"loss": 0.0675, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 3.2265625, |
|
"grad_norm": 0.28992605209350586, |
|
"learning_rate": 1.3022888713496448e-06, |
|
"loss": 0.0744, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 3.2301136363636362, |
|
"grad_norm": 0.17767734825611115, |
|
"learning_rate": 1.1049723756906078e-06, |
|
"loss": 0.1551, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 3.233664772727273, |
|
"grad_norm": 0.43689748644828796, |
|
"learning_rate": 9.076558800315707e-07, |
|
"loss": 0.1754, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 3.237215909090909, |
|
"grad_norm": 0.956597089767456, |
|
"learning_rate": 7.103393843725336e-07, |
|
"loss": 0.0069, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 3.2407670454545454, |
|
"grad_norm": 25.514036178588867, |
|
"learning_rate": 5.130228887134964e-07, |
|
"loss": 0.0997, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 3.2443181818181817, |
|
"grad_norm": 3.5862650871276855, |
|
"learning_rate": 3.1570639305445935e-07, |
|
"loss": 0.1085, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 3.2478693181818183, |
|
"grad_norm": 0.035997312515974045, |
|
"learning_rate": 1.1838989739542227e-07, |
|
"loss": 0.0289, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"eval_accuracy": 0.9879931389365352, |
|
"eval_f1": 0.9879774481091409, |
|
"eval_loss": 0.037413839250802994, |
|
"eval_precision": 0.9880848996629442, |
|
"eval_recall": 0.9879997453526865, |
|
"eval_runtime": 692.1711, |
|
"eval_samples_per_second": 0.842, |
|
"eval_steps_per_second": 0.211, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"step": 2816, |
|
"total_flos": 1.4036798084625727e+19, |
|
"train_loss": 0.6578278151044453, |
|
"train_runtime": 18231.9061, |
|
"train_samples_per_second": 0.618, |
|
"train_steps_per_second": 0.154 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"eval_accuracy": 0.9558620689655173, |
|
"eval_f1": 0.9562337987051189, |
|
"eval_loss": 0.14939993619918823, |
|
"eval_precision": 0.9567608190198278, |
|
"eval_recall": 0.956465718392097, |
|
"eval_runtime": 911.348, |
|
"eval_samples_per_second": 0.796, |
|
"eval_steps_per_second": 0.2, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"eval_accuracy": 0.9558620689655173, |
|
"eval_f1": 0.9562337987051189, |
|
"eval_loss": 0.14939993619918823, |
|
"eval_precision": 0.9567608190198278, |
|
"eval_recall": 0.956465718392097, |
|
"eval_runtime": 930.9063, |
|
"eval_samples_per_second": 0.779, |
|
"eval_steps_per_second": 0.196, |
|
"step": 2816 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2816, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 9223372036854775807, |
|
"save_steps": 500, |
|
"total_flos": 1.4036798084625727e+19, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|