antalvdb's picture
Training in progress, epoch 4
0522b09 verified
raw
history blame
No virus
1.38 kB
{
"best_metric": 0.06558874629318973,
"best_model_checkpoint": "distilbert-base-uncased-finetuned-cola/run-6/checkpoint-42",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 42,
"is_hyper_param_search": true,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.0,
"eval_loss": 0.6143483519554138,
"eval_matthews_correlation": 0.0,
"eval_runtime": 32.3274,
"eval_samples_per_second": 32.264,
"eval_steps_per_second": 2.042,
"step": 14
},
{
"epoch": 2.0,
"eval_loss": 0.6015313863754272,
"eval_matthews_correlation": 0.0,
"eval_runtime": 31.2134,
"eval_samples_per_second": 33.415,
"eval_steps_per_second": 2.114,
"step": 28
},
{
"epoch": 3.0,
"eval_loss": 0.5886062383651733,
"eval_matthews_correlation": 0.06558874629318973,
"eval_runtime": 34.8082,
"eval_samples_per_second": 29.964,
"eval_steps_per_second": 1.896,
"step": 42
}
],
"logging_steps": 500,
"max_steps": 56,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"total_flos": 0,
"train_batch_size": 64,
"trial_name": null,
"trial_params": {
"learning_rate": 3.3647806166177066e-05,
"num_train_epochs": 4,
"per_device_train_batch_size": 64,
"seed": 8
}
}