|
{
|
|
"best_metric": 0.6355140186915887,
|
|
"best_model_checkpoint": "BEiT-RHS-DA\\checkpoint-360",
|
|
"epoch": 39.111111111111114,
|
|
"eval_steps": 500,
|
|
"global_step": 880,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 1.1363636363636365e-05,
|
|
"loss": 1.3359,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 2.272727272727273e-05,
|
|
"loss": 1.2357,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 0.7114280462265015,
|
|
"eval_runtime": 2.6053,
|
|
"eval_samples_per_second": 41.07,
|
|
"eval_steps_per_second": 2.687,
|
|
"step": 22
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 3.409090909090909e-05,
|
|
"loss": 0.9874,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 4.545454545454546e-05,
|
|
"loss": 0.6596,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 0.7058635950088501,
|
|
"eval_runtime": 2.8903,
|
|
"eval_samples_per_second": 37.02,
|
|
"eval_steps_per_second": 2.422,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 4.964114832535885e-05,
|
|
"loss": 0.3724,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 4.904306220095694e-05,
|
|
"loss": 0.206,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 1.14487624168396,
|
|
"eval_runtime": 2.567,
|
|
"eval_samples_per_second": 41.682,
|
|
"eval_steps_per_second": 2.727,
|
|
"step": 67
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 4.844497607655503e-05,
|
|
"loss": 0.1872,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 4.784688995215311e-05,
|
|
"loss": 0.1164,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"learning_rate": 4.72488038277512e-05,
|
|
"loss": 0.1664,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"eval_accuracy": 0.3925233644859813,
|
|
"eval_loss": 2.2061808109283447,
|
|
"eval_runtime": 2.5208,
|
|
"eval_samples_per_second": 42.448,
|
|
"eval_steps_per_second": 2.777,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 4.44,
|
|
"learning_rate": 4.665071770334928e-05,
|
|
"loss": 0.1706,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 4.89,
|
|
"learning_rate": 4.605263157894737e-05,
|
|
"loss": 0.1011,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"eval_accuracy": 0.4672897196261682,
|
|
"eval_loss": 2.040870189666748,
|
|
"eval_runtime": 2.5598,
|
|
"eval_samples_per_second": 41.8,
|
|
"eval_steps_per_second": 2.735,
|
|
"step": 112
|
|
},
|
|
{
|
|
"epoch": 5.33,
|
|
"learning_rate": 4.545454545454546e-05,
|
|
"loss": 0.2439,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 5.78,
|
|
"learning_rate": 4.485645933014354e-05,
|
|
"loss": 0.0653,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"eval_accuracy": 0.6261682242990654,
|
|
"eval_loss": 1.3038437366485596,
|
|
"eval_runtime": 2.6986,
|
|
"eval_samples_per_second": 39.65,
|
|
"eval_steps_per_second": 2.594,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 6.22,
|
|
"learning_rate": 4.425837320574163e-05,
|
|
"loss": 0.2341,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 6.67,
|
|
"learning_rate": 4.366028708133971e-05,
|
|
"loss": 0.2843,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 6.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 1.7210038900375366,
|
|
"eval_runtime": 2.6158,
|
|
"eval_samples_per_second": 40.905,
|
|
"eval_steps_per_second": 2.676,
|
|
"step": 157
|
|
},
|
|
{
|
|
"epoch": 7.11,
|
|
"learning_rate": 4.3062200956937806e-05,
|
|
"loss": 0.0424,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 7.56,
|
|
"learning_rate": 4.246411483253589e-05,
|
|
"loss": 0.0082,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"learning_rate": 4.1866028708133976e-05,
|
|
"loss": 0.059,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"eval_accuracy": 0.4672897196261682,
|
|
"eval_loss": 2.870556116104126,
|
|
"eval_runtime": 2.8378,
|
|
"eval_samples_per_second": 37.705,
|
|
"eval_steps_per_second": 2.467,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 8.44,
|
|
"learning_rate": 4.1267942583732064e-05,
|
|
"loss": 0.1224,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 8.89,
|
|
"learning_rate": 4.0669856459330146e-05,
|
|
"loss": 0.1318,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 8.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.451904058456421,
|
|
"eval_runtime": 2.6549,
|
|
"eval_samples_per_second": 40.303,
|
|
"eval_steps_per_second": 2.637,
|
|
"step": 202
|
|
},
|
|
{
|
|
"epoch": 9.33,
|
|
"learning_rate": 4.0071770334928235e-05,
|
|
"loss": 0.0267,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 9.78,
|
|
"learning_rate": 3.9473684210526316e-05,
|
|
"loss": 0.0501,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 10.0,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.2037127017974854,
|
|
"eval_runtime": 2.6808,
|
|
"eval_samples_per_second": 39.913,
|
|
"eval_steps_per_second": 2.611,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 10.22,
|
|
"learning_rate": 3.8875598086124405e-05,
|
|
"loss": 0.0481,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 10.67,
|
|
"learning_rate": 3.8277511961722486e-05,
|
|
"loss": 0.054,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 10.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.646747589111328,
|
|
"eval_runtime": 2.8537,
|
|
"eval_samples_per_second": 37.495,
|
|
"eval_steps_per_second": 2.453,
|
|
"step": 247
|
|
},
|
|
{
|
|
"epoch": 11.11,
|
|
"learning_rate": 3.7679425837320575e-05,
|
|
"loss": 0.0495,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 11.56,
|
|
"learning_rate": 3.7081339712918663e-05,
|
|
"loss": 0.0367,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 12.0,
|
|
"learning_rate": 3.6483253588516745e-05,
|
|
"loss": 0.0263,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 12.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 2.4033188819885254,
|
|
"eval_runtime": 2.6437,
|
|
"eval_samples_per_second": 40.474,
|
|
"eval_steps_per_second": 2.648,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 12.44,
|
|
"learning_rate": 3.5885167464114834e-05,
|
|
"loss": 0.0138,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 12.89,
|
|
"learning_rate": 3.5287081339712915e-05,
|
|
"loss": 0.0553,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 12.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 1.6589338779449463,
|
|
"eval_runtime": 2.6219,
|
|
"eval_samples_per_second": 40.811,
|
|
"eval_steps_per_second": 2.67,
|
|
"step": 292
|
|
},
|
|
{
|
|
"epoch": 13.33,
|
|
"learning_rate": 3.4688995215311004e-05,
|
|
"loss": 0.0092,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 13.78,
|
|
"learning_rate": 3.409090909090909e-05,
|
|
"loss": 0.0898,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 14.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 1.7657066583633423,
|
|
"eval_runtime": 2.6872,
|
|
"eval_samples_per_second": 39.819,
|
|
"eval_steps_per_second": 2.605,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 14.22,
|
|
"learning_rate": 3.349282296650718e-05,
|
|
"loss": 0.0304,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 14.67,
|
|
"learning_rate": 3.289473684210527e-05,
|
|
"loss": 0.0324,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 14.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.8265936374664307,
|
|
"eval_runtime": 2.6021,
|
|
"eval_samples_per_second": 41.121,
|
|
"eval_steps_per_second": 2.69,
|
|
"step": 337
|
|
},
|
|
{
|
|
"epoch": 15.11,
|
|
"learning_rate": 3.229665071770335e-05,
|
|
"loss": 0.0424,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 15.56,
|
|
"learning_rate": 3.169856459330144e-05,
|
|
"loss": 0.0556,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 16.0,
|
|
"learning_rate": 3.110047846889952e-05,
|
|
"loss": 0.0322,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 16.0,
|
|
"eval_accuracy": 0.6355140186915887,
|
|
"eval_loss": 1.7194138765335083,
|
|
"eval_runtime": 2.6447,
|
|
"eval_samples_per_second": 40.458,
|
|
"eval_steps_per_second": 2.647,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 16.44,
|
|
"learning_rate": 3.050239234449761e-05,
|
|
"loss": 0.0236,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 16.89,
|
|
"learning_rate": 2.9904306220095695e-05,
|
|
"loss": 0.03,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 16.98,
|
|
"eval_accuracy": 0.616822429906542,
|
|
"eval_loss": 2.035153865814209,
|
|
"eval_runtime": 2.6411,
|
|
"eval_samples_per_second": 40.513,
|
|
"eval_steps_per_second": 2.65,
|
|
"step": 382
|
|
},
|
|
{
|
|
"epoch": 17.33,
|
|
"learning_rate": 2.9306220095693783e-05,
|
|
"loss": 0.0244,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 17.78,
|
|
"learning_rate": 2.8708133971291868e-05,
|
|
"loss": 0.0392,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 18.0,
|
|
"eval_accuracy": 0.616822429906542,
|
|
"eval_loss": 2.412965774536133,
|
|
"eval_runtime": 2.7163,
|
|
"eval_samples_per_second": 39.392,
|
|
"eval_steps_per_second": 2.577,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 18.22,
|
|
"learning_rate": 2.8110047846889953e-05,
|
|
"loss": 0.0096,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 18.67,
|
|
"learning_rate": 2.751196172248804e-05,
|
|
"loss": 0.0428,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 18.98,
|
|
"eval_accuracy": 0.6074766355140186,
|
|
"eval_loss": 2.0628373622894287,
|
|
"eval_runtime": 2.5916,
|
|
"eval_samples_per_second": 41.288,
|
|
"eval_steps_per_second": 2.701,
|
|
"step": 427
|
|
},
|
|
{
|
|
"epoch": 19.11,
|
|
"learning_rate": 2.6913875598086123e-05,
|
|
"loss": 0.0189,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 19.56,
|
|
"learning_rate": 2.6315789473684212e-05,
|
|
"loss": 0.0119,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 20.0,
|
|
"learning_rate": 2.5717703349282297e-05,
|
|
"loss": 0.0127,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 20.0,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.743129253387451,
|
|
"eval_runtime": 2.5383,
|
|
"eval_samples_per_second": 42.154,
|
|
"eval_steps_per_second": 2.758,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 20.44,
|
|
"learning_rate": 2.5119617224880382e-05,
|
|
"loss": 0.0164,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 20.89,
|
|
"learning_rate": 2.452153110047847e-05,
|
|
"loss": 0.0187,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 20.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 2.700892686843872,
|
|
"eval_runtime": 2.592,
|
|
"eval_samples_per_second": 41.281,
|
|
"eval_steps_per_second": 2.701,
|
|
"step": 472
|
|
},
|
|
{
|
|
"epoch": 21.33,
|
|
"learning_rate": 2.3923444976076556e-05,
|
|
"loss": 0.0349,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 21.78,
|
|
"learning_rate": 2.332535885167464e-05,
|
|
"loss": 0.0469,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 22.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 2.578273057937622,
|
|
"eval_runtime": 2.6372,
|
|
"eval_samples_per_second": 40.573,
|
|
"eval_steps_per_second": 2.654,
|
|
"step": 495
|
|
},
|
|
{
|
|
"epoch": 22.22,
|
|
"learning_rate": 2.272727272727273e-05,
|
|
"loss": 0.0196,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 22.67,
|
|
"learning_rate": 2.2129186602870814e-05,
|
|
"loss": 0.0095,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 22.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 2.304006338119507,
|
|
"eval_runtime": 2.5827,
|
|
"eval_samples_per_second": 41.429,
|
|
"eval_steps_per_second": 2.71,
|
|
"step": 517
|
|
},
|
|
{
|
|
"epoch": 23.11,
|
|
"learning_rate": 2.1531100478468903e-05,
|
|
"loss": 0.0428,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 23.56,
|
|
"learning_rate": 2.0933014354066988e-05,
|
|
"loss": 0.0007,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 24.0,
|
|
"learning_rate": 2.0334928229665073e-05,
|
|
"loss": 0.0025,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 24.0,
|
|
"eval_accuracy": 0.616822429906542,
|
|
"eval_loss": 2.521818161010742,
|
|
"eval_runtime": 2.612,
|
|
"eval_samples_per_second": 40.964,
|
|
"eval_steps_per_second": 2.68,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 24.44,
|
|
"learning_rate": 1.9736842105263158e-05,
|
|
"loss": 0.0005,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 24.89,
|
|
"learning_rate": 1.9138755980861243e-05,
|
|
"loss": 0.0281,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 24.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 3.231044292449951,
|
|
"eval_runtime": 2.5813,
|
|
"eval_samples_per_second": 41.451,
|
|
"eval_steps_per_second": 2.712,
|
|
"step": 562
|
|
},
|
|
{
|
|
"epoch": 25.33,
|
|
"learning_rate": 1.8540669856459332e-05,
|
|
"loss": 0.0084,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 25.78,
|
|
"learning_rate": 1.7942583732057417e-05,
|
|
"loss": 0.0004,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 26.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 3.273077964782715,
|
|
"eval_runtime": 2.6145,
|
|
"eval_samples_per_second": 40.925,
|
|
"eval_steps_per_second": 2.677,
|
|
"step": 585
|
|
},
|
|
{
|
|
"epoch": 26.22,
|
|
"learning_rate": 1.7344497607655502e-05,
|
|
"loss": 0.001,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 26.67,
|
|
"learning_rate": 1.674641148325359e-05,
|
|
"loss": 0.0109,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 26.98,
|
|
"eval_accuracy": 0.6261682242990654,
|
|
"eval_loss": 2.480931282043457,
|
|
"eval_runtime": 2.6486,
|
|
"eval_samples_per_second": 40.399,
|
|
"eval_steps_per_second": 2.643,
|
|
"step": 607
|
|
},
|
|
{
|
|
"epoch": 27.11,
|
|
"learning_rate": 1.6148325358851675e-05,
|
|
"loss": 0.0158,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 27.56,
|
|
"learning_rate": 1.555023923444976e-05,
|
|
"loss": 0.0375,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 28.0,
|
|
"learning_rate": 1.4952153110047847e-05,
|
|
"loss": 0.0191,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 28.0,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 2.782459259033203,
|
|
"eval_runtime": 2.7194,
|
|
"eval_samples_per_second": 39.347,
|
|
"eval_steps_per_second": 2.574,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 28.44,
|
|
"learning_rate": 1.4354066985645934e-05,
|
|
"loss": 0.0057,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 28.89,
|
|
"learning_rate": 1.375598086124402e-05,
|
|
"loss": 0.0005,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 28.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.528034210205078,
|
|
"eval_runtime": 2.584,
|
|
"eval_samples_per_second": 41.408,
|
|
"eval_steps_per_second": 2.709,
|
|
"step": 652
|
|
},
|
|
{
|
|
"epoch": 29.33,
|
|
"learning_rate": 1.3157894736842106e-05,
|
|
"loss": 0.0136,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 29.78,
|
|
"learning_rate": 1.2559808612440191e-05,
|
|
"loss": 0.0093,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 30.0,
|
|
"eval_accuracy": 0.6074766355140186,
|
|
"eval_loss": 2.828961133956909,
|
|
"eval_runtime": 2.5935,
|
|
"eval_samples_per_second": 41.257,
|
|
"eval_steps_per_second": 2.699,
|
|
"step": 675
|
|
},
|
|
{
|
|
"epoch": 30.22,
|
|
"learning_rate": 1.1961722488038278e-05,
|
|
"loss": 0.0004,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 30.67,
|
|
"learning_rate": 1.1363636363636365e-05,
|
|
"loss": 0.0224,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 30.98,
|
|
"eval_accuracy": 0.5794392523364486,
|
|
"eval_loss": 2.9545552730560303,
|
|
"eval_runtime": 2.7643,
|
|
"eval_samples_per_second": 38.707,
|
|
"eval_steps_per_second": 2.532,
|
|
"step": 697
|
|
},
|
|
{
|
|
"epoch": 31.11,
|
|
"learning_rate": 1.0765550239234451e-05,
|
|
"loss": 0.0325,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 31.56,
|
|
"learning_rate": 1.0167464114832537e-05,
|
|
"loss": 0.0102,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 32.0,
|
|
"learning_rate": 9.569377990430622e-06,
|
|
"loss": 0.0011,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 32.0,
|
|
"eval_accuracy": 0.6074766355140186,
|
|
"eval_loss": 3.0147745609283447,
|
|
"eval_runtime": 2.6,
|
|
"eval_samples_per_second": 41.155,
|
|
"eval_steps_per_second": 2.692,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 32.44,
|
|
"learning_rate": 8.971291866028708e-06,
|
|
"loss": 0.0014,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 32.89,
|
|
"learning_rate": 8.373205741626795e-06,
|
|
"loss": 0.003,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 32.98,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 3.2915587425231934,
|
|
"eval_runtime": 2.5651,
|
|
"eval_samples_per_second": 41.714,
|
|
"eval_steps_per_second": 2.729,
|
|
"step": 742
|
|
},
|
|
{
|
|
"epoch": 33.33,
|
|
"learning_rate": 7.77511961722488e-06,
|
|
"loss": 0.0005,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 33.78,
|
|
"learning_rate": 7.177033492822967e-06,
|
|
"loss": 0.0003,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 34.0,
|
|
"eval_accuracy": 0.5981308411214953,
|
|
"eval_loss": 3.292980909347534,
|
|
"eval_runtime": 2.5274,
|
|
"eval_samples_per_second": 42.337,
|
|
"eval_steps_per_second": 2.77,
|
|
"step": 765
|
|
},
|
|
{
|
|
"epoch": 34.22,
|
|
"learning_rate": 6.578947368421053e-06,
|
|
"loss": 0.0003,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 34.67,
|
|
"learning_rate": 5.980861244019139e-06,
|
|
"loss": 0.0003,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 34.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.628692626953125,
|
|
"eval_runtime": 2.6394,
|
|
"eval_samples_per_second": 40.539,
|
|
"eval_steps_per_second": 2.652,
|
|
"step": 787
|
|
},
|
|
{
|
|
"epoch": 35.11,
|
|
"learning_rate": 5.382775119617226e-06,
|
|
"loss": 0.0029,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 35.56,
|
|
"learning_rate": 4.784688995215311e-06,
|
|
"loss": 0.0003,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 36.0,
|
|
"learning_rate": 4.186602870813398e-06,
|
|
"loss": 0.0002,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 36.0,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.6918015480041504,
|
|
"eval_runtime": 2.7357,
|
|
"eval_samples_per_second": 39.112,
|
|
"eval_steps_per_second": 2.559,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 36.44,
|
|
"learning_rate": 3.5885167464114835e-06,
|
|
"loss": 0.0003,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 36.89,
|
|
"learning_rate": 2.9904306220095695e-06,
|
|
"loss": 0.0004,
|
|
"step": 830
|
|
},
|
|
{
|
|
"epoch": 36.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.659736156463623,
|
|
"eval_runtime": 2.6137,
|
|
"eval_samples_per_second": 40.938,
|
|
"eval_steps_per_second": 2.678,
|
|
"step": 832
|
|
},
|
|
{
|
|
"epoch": 37.33,
|
|
"learning_rate": 2.3923444976076554e-06,
|
|
"loss": 0.0002,
|
|
"step": 840
|
|
},
|
|
{
|
|
"epoch": 37.78,
|
|
"learning_rate": 1.7942583732057418e-06,
|
|
"loss": 0.0003,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 38.0,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.65986967086792,
|
|
"eval_runtime": 2.5356,
|
|
"eval_samples_per_second": 42.198,
|
|
"eval_steps_per_second": 2.761,
|
|
"step": 855
|
|
},
|
|
{
|
|
"epoch": 38.22,
|
|
"learning_rate": 1.1961722488038277e-06,
|
|
"loss": 0.0022,
|
|
"step": 860
|
|
},
|
|
{
|
|
"epoch": 38.67,
|
|
"learning_rate": 5.980861244019139e-07,
|
|
"loss": 0.0002,
|
|
"step": 870
|
|
},
|
|
{
|
|
"epoch": 38.98,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.673996925354004,
|
|
"eval_runtime": 2.6372,
|
|
"eval_samples_per_second": 40.573,
|
|
"eval_steps_per_second": 2.654,
|
|
"step": 877
|
|
},
|
|
{
|
|
"epoch": 39.11,
|
|
"learning_rate": 0.0,
|
|
"loss": 0.0002,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 39.11,
|
|
"eval_accuracy": 0.5887850467289719,
|
|
"eval_loss": 3.674065351486206,
|
|
"eval_runtime": 2.9557,
|
|
"eval_samples_per_second": 36.201,
|
|
"eval_steps_per_second": 2.368,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 39.11,
|
|
"step": 880,
|
|
"total_flos": 4.350399898167558e+18,
|
|
"train_loss": 0.09111572096166624,
|
|
"train_runtime": 1452.6155,
|
|
"train_samples_per_second": 39.542,
|
|
"train_steps_per_second": 0.606
|
|
}
|
|
],
|
|
"logging_steps": 10,
|
|
"max_steps": 880,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 40,
|
|
"save_steps": 500,
|
|
"total_flos": 4.350399898167558e+18,
|
|
"train_batch_size": 16,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|