error577's picture
Training in progress, step 400, checkpoint
bba3d24 verified
{
"best_metric": 0.5701812505722046,
"best_model_checkpoint": "miner_id_24/checkpoint-250",
"epoch": 2.9304029304029307,
"eval_steps": 50,
"global_step": 400,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.007326007326007326,
"grad_norm": 1.0746264457702637,
"learning_rate": 2.9999999999999997e-05,
"loss": 0.7495,
"step": 1
},
{
"epoch": 0.007326007326007326,
"eval_loss": 1.1676084995269775,
"eval_runtime": 73.5757,
"eval_samples_per_second": 7.435,
"eval_steps_per_second": 7.435,
"step": 1
},
{
"epoch": 0.014652014652014652,
"grad_norm": 0.2758323550224304,
"learning_rate": 5.9999999999999995e-05,
"loss": 0.9081,
"step": 2
},
{
"epoch": 0.02197802197802198,
"grad_norm": 0.29802656173706055,
"learning_rate": 8.999999999999999e-05,
"loss": 0.8211,
"step": 3
},
{
"epoch": 0.029304029304029304,
"grad_norm": 0.3587598502635956,
"learning_rate": 0.00011999999999999999,
"loss": 0.8768,
"step": 4
},
{
"epoch": 0.03663003663003663,
"grad_norm": 0.5663146376609802,
"learning_rate": 0.00015,
"loss": 0.7956,
"step": 5
},
{
"epoch": 0.04395604395604396,
"grad_norm": 0.49756062030792236,
"learning_rate": 0.00017999999999999998,
"loss": 0.9633,
"step": 6
},
{
"epoch": 0.05128205128205128,
"grad_norm": 0.5426256656646729,
"learning_rate": 0.00020999999999999998,
"loss": 0.9303,
"step": 7
},
{
"epoch": 0.05860805860805861,
"grad_norm": 0.44440993666648865,
"learning_rate": 0.00023999999999999998,
"loss": 0.8358,
"step": 8
},
{
"epoch": 0.06593406593406594,
"grad_norm": 0.3511325418949127,
"learning_rate": 0.00027,
"loss": 0.8668,
"step": 9
},
{
"epoch": 0.07326007326007326,
"grad_norm": 0.31668755412101746,
"learning_rate": 0.0003,
"loss": 0.7562,
"step": 10
},
{
"epoch": 0.08058608058608059,
"grad_norm": 0.35746482014656067,
"learning_rate": 0.0002999953270341234,
"loss": 0.779,
"step": 11
},
{
"epoch": 0.08791208791208792,
"grad_norm": 0.3406325876712799,
"learning_rate": 0.00029998130842764855,
"loss": 0.7256,
"step": 12
},
{
"epoch": 0.09523809523809523,
"grad_norm": 0.37537914514541626,
"learning_rate": 0.00029995794505402164,
"loss": 0.7405,
"step": 13
},
{
"epoch": 0.10256410256410256,
"grad_norm": 0.4480861723423004,
"learning_rate": 0.00029992523836892604,
"loss": 0.7201,
"step": 14
},
{
"epoch": 0.10989010989010989,
"grad_norm": 0.5216184854507446,
"learning_rate": 0.00029988319041019133,
"loss": 0.7125,
"step": 15
},
{
"epoch": 0.11721611721611722,
"grad_norm": 0.43949222564697266,
"learning_rate": 0.00029983180379766647,
"loss": 0.7151,
"step": 16
},
{
"epoch": 0.12454212454212454,
"grad_norm": 0.3697758615016937,
"learning_rate": 0.0002997710817330568,
"loss": 0.6527,
"step": 17
},
{
"epoch": 0.13186813186813187,
"grad_norm": 0.29461607336997986,
"learning_rate": 0.000299701027999724,
"loss": 0.5593,
"step": 18
},
{
"epoch": 0.1391941391941392,
"grad_norm": 0.3185078203678131,
"learning_rate": 0.00029962164696245076,
"loss": 0.6053,
"step": 19
},
{
"epoch": 0.14652014652014653,
"grad_norm": 0.4523352086544037,
"learning_rate": 0.000299532943567169,
"loss": 0.8082,
"step": 20
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.3355499505996704,
"learning_rate": 0.000299434923340651,
"loss": 0.4784,
"step": 21
},
{
"epoch": 0.16117216117216118,
"grad_norm": 0.38593626022338867,
"learning_rate": 0.0002993275923901659,
"loss": 0.7062,
"step": 22
},
{
"epoch": 0.1684981684981685,
"grad_norm": 0.3636136054992676,
"learning_rate": 0.00029921095740309854,
"loss": 0.689,
"step": 23
},
{
"epoch": 0.17582417582417584,
"grad_norm": 0.37379103899002075,
"learning_rate": 0.0002990850256465331,
"loss": 0.6864,
"step": 24
},
{
"epoch": 0.18315018315018314,
"grad_norm": 0.44014841318130493,
"learning_rate": 0.0002989498049668004,
"loss": 0.5692,
"step": 25
},
{
"epoch": 0.19047619047619047,
"grad_norm": 0.36174437403678894,
"learning_rate": 0.0002988053037889886,
"loss": 0.6588,
"step": 26
},
{
"epoch": 0.1978021978021978,
"grad_norm": 0.3935069143772125,
"learning_rate": 0.0002986515311164188,
"loss": 0.6105,
"step": 27
},
{
"epoch": 0.20512820512820512,
"grad_norm": 0.4123885929584503,
"learning_rate": 0.00029848849653008356,
"loss": 0.5963,
"step": 28
},
{
"epoch": 0.21245421245421245,
"grad_norm": 0.4219466745853424,
"learning_rate": 0.00029831621018805036,
"loss": 0.598,
"step": 29
},
{
"epoch": 0.21978021978021978,
"grad_norm": 0.40744897723197937,
"learning_rate": 0.00029813468282482856,
"loss": 0.6318,
"step": 30
},
{
"epoch": 0.2271062271062271,
"grad_norm": 0.41496506333351135,
"learning_rate": 0.0002979439257507004,
"loss": 0.5648,
"step": 31
},
{
"epoch": 0.23443223443223443,
"grad_norm": 0.4135119318962097,
"learning_rate": 0.0002977439508510166,
"loss": 0.5706,
"step": 32
},
{
"epoch": 0.24175824175824176,
"grad_norm": 0.5413480997085571,
"learning_rate": 0.00029753477058545543,
"loss": 0.6367,
"step": 33
},
{
"epoch": 0.2490842490842491,
"grad_norm": 0.7710244059562683,
"learning_rate": 0.00029731639798724705,
"loss": 0.5927,
"step": 34
},
{
"epoch": 0.2564102564102564,
"grad_norm": 0.29940611124038696,
"learning_rate": 0.0002970888466623606,
"loss": 0.6232,
"step": 35
},
{
"epoch": 0.26373626373626374,
"grad_norm": 0.3256644606590271,
"learning_rate": 0.00029685213078865715,
"loss": 0.6949,
"step": 36
},
{
"epoch": 0.27106227106227104,
"grad_norm": 0.2492157518863678,
"learning_rate": 0.00029660626511500605,
"loss": 0.6002,
"step": 37
},
{
"epoch": 0.2783882783882784,
"grad_norm": 0.25613823533058167,
"learning_rate": 0.000296351264960366,
"loss": 0.6237,
"step": 38
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.24584399163722992,
"learning_rate": 0.00029608714621283063,
"loss": 0.688,
"step": 39
},
{
"epoch": 0.29304029304029305,
"grad_norm": 0.23079101741313934,
"learning_rate": 0.0002958139253286385,
"loss": 0.7131,
"step": 40
},
{
"epoch": 0.30036630036630035,
"grad_norm": 0.28000855445861816,
"learning_rate": 0.0002955316193311479,
"loss": 0.6817,
"step": 41
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.2555966377258301,
"learning_rate": 0.000295240245809776,
"loss": 0.6851,
"step": 42
},
{
"epoch": 0.315018315018315,
"grad_norm": 0.24511605501174927,
"learning_rate": 0.00029493982291890324,
"loss": 0.6946,
"step": 43
},
{
"epoch": 0.32234432234432236,
"grad_norm": 0.24352319538593292,
"learning_rate": 0.00029463036937674175,
"loss": 0.6592,
"step": 44
},
{
"epoch": 0.32967032967032966,
"grad_norm": 0.27086544036865234,
"learning_rate": 0.0002943119044641694,
"loss": 0.6949,
"step": 45
},
{
"epoch": 0.336996336996337,
"grad_norm": 0.25139376521110535,
"learning_rate": 0.00029398444802352855,
"loss": 0.6856,
"step": 46
},
{
"epoch": 0.3443223443223443,
"grad_norm": 0.23585085570812225,
"learning_rate": 0.0002936480204573894,
"loss": 0.6426,
"step": 47
},
{
"epoch": 0.3516483516483517,
"grad_norm": 0.24448701739311218,
"learning_rate": 0.00029330264272727917,
"loss": 0.5887,
"step": 48
},
{
"epoch": 0.358974358974359,
"grad_norm": 0.2956358790397644,
"learning_rate": 0.00029294833635237587,
"loss": 0.6925,
"step": 49
},
{
"epoch": 0.3663003663003663,
"grad_norm": 0.242750346660614,
"learning_rate": 0.0002925851234081674,
"loss": 0.5302,
"step": 50
},
{
"epoch": 0.3663003663003663,
"eval_loss": 0.6130701303482056,
"eval_runtime": 73.3992,
"eval_samples_per_second": 7.452,
"eval_steps_per_second": 7.452,
"step": 50
},
{
"epoch": 0.37362637362637363,
"grad_norm": 0.2964664399623871,
"learning_rate": 0.0002922130265250764,
"loss": 0.6552,
"step": 51
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.3150418996810913,
"learning_rate": 0.00029183206888705004,
"loss": 0.6041,
"step": 52
},
{
"epoch": 0.3882783882783883,
"grad_norm": 0.2626626491546631,
"learning_rate": 0.00029144227423011555,
"loss": 0.6236,
"step": 53
},
{
"epoch": 0.3956043956043956,
"grad_norm": 0.31966981291770935,
"learning_rate": 0.0002910436668409013,
"loss": 0.6936,
"step": 54
},
{
"epoch": 0.40293040293040294,
"grad_norm": 0.3115909695625305,
"learning_rate": 0.00029063627155512375,
"loss": 0.5992,
"step": 55
},
{
"epoch": 0.41025641025641024,
"grad_norm": 0.29618480801582336,
"learning_rate": 0.0002902201137560397,
"loss": 0.5781,
"step": 56
},
{
"epoch": 0.4175824175824176,
"grad_norm": 0.2827432453632355,
"learning_rate": 0.0002897952193728652,
"loss": 0.5961,
"step": 57
},
{
"epoch": 0.4249084249084249,
"grad_norm": 0.30941715836524963,
"learning_rate": 0.00028936161487915955,
"loss": 0.6069,
"step": 58
},
{
"epoch": 0.43223443223443225,
"grad_norm": 0.34647953510284424,
"learning_rate": 0.00028891932729117615,
"loss": 0.6544,
"step": 59
},
{
"epoch": 0.43956043956043955,
"grad_norm": 0.34131157398223877,
"learning_rate": 0.0002884683841661791,
"loss": 0.5973,
"step": 60
},
{
"epoch": 0.4468864468864469,
"grad_norm": 0.3378750681877136,
"learning_rate": 0.00028800881360072615,
"loss": 0.5805,
"step": 61
},
{
"epoch": 0.4542124542124542,
"grad_norm": 0.3864741027355194,
"learning_rate": 0.0002875406442289183,
"loss": 0.5829,
"step": 62
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.31040650606155396,
"learning_rate": 0.0002870639052206155,
"loss": 0.5408,
"step": 63
},
{
"epoch": 0.46886446886446886,
"grad_norm": 0.3510590195655823,
"learning_rate": 0.0002865786262796193,
"loss": 0.5623,
"step": 64
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.34330350160598755,
"learning_rate": 0.00028608483764182215,
"loss": 0.5304,
"step": 65
},
{
"epoch": 0.4835164835164835,
"grad_norm": 0.3788071274757385,
"learning_rate": 0.0002855825700733235,
"loss": 0.4966,
"step": 66
},
{
"epoch": 0.4908424908424908,
"grad_norm": 0.38324546813964844,
"learning_rate": 0.00028507185486851275,
"loss": 0.44,
"step": 67
},
{
"epoch": 0.4981684981684982,
"grad_norm": 0.4564480185508728,
"learning_rate": 0.0002845527238481195,
"loss": 0.4097,
"step": 68
},
{
"epoch": 0.5054945054945055,
"grad_norm": 0.238909512758255,
"learning_rate": 0.0002840252093572311,
"loss": 0.6646,
"step": 69
},
{
"epoch": 0.5128205128205128,
"grad_norm": 0.2642434239387512,
"learning_rate": 0.000283489344263277,
"loss": 0.6419,
"step": 70
},
{
"epoch": 0.5201465201465202,
"grad_norm": 0.2626118063926697,
"learning_rate": 0.00028294516195398125,
"loss": 0.6318,
"step": 71
},
{
"epoch": 0.5274725274725275,
"grad_norm": 0.23919053375720978,
"learning_rate": 0.00028239269633528204,
"loss": 0.6266,
"step": 72
},
{
"epoch": 0.5347985347985348,
"grad_norm": 0.25669869780540466,
"learning_rate": 0.000281831981829219,
"loss": 0.6594,
"step": 73
},
{
"epoch": 0.5421245421245421,
"grad_norm": 0.21138758957386017,
"learning_rate": 0.00028126305337178905,
"loss": 0.5986,
"step": 74
},
{
"epoch": 0.5494505494505495,
"grad_norm": 0.24748112261295319,
"learning_rate": 0.0002806859464107689,
"loss": 0.6108,
"step": 75
},
{
"epoch": 0.5567765567765568,
"grad_norm": 0.23385199904441833,
"learning_rate": 0.00028010069690350716,
"loss": 0.6049,
"step": 76
},
{
"epoch": 0.5641025641025641,
"grad_norm": 0.2671392560005188,
"learning_rate": 0.00027950734131468346,
"loss": 0.6324,
"step": 77
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.20720772445201874,
"learning_rate": 0.00027890591661403676,
"loss": 0.538,
"step": 78
},
{
"epoch": 0.5787545787545788,
"grad_norm": 0.26165518164634705,
"learning_rate": 0.00027829646027406174,
"loss": 0.6713,
"step": 79
},
{
"epoch": 0.5860805860805861,
"grad_norm": 0.285157173871994,
"learning_rate": 0.00027767901026767416,
"loss": 0.7407,
"step": 80
},
{
"epoch": 0.5934065934065934,
"grad_norm": 0.24118804931640625,
"learning_rate": 0.00027705360506584484,
"loss": 0.6414,
"step": 81
},
{
"epoch": 0.6007326007326007,
"grad_norm": 0.2540411651134491,
"learning_rate": 0.00027642028363520255,
"loss": 0.6584,
"step": 82
},
{
"epoch": 0.608058608058608,
"grad_norm": 0.2713707685470581,
"learning_rate": 0.0002757790854356066,
"loss": 0.6356,
"step": 83
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.25061967968940735,
"learning_rate": 0.0002751300504176876,
"loss": 0.5919,
"step": 84
},
{
"epoch": 0.6227106227106227,
"grad_norm": 0.2813633382320404,
"learning_rate": 0.0002744732190203589,
"loss": 0.6507,
"step": 85
},
{
"epoch": 0.63003663003663,
"grad_norm": 0.36446070671081543,
"learning_rate": 0.00027380863216829645,
"loss": 0.6146,
"step": 86
},
{
"epoch": 0.6373626373626373,
"grad_norm": 0.2936757802963257,
"learning_rate": 0.00027313633126938936,
"loss": 0.5807,
"step": 87
},
{
"epoch": 0.6446886446886447,
"grad_norm": 0.32543811202049255,
"learning_rate": 0.0002724563582121598,
"loss": 0.633,
"step": 88
},
{
"epoch": 0.652014652014652,
"grad_norm": 0.26980024576187134,
"learning_rate": 0.0002717687553631531,
"loss": 0.5771,
"step": 89
},
{
"epoch": 0.6593406593406593,
"grad_norm": 0.3081660568714142,
"learning_rate": 0.0002710735655642978,
"loss": 0.5795,
"step": 90
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.30959567427635193,
"learning_rate": 0.0002703708321302367,
"loss": 0.5278,
"step": 91
},
{
"epoch": 0.673992673992674,
"grad_norm": 0.30625343322753906,
"learning_rate": 0.0002696605988456279,
"loss": 0.5816,
"step": 92
},
{
"epoch": 0.6813186813186813,
"grad_norm": 0.31666725873947144,
"learning_rate": 0.00026894290996241677,
"loss": 0.5393,
"step": 93
},
{
"epoch": 0.6886446886446886,
"grad_norm": 0.31827884912490845,
"learning_rate": 0.0002682178101970788,
"loss": 0.5869,
"step": 94
},
{
"epoch": 0.6959706959706959,
"grad_norm": 0.3418588936328888,
"learning_rate": 0.00026748534472783355,
"loss": 0.5939,
"step": 95
},
{
"epoch": 0.7032967032967034,
"grad_norm": 0.35669511556625366,
"learning_rate": 0.00026674555919182943,
"loss": 0.6145,
"step": 96
},
{
"epoch": 0.7106227106227107,
"grad_norm": 0.3676556348800659,
"learning_rate": 0.00026599849968230084,
"loss": 0.5306,
"step": 97
},
{
"epoch": 0.717948717948718,
"grad_norm": 0.3910490870475769,
"learning_rate": 0.00026524421274569556,
"loss": 0.6035,
"step": 98
},
{
"epoch": 0.7252747252747253,
"grad_norm": 0.3907155394554138,
"learning_rate": 0.00026448274537877527,
"loss": 0.5149,
"step": 99
},
{
"epoch": 0.7326007326007326,
"grad_norm": 0.3669157326221466,
"learning_rate": 0.0002637141450256868,
"loss": 0.4153,
"step": 100
},
{
"epoch": 0.7326007326007326,
"eval_loss": 0.5930299162864685,
"eval_runtime": 73.3788,
"eval_samples_per_second": 7.454,
"eval_steps_per_second": 7.454,
"step": 100
},
{
"epoch": 0.73992673992674,
"grad_norm": 0.5483953952789307,
"learning_rate": 0.0002629384595750065,
"loss": 0.4924,
"step": 101
},
{
"epoch": 0.7472527472527473,
"grad_norm": 0.44320136308670044,
"learning_rate": 0.00026215573735675635,
"loss": 0.4333,
"step": 102
},
{
"epoch": 0.7545787545787546,
"grad_norm": 0.2921634316444397,
"learning_rate": 0.0002613660271393924,
"loss": 0.5333,
"step": 103
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.3033207654953003,
"learning_rate": 0.0002605693781267668,
"loss": 0.6723,
"step": 104
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.21631714701652527,
"learning_rate": 0.00025976583995506147,
"loss": 0.7413,
"step": 105
},
{
"epoch": 0.7765567765567766,
"grad_norm": 0.2756958305835724,
"learning_rate": 0.0002589554626896959,
"loss": 0.6334,
"step": 106
},
{
"epoch": 0.7838827838827839,
"grad_norm": 0.23255576193332672,
"learning_rate": 0.0002581382968222075,
"loss": 0.6241,
"step": 107
},
{
"epoch": 0.7912087912087912,
"grad_norm": 0.2933904230594635,
"learning_rate": 0.00025731439326710564,
"loss": 0.6238,
"step": 108
},
{
"epoch": 0.7985347985347986,
"grad_norm": 0.24506813287734985,
"learning_rate": 0.0002564838033586999,
"loss": 0.7049,
"step": 109
},
{
"epoch": 0.8058608058608059,
"grad_norm": 0.21781164407730103,
"learning_rate": 0.0002556465788479008,
"loss": 0.5831,
"step": 110
},
{
"epoch": 0.8131868131868132,
"grad_norm": 0.23479105532169342,
"learning_rate": 0.00025480277189899594,
"loss": 0.6347,
"step": 111
},
{
"epoch": 0.8205128205128205,
"grad_norm": 0.22827807068824768,
"learning_rate": 0.00025395243508639974,
"loss": 0.5881,
"step": 112
},
{
"epoch": 0.8278388278388278,
"grad_norm": 0.24758267402648926,
"learning_rate": 0.00025309562139137765,
"loss": 0.6288,
"step": 113
},
{
"epoch": 0.8351648351648352,
"grad_norm": 0.23672336339950562,
"learning_rate": 0.00025223238419874505,
"loss": 0.623,
"step": 114
},
{
"epoch": 0.8424908424908425,
"grad_norm": 0.25866055488586426,
"learning_rate": 0.0002513627772935413,
"loss": 0.6076,
"step": 115
},
{
"epoch": 0.8498168498168498,
"grad_norm": 0.23350456357002258,
"learning_rate": 0.00025048685485767826,
"loss": 0.6124,
"step": 116
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.26988059282302856,
"learning_rate": 0.00024960467146656465,
"loss": 0.667,
"step": 117
},
{
"epoch": 0.8644688644688645,
"grad_norm": 0.2606026530265808,
"learning_rate": 0.0002487162820857056,
"loss": 0.514,
"step": 118
},
{
"epoch": 0.8717948717948718,
"grad_norm": 0.26499950885772705,
"learning_rate": 0.00024782174206727797,
"loss": 0.5818,
"step": 119
},
{
"epoch": 0.8791208791208791,
"grad_norm": 0.2778685390949249,
"learning_rate": 0.00024692110714668144,
"loss": 0.6545,
"step": 120
},
{
"epoch": 0.8864468864468864,
"grad_norm": 0.2645077705383301,
"learning_rate": 0.00024601443343906607,
"loss": 0.5948,
"step": 121
},
{
"epoch": 0.8937728937728938,
"grad_norm": 0.2799476385116577,
"learning_rate": 0.00024510177743583577,
"loss": 0.6282,
"step": 122
},
{
"epoch": 0.9010989010989011,
"grad_norm": 0.3192484676837921,
"learning_rate": 0.00024418319600112877,
"loss": 0.6482,
"step": 123
},
{
"epoch": 0.9084249084249084,
"grad_norm": 0.28705793619155884,
"learning_rate": 0.00024325874636827428,
"loss": 0.6307,
"step": 124
},
{
"epoch": 0.9157509157509157,
"grad_norm": 0.30531764030456543,
"learning_rate": 0.00024232848613622686,
"loss": 0.5717,
"step": 125
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.31991544365882874,
"learning_rate": 0.00024139247326597748,
"loss": 0.5068,
"step": 126
},
{
"epoch": 0.9304029304029304,
"grad_norm": 0.40226104855537415,
"learning_rate": 0.00024045076607694216,
"loss": 0.7029,
"step": 127
},
{
"epoch": 0.9377289377289377,
"grad_norm": 0.38639435172080994,
"learning_rate": 0.0002395034232433284,
"loss": 0.5574,
"step": 128
},
{
"epoch": 0.945054945054945,
"grad_norm": 0.35603398084640503,
"learning_rate": 0.00023855050379047922,
"loss": 0.568,
"step": 129
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.3355114459991455,
"learning_rate": 0.00023759206709119595,
"loss": 0.4997,
"step": 130
},
{
"epoch": 0.9597069597069597,
"grad_norm": 0.39614033699035645,
"learning_rate": 0.0002366281728620383,
"loss": 0.5225,
"step": 131
},
{
"epoch": 0.967032967032967,
"grad_norm": 0.33072763681411743,
"learning_rate": 0.0002356588811596042,
"loss": 0.427,
"step": 132
},
{
"epoch": 0.9743589743589743,
"grad_norm": 0.5783118605613708,
"learning_rate": 0.00023468425237678769,
"loss": 0.5669,
"step": 133
},
{
"epoch": 0.9816849816849816,
"grad_norm": 0.44978731870651245,
"learning_rate": 0.000233704347239016,
"loss": 0.5213,
"step": 134
},
{
"epoch": 0.989010989010989,
"grad_norm": 0.4447237551212311,
"learning_rate": 0.0002327192268004661,
"loss": 0.4526,
"step": 135
},
{
"epoch": 0.9963369963369964,
"grad_norm": 0.5178239941596985,
"learning_rate": 0.00023172895244026056,
"loss": 0.419,
"step": 136
},
{
"epoch": 1.0036630036630036,
"grad_norm": 0.5518823266029358,
"learning_rate": 0.0002307335858586433,
"loss": 0.9609,
"step": 137
},
{
"epoch": 1.010989010989011,
"grad_norm": 0.19367220997810364,
"learning_rate": 0.0002297331890731352,
"loss": 0.6718,
"step": 138
},
{
"epoch": 1.0183150183150182,
"grad_norm": 0.2209685742855072,
"learning_rate": 0.0002287278244146702,
"loss": 0.6591,
"step": 139
},
{
"epoch": 1.0256410256410255,
"grad_norm": 0.2094419300556183,
"learning_rate": 0.00022771755452371162,
"loss": 0.5926,
"step": 140
},
{
"epoch": 1.032967032967033,
"grad_norm": 0.22118069231510162,
"learning_rate": 0.00022670244234634902,
"loss": 0.6281,
"step": 141
},
{
"epoch": 1.0402930402930404,
"grad_norm": 0.23426313698291779,
"learning_rate": 0.00022568255113037658,
"loss": 0.5911,
"step": 142
},
{
"epoch": 1.0476190476190477,
"grad_norm": 0.22640864551067352,
"learning_rate": 0.00022465794442135244,
"loss": 0.6435,
"step": 143
},
{
"epoch": 1.054945054945055,
"grad_norm": 0.2409924864768982,
"learning_rate": 0.000223628686058639,
"loss": 0.594,
"step": 144
},
{
"epoch": 1.0622710622710623,
"grad_norm": 0.2898027002811432,
"learning_rate": 0.00022259484017142582,
"loss": 0.578,
"step": 145
},
{
"epoch": 1.0695970695970696,
"grad_norm": 0.2668663561344147,
"learning_rate": 0.00022155647117473343,
"loss": 0.6175,
"step": 146
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.25386887788772583,
"learning_rate": 0.00022051364376540045,
"loss": 0.5452,
"step": 147
},
{
"epoch": 1.0842490842490842,
"grad_norm": 0.2420186847448349,
"learning_rate": 0.00021946642291805214,
"loss": 0.5267,
"step": 148
},
{
"epoch": 1.0915750915750915,
"grad_norm": 0.26154059171676636,
"learning_rate": 0.00021841487388105235,
"loss": 0.5904,
"step": 149
},
{
"epoch": 1.098901098901099,
"grad_norm": 0.25012269616127014,
"learning_rate": 0.00021735906217243802,
"loss": 0.6004,
"step": 150
},
{
"epoch": 1.098901098901099,
"eval_loss": 0.5869444012641907,
"eval_runtime": 73.5969,
"eval_samples_per_second": 7.432,
"eval_steps_per_second": 7.432,
"step": 150
},
{
"epoch": 1.1062271062271063,
"grad_norm": 0.2936391234397888,
"learning_rate": 0.00021629905357583687,
"loss": 0.6069,
"step": 151
},
{
"epoch": 1.1135531135531136,
"grad_norm": 0.3034394383430481,
"learning_rate": 0.00021523491413636894,
"loss": 0.4688,
"step": 152
},
{
"epoch": 1.120879120879121,
"grad_norm": 0.2803559899330139,
"learning_rate": 0.00021416671015653144,
"loss": 0.5647,
"step": 153
},
{
"epoch": 1.1282051282051282,
"grad_norm": 0.3114618957042694,
"learning_rate": 0.00021309450819206763,
"loss": 0.4767,
"step": 154
},
{
"epoch": 1.1355311355311355,
"grad_norm": 0.3169534206390381,
"learning_rate": 0.0002120183750478201,
"loss": 0.5048,
"step": 155
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.3803180754184723,
"learning_rate": 0.00021093837777356835,
"loss": 0.5439,
"step": 156
},
{
"epoch": 1.15018315018315,
"grad_norm": 0.3590928614139557,
"learning_rate": 0.00020985458365985112,
"loss": 0.6012,
"step": 157
},
{
"epoch": 1.1575091575091574,
"grad_norm": 0.3438403606414795,
"learning_rate": 0.00020876706023377394,
"loss": 0.4902,
"step": 158
},
{
"epoch": 1.164835164835165,
"grad_norm": 0.35391849279403687,
"learning_rate": 0.0002076758752548016,
"loss": 0.4655,
"step": 159
},
{
"epoch": 1.1721611721611722,
"grad_norm": 0.3842090666294098,
"learning_rate": 0.0002065810967105364,
"loss": 0.4753,
"step": 160
},
{
"epoch": 1.1794871794871795,
"grad_norm": 0.35003578662872314,
"learning_rate": 0.000205482792812482,
"loss": 0.4344,
"step": 161
},
{
"epoch": 1.1868131868131868,
"grad_norm": 0.3732411563396454,
"learning_rate": 0.0002043810319917937,
"loss": 0.4107,
"step": 162
},
{
"epoch": 1.1941391941391941,
"grad_norm": 0.4202706515789032,
"learning_rate": 0.00020327588289501425,
"loss": 0.5001,
"step": 163
},
{
"epoch": 1.2014652014652014,
"grad_norm": 0.41824397444725037,
"learning_rate": 0.00020216741437979735,
"loss": 0.4344,
"step": 164
},
{
"epoch": 1.2087912087912087,
"grad_norm": 0.5380794405937195,
"learning_rate": 0.00020105569551061693,
"loss": 0.4992,
"step": 165
},
{
"epoch": 1.2161172161172162,
"grad_norm": 0.557085394859314,
"learning_rate": 0.00019994079555446417,
"loss": 0.4411,
"step": 166
},
{
"epoch": 1.2234432234432235,
"grad_norm": 0.485733300447464,
"learning_rate": 0.00019882278397653175,
"loss": 0.4401,
"step": 167
},
{
"epoch": 1.2307692307692308,
"grad_norm": 1.3742977380752563,
"learning_rate": 0.0001977017304358857,
"loss": 0.3275,
"step": 168
},
{
"epoch": 1.2380952380952381,
"grad_norm": 0.5038601756095886,
"learning_rate": 0.00019657770478112533,
"loss": 0.3036,
"step": 169
},
{
"epoch": 1.2454212454212454,
"grad_norm": 0.615625262260437,
"learning_rate": 0.00019545077704603088,
"loss": 0.3079,
"step": 170
},
{
"epoch": 1.2527472527472527,
"grad_norm": 0.4520106315612793,
"learning_rate": 0.00019432101744520052,
"loss": 0.4792,
"step": 171
},
{
"epoch": 1.26007326007326,
"grad_norm": 0.4422815442085266,
"learning_rate": 0.00019318849636967497,
"loss": 0.5819,
"step": 172
},
{
"epoch": 1.2673992673992673,
"grad_norm": 0.39762523770332336,
"learning_rate": 0.00019205328438255215,
"loss": 0.6133,
"step": 173
},
{
"epoch": 1.2747252747252746,
"grad_norm": 0.308088481426239,
"learning_rate": 0.00019091545221459048,
"loss": 0.5021,
"step": 174
},
{
"epoch": 1.282051282051282,
"grad_norm": 0.3217431902885437,
"learning_rate": 0.0001897750707598018,
"loss": 0.5656,
"step": 175
},
{
"epoch": 1.2893772893772895,
"grad_norm": 0.34072449803352356,
"learning_rate": 0.00018863221107103453,
"loss": 0.6147,
"step": 176
},
{
"epoch": 1.2967032967032968,
"grad_norm": 0.3127173185348511,
"learning_rate": 0.00018748694435554626,
"loss": 0.5093,
"step": 177
},
{
"epoch": 1.304029304029304,
"grad_norm": 0.28147441148757935,
"learning_rate": 0.00018633934197056747,
"loss": 0.5148,
"step": 178
},
{
"epoch": 1.3113553113553114,
"grad_norm": 0.32677727937698364,
"learning_rate": 0.00018518947541885532,
"loss": 0.5213,
"step": 179
},
{
"epoch": 1.3186813186813187,
"grad_norm": 0.3397771120071411,
"learning_rate": 0.0001840374163442385,
"loss": 0.5752,
"step": 180
},
{
"epoch": 1.326007326007326,
"grad_norm": 0.32977497577667236,
"learning_rate": 0.00018288323652715372,
"loss": 0.6049,
"step": 181
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.3196227550506592,
"learning_rate": 0.00018172700788017285,
"loss": 0.6103,
"step": 182
},
{
"epoch": 1.3406593406593408,
"grad_norm": 0.3648357689380646,
"learning_rate": 0.0001805688024435228,
"loss": 0.6024,
"step": 183
},
{
"epoch": 1.347985347985348,
"grad_norm": 0.3467482924461365,
"learning_rate": 0.00017940869238059672,
"loss": 0.6184,
"step": 184
},
{
"epoch": 1.3553113553113554,
"grad_norm": 0.34417080879211426,
"learning_rate": 0.00017824674997345765,
"loss": 0.5201,
"step": 185
},
{
"epoch": 1.3626373626373627,
"grad_norm": 0.32738953828811646,
"learning_rate": 0.0001770830476183353,
"loss": 0.4373,
"step": 186
},
{
"epoch": 1.36996336996337,
"grad_norm": 0.3495752513408661,
"learning_rate": 0.00017591765782111493,
"loss": 0.5156,
"step": 187
},
{
"epoch": 1.3772893772893773,
"grad_norm": 0.38357013463974,
"learning_rate": 0.0001747506531928199,
"loss": 0.5689,
"step": 188
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.3947698473930359,
"learning_rate": 0.00017358210644508761,
"loss": 0.5288,
"step": 189
},
{
"epoch": 1.3919413919413919,
"grad_norm": 0.39951273798942566,
"learning_rate": 0.00017241209038563915,
"loss": 0.4936,
"step": 190
},
{
"epoch": 1.3992673992673992,
"grad_norm": 0.46843069791793823,
"learning_rate": 0.0001712406779137427,
"loss": 0.5193,
"step": 191
},
{
"epoch": 1.4065934065934065,
"grad_norm": 0.4092109799385071,
"learning_rate": 0.00017006794201567186,
"loss": 0.4587,
"step": 192
},
{
"epoch": 1.4139194139194138,
"grad_norm": 0.5061143636703491,
"learning_rate": 0.00016889395576015763,
"loss": 0.4816,
"step": 193
},
{
"epoch": 1.4212454212454213,
"grad_norm": 0.4533427655696869,
"learning_rate": 0.0001677187922938362,
"loss": 0.5173,
"step": 194
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.4554060399532318,
"learning_rate": 0.00016654252483669124,
"loss": 0.4767,
"step": 195
},
{
"epoch": 1.435897435897436,
"grad_norm": 0.4680671691894531,
"learning_rate": 0.00016536522667749196,
"loss": 0.4486,
"step": 196
},
{
"epoch": 1.4432234432234432,
"grad_norm": 0.5479559898376465,
"learning_rate": 0.00016418697116922672,
"loss": 0.4627,
"step": 197
},
{
"epoch": 1.4505494505494505,
"grad_norm": 0.5161874890327454,
"learning_rate": 0.00016300783172453265,
"loss": 0.4527,
"step": 198
},
{
"epoch": 1.4578754578754578,
"grad_norm": 0.4987958073616028,
"learning_rate": 0.00016182788181112146,
"loss": 0.43,
"step": 199
},
{
"epoch": 1.4652014652014653,
"grad_norm": 0.5920132994651794,
"learning_rate": 0.00016064719494720238,
"loss": 0.4271,
"step": 200
},
{
"epoch": 1.4652014652014653,
"eval_loss": 0.5827460885047913,
"eval_runtime": 73.578,
"eval_samples_per_second": 7.434,
"eval_steps_per_second": 7.434,
"step": 200
},
{
"epoch": 1.4725274725274726,
"grad_norm": 0.5107991695404053,
"learning_rate": 0.000159465844696901,
"loss": 0.3252,
"step": 201
},
{
"epoch": 1.47985347985348,
"grad_norm": 0.5588508248329163,
"learning_rate": 0.0001582839046656762,
"loss": 0.3086,
"step": 202
},
{
"epoch": 1.4871794871794872,
"grad_norm": 0.5706549882888794,
"learning_rate": 0.0001571014484957337,
"loss": 0.2613,
"step": 203
},
{
"epoch": 1.4945054945054945,
"grad_norm": 0.8393925428390503,
"learning_rate": 0.0001559185498614379,
"loss": 0.3406,
"step": 204
},
{
"epoch": 1.5018315018315018,
"grad_norm": 0.6459693312644958,
"learning_rate": 0.0001547352824647216,
"loss": 0.3553,
"step": 205
},
{
"epoch": 1.5091575091575091,
"grad_norm": 0.313265323638916,
"learning_rate": 0.00015355172003049357,
"loss": 0.4846,
"step": 206
},
{
"epoch": 1.5164835164835164,
"grad_norm": 0.3253680169582367,
"learning_rate": 0.00015236793630204554,
"loss": 0.5412,
"step": 207
},
{
"epoch": 1.5238095238095237,
"grad_norm": 0.35269081592559814,
"learning_rate": 0.000151184005036457,
"loss": 0.6062,
"step": 208
},
{
"epoch": 1.531135531135531,
"grad_norm": 0.3606890141963959,
"learning_rate": 0.00015,
"loss": 0.5544,
"step": 209
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.36298248171806335,
"learning_rate": 0.000148815994963543,
"loss": 0.542,
"step": 210
},
{
"epoch": 1.5457875457875456,
"grad_norm": 0.36222681403160095,
"learning_rate": 0.00014763206369795446,
"loss": 0.6695,
"step": 211
},
{
"epoch": 1.5531135531135531,
"grad_norm": 0.4112010598182678,
"learning_rate": 0.00014644827996950643,
"loss": 0.6308,
"step": 212
},
{
"epoch": 1.5604395604395604,
"grad_norm": 0.33088386058807373,
"learning_rate": 0.00014526471753527838,
"loss": 0.5061,
"step": 213
},
{
"epoch": 1.5677655677655677,
"grad_norm": 0.3446553349494934,
"learning_rate": 0.0001440814501385621,
"loss": 0.5424,
"step": 214
},
{
"epoch": 1.575091575091575,
"grad_norm": 0.3574959933757782,
"learning_rate": 0.0001428985515042663,
"loss": 0.448,
"step": 215
},
{
"epoch": 1.5824175824175826,
"grad_norm": 0.35467544198036194,
"learning_rate": 0.00014171609533432378,
"loss": 0.5307,
"step": 216
},
{
"epoch": 1.5897435897435899,
"grad_norm": 0.34874802827835083,
"learning_rate": 0.00014053415530309896,
"loss": 0.5321,
"step": 217
},
{
"epoch": 1.5970695970695972,
"grad_norm": 0.37879034876823425,
"learning_rate": 0.0001393528050527976,
"loss": 0.4935,
"step": 218
},
{
"epoch": 1.6043956043956045,
"grad_norm": 0.3754430115222931,
"learning_rate": 0.00013817211818887852,
"loss": 0.5807,
"step": 219
},
{
"epoch": 1.6117216117216118,
"grad_norm": 0.40496763586997986,
"learning_rate": 0.0001369921682754674,
"loss": 0.6015,
"step": 220
},
{
"epoch": 1.619047619047619,
"grad_norm": 0.38105273246765137,
"learning_rate": 0.00013581302883077325,
"loss": 0.512,
"step": 221
},
{
"epoch": 1.6263736263736264,
"grad_norm": 0.45323583483695984,
"learning_rate": 0.000134634773322508,
"loss": 0.5284,
"step": 222
},
{
"epoch": 1.6336996336996337,
"grad_norm": 0.4019657075405121,
"learning_rate": 0.00013345747516330873,
"loss": 0.5855,
"step": 223
},
{
"epoch": 1.641025641025641,
"grad_norm": 0.40215152502059937,
"learning_rate": 0.0001322812077061638,
"loss": 0.472,
"step": 224
},
{
"epoch": 1.6483516483516483,
"grad_norm": 0.4318372309207916,
"learning_rate": 0.0001311060442398424,
"loss": 0.4597,
"step": 225
},
{
"epoch": 1.6556776556776556,
"grad_norm": 0.39995384216308594,
"learning_rate": 0.00012993205798432814,
"loss": 0.4269,
"step": 226
},
{
"epoch": 1.6630036630036629,
"grad_norm": 0.44160038232803345,
"learning_rate": 0.0001287593220862573,
"loss": 0.495,
"step": 227
},
{
"epoch": 1.6703296703296702,
"grad_norm": 0.4725303649902344,
"learning_rate": 0.00012758790961436083,
"loss": 0.4705,
"step": 228
},
{
"epoch": 1.6776556776556777,
"grad_norm": 0.7378038763999939,
"learning_rate": 0.0001264178935549124,
"loss": 0.4881,
"step": 229
},
{
"epoch": 1.684981684981685,
"grad_norm": 0.5113095045089722,
"learning_rate": 0.0001252493468071801,
"loss": 0.4921,
"step": 230
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.5403740406036377,
"learning_rate": 0.00012408234217888508,
"loss": 0.4288,
"step": 231
},
{
"epoch": 1.6996336996336996,
"grad_norm": 0.5715883374214172,
"learning_rate": 0.0001229169523816647,
"loss": 0.3966,
"step": 232
},
{
"epoch": 1.7069597069597071,
"grad_norm": 0.6109308004379272,
"learning_rate": 0.00012175325002654229,
"loss": 0.4353,
"step": 233
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.5063544511795044,
"learning_rate": 0.00012059130761940328,
"loss": 0.3582,
"step": 234
},
{
"epoch": 1.7216117216117217,
"grad_norm": 0.5469954609870911,
"learning_rate": 0.00011943119755647719,
"loss": 0.3257,
"step": 235
},
{
"epoch": 1.728937728937729,
"grad_norm": 0.6388989686965942,
"learning_rate": 0.00011827299211982714,
"loss": 0.3494,
"step": 236
},
{
"epoch": 1.7362637362637363,
"grad_norm": 0.6123701333999634,
"learning_rate": 0.00011711676347284631,
"loss": 0.2883,
"step": 237
},
{
"epoch": 1.7435897435897436,
"grad_norm": 0.8964057564735413,
"learning_rate": 0.00011596258365576144,
"loss": 0.3418,
"step": 238
},
{
"epoch": 1.750915750915751,
"grad_norm": 0.5625249147415161,
"learning_rate": 0.0001148105245811447,
"loss": 0.3956,
"step": 239
},
{
"epoch": 1.7582417582417582,
"grad_norm": 0.3839769661426544,
"learning_rate": 0.00011366065802943254,
"loss": 0.5822,
"step": 240
},
{
"epoch": 1.7655677655677655,
"grad_norm": 0.38849931955337524,
"learning_rate": 0.00011251305564445375,
"loss": 0.7068,
"step": 241
},
{
"epoch": 1.7728937728937728,
"grad_norm": 0.3844453990459442,
"learning_rate": 0.00011136778892896552,
"loss": 0.5259,
"step": 242
},
{
"epoch": 1.7802197802197801,
"grad_norm": 0.39755213260650635,
"learning_rate": 0.00011022492924019817,
"loss": 0.5587,
"step": 243
},
{
"epoch": 1.7875457875457874,
"grad_norm": 0.4066022038459778,
"learning_rate": 0.00010908454778540952,
"loss": 0.5985,
"step": 244
},
{
"epoch": 1.7948717948717947,
"grad_norm": 0.453279048204422,
"learning_rate": 0.00010794671561744785,
"loss": 0.6355,
"step": 245
},
{
"epoch": 1.8021978021978022,
"grad_norm": 0.44758322834968567,
"learning_rate": 0.00010681150363032503,
"loss": 0.5131,
"step": 246
},
{
"epoch": 1.8095238095238095,
"grad_norm": 0.39223358035087585,
"learning_rate": 0.00010567898255479951,
"loss": 0.5434,
"step": 247
},
{
"epoch": 1.8168498168498168,
"grad_norm": 0.38072723150253296,
"learning_rate": 0.00010454922295396907,
"loss": 0.5052,
"step": 248
},
{
"epoch": 1.8241758241758241,
"grad_norm": 0.3893994092941284,
"learning_rate": 0.00010342229521887467,
"loss": 0.5188,
"step": 249
},
{
"epoch": 1.8315018315018317,
"grad_norm": 0.3957141041755676,
"learning_rate": 0.00010229826956411426,
"loss": 0.4949,
"step": 250
},
{
"epoch": 1.8315018315018317,
"eval_loss": 0.5701812505722046,
"eval_runtime": 73.3521,
"eval_samples_per_second": 7.457,
"eval_steps_per_second": 7.457,
"step": 250
},
{
"epoch": 1.838827838827839,
"grad_norm": 0.3748125433921814,
"learning_rate": 0.00010117721602346823,
"loss": 0.482,
"step": 251
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.35637104511260986,
"learning_rate": 0.00010005920444553586,
"loss": 0.4196,
"step": 252
},
{
"epoch": 1.8534798534798536,
"grad_norm": 0.45035073161125183,
"learning_rate": 9.894430448938305e-05,
"loss": 0.5593,
"step": 253
},
{
"epoch": 1.8608058608058609,
"grad_norm": 0.42238089442253113,
"learning_rate": 9.783258562020263e-05,
"loss": 0.4647,
"step": 254
},
{
"epoch": 1.8681318681318682,
"grad_norm": 0.3975355625152588,
"learning_rate": 9.672411710498575e-05,
"loss": 0.5369,
"step": 255
},
{
"epoch": 1.8754578754578755,
"grad_norm": 0.4017747640609741,
"learning_rate": 9.561896800820633e-05,
"loss": 0.4813,
"step": 256
},
{
"epoch": 1.8827838827838828,
"grad_norm": 0.3966550827026367,
"learning_rate": 9.4517207187518e-05,
"loss": 0.425,
"step": 257
},
{
"epoch": 1.89010989010989,
"grad_norm": 0.4850409924983978,
"learning_rate": 9.341890328946358e-05,
"loss": 0.6023,
"step": 258
},
{
"epoch": 1.8974358974358974,
"grad_norm": 0.4538116753101349,
"learning_rate": 9.23241247451984e-05,
"loss": 0.4504,
"step": 259
},
{
"epoch": 1.9047619047619047,
"grad_norm": 0.42693740129470825,
"learning_rate": 9.123293976622602e-05,
"loss": 0.4383,
"step": 260
},
{
"epoch": 1.912087912087912,
"grad_norm": 0.529269814491272,
"learning_rate": 9.014541634014885e-05,
"loss": 0.5292,
"step": 261
},
{
"epoch": 1.9194139194139193,
"grad_norm": 0.520923912525177,
"learning_rate": 8.906162222643167e-05,
"loss": 0.5155,
"step": 262
},
{
"epoch": 1.9267399267399268,
"grad_norm": 0.47154876589775085,
"learning_rate": 8.798162495217989e-05,
"loss": 0.3989,
"step": 263
},
{
"epoch": 1.934065934065934,
"grad_norm": 0.5603944659233093,
"learning_rate": 8.690549180793239e-05,
"loss": 0.4806,
"step": 264
},
{
"epoch": 1.9413919413919414,
"grad_norm": 0.5125073194503784,
"learning_rate": 8.583328984346854e-05,
"loss": 0.4296,
"step": 265
},
{
"epoch": 1.9487179487179487,
"grad_norm": 0.519232988357544,
"learning_rate": 8.476508586363106e-05,
"loss": 0.4025,
"step": 266
},
{
"epoch": 1.9560439560439562,
"grad_norm": 0.5697182416915894,
"learning_rate": 8.370094642416314e-05,
"loss": 0.3478,
"step": 267
},
{
"epoch": 1.9633699633699635,
"grad_norm": 0.516194760799408,
"learning_rate": 8.264093782756195e-05,
"loss": 0.3657,
"step": 268
},
{
"epoch": 1.9706959706959708,
"grad_norm": 0.6397049427032471,
"learning_rate": 8.158512611894759e-05,
"loss": 0.4872,
"step": 269
},
{
"epoch": 1.978021978021978,
"grad_norm": 0.6545494794845581,
"learning_rate": 8.053357708194783e-05,
"loss": 0.3736,
"step": 270
},
{
"epoch": 1.9853479853479854,
"grad_norm": 0.6731343269348145,
"learning_rate": 7.948635623459958e-05,
"loss": 0.3911,
"step": 271
},
{
"epoch": 1.9926739926739927,
"grad_norm": 0.8151178359985352,
"learning_rate": 7.844352882526661e-05,
"loss": 0.2981,
"step": 272
},
{
"epoch": 2.0,
"grad_norm": 0.9851661920547485,
"learning_rate": 7.740515982857419e-05,
"loss": 0.5828,
"step": 273
},
{
"epoch": 2.0073260073260073,
"grad_norm": 0.26004815101623535,
"learning_rate": 7.637131394136096e-05,
"loss": 0.45,
"step": 274
},
{
"epoch": 2.0146520146520146,
"grad_norm": 0.27779507637023926,
"learning_rate": 7.534205557864752e-05,
"loss": 0.5448,
"step": 275
},
{
"epoch": 2.021978021978022,
"grad_norm": 0.31077465415000916,
"learning_rate": 7.431744886962338e-05,
"loss": 0.5772,
"step": 276
},
{
"epoch": 2.029304029304029,
"grad_norm": 0.3218604326248169,
"learning_rate": 7.329755765365101e-05,
"loss": 0.4763,
"step": 277
},
{
"epoch": 2.0366300366300365,
"grad_norm": 0.35660311579704285,
"learning_rate": 7.228244547628837e-05,
"loss": 0.5446,
"step": 278
},
{
"epoch": 2.043956043956044,
"grad_norm": 0.369620680809021,
"learning_rate": 7.127217558532974e-05,
"loss": 0.5142,
"step": 279
},
{
"epoch": 2.051282051282051,
"grad_norm": 0.374860942363739,
"learning_rate": 7.02668109268648e-05,
"loss": 0.4916,
"step": 280
},
{
"epoch": 2.0586080586080584,
"grad_norm": 0.38749629259109497,
"learning_rate": 6.926641414135674e-05,
"loss": 0.4357,
"step": 281
},
{
"epoch": 2.065934065934066,
"grad_norm": 0.405636191368103,
"learning_rate": 6.827104755973947e-05,
"loss": 0.4206,
"step": 282
},
{
"epoch": 2.0732600732600734,
"grad_norm": 0.40589308738708496,
"learning_rate": 6.728077319953388e-05,
"loss": 0.4861,
"step": 283
},
{
"epoch": 2.0805860805860807,
"grad_norm": 0.4325176775455475,
"learning_rate": 6.629565276098398e-05,
"loss": 0.4732,
"step": 284
},
{
"epoch": 2.087912087912088,
"grad_norm": 0.4640852212905884,
"learning_rate": 6.531574762321226e-05,
"loss": 0.4558,
"step": 285
},
{
"epoch": 2.0952380952380953,
"grad_norm": 0.43299245834350586,
"learning_rate": 6.434111884039579e-05,
"loss": 0.4278,
"step": 286
},
{
"epoch": 2.1025641025641026,
"grad_norm": 0.4622572660446167,
"learning_rate": 6.337182713796172e-05,
"loss": 0.4749,
"step": 287
},
{
"epoch": 2.10989010989011,
"grad_norm": 0.47975558042526245,
"learning_rate": 6.240793290880404e-05,
"loss": 0.3462,
"step": 288
},
{
"epoch": 2.1172161172161172,
"grad_norm": 0.48463815450668335,
"learning_rate": 6.144949620952074e-05,
"loss": 0.4606,
"step": 289
},
{
"epoch": 2.1245421245421245,
"grad_norm": 0.5988311767578125,
"learning_rate": 6.049657675667161e-05,
"loss": 0.3281,
"step": 290
},
{
"epoch": 2.131868131868132,
"grad_norm": 0.43833500146865845,
"learning_rate": 5.954923392305783e-05,
"loss": 0.3536,
"step": 291
},
{
"epoch": 2.139194139194139,
"grad_norm": 0.5085727572441101,
"learning_rate": 5.860752673402253e-05,
"loss": 0.3745,
"step": 292
},
{
"epoch": 2.1465201465201464,
"grad_norm": 0.5323147177696228,
"learning_rate": 5.767151386377313e-05,
"loss": 0.4184,
"step": 293
},
{
"epoch": 2.1538461538461537,
"grad_norm": 0.5333714485168457,
"learning_rate": 5.6741253631725734e-05,
"loss": 0.2953,
"step": 294
},
{
"epoch": 2.161172161172161,
"grad_norm": 0.5652948021888733,
"learning_rate": 5.581680399887123e-05,
"loss": 0.3118,
"step": 295
},
{
"epoch": 2.1684981684981683,
"grad_norm": 0.6583665609359741,
"learning_rate": 5.4898222564164196e-05,
"loss": 0.2828,
"step": 296
},
{
"epoch": 2.1758241758241756,
"grad_norm": 0.603283703327179,
"learning_rate": 5.398556656093393e-05,
"loss": 0.31,
"step": 297
},
{
"epoch": 2.183150183150183,
"grad_norm": 0.770128607749939,
"learning_rate": 5.307889285331851e-05,
"loss": 0.3279,
"step": 298
},
{
"epoch": 2.1904761904761907,
"grad_norm": 0.6317210793495178,
"learning_rate": 5.2178257932721996e-05,
"loss": 0.3703,
"step": 299
},
{
"epoch": 2.197802197802198,
"grad_norm": 0.6448614597320557,
"learning_rate": 5.128371791429436e-05,
"loss": 0.3123,
"step": 300
},
{
"epoch": 2.197802197802198,
"eval_loss": 0.6170002222061157,
"eval_runtime": 73.4287,
"eval_samples_per_second": 7.449,
"eval_steps_per_second": 7.449,
"step": 300
},
{
"epoch": 2.2051282051282053,
"grad_norm": 0.599308431148529,
"learning_rate": 5.039532853343533e-05,
"loss": 0.249,
"step": 301
},
{
"epoch": 2.2124542124542126,
"grad_norm": 0.7922989130020142,
"learning_rate": 4.951314514232175e-05,
"loss": 0.2308,
"step": 302
},
{
"epoch": 2.21978021978022,
"grad_norm": 0.8403403162956238,
"learning_rate": 4.863722270645869e-05,
"loss": 0.2293,
"step": 303
},
{
"epoch": 2.227106227106227,
"grad_norm": 0.8893972039222717,
"learning_rate": 4.776761580125495e-05,
"loss": 0.2041,
"step": 304
},
{
"epoch": 2.2344322344322345,
"grad_norm": 0.6155855059623718,
"learning_rate": 4.690437860862234e-05,
"loss": 0.1374,
"step": 305
},
{
"epoch": 2.241758241758242,
"grad_norm": 0.7354236245155334,
"learning_rate": 4.6047564913600234e-05,
"loss": 0.1482,
"step": 306
},
{
"epoch": 2.249084249084249,
"grad_norm": 2.1080105304718018,
"learning_rate": 4.519722810100403e-05,
"loss": 0.1127,
"step": 307
},
{
"epoch": 2.2564102564102564,
"grad_norm": 0.4302568733692169,
"learning_rate": 4.435342115209916e-05,
"loss": 0.5135,
"step": 308
},
{
"epoch": 2.2637362637362637,
"grad_norm": 0.8538389205932617,
"learning_rate": 4.35161966413001e-05,
"loss": 0.6252,
"step": 309
},
{
"epoch": 2.271062271062271,
"grad_norm": 0.7556006908416748,
"learning_rate": 4.2685606732894316e-05,
"loss": 0.4975,
"step": 310
},
{
"epoch": 2.2783882783882783,
"grad_norm": 0.58868408203125,
"learning_rate": 4.186170317779257e-05,
"loss": 0.4612,
"step": 311
},
{
"epoch": 2.2857142857142856,
"grad_norm": 0.7426422238349915,
"learning_rate": 4.1044537310304135e-05,
"loss": 0.458,
"step": 312
},
{
"epoch": 2.293040293040293,
"grad_norm": 0.6092411279678345,
"learning_rate": 4.023416004493849e-05,
"loss": 0.5098,
"step": 313
},
{
"epoch": 2.3003663003663,
"grad_norm": 0.6772862076759338,
"learning_rate": 3.943062187323317e-05,
"loss": 0.5537,
"step": 314
},
{
"epoch": 2.3076923076923075,
"grad_norm": 0.5766953229904175,
"learning_rate": 3.863397286060752e-05,
"loss": 0.4449,
"step": 315
},
{
"epoch": 2.315018315018315,
"grad_norm": 0.5632515549659729,
"learning_rate": 3.784426264324364e-05,
"loss": 0.4455,
"step": 316
},
{
"epoch": 2.3223443223443225,
"grad_norm": 0.4638958275318146,
"learning_rate": 3.7061540424993455e-05,
"loss": 0.3575,
"step": 317
},
{
"epoch": 2.32967032967033,
"grad_norm": 0.5185256004333496,
"learning_rate": 3.628585497431319e-05,
"loss": 0.4092,
"step": 318
},
{
"epoch": 2.336996336996337,
"grad_norm": 0.4830188751220703,
"learning_rate": 3.551725462122475e-05,
"loss": 0.4283,
"step": 319
},
{
"epoch": 2.3443223443223444,
"grad_norm": 0.5283904075622559,
"learning_rate": 3.47557872543044e-05,
"loss": 0.4279,
"step": 320
},
{
"epoch": 2.3516483516483517,
"grad_norm": 0.48210349678993225,
"learning_rate": 3.400150031769916e-05,
"loss": 0.3868,
"step": 321
},
{
"epoch": 2.358974358974359,
"grad_norm": 0.49750009179115295,
"learning_rate": 3.325444080817054e-05,
"loss": 0.4265,
"step": 322
},
{
"epoch": 2.3663003663003663,
"grad_norm": 0.5228102803230286,
"learning_rate": 3.251465527216644e-05,
"loss": 0.3988,
"step": 323
},
{
"epoch": 2.3736263736263736,
"grad_norm": 0.4787524938583374,
"learning_rate": 3.178218980292116e-05,
"loss": 0.4338,
"step": 324
},
{
"epoch": 2.380952380952381,
"grad_norm": 0.5247374773025513,
"learning_rate": 3.1057090037583195e-05,
"loss": 0.457,
"step": 325
},
{
"epoch": 2.3882783882783882,
"grad_norm": 0.49824175238609314,
"learning_rate": 3.03394011543721e-05,
"loss": 0.3382,
"step": 326
},
{
"epoch": 2.3956043956043955,
"grad_norm": 0.5834226012229919,
"learning_rate": 2.9629167869763314e-05,
"loss": 0.4105,
"step": 327
},
{
"epoch": 2.402930402930403,
"grad_norm": 0.5177899599075317,
"learning_rate": 2.8926434435702213e-05,
"loss": 0.3676,
"step": 328
},
{
"epoch": 2.41025641025641,
"grad_norm": 0.5421662926673889,
"learning_rate": 2.823124463684692e-05,
"loss": 0.3891,
"step": 329
},
{
"epoch": 2.4175824175824174,
"grad_norm": 0.6656437516212463,
"learning_rate": 2.7543641787840137e-05,
"loss": 0.4003,
"step": 330
},
{
"epoch": 2.4249084249084247,
"grad_norm": 0.5354308485984802,
"learning_rate": 2.6863668730610628e-05,
"loss": 0.334,
"step": 331
},
{
"epoch": 2.4322344322344325,
"grad_norm": 0.5242961049079895,
"learning_rate": 2.6191367831703597e-05,
"loss": 0.3575,
"step": 332
},
{
"epoch": 2.4395604395604398,
"grad_norm": 0.5668814182281494,
"learning_rate": 2.5526780979641132e-05,
"loss": 0.251,
"step": 333
},
{
"epoch": 2.446886446886447,
"grad_norm": 0.6647098064422607,
"learning_rate": 2.486994958231238e-05,
"loss": 0.3012,
"step": 334
},
{
"epoch": 2.4542124542124544,
"grad_norm": 0.590238630771637,
"learning_rate": 2.422091456439338e-05,
"loss": 0.2125,
"step": 335
},
{
"epoch": 2.4615384615384617,
"grad_norm": 0.6544525027275085,
"learning_rate": 2.3579716364797406e-05,
"loss": 0.2345,
"step": 336
},
{
"epoch": 2.468864468864469,
"grad_norm": 0.6543833017349243,
"learning_rate": 2.294639493415517e-05,
"loss": 0.2455,
"step": 337
},
{
"epoch": 2.4761904761904763,
"grad_norm": 0.8714612126350403,
"learning_rate": 2.2320989732325816e-05,
"loss": 0.2462,
"step": 338
},
{
"epoch": 2.4835164835164836,
"grad_norm": 0.8587841391563416,
"learning_rate": 2.170353972593825e-05,
"loss": 0.1519,
"step": 339
},
{
"epoch": 2.490842490842491,
"grad_norm": 0.6549997329711914,
"learning_rate": 2.1094083385963202e-05,
"loss": 0.1206,
"step": 340
},
{
"epoch": 2.498168498168498,
"grad_norm": 0.5828280448913574,
"learning_rate": 2.049265868531651e-05,
"loss": 0.0964,
"step": 341
},
{
"epoch": 2.5054945054945055,
"grad_norm": 0.29954391717910767,
"learning_rate": 1.989930309649282e-05,
"loss": 0.4719,
"step": 342
},
{
"epoch": 2.5128205128205128,
"grad_norm": 0.41014474630355835,
"learning_rate": 1.9314053589231067e-05,
"loss": 0.5926,
"step": 343
},
{
"epoch": 2.52014652014652,
"grad_norm": 0.42200765013694763,
"learning_rate": 1.873694662821096e-05,
"loss": 0.5305,
"step": 344
},
{
"epoch": 2.5274725274725274,
"grad_norm": 0.4672837257385254,
"learning_rate": 1.816801817078093e-05,
"loss": 0.5177,
"step": 345
},
{
"epoch": 2.5347985347985347,
"grad_norm": 0.45549672842025757,
"learning_rate": 1.760730366471796e-05,
"loss": 0.4906,
"step": 346
},
{
"epoch": 2.542124542124542,
"grad_norm": 0.42634275555610657,
"learning_rate": 1.705483804601871e-05,
"loss": 0.4466,
"step": 347
},
{
"epoch": 2.5494505494505493,
"grad_norm": 0.5157731175422668,
"learning_rate": 1.6510655736722967e-05,
"loss": 0.5153,
"step": 348
},
{
"epoch": 2.5567765567765566,
"grad_norm": 0.6015162467956543,
"learning_rate": 1.5974790642768903e-05,
"loss": 0.4401,
"step": 349
},
{
"epoch": 2.564102564102564,
"grad_norm": 0.5981739163398743,
"learning_rate": 1.5447276151880473e-05,
"loss": 0.5157,
"step": 350
},
{
"epoch": 2.564102564102564,
"eval_loss": 0.6206538081169128,
"eval_runtime": 73.5137,
"eval_samples_per_second": 7.441,
"eval_steps_per_second": 7.441,
"step": 350
},
{
"epoch": 2.571428571428571,
"grad_norm": 0.5820057988166809,
"learning_rate": 1.4928145131487267e-05,
"loss": 0.4702,
"step": 351
},
{
"epoch": 2.578754578754579,
"grad_norm": 0.5424510836601257,
"learning_rate": 1.4417429926676482e-05,
"loss": 0.4174,
"step": 352
},
{
"epoch": 2.586080586080586,
"grad_norm": 0.6577489376068115,
"learning_rate": 1.39151623581778e-05,
"loss": 0.4169,
"step": 353
},
{
"epoch": 2.5934065934065935,
"grad_norm": 0.5944687128067017,
"learning_rate": 1.3421373720380669e-05,
"loss": 0.4272,
"step": 354
},
{
"epoch": 2.600732600732601,
"grad_norm": 0.7125936150550842,
"learning_rate": 1.2936094779384486e-05,
"loss": 0.4438,
"step": 355
},
{
"epoch": 2.608058608058608,
"grad_norm": 0.679352879524231,
"learning_rate": 1.245935577108168e-05,
"loss": 0.4949,
"step": 356
},
{
"epoch": 2.6153846153846154,
"grad_norm": 1.0750776529312134,
"learning_rate": 1.199118639927385e-05,
"loss": 0.3986,
"step": 357
},
{
"epoch": 2.6227106227106227,
"grad_norm": 0.8349664211273193,
"learning_rate": 1.1531615833820906e-05,
"loss": 0.4152,
"step": 358
},
{
"epoch": 2.63003663003663,
"grad_norm": 0.770355761051178,
"learning_rate": 1.108067270882384e-05,
"loss": 0.4878,
"step": 359
},
{
"epoch": 2.6373626373626373,
"grad_norm": 0.7165803909301758,
"learning_rate": 1.0638385120840414e-05,
"loss": 0.3852,
"step": 360
},
{
"epoch": 2.6446886446886446,
"grad_norm": 0.7379334568977356,
"learning_rate": 1.0204780627134784e-05,
"loss": 0.4758,
"step": 361
},
{
"epoch": 2.652014652014652,
"grad_norm": 0.6477860808372498,
"learning_rate": 9.77988624396025e-06,
"loss": 0.3718,
"step": 362
},
{
"epoch": 2.659340659340659,
"grad_norm": 0.6215366721153259,
"learning_rate": 9.363728444876239e-06,
"loss": 0.3146,
"step": 363
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.7191038727760315,
"learning_rate": 8.956333159098677e-06,
"loss": 0.2567,
"step": 364
},
{
"epoch": 2.6739926739926743,
"grad_norm": 0.6438663005828857,
"learning_rate": 8.557725769884444e-06,
"loss": 0.3366,
"step": 365
},
{
"epoch": 2.6813186813186816,
"grad_norm": 0.7834159135818481,
"learning_rate": 8.167931112949955e-06,
"loss": 0.2802,
"step": 366
},
{
"epoch": 2.688644688644689,
"grad_norm": 0.687300443649292,
"learning_rate": 7.786973474923569e-06,
"loss": 0.3275,
"step": 367
},
{
"epoch": 2.695970695970696,
"grad_norm": 0.633350133895874,
"learning_rate": 7.41487659183258e-06,
"loss": 0.2462,
"step": 368
},
{
"epoch": 2.7032967032967035,
"grad_norm": 0.8299592137336731,
"learning_rate": 7.051663647624117e-06,
"loss": 0.2847,
"step": 369
},
{
"epoch": 2.7106227106227108,
"grad_norm": 0.6400296688079834,
"learning_rate": 6.697357272720782e-06,
"loss": 0.239,
"step": 370
},
{
"epoch": 2.717948717948718,
"grad_norm": 0.620663583278656,
"learning_rate": 6.35197954261058e-06,
"loss": 0.183,
"step": 371
},
{
"epoch": 2.7252747252747254,
"grad_norm": 0.7446674108505249,
"learning_rate": 6.015551976471433e-06,
"loss": 0.2334,
"step": 372
},
{
"epoch": 2.7326007326007327,
"grad_norm": 0.8476528525352478,
"learning_rate": 5.688095535830573e-06,
"loss": 0.1747,
"step": 373
},
{
"epoch": 2.73992673992674,
"grad_norm": 0.8313474655151367,
"learning_rate": 5.369630623258248e-06,
"loss": 0.189,
"step": 374
},
{
"epoch": 2.7472527472527473,
"grad_norm": 0.9323949217796326,
"learning_rate": 5.060177081096728e-06,
"loss": 0.0911,
"step": 375
},
{
"epoch": 2.7545787545787546,
"grad_norm": 0.33731380105018616,
"learning_rate": 4.759754190223925e-06,
"loss": 0.4471,
"step": 376
},
{
"epoch": 2.761904761904762,
"grad_norm": 0.37849536538124084,
"learning_rate": 4.468380668852068e-06,
"loss": 0.6115,
"step": 377
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.4009314775466919,
"learning_rate": 4.186074671361456e-06,
"loss": 0.575,
"step": 378
},
{
"epoch": 2.7765567765567765,
"grad_norm": 0.3839149475097656,
"learning_rate": 3.912853787169345e-06,
"loss": 0.5056,
"step": 379
},
{
"epoch": 2.7838827838827838,
"grad_norm": 0.40725114941596985,
"learning_rate": 3.6487350396339597e-06,
"loss": 0.4408,
"step": 380
},
{
"epoch": 2.791208791208791,
"grad_norm": 0.4696556329727173,
"learning_rate": 3.3937348849939204e-06,
"loss": 0.5321,
"step": 381
},
{
"epoch": 2.7985347985347984,
"grad_norm": 0.43771475553512573,
"learning_rate": 3.147869211342818e-06,
"loss": 0.4356,
"step": 382
},
{
"epoch": 2.8058608058608057,
"grad_norm": 0.5101984739303589,
"learning_rate": 2.911153337639388e-06,
"loss": 0.4499,
"step": 383
},
{
"epoch": 2.813186813186813,
"grad_norm": 0.5081220865249634,
"learning_rate": 2.683602012752939e-06,
"loss": 0.4756,
"step": 384
},
{
"epoch": 2.8205128205128203,
"grad_norm": 0.5413169264793396,
"learning_rate": 2.4652294145445226e-06,
"loss": 0.4866,
"step": 385
},
{
"epoch": 2.8278388278388276,
"grad_norm": 0.5008475184440613,
"learning_rate": 2.256049148983441e-06,
"loss": 0.4641,
"step": 386
},
{
"epoch": 2.8351648351648353,
"grad_norm": 0.5848665833473206,
"learning_rate": 2.0560742492995885e-06,
"loss": 0.4572,
"step": 387
},
{
"epoch": 2.8424908424908426,
"grad_norm": 0.6183776259422302,
"learning_rate": 1.8653171751714379e-06,
"loss": 0.5198,
"step": 388
},
{
"epoch": 2.84981684981685,
"grad_norm": 0.5888795852661133,
"learning_rate": 1.6837898119496263e-06,
"loss": 0.451,
"step": 389
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.5763834118843079,
"learning_rate": 1.5115034699164308e-06,
"loss": 0.4926,
"step": 390
},
{
"epoch": 2.8644688644688645,
"grad_norm": 0.6053770184516907,
"learning_rate": 1.348468883581183e-06,
"loss": 0.443,
"step": 391
},
{
"epoch": 2.871794871794872,
"grad_norm": 0.5707501769065857,
"learning_rate": 1.19469621101132e-06,
"loss": 0.4184,
"step": 392
},
{
"epoch": 2.879120879120879,
"grad_norm": 0.5886608958244324,
"learning_rate": 1.0501950331995578e-06,
"loss": 0.4002,
"step": 393
},
{
"epoch": 2.8864468864468864,
"grad_norm": 0.5797408223152161,
"learning_rate": 9.149743534668353e-07,
"loss": 0.4117,
"step": 394
},
{
"epoch": 2.8937728937728937,
"grad_norm": 0.7185364961624146,
"learning_rate": 7.890425969014625e-07,
"loss": 0.4059,
"step": 395
},
{
"epoch": 2.901098901098901,
"grad_norm": 0.7297194004058838,
"learning_rate": 6.724076098341247e-07,
"loss": 0.3407,
"step": 396
},
{
"epoch": 2.9084249084249083,
"grad_norm": 0.6948420405387878,
"learning_rate": 5.650766593489897e-07,
"loss": 0.3564,
"step": 397
},
{
"epoch": 2.9157509157509156,
"grad_norm": 0.6854040026664734,
"learning_rate": 4.6705643283102003e-07,
"loss": 0.4018,
"step": 398
},
{
"epoch": 2.9230769230769234,
"grad_norm": 0.7185117602348328,
"learning_rate": 3.7835303754918943e-07,
"loss": 0.3607,
"step": 399
},
{
"epoch": 2.9304029304029307,
"grad_norm": 0.6749255061149597,
"learning_rate": 2.9897200027598767e-07,
"loss": 0.276,
"step": 400
},
{
"epoch": 2.9304029304029307,
"eval_loss": 0.60358726978302,
"eval_runtime": 73.527,
"eval_samples_per_second": 7.439,
"eval_steps_per_second": 7.439,
"step": 400
}
],
"logging_steps": 1,
"max_steps": 408,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 3,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 3
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.659682780951347e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}