sp-embraceable's picture
Upload 10 files
6c7ba25 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.1359452956130453,
"eval_steps": 500,
"global_step": 10000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0013594529561304532,
"grad_norm": 2.6843507289886475,
"learning_rate": 2.9959624247202926e-05,
"loss": 0.791,
"step": 100
},
{
"epoch": 0.0027189059122609063,
"grad_norm": 7.921170711517334,
"learning_rate": 2.9918840658519014e-05,
"loss": 0.7116,
"step": 200
},
{
"epoch": 0.0040783588683913595,
"grad_norm": 1.703231692314148,
"learning_rate": 2.98780570698351e-05,
"loss": 0.7025,
"step": 300
},
{
"epoch": 0.005437811824521813,
"grad_norm": 1.4689126014709473,
"learning_rate": 2.9837273481151187e-05,
"loss": 0.7026,
"step": 400
},
{
"epoch": 0.006797264780652266,
"grad_norm": 6.632850646972656,
"learning_rate": 2.979648989246727e-05,
"loss": 0.7006,
"step": 500
},
{
"epoch": 0.008156717736782719,
"grad_norm": 1.7937493324279785,
"learning_rate": 2.975570630378336e-05,
"loss": 0.6996,
"step": 600
},
{
"epoch": 0.009516170692913172,
"grad_norm": 1.545050024986267,
"learning_rate": 2.9714922715099447e-05,
"loss": 0.6988,
"step": 700
},
{
"epoch": 0.010875623649043625,
"grad_norm": 6.718420505523682,
"learning_rate": 2.967413912641553e-05,
"loss": 0.6934,
"step": 800
},
{
"epoch": 0.012235076605174078,
"grad_norm": 2.417815923690796,
"learning_rate": 2.9633355537731616e-05,
"loss": 0.7031,
"step": 900
},
{
"epoch": 0.013594529561304532,
"grad_norm": 3.016540765762329,
"learning_rate": 2.9592571949047704e-05,
"loss": 0.7003,
"step": 1000
},
{
"epoch": 0.014953982517434985,
"grad_norm": 1.3102728128433228,
"learning_rate": 2.9551788360363792e-05,
"loss": 0.6939,
"step": 1100
},
{
"epoch": 0.016313435473565438,
"grad_norm": 3.511146306991577,
"learning_rate": 2.9511004771679877e-05,
"loss": 0.7018,
"step": 1200
},
{
"epoch": 0.01767288842969589,
"grad_norm": 2.3049850463867188,
"learning_rate": 2.9470221182995965e-05,
"loss": 0.7093,
"step": 1300
},
{
"epoch": 0.019032341385826344,
"grad_norm": 1.0154485702514648,
"learning_rate": 2.942943759431205e-05,
"loss": 0.7035,
"step": 1400
},
{
"epoch": 0.020391794341956797,
"grad_norm": 2.085158348083496,
"learning_rate": 2.9388654005628134e-05,
"loss": 0.7015,
"step": 1500
},
{
"epoch": 0.02175124729808725,
"grad_norm": 3.9277114868164062,
"learning_rate": 2.9347870416944225e-05,
"loss": 0.6994,
"step": 1600
},
{
"epoch": 0.023110700254217704,
"grad_norm": 1.8050942420959473,
"learning_rate": 2.930708682826031e-05,
"loss": 0.7027,
"step": 1700
},
{
"epoch": 0.024470153210348157,
"grad_norm": 1.5299062728881836,
"learning_rate": 2.9266303239576394e-05,
"loss": 0.6989,
"step": 1800
},
{
"epoch": 0.02582960616647861,
"grad_norm": 3.9614956378936768,
"learning_rate": 2.9225519650892482e-05,
"loss": 0.6968,
"step": 1900
},
{
"epoch": 0.027189059122609063,
"grad_norm": 1.5142548084259033,
"learning_rate": 2.9184736062208567e-05,
"loss": 0.6996,
"step": 2000
},
{
"epoch": 0.028548512078739516,
"grad_norm": 2.0255579948425293,
"learning_rate": 2.9143952473524655e-05,
"loss": 0.6992,
"step": 2100
},
{
"epoch": 0.02990796503486997,
"grad_norm": 3.3495869636535645,
"learning_rate": 2.9103168884840743e-05,
"loss": 0.6986,
"step": 2200
},
{
"epoch": 0.03126741799100042,
"grad_norm": 4.41119909286499,
"learning_rate": 2.9062385296156827e-05,
"loss": 0.6963,
"step": 2300
},
{
"epoch": 0.032626870947130876,
"grad_norm": 2.923621416091919,
"learning_rate": 2.902160170747291e-05,
"loss": 0.6935,
"step": 2400
},
{
"epoch": 0.033986323903261326,
"grad_norm": 6.4013285636901855,
"learning_rate": 2.8980818118789e-05,
"loss": 0.6984,
"step": 2500
},
{
"epoch": 0.03534577685939178,
"grad_norm": 2.191399097442627,
"learning_rate": 2.8940034530105088e-05,
"loss": 0.6955,
"step": 2600
},
{
"epoch": 0.03670522981552223,
"grad_norm": 1.6993489265441895,
"learning_rate": 2.8899250941421172e-05,
"loss": 0.7002,
"step": 2700
},
{
"epoch": 0.03806468277165269,
"grad_norm": 3.364286184310913,
"learning_rate": 2.885846735273726e-05,
"loss": 0.6964,
"step": 2800
},
{
"epoch": 0.03942413572778314,
"grad_norm": 1.9982099533081055,
"learning_rate": 2.8817683764053345e-05,
"loss": 0.7069,
"step": 2900
},
{
"epoch": 0.040783588683913595,
"grad_norm": 1.524899959564209,
"learning_rate": 2.877690017536943e-05,
"loss": 0.6961,
"step": 3000
},
{
"epoch": 0.042143041640044045,
"grad_norm": 2.2762181758880615,
"learning_rate": 2.873611658668552e-05,
"loss": 0.6991,
"step": 3100
},
{
"epoch": 0.0435024945961745,
"grad_norm": 2.0052831172943115,
"learning_rate": 2.8695332998001605e-05,
"loss": 0.6987,
"step": 3200
},
{
"epoch": 0.04486194755230495,
"grad_norm": 1.6109389066696167,
"learning_rate": 2.865454940931769e-05,
"loss": 0.6974,
"step": 3300
},
{
"epoch": 0.04622140050843541,
"grad_norm": 1.7458642721176147,
"learning_rate": 2.8613765820633778e-05,
"loss": 0.6984,
"step": 3400
},
{
"epoch": 0.04758085346456586,
"grad_norm": 5.7916083335876465,
"learning_rate": 2.8572982231949862e-05,
"loss": 0.6986,
"step": 3500
},
{
"epoch": 0.048940306420696314,
"grad_norm": 2.8352317810058594,
"learning_rate": 2.8532198643265954e-05,
"loss": 0.7004,
"step": 3600
},
{
"epoch": 0.050299759376826764,
"grad_norm": 6.102287769317627,
"learning_rate": 2.8491415054582038e-05,
"loss": 0.6975,
"step": 3700
},
{
"epoch": 0.05165921233295722,
"grad_norm": 2.8221065998077393,
"learning_rate": 2.8450631465898123e-05,
"loss": 0.7092,
"step": 3800
},
{
"epoch": 0.05301866528908767,
"grad_norm": 2.0802714824676514,
"learning_rate": 2.840984787721421e-05,
"loss": 0.7097,
"step": 3900
},
{
"epoch": 0.05437811824521813,
"grad_norm": 4.515227794647217,
"learning_rate": 2.8369064288530295e-05,
"loss": 0.6991,
"step": 4000
},
{
"epoch": 0.055737571201348576,
"grad_norm": 4.737243175506592,
"learning_rate": 2.8328280699846383e-05,
"loss": 0.7042,
"step": 4100
},
{
"epoch": 0.05709702415747903,
"grad_norm": 5.016382694244385,
"learning_rate": 2.828749711116247e-05,
"loss": 0.7008,
"step": 4200
},
{
"epoch": 0.05845647711360948,
"grad_norm": 2.7722420692443848,
"learning_rate": 2.8246713522478556e-05,
"loss": 0.7038,
"step": 4300
},
{
"epoch": 0.05981593006973994,
"grad_norm": 3.596283435821533,
"learning_rate": 2.820592993379464e-05,
"loss": 0.7008,
"step": 4400
},
{
"epoch": 0.06117538302587039,
"grad_norm": 3.8389534950256348,
"learning_rate": 2.8165146345110728e-05,
"loss": 0.6986,
"step": 4500
},
{
"epoch": 0.06253483598200084,
"grad_norm": 2.992204189300537,
"learning_rate": 2.8124362756426816e-05,
"loss": 0.7027,
"step": 4600
},
{
"epoch": 0.0638942889381313,
"grad_norm": 1.4091521501541138,
"learning_rate": 2.80835791677429e-05,
"loss": 0.7032,
"step": 4700
},
{
"epoch": 0.06525374189426175,
"grad_norm": 1.7205729484558105,
"learning_rate": 2.804279557905899e-05,
"loss": 0.7061,
"step": 4800
},
{
"epoch": 0.0666131948503922,
"grad_norm": 2.666346549987793,
"learning_rate": 2.8002011990375073e-05,
"loss": 0.7016,
"step": 4900
},
{
"epoch": 0.06797264780652265,
"grad_norm": 2.0234336853027344,
"learning_rate": 2.7961228401691158e-05,
"loss": 0.6986,
"step": 5000
},
{
"epoch": 0.06933210076265311,
"grad_norm": 1.3103491067886353,
"learning_rate": 2.792044481300725e-05,
"loss": 0.6995,
"step": 5100
},
{
"epoch": 0.07069155371878356,
"grad_norm": 3.3153600692749023,
"learning_rate": 2.7879661224323334e-05,
"loss": 0.6968,
"step": 5200
},
{
"epoch": 0.07205100667491401,
"grad_norm": 3.262269973754883,
"learning_rate": 2.7838877635639418e-05,
"loss": 0.7041,
"step": 5300
},
{
"epoch": 0.07341045963104446,
"grad_norm": 5.839659214019775,
"learning_rate": 2.7798094046955506e-05,
"loss": 0.692,
"step": 5400
},
{
"epoch": 0.07476991258717493,
"grad_norm": 6.034287929534912,
"learning_rate": 2.775731045827159e-05,
"loss": 0.7034,
"step": 5500
},
{
"epoch": 0.07612936554330538,
"grad_norm": 3.7333922386169434,
"learning_rate": 2.771652686958768e-05,
"loss": 0.6986,
"step": 5600
},
{
"epoch": 0.07748881849943583,
"grad_norm": 2.81872820854187,
"learning_rate": 2.7675743280903767e-05,
"loss": 0.7021,
"step": 5700
},
{
"epoch": 0.07884827145556628,
"grad_norm": 2.6562986373901367,
"learning_rate": 2.763495969221985e-05,
"loss": 0.6999,
"step": 5800
},
{
"epoch": 0.08020772441169673,
"grad_norm": 4.191847801208496,
"learning_rate": 2.7594176103535936e-05,
"loss": 0.704,
"step": 5900
},
{
"epoch": 0.08156717736782719,
"grad_norm": 1.9238234758377075,
"learning_rate": 2.7553392514852024e-05,
"loss": 0.6996,
"step": 6000
},
{
"epoch": 0.08292663032395764,
"grad_norm": 1.6448299884796143,
"learning_rate": 2.751260892616811e-05,
"loss": 0.7033,
"step": 6100
},
{
"epoch": 0.08428608328008809,
"grad_norm": 2.8520469665527344,
"learning_rate": 2.7471825337484196e-05,
"loss": 0.707,
"step": 6200
},
{
"epoch": 0.08564553623621854,
"grad_norm": 4.698349952697754,
"learning_rate": 2.7431041748800284e-05,
"loss": 0.6995,
"step": 6300
},
{
"epoch": 0.087004989192349,
"grad_norm": 3.2636826038360596,
"learning_rate": 2.739025816011637e-05,
"loss": 0.697,
"step": 6400
},
{
"epoch": 0.08836444214847945,
"grad_norm": 5.062309741973877,
"learning_rate": 2.7349474571432457e-05,
"loss": 0.7068,
"step": 6500
},
{
"epoch": 0.0897238951046099,
"grad_norm": 1.9477702379226685,
"learning_rate": 2.7308690982748545e-05,
"loss": 0.6946,
"step": 6600
},
{
"epoch": 0.09108334806074035,
"grad_norm": 6.437952518463135,
"learning_rate": 2.726790739406463e-05,
"loss": 0.6968,
"step": 6700
},
{
"epoch": 0.09244280101687082,
"grad_norm": 1.488918423652649,
"learning_rate": 2.7227123805380717e-05,
"loss": 0.6964,
"step": 6800
},
{
"epoch": 0.09380225397300126,
"grad_norm": 1.5171183347702026,
"learning_rate": 2.7186340216696802e-05,
"loss": 0.6948,
"step": 6900
},
{
"epoch": 0.09516170692913171,
"grad_norm": 5.248293876647949,
"learning_rate": 2.7145556628012886e-05,
"loss": 0.6914,
"step": 7000
},
{
"epoch": 0.09652115988526216,
"grad_norm": 3.5038247108459473,
"learning_rate": 2.7104773039328978e-05,
"loss": 0.7016,
"step": 7100
},
{
"epoch": 0.09788061284139263,
"grad_norm": 2.2439801692962646,
"learning_rate": 2.7063989450645062e-05,
"loss": 0.6976,
"step": 7200
},
{
"epoch": 0.09924006579752308,
"grad_norm": 5.262351036071777,
"learning_rate": 2.7023205861961147e-05,
"loss": 0.7015,
"step": 7300
},
{
"epoch": 0.10059951875365353,
"grad_norm": 2.734067916870117,
"learning_rate": 2.6982422273277235e-05,
"loss": 0.7012,
"step": 7400
},
{
"epoch": 0.10195897170978398,
"grad_norm": 7.341092586517334,
"learning_rate": 2.694163868459332e-05,
"loss": 0.6977,
"step": 7500
},
{
"epoch": 0.10331842466591444,
"grad_norm": 2.047778367996216,
"learning_rate": 2.6900855095909407e-05,
"loss": 0.6934,
"step": 7600
},
{
"epoch": 0.10467787762204489,
"grad_norm": 5.612318515777588,
"learning_rate": 2.6860071507225495e-05,
"loss": 0.696,
"step": 7700
},
{
"epoch": 0.10603733057817534,
"grad_norm": 3.8864567279815674,
"learning_rate": 2.681928791854158e-05,
"loss": 0.6958,
"step": 7800
},
{
"epoch": 0.10739678353430579,
"grad_norm": 2.456672191619873,
"learning_rate": 2.6778504329857664e-05,
"loss": 0.7014,
"step": 7900
},
{
"epoch": 0.10875623649043625,
"grad_norm": 1.5562827587127686,
"learning_rate": 2.6737720741173752e-05,
"loss": 0.7,
"step": 8000
},
{
"epoch": 0.1101156894465667,
"grad_norm": 1.646262764930725,
"learning_rate": 2.669693715248984e-05,
"loss": 0.7028,
"step": 8100
},
{
"epoch": 0.11147514240269715,
"grad_norm": 2.5881056785583496,
"learning_rate": 2.6656153563805925e-05,
"loss": 0.697,
"step": 8200
},
{
"epoch": 0.1128345953588276,
"grad_norm": 3.108797788619995,
"learning_rate": 2.6615369975122013e-05,
"loss": 0.7,
"step": 8300
},
{
"epoch": 0.11419404831495807,
"grad_norm": 1.5215388536453247,
"learning_rate": 2.6574586386438097e-05,
"loss": 0.6956,
"step": 8400
},
{
"epoch": 0.11555350127108852,
"grad_norm": 1.2557023763656616,
"learning_rate": 2.6533802797754182e-05,
"loss": 0.7032,
"step": 8500
},
{
"epoch": 0.11691295422721897,
"grad_norm": 3.6592652797698975,
"learning_rate": 2.6493019209070273e-05,
"loss": 0.6957,
"step": 8600
},
{
"epoch": 0.11827240718334942,
"grad_norm": 1.6143642663955688,
"learning_rate": 2.6452235620386358e-05,
"loss": 0.6961,
"step": 8700
},
{
"epoch": 0.11963186013947988,
"grad_norm": 1.7212355136871338,
"learning_rate": 2.6411452031702442e-05,
"loss": 0.7041,
"step": 8800
},
{
"epoch": 0.12099131309561033,
"grad_norm": 1.2407207489013672,
"learning_rate": 2.637066844301853e-05,
"loss": 0.6938,
"step": 8900
},
{
"epoch": 0.12235076605174078,
"grad_norm": 4.306702613830566,
"learning_rate": 2.6329884854334615e-05,
"loss": 0.704,
"step": 9000
},
{
"epoch": 0.12371021900787123,
"grad_norm": 4.0667219161987305,
"learning_rate": 2.6289101265650703e-05,
"loss": 0.6966,
"step": 9100
},
{
"epoch": 0.12506967196400168,
"grad_norm": 2.244699478149414,
"learning_rate": 2.624831767696679e-05,
"loss": 0.6976,
"step": 9200
},
{
"epoch": 0.12642912492013214,
"grad_norm": 4.839937210083008,
"learning_rate": 2.6207534088282875e-05,
"loss": 0.6972,
"step": 9300
},
{
"epoch": 0.1277885778762626,
"grad_norm": 5.436954021453857,
"learning_rate": 2.6166750499598963e-05,
"loss": 0.698,
"step": 9400
},
{
"epoch": 0.12914803083239304,
"grad_norm": 5.324636459350586,
"learning_rate": 2.6125966910915048e-05,
"loss": 0.699,
"step": 9500
},
{
"epoch": 0.1305074837885235,
"grad_norm": 2.4143927097320557,
"learning_rate": 2.6085183322231136e-05,
"loss": 0.7033,
"step": 9600
},
{
"epoch": 0.13186693674465394,
"grad_norm": 3.336245059967041,
"learning_rate": 2.6044399733547224e-05,
"loss": 0.6993,
"step": 9700
},
{
"epoch": 0.1332263897007844,
"grad_norm": 1.2645655870437622,
"learning_rate": 2.600361614486331e-05,
"loss": 0.6977,
"step": 9800
},
{
"epoch": 0.13458584265691487,
"grad_norm": 3.2803938388824463,
"learning_rate": 2.5962832556179393e-05,
"loss": 0.6993,
"step": 9900
},
{
"epoch": 0.1359452956130453,
"grad_norm": 2.2295751571655273,
"learning_rate": 2.592204896749548e-05,
"loss": 0.6923,
"step": 10000
}
],
"logging_steps": 100,
"max_steps": 73559,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}