| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 45.96219931271477, | |
| "global_step": 53500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.97852233676976e-05, | |
| "loss": 2.4481, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.957044673539519e-05, | |
| "loss": 2.0258, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.935567010309279e-05, | |
| "loss": 1.8794, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.9140893470790375e-05, | |
| "loss": 1.8082, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.892611683848797e-05, | |
| "loss": 1.7553, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.871134020618557e-05, | |
| "loss": 1.6993, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.849656357388316e-05, | |
| "loss": 1.6629, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.828178694158076e-05, | |
| "loss": 1.6249, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.806701030927835e-05, | |
| "loss": 1.6062, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 4.785223367697595e-05, | |
| "loss": 1.5661, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 4.763745704467354e-05, | |
| "loss": 1.5406, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 5.15, | |
| "learning_rate": 4.7422680412371134e-05, | |
| "loss": 1.5285, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 5.58, | |
| "learning_rate": 4.720790378006873e-05, | |
| "loss": 1.5024, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 4.6993127147766324e-05, | |
| "loss": 1.4933, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 6.44, | |
| "learning_rate": 4.677835051546392e-05, | |
| "loss": 1.474, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 4.656357388316151e-05, | |
| "loss": 1.46, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 7.3, | |
| "learning_rate": 4.634879725085911e-05, | |
| "loss": 1.4409, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 7.73, | |
| "learning_rate": 4.61340206185567e-05, | |
| "loss": 1.4251, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 4.5919243986254295e-05, | |
| "loss": 1.4157, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 8.59, | |
| "learning_rate": 4.570446735395189e-05, | |
| "loss": 1.3975, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 4.5489690721649484e-05, | |
| "loss": 1.3924, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 4.527491408934708e-05, | |
| "loss": 1.3699, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 9.88, | |
| "learning_rate": 4.5060137457044674e-05, | |
| "loss": 1.3711, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 4.484536082474227e-05, | |
| "loss": 1.3514, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 4.463058419243987e-05, | |
| "loss": 1.3502, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "learning_rate": 4.4415807560137455e-05, | |
| "loss": 1.3308, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 4.4201030927835053e-05, | |
| "loss": 1.3288, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 12.03, | |
| "learning_rate": 4.3986254295532645e-05, | |
| "loss": 1.3197, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 12.46, | |
| "learning_rate": 4.377147766323024e-05, | |
| "loss": 1.3022, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 4.3556701030927835e-05, | |
| "loss": 1.3039, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 13.32, | |
| "learning_rate": 4.334192439862543e-05, | |
| "loss": 1.2948, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 13.75, | |
| "learning_rate": 4.312714776632303e-05, | |
| "loss": 1.2842, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 14.18, | |
| "learning_rate": 4.2912371134020616e-05, | |
| "loss": 1.2804, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 4.2697594501718214e-05, | |
| "loss": 1.2646, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 15.03, | |
| "learning_rate": 4.2482817869415805e-05, | |
| "loss": 1.2672, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 15.46, | |
| "learning_rate": 4.2268041237113404e-05, | |
| "loss": 1.2501, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 15.89, | |
| "learning_rate": 4.2053264604811e-05, | |
| "loss": 1.2566, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 16.32, | |
| "learning_rate": 4.1838487972508593e-05, | |
| "loss": 1.2406, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 16.75, | |
| "learning_rate": 4.162371134020619e-05, | |
| "loss": 1.2383, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 17.18, | |
| "learning_rate": 4.140893470790378e-05, | |
| "loss": 1.2246, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 17.61, | |
| "learning_rate": 4.1194158075601375e-05, | |
| "loss": 1.2219, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 18.04, | |
| "learning_rate": 4.097938144329897e-05, | |
| "loss": 1.2174, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 18.47, | |
| "learning_rate": 4.0764604810996564e-05, | |
| "loss": 1.2089, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 18.9, | |
| "learning_rate": 4.054982817869416e-05, | |
| "loss": 1.2049, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 19.33, | |
| "learning_rate": 4.0335051546391754e-05, | |
| "loss": 1.2014, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 19.76, | |
| "learning_rate": 4.012027491408935e-05, | |
| "loss": 1.1965, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 20.19, | |
| "learning_rate": 3.9905498281786944e-05, | |
| "loss": 1.184, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 20.62, | |
| "learning_rate": 3.9690721649484535e-05, | |
| "loss": 1.178, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 21.05, | |
| "learning_rate": 3.9475945017182134e-05, | |
| "loss": 1.1801, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 21.48, | |
| "learning_rate": 3.9261168384879725e-05, | |
| "loss": 1.1661, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 21.91, | |
| "learning_rate": 3.904639175257732e-05, | |
| "loss": 1.1687, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 22.34, | |
| "learning_rate": 3.8831615120274915e-05, | |
| "loss": 1.1636, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 22.77, | |
| "learning_rate": 3.861683848797251e-05, | |
| "loss": 1.1578, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 23.2, | |
| "learning_rate": 3.8402061855670104e-05, | |
| "loss": 1.1456, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 23.63, | |
| "learning_rate": 3.8187285223367696e-05, | |
| "loss": 1.148, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 24.05, | |
| "learning_rate": 3.7972508591065294e-05, | |
| "loss": 1.1458, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 24.48, | |
| "learning_rate": 3.7757731958762886e-05, | |
| "loss": 1.1348, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 24.91, | |
| "learning_rate": 3.7542955326460484e-05, | |
| "loss": 1.1359, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 25.34, | |
| "learning_rate": 3.7328178694158075e-05, | |
| "loss": 1.1266, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 25.77, | |
| "learning_rate": 3.7113402061855674e-05, | |
| "loss": 1.1193, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 26.2, | |
| "learning_rate": 3.689862542955327e-05, | |
| "loss": 1.1201, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 26.63, | |
| "learning_rate": 3.6683848797250856e-05, | |
| "loss": 1.1177, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 27.06, | |
| "learning_rate": 3.6469072164948455e-05, | |
| "loss": 1.1102, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 27.49, | |
| "learning_rate": 3.6254295532646046e-05, | |
| "loss": 1.1038, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 27.92, | |
| "learning_rate": 3.6039518900343644e-05, | |
| "loss": 1.1086, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 28.35, | |
| "learning_rate": 3.5824742268041236e-05, | |
| "loss": 1.0972, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 28.78, | |
| "learning_rate": 3.5609965635738834e-05, | |
| "loss": 1.1021, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 29.21, | |
| "learning_rate": 3.539518900343643e-05, | |
| "loss": 1.0974, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 29.64, | |
| "learning_rate": 3.5180412371134024e-05, | |
| "loss": 1.0839, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 30.07, | |
| "learning_rate": 3.4965635738831615e-05, | |
| "loss": 1.0918, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 30.5, | |
| "learning_rate": 3.475085910652921e-05, | |
| "loss": 1.0809, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 30.93, | |
| "learning_rate": 3.4536082474226805e-05, | |
| "loss": 1.0789, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 31.36, | |
| "learning_rate": 3.43213058419244e-05, | |
| "loss": 1.0741, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 31.79, | |
| "learning_rate": 3.4106529209621995e-05, | |
| "loss": 1.0701, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 32.22, | |
| "learning_rate": 3.389175257731959e-05, | |
| "loss": 1.0648, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 32.65, | |
| "learning_rate": 3.3676975945017185e-05, | |
| "loss": 1.0627, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 33.08, | |
| "learning_rate": 3.3462199312714776e-05, | |
| "loss": 1.0659, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 33.51, | |
| "learning_rate": 3.3247422680412374e-05, | |
| "loss": 1.05, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 33.93, | |
| "learning_rate": 3.3032646048109966e-05, | |
| "loss": 1.0496, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 34.36, | |
| "learning_rate": 3.2817869415807564e-05, | |
| "loss": 1.0472, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 34.79, | |
| "learning_rate": 3.2603092783505155e-05, | |
| "loss": 1.0439, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 35.22, | |
| "learning_rate": 3.2388316151202754e-05, | |
| "loss": 1.0453, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 35.65, | |
| "learning_rate": 3.2173539518900345e-05, | |
| "loss": 1.0418, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 36.08, | |
| "learning_rate": 3.1958762886597937e-05, | |
| "loss": 1.0342, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 36.51, | |
| "learning_rate": 3.1743986254295535e-05, | |
| "loss": 1.0258, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 36.94, | |
| "learning_rate": 3.1529209621993126e-05, | |
| "loss": 1.0319, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 37.37, | |
| "learning_rate": 3.1314432989690725e-05, | |
| "loss": 1.0238, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 37.8, | |
| "learning_rate": 3.1099656357388316e-05, | |
| "loss": 1.0274, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 38.23, | |
| "learning_rate": 3.0884879725085914e-05, | |
| "loss": 1.0233, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 38.66, | |
| "learning_rate": 3.0670103092783506e-05, | |
| "loss": 1.0209, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 39.09, | |
| "learning_rate": 3.0455326460481097e-05, | |
| "loss": 1.0164, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 39.52, | |
| "learning_rate": 3.0240549828178692e-05, | |
| "loss": 1.0027, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 39.95, | |
| "learning_rate": 3.002577319587629e-05, | |
| "loss": 1.0168, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 40.38, | |
| "learning_rate": 2.9810996563573885e-05, | |
| "loss": 1.0003, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 40.81, | |
| "learning_rate": 2.959621993127148e-05, | |
| "loss": 1.0067, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 41.24, | |
| "learning_rate": 2.9381443298969075e-05, | |
| "loss": 0.9941, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 41.67, | |
| "learning_rate": 2.916666666666667e-05, | |
| "loss": 1.0014, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 42.1, | |
| "learning_rate": 2.8951890034364265e-05, | |
| "loss": 0.9963, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 42.53, | |
| "learning_rate": 2.8737113402061856e-05, | |
| "loss": 0.9899, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 42.96, | |
| "learning_rate": 2.852233676975945e-05, | |
| "loss": 0.9882, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 43.38, | |
| "learning_rate": 2.8307560137457046e-05, | |
| "loss": 0.9832, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 43.81, | |
| "learning_rate": 2.809278350515464e-05, | |
| "loss": 0.9817, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 44.24, | |
| "learning_rate": 2.7878006872852236e-05, | |
| "loss": 0.9816, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 44.67, | |
| "learning_rate": 2.766323024054983e-05, | |
| "loss": 0.9795, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 45.1, | |
| "learning_rate": 2.7448453608247425e-05, | |
| "loss": 0.9737, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 45.53, | |
| "learning_rate": 2.7233676975945017e-05, | |
| "loss": 0.9693, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 45.96, | |
| "learning_rate": 2.701890034364261e-05, | |
| "loss": 0.9735, | |
| "step": 53500 | |
| } | |
| ], | |
| "max_steps": 116400, | |
| "num_train_epochs": 100, | |
| "total_flos": 4.504557638270131e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |