Folder / trainer_state.json
Anonumous's picture
Upload folder using huggingface_hub
a582bba verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.402861860209136,
"eval_steps": 500,
"global_step": 40000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00550357732526142,
"grad_norm": 5.828019142150879,
"learning_rate": 6.47007042253521e-06,
"loss": 8.235,
"step": 50
},
{
"epoch": 0.01100715465052284,
"grad_norm": 15.358248710632324,
"learning_rate": 1.3072183098591547e-05,
"loss": 6.0857,
"step": 100
},
{
"epoch": 0.01651073197578426,
"grad_norm": 7.99273681640625,
"learning_rate": 1.9674295774647885e-05,
"loss": 4.4315,
"step": 150
},
{
"epoch": 0.02201430930104568,
"grad_norm": 2.8550803661346436,
"learning_rate": 2.6276408450704222e-05,
"loss": 3.681,
"step": 200
},
{
"epoch": 0.0275178866263071,
"grad_norm": 2.3769114017486572,
"learning_rate": 3.2878521126760565e-05,
"loss": 3.4729,
"step": 250
},
{
"epoch": 0.03302146395156852,
"grad_norm": 5.98811149597168,
"learning_rate": 3.94806338028169e-05,
"loss": 3.3829,
"step": 300
},
{
"epoch": 0.03852504127682994,
"grad_norm": 3.616163492202759,
"learning_rate": 4.608274647887324e-05,
"loss": 3.4119,
"step": 350
},
{
"epoch": 0.04402861860209136,
"grad_norm": 2.1746344566345215,
"learning_rate": 5.268485915492957e-05,
"loss": 3.3955,
"step": 400
},
{
"epoch": 0.04953219592735278,
"grad_norm": 1.3189276456832886,
"learning_rate": 5.928697183098591e-05,
"loss": 3.3177,
"step": 450
},
{
"epoch": 0.0550357732526142,
"grad_norm": 1.2459770441055298,
"learning_rate": 6.588908450704225e-05,
"loss": 3.3095,
"step": 500
},
{
"epoch": 0.06053935057787562,
"grad_norm": 1.2855902910232544,
"learning_rate": 7.249119718309858e-05,
"loss": 3.2437,
"step": 550
},
{
"epoch": 0.06604292790313704,
"grad_norm": 1.2871235609054565,
"learning_rate": 7.909330985915493e-05,
"loss": 3.2069,
"step": 600
},
{
"epoch": 0.07154650522839846,
"grad_norm": 1.0846847295761108,
"learning_rate": 8.569542253521127e-05,
"loss": 3.2512,
"step": 650
},
{
"epoch": 0.07705008255365987,
"grad_norm": 1.5505499839782715,
"learning_rate": 9.22975352112676e-05,
"loss": 3.1704,
"step": 700
},
{
"epoch": 0.0825536598789213,
"grad_norm": 1.178614616394043,
"learning_rate": 9.889964788732394e-05,
"loss": 3.1747,
"step": 750
},
{
"epoch": 0.08805723720418272,
"grad_norm": 1.1504727602005005,
"learning_rate": 0.00010550176056338028,
"loss": 3.1241,
"step": 800
},
{
"epoch": 0.09356081452944413,
"grad_norm": 1.02865469455719,
"learning_rate": 0.00011210387323943662,
"loss": 3.0481,
"step": 850
},
{
"epoch": 0.09906439185470556,
"grad_norm": 1.1366077661514282,
"learning_rate": 0.00011870598591549295,
"loss": 3.0201,
"step": 900
},
{
"epoch": 0.10456796917996698,
"grad_norm": 0.9753648042678833,
"learning_rate": 0.00012530809859154929,
"loss": 3.0145,
"step": 950
},
{
"epoch": 0.1100715465052284,
"grad_norm": 0.6859256625175476,
"learning_rate": 0.00013191021126760563,
"loss": 2.9671,
"step": 1000
},
{
"epoch": 0.11557512383048982,
"grad_norm": 0.8368203043937683,
"learning_rate": 0.00013851232394366197,
"loss": 2.992,
"step": 1050
},
{
"epoch": 0.12107870115575124,
"grad_norm": 1.1109174489974976,
"learning_rate": 0.0001451144366197183,
"loss": 2.967,
"step": 1100
},
{
"epoch": 0.12658227848101267,
"grad_norm": 0.6072912812232971,
"learning_rate": 0.00015171654929577465,
"loss": 2.9501,
"step": 1150
},
{
"epoch": 0.13208585580627408,
"grad_norm": 0.7659889459609985,
"learning_rate": 0.00015831866197183099,
"loss": 2.9026,
"step": 1200
},
{
"epoch": 0.1375894331315355,
"grad_norm": 0.5841110348701477,
"learning_rate": 0.0001649207746478873,
"loss": 2.8972,
"step": 1250
},
{
"epoch": 0.14309301045679693,
"grad_norm": 0.5997458100318909,
"learning_rate": 0.00017152288732394364,
"loss": 2.9189,
"step": 1300
},
{
"epoch": 0.14859658778205834,
"grad_norm": 0.6082264184951782,
"learning_rate": 0.00017812499999999998,
"loss": 2.8877,
"step": 1350
},
{
"epoch": 0.15410016510731975,
"grad_norm": 0.696685254573822,
"learning_rate": 0.00018472711267605632,
"loss": 2.8752,
"step": 1400
},
{
"epoch": 0.15960374243258119,
"grad_norm": 0.6795832514762878,
"learning_rate": 0.00019132922535211266,
"loss": 2.851,
"step": 1450
},
{
"epoch": 0.1651073197578426,
"grad_norm": 0.5588585734367371,
"learning_rate": 0.000197931338028169,
"loss": 2.8546,
"step": 1500
},
{
"epoch": 0.170610897083104,
"grad_norm": 0.5267760157585144,
"learning_rate": 0.00020453345070422534,
"loss": 2.8049,
"step": 1550
},
{
"epoch": 0.17611447440836545,
"grad_norm": 0.591826856136322,
"learning_rate": 0.00021113556338028168,
"loss": 2.8071,
"step": 1600
},
{
"epoch": 0.18161805173362686,
"grad_norm": 0.5463298559188843,
"learning_rate": 0.00021773767605633802,
"loss": 2.7667,
"step": 1650
},
{
"epoch": 0.18712162905888827,
"grad_norm": 0.5745858550071716,
"learning_rate": 0.00022433978873239433,
"loss": 2.7968,
"step": 1700
},
{
"epoch": 0.1926252063841497,
"grad_norm": 0.6239858865737915,
"learning_rate": 0.00023094190140845067,
"loss": 2.8206,
"step": 1750
},
{
"epoch": 0.19812878370941112,
"grad_norm": 0.46496015787124634,
"learning_rate": 0.000237544014084507,
"loss": 2.7944,
"step": 1800
},
{
"epoch": 0.20363236103467253,
"grad_norm": 0.5664075016975403,
"learning_rate": 0.00024414612676056335,
"loss": 2.7837,
"step": 1850
},
{
"epoch": 0.20913593835993396,
"grad_norm": 0.537627100944519,
"learning_rate": 0.0002507482394366197,
"loss": 2.7179,
"step": 1900
},
{
"epoch": 0.21463951568519538,
"grad_norm": 0.544585645198822,
"learning_rate": 0.00025735035211267603,
"loss": 2.7552,
"step": 1950
},
{
"epoch": 0.2201430930104568,
"grad_norm": 0.5067969560623169,
"learning_rate": 0.0002639524647887324,
"loss": 2.7309,
"step": 2000
},
{
"epoch": 0.22564667033571822,
"grad_norm": 0.462003618478775,
"learning_rate": 0.0002705545774647887,
"loss": 2.716,
"step": 2050
},
{
"epoch": 0.23115024766097964,
"grad_norm": 0.44174736738204956,
"learning_rate": 0.00027715669014084505,
"loss": 2.7319,
"step": 2100
},
{
"epoch": 0.23665382498624105,
"grad_norm": 0.5010894536972046,
"learning_rate": 0.0002837588028169014,
"loss": 2.7199,
"step": 2150
},
{
"epoch": 0.24215740231150248,
"grad_norm": 0.44113022089004517,
"learning_rate": 0.00029036091549295773,
"loss": 2.7146,
"step": 2200
},
{
"epoch": 0.2476609796367639,
"grad_norm": 0.4972345232963562,
"learning_rate": 0.0002969630281690141,
"loss": 2.6471,
"step": 2250
},
{
"epoch": 0.25316455696202533,
"grad_norm": 0.5614003539085388,
"learning_rate": 0.0002999997102213327,
"loss": 2.687,
"step": 2300
},
{
"epoch": 0.25866813428728674,
"grad_norm": 0.5223066806793213,
"learning_rate": 0.0002999976432184194,
"loss": 2.6279,
"step": 2350
},
{
"epoch": 0.26417171161254815,
"grad_norm": 0.4381965100765228,
"learning_rate": 0.0002999935887402823,
"loss": 2.6889,
"step": 2400
},
{
"epoch": 0.26967528893780957,
"grad_norm": 0.3969985842704773,
"learning_rate": 0.00029998754684064345,
"loss": 2.6565,
"step": 2450
},
{
"epoch": 0.275178866263071,
"grad_norm": 0.44786953926086426,
"learning_rate": 0.00029997951759955823,
"loss": 2.6736,
"step": 2500
},
{
"epoch": 0.28068244358833244,
"grad_norm": 0.3951723873615265,
"learning_rate": 0.0002999695011234145,
"loss": 2.6173,
"step": 2550
},
{
"epoch": 0.28618602091359385,
"grad_norm": 0.4258750081062317,
"learning_rate": 0.00029995749754493093,
"loss": 2.5992,
"step": 2600
},
{
"epoch": 0.29168959823885526,
"grad_norm": 0.41892707347869873,
"learning_rate": 0.0002999435070231555,
"loss": 2.6194,
"step": 2650
},
{
"epoch": 0.2971931755641167,
"grad_norm": 0.4064221978187561,
"learning_rate": 0.0002999275297434632,
"loss": 2.6265,
"step": 2700
},
{
"epoch": 0.3026967528893781,
"grad_norm": 0.3920956254005432,
"learning_rate": 0.00029990956591755365,
"loss": 2.6035,
"step": 2750
},
{
"epoch": 0.3082003302146395,
"grad_norm": 0.366318941116333,
"learning_rate": 0.0002998896157834484,
"loss": 2.6352,
"step": 2800
},
{
"epoch": 0.31370390753990096,
"grad_norm": 0.40867865085601807,
"learning_rate": 0.0002998676796054875,
"loss": 2.5674,
"step": 2850
},
{
"epoch": 0.31920748486516237,
"grad_norm": 0.3685750365257263,
"learning_rate": 0.00029984375767432627,
"loss": 2.5748,
"step": 2900
},
{
"epoch": 0.3247110621904238,
"grad_norm": 0.38134968280792236,
"learning_rate": 0.0002998178503069314,
"loss": 2.5703,
"step": 2950
},
{
"epoch": 0.3302146395156852,
"grad_norm": 0.36195963621139526,
"learning_rate": 0.00029978995784657643,
"loss": 2.5367,
"step": 3000
},
{
"epoch": 0.3357182168409466,
"grad_norm": 0.3696858584880829,
"learning_rate": 0.0002997600806628379,
"loss": 2.5693,
"step": 3050
},
{
"epoch": 0.341221794166208,
"grad_norm": 0.3404608964920044,
"learning_rate": 0.00029972821915158964,
"loss": 2.5454,
"step": 3100
},
{
"epoch": 0.3467253714914695,
"grad_norm": 0.33220136165618896,
"learning_rate": 0.0002996943737349981,
"loss": 2.5675,
"step": 3150
},
{
"epoch": 0.3522289488167309,
"grad_norm": 0.3744208812713623,
"learning_rate": 0.00029965854486151643,
"loss": 2.5453,
"step": 3200
},
{
"epoch": 0.3577325261419923,
"grad_norm": 0.36920756101608276,
"learning_rate": 0.0002996207330058788,
"loss": 2.5314,
"step": 3250
},
{
"epoch": 0.3632361034672537,
"grad_norm": 0.36301830410957336,
"learning_rate": 0.00029958093866909403,
"loss": 2.5361,
"step": 3300
},
{
"epoch": 0.3687396807925151,
"grad_norm": 0.3135142922401428,
"learning_rate": 0.00029953916237843853,
"loss": 2.5222,
"step": 3350
},
{
"epoch": 0.37424325811777653,
"grad_norm": 0.3634016513824463,
"learning_rate": 0.00029949540468745015,
"loss": 2.5179,
"step": 3400
},
{
"epoch": 0.379746835443038,
"grad_norm": 0.370914101600647,
"learning_rate": 0.00029944966617592017,
"loss": 2.5077,
"step": 3450
},
{
"epoch": 0.3852504127682994,
"grad_norm": 0.32538744807243347,
"learning_rate": 0.0002994019474498858,
"loss": 2.4873,
"step": 3500
},
{
"epoch": 0.3907539900935608,
"grad_norm": 0.3077157139778137,
"learning_rate": 0.0002993522491416221,
"loss": 2.5149,
"step": 3550
},
{
"epoch": 0.39625756741882223,
"grad_norm": 0.32142704725265503,
"learning_rate": 0.000299300571909634,
"loss": 2.4989,
"step": 3600
},
{
"epoch": 0.40176114474408364,
"grad_norm": 0.3485497832298279,
"learning_rate": 0.00029924691643864684,
"loss": 2.5114,
"step": 3650
},
{
"epoch": 0.40726472206934505,
"grad_norm": 0.34483280777931213,
"learning_rate": 0.0002991912834395981,
"loss": 2.4863,
"step": 3700
},
{
"epoch": 0.4127682993946065,
"grad_norm": 0.30801650881767273,
"learning_rate": 0.00029913367364962733,
"loss": 2.4935,
"step": 3750
},
{
"epoch": 0.41827187671986793,
"grad_norm": 0.3099140524864197,
"learning_rate": 0.00029907408783206674,
"loss": 2.5122,
"step": 3800
},
{
"epoch": 0.42377545404512934,
"grad_norm": 0.33073538541793823,
"learning_rate": 0.0002990125267764309,
"loss": 2.4613,
"step": 3850
},
{
"epoch": 0.42927903137039075,
"grad_norm": 0.3025052845478058,
"learning_rate": 0.00029894899129840653,
"loss": 2.4855,
"step": 3900
},
{
"epoch": 0.43478260869565216,
"grad_norm": 0.3200172483921051,
"learning_rate": 0.00029888348223984143,
"loss": 2.4542,
"step": 3950
},
{
"epoch": 0.4402861860209136,
"grad_norm": 0.31750544905662537,
"learning_rate": 0.0002988160004687335,
"loss": 2.4714,
"step": 4000
},
{
"epoch": 0.44578976334617504,
"grad_norm": 0.28732484579086304,
"learning_rate": 0.00029874654687921895,
"loss": 2.4595,
"step": 4050
},
{
"epoch": 0.45129334067143645,
"grad_norm": 0.33312809467315674,
"learning_rate": 0.0002986751223915609,
"loss": 2.4797,
"step": 4100
},
{
"epoch": 0.45679691799669786,
"grad_norm": 0.3109307289123535,
"learning_rate": 0.00029860172795213695,
"loss": 2.4506,
"step": 4150
},
{
"epoch": 0.46230049532195927,
"grad_norm": 0.31127694249153137,
"learning_rate": 0.0002985263645334266,
"loss": 2.4544,
"step": 4200
},
{
"epoch": 0.4678040726472207,
"grad_norm": 0.34132882952690125,
"learning_rate": 0.0002984490331339982,
"loss": 2.455,
"step": 4250
},
{
"epoch": 0.4733076499724821,
"grad_norm": 0.3107958436012268,
"learning_rate": 0.00029836973477849634,
"loss": 2.4733,
"step": 4300
},
{
"epoch": 0.47881122729774356,
"grad_norm": 0.27321889996528625,
"learning_rate": 0.00029828847051762753,
"loss": 2.4381,
"step": 4350
},
{
"epoch": 0.48431480462300497,
"grad_norm": 0.3114171326160431,
"learning_rate": 0.0002982052414281467,
"loss": 2.4612,
"step": 4400
},
{
"epoch": 0.4898183819482664,
"grad_norm": 0.2926501929759979,
"learning_rate": 0.00029812004861284294,
"loss": 2.4364,
"step": 4450
},
{
"epoch": 0.4953219592735278,
"grad_norm": 0.2824021279811859,
"learning_rate": 0.00029803289320052466,
"loss": 2.4453,
"step": 4500
},
{
"epoch": 0.5008255365987893,
"grad_norm": 0.2710123062133789,
"learning_rate": 0.00029794377634600485,
"loss": 2.4207,
"step": 4550
},
{
"epoch": 0.5063291139240507,
"grad_norm": 0.28232479095458984,
"learning_rate": 0.0002978526992300856,
"loss": 2.4332,
"step": 4600
},
{
"epoch": 0.5118326912493121,
"grad_norm": 0.2661692798137665,
"learning_rate": 0.0002977596630595427,
"loss": 2.4469,
"step": 4650
},
{
"epoch": 0.5173362685745735,
"grad_norm": 0.317094087600708,
"learning_rate": 0.0002976646690671094,
"loss": 2.4486,
"step": 4700
},
{
"epoch": 0.5228398458998349,
"grad_norm": 0.268915593624115,
"learning_rate": 0.0002975677185114602,
"loss": 2.4447,
"step": 4750
},
{
"epoch": 0.5283434232250963,
"grad_norm": 0.2714874744415283,
"learning_rate": 0.000297468812677194,
"loss": 2.4065,
"step": 4800
},
{
"epoch": 0.5338470005503577,
"grad_norm": 0.2872399389743805,
"learning_rate": 0.0002973679528748175,
"loss": 2.3943,
"step": 4850
},
{
"epoch": 0.5393505778756191,
"grad_norm": 0.24350598454475403,
"learning_rate": 0.00029726514044072736,
"loss": 2.3888,
"step": 4900
},
{
"epoch": 0.5448541552008805,
"grad_norm": 0.2634688913822174,
"learning_rate": 0.00029716037673719275,
"loss": 2.3953,
"step": 4950
},
{
"epoch": 0.550357732526142,
"grad_norm": 0.2842876613140106,
"learning_rate": 0.0002970536631523373,
"loss": 2.403,
"step": 5000
},
{
"epoch": 0.5558613098514034,
"grad_norm": 0.30130457878112793,
"learning_rate": 0.00029694500110012055,
"loss": 2.4393,
"step": 5050
},
{
"epoch": 0.5613648871766649,
"grad_norm": 0.2594560980796814,
"learning_rate": 0.00029683439202031936,
"loss": 2.3827,
"step": 5100
},
{
"epoch": 0.5668684645019263,
"grad_norm": 0.25417500734329224,
"learning_rate": 0.0002967218373785088,
"loss": 2.3789,
"step": 5150
},
{
"epoch": 0.5723720418271877,
"grad_norm": 0.2533874213695526,
"learning_rate": 0.0002966073386660428,
"loss": 2.4211,
"step": 5200
},
{
"epoch": 0.5778756191524491,
"grad_norm": 0.2343342900276184,
"learning_rate": 0.0002964908974000341,
"loss": 2.4332,
"step": 5250
},
{
"epoch": 0.5833791964777105,
"grad_norm": 0.25453534722328186,
"learning_rate": 0.0002963725151233345,
"loss": 2.3984,
"step": 5300
},
{
"epoch": 0.5888827738029719,
"grad_norm": 0.2670257091522217,
"learning_rate": 0.00029625219340451435,
"loss": 2.4126,
"step": 5350
},
{
"epoch": 0.5943863511282333,
"grad_norm": 0.2728422284126282,
"learning_rate": 0.00029612993383784154,
"loss": 2.4152,
"step": 5400
},
{
"epoch": 0.5998899284534948,
"grad_norm": 0.27964428067207336,
"learning_rate": 0.0002960057380432606,
"loss": 2.3675,
"step": 5450
},
{
"epoch": 0.6053935057787562,
"grad_norm": 0.27354755997657776,
"learning_rate": 0.00029587960766637103,
"loss": 2.4109,
"step": 5500
},
{
"epoch": 0.6108970831040176,
"grad_norm": 0.26175597310066223,
"learning_rate": 0.000295751544378406,
"loss": 2.3953,
"step": 5550
},
{
"epoch": 0.616400660429279,
"grad_norm": 0.2590219974517822,
"learning_rate": 0.0002956215498762093,
"loss": 2.3699,
"step": 5600
},
{
"epoch": 0.6219042377545404,
"grad_norm": 0.2495882660150528,
"learning_rate": 0.0002954896258822139,
"loss": 2.3949,
"step": 5650
},
{
"epoch": 0.6274078150798019,
"grad_norm": 0.26052239537239075,
"learning_rate": 0.0002953557741444183,
"loss": 2.3642,
"step": 5700
},
{
"epoch": 0.6329113924050633,
"grad_norm": 0.2627600431442261,
"learning_rate": 0.0002952199964363638,
"loss": 2.3895,
"step": 5750
},
{
"epoch": 0.6384149697303247,
"grad_norm": 0.2417898327112198,
"learning_rate": 0.00029508229455711086,
"loss": 2.4087,
"step": 5800
},
{
"epoch": 0.6439185470555862,
"grad_norm": 0.24891215562820435,
"learning_rate": 0.00029494267033121525,
"loss": 2.3642,
"step": 5850
},
{
"epoch": 0.6494221243808476,
"grad_norm": 0.2453078180551529,
"learning_rate": 0.0002948011256087041,
"loss": 2.3761,
"step": 5900
},
{
"epoch": 0.654925701706109,
"grad_norm": 0.24607980251312256,
"learning_rate": 0.0002946576622650509,
"loss": 2.3596,
"step": 5950
},
{
"epoch": 0.6604292790313704,
"grad_norm": 0.2540590763092041,
"learning_rate": 0.0002945122822011513,
"loss": 2.3666,
"step": 6000
},
{
"epoch": 0.6659328563566318,
"grad_norm": 0.25918814539909363,
"learning_rate": 0.00029436498734329727,
"loss": 2.3494,
"step": 6050
},
{
"epoch": 0.6714364336818932,
"grad_norm": 0.27554330229759216,
"learning_rate": 0.0002942157796431521,
"loss": 2.3793,
"step": 6100
},
{
"epoch": 0.6769400110071546,
"grad_norm": 0.22879928350448608,
"learning_rate": 0.00029406466107772416,
"loss": 2.3898,
"step": 6150
},
{
"epoch": 0.682443588332416,
"grad_norm": 0.2851306200027466,
"learning_rate": 0.00029391163364934095,
"loss": 2.3493,
"step": 6200
},
{
"epoch": 0.6879471656576774,
"grad_norm": 0.255014568567276,
"learning_rate": 0.0002937566993856225,
"loss": 2.3959,
"step": 6250
},
{
"epoch": 0.693450742982939,
"grad_norm": 0.23992781341075897,
"learning_rate": 0.00029359986033945454,
"loss": 2.3588,
"step": 6300
},
{
"epoch": 0.6989543203082004,
"grad_norm": 0.24902793765068054,
"learning_rate": 0.000293441118588961,
"loss": 2.3908,
"step": 6350
},
{
"epoch": 0.7044578976334618,
"grad_norm": 0.2294873744249344,
"learning_rate": 0.0002932804762374771,
"loss": 2.398,
"step": 6400
},
{
"epoch": 0.7099614749587232,
"grad_norm": 0.23560309410095215,
"learning_rate": 0.00029311793541352075,
"loss": 2.3481,
"step": 6450
},
{
"epoch": 0.7154650522839846,
"grad_norm": 0.27493488788604736,
"learning_rate": 0.00029295349827076497,
"loss": 2.3248,
"step": 6500
},
{
"epoch": 0.720968629609246,
"grad_norm": 0.22731854021549225,
"learning_rate": 0.0002927871669880089,
"loss": 2.3444,
"step": 6550
},
{
"epoch": 0.7264722069345074,
"grad_norm": 0.22198539972305298,
"learning_rate": 0.0002926189437691492,
"loss": 2.3148,
"step": 6600
},
{
"epoch": 0.7319757842597688,
"grad_norm": 0.25049686431884766,
"learning_rate": 0.0002924488308431508,
"loss": 2.3134,
"step": 6650
},
{
"epoch": 0.7374793615850302,
"grad_norm": 0.2354484349489212,
"learning_rate": 0.0002922768304640172,
"loss": 2.348,
"step": 6700
},
{
"epoch": 0.7429829389102917,
"grad_norm": 0.23955915868282318,
"learning_rate": 0.00029210294491076094,
"loss": 2.3172,
"step": 6750
},
{
"epoch": 0.7484865162355531,
"grad_norm": 0.2476130574941635,
"learning_rate": 0.000291927176487373,
"loss": 2.3842,
"step": 6800
},
{
"epoch": 0.7539900935608145,
"grad_norm": 0.2247840166091919,
"learning_rate": 0.00029174952752279254,
"loss": 2.3302,
"step": 6850
},
{
"epoch": 0.759493670886076,
"grad_norm": 0.22530816495418549,
"learning_rate": 0.000291570000370876,
"loss": 2.3707,
"step": 6900
},
{
"epoch": 0.7649972482113374,
"grad_norm": 0.2242722362279892,
"learning_rate": 0.00029138859741036587,
"loss": 2.312,
"step": 6950
},
{
"epoch": 0.7705008255365988,
"grad_norm": 0.23376357555389404,
"learning_rate": 0.0002912053210448592,
"loss": 2.3334,
"step": 7000
},
{
"epoch": 0.7760044028618602,
"grad_norm": 0.23131632804870605,
"learning_rate": 0.0002910201737027757,
"loss": 2.3293,
"step": 7050
},
{
"epoch": 0.7815079801871216,
"grad_norm": 0.21440596878528595,
"learning_rate": 0.0002908331578373256,
"loss": 2.3593,
"step": 7100
},
{
"epoch": 0.787011557512383,
"grad_norm": 0.225584477186203,
"learning_rate": 0.00029064427592647715,
"loss": 2.3464,
"step": 7150
},
{
"epoch": 0.7925151348376445,
"grad_norm": 0.22819140553474426,
"learning_rate": 0.0002904535304729238,
"loss": 2.3145,
"step": 7200
},
{
"epoch": 0.7980187121629059,
"grad_norm": 0.2503032982349396,
"learning_rate": 0.00029026092400405115,
"loss": 2.3376,
"step": 7250
},
{
"epoch": 0.8035222894881673,
"grad_norm": 0.2061757743358612,
"learning_rate": 0.00029006645907190295,
"loss": 2.3252,
"step": 7300
},
{
"epoch": 0.8090258668134287,
"grad_norm": 0.2099982351064682,
"learning_rate": 0.00028987013825314804,
"loss": 2.35,
"step": 7350
},
{
"epoch": 0.8145294441386901,
"grad_norm": 0.212050199508667,
"learning_rate": 0.00028967196414904573,
"loss": 2.3304,
"step": 7400
},
{
"epoch": 0.8200330214639515,
"grad_norm": 0.22283074259757996,
"learning_rate": 0.0002894719393854114,
"loss": 2.3389,
"step": 7450
},
{
"epoch": 0.825536598789213,
"grad_norm": 0.2060365378856659,
"learning_rate": 0.0002892700666125817,
"loss": 2.3215,
"step": 7500
},
{
"epoch": 0.8310401761144744,
"grad_norm": 0.22510799765586853,
"learning_rate": 0.0002890663485053797,
"loss": 2.3295,
"step": 7550
},
{
"epoch": 0.8365437534397359,
"grad_norm": 0.21860332787036896,
"learning_rate": 0.00028886078776307905,
"loss": 2.3286,
"step": 7600
},
{
"epoch": 0.8420473307649973,
"grad_norm": 0.20771068334579468,
"learning_rate": 0.00028865338710936826,
"loss": 2.3143,
"step": 7650
},
{
"epoch": 0.8475509080902587,
"grad_norm": 0.23098550736904144,
"learning_rate": 0.0002884441492923151,
"loss": 2.3583,
"step": 7700
},
{
"epoch": 0.8530544854155201,
"grad_norm": 0.204328715801239,
"learning_rate": 0.00028823307708432963,
"loss": 2.3425,
"step": 7750
},
{
"epoch": 0.8585580627407815,
"grad_norm": 0.20610789954662323,
"learning_rate": 0.0002880201732821275,
"loss": 2.3269,
"step": 7800
},
{
"epoch": 0.8640616400660429,
"grad_norm": 0.22584037482738495,
"learning_rate": 0.0002878054407066935,
"loss": 2.3061,
"step": 7850
},
{
"epoch": 0.8695652173913043,
"grad_norm": 0.22317476570606232,
"learning_rate": 0.0002875888822032433,
"loss": 2.3047,
"step": 7900
},
{
"epoch": 0.8750687947165657,
"grad_norm": 0.2132745087146759,
"learning_rate": 0.00028737050064118645,
"loss": 2.2925,
"step": 7950
},
{
"epoch": 0.8805723720418271,
"grad_norm": 0.21960894763469696,
"learning_rate": 0.00028715029891408805,
"loss": 2.3365,
"step": 8000
},
{
"epoch": 0.8860759493670886,
"grad_norm": 0.2113264799118042,
"learning_rate": 0.0002869282799396305,
"loss": 2.2814,
"step": 8050
},
{
"epoch": 0.8915795266923501,
"grad_norm": 0.19959582388401031,
"learning_rate": 0.00028670444665957465,
"loss": 2.2787,
"step": 8100
},
{
"epoch": 0.8970831040176115,
"grad_norm": 0.24770890176296234,
"learning_rate": 0.00028647880203972115,
"loss": 2.3312,
"step": 8150
},
{
"epoch": 0.9025866813428729,
"grad_norm": 0.2229924350976944,
"learning_rate": 0.0002862513490698709,
"loss": 2.2825,
"step": 8200
},
{
"epoch": 0.9080902586681343,
"grad_norm": 0.23118580877780914,
"learning_rate": 0.0002860220907637856,
"loss": 2.2973,
"step": 8250
},
{
"epoch": 0.9135938359933957,
"grad_norm": 0.1978590488433838,
"learning_rate": 0.0002857910301591475,
"loss": 2.3106,
"step": 8300
},
{
"epoch": 0.9190974133186571,
"grad_norm": 0.20972274243831635,
"learning_rate": 0.0002855581703175198,
"loss": 2.3489,
"step": 8350
},
{
"epoch": 0.9246009906439185,
"grad_norm": 0.20369485020637512,
"learning_rate": 0.0002853235143243052,
"loss": 2.3135,
"step": 8400
},
{
"epoch": 0.93010456796918,
"grad_norm": 0.19085292518138885,
"learning_rate": 0.00028508706528870576,
"loss": 2.2969,
"step": 8450
},
{
"epoch": 0.9356081452944414,
"grad_norm": 0.21080902218818665,
"learning_rate": 0.0002848488263436814,
"loss": 2.2979,
"step": 8500
},
{
"epoch": 0.9411117226197028,
"grad_norm": 0.1939728856086731,
"learning_rate": 0.00028460880064590835,
"loss": 2.31,
"step": 8550
},
{
"epoch": 0.9466152999449642,
"grad_norm": 0.23909969627857208,
"learning_rate": 0.0002843669913757375,
"loss": 2.2585,
"step": 8600
},
{
"epoch": 0.9521188772702256,
"grad_norm": 0.22809971868991852,
"learning_rate": 0.000284123401737152,
"loss": 2.3058,
"step": 8650
},
{
"epoch": 0.9576224545954871,
"grad_norm": 0.2066497504711151,
"learning_rate": 0.00028387803495772513,
"loss": 2.2736,
"step": 8700
},
{
"epoch": 0.9631260319207485,
"grad_norm": 0.1921154260635376,
"learning_rate": 0.00028363089428857716,
"loss": 2.2813,
"step": 8750
},
{
"epoch": 0.9686296092460099,
"grad_norm": 0.19423851370811462,
"learning_rate": 0.00028338198300433264,
"loss": 2.314,
"step": 8800
},
{
"epoch": 0.9741331865712713,
"grad_norm": 0.224751815199852,
"learning_rate": 0.0002831313044030768,
"loss": 2.3335,
"step": 8850
},
{
"epoch": 0.9796367638965328,
"grad_norm": 0.21634523570537567,
"learning_rate": 0.00028287886180631183,
"loss": 2.3021,
"step": 8900
},
{
"epoch": 0.9851403412217942,
"grad_norm": 0.19320085644721985,
"learning_rate": 0.00028262465855891306,
"loss": 2.2635,
"step": 8950
},
{
"epoch": 0.9906439185470556,
"grad_norm": 0.20169596374034882,
"learning_rate": 0.00028236869802908453,
"loss": 2.3013,
"step": 9000
},
{
"epoch": 0.996147495872317,
"grad_norm": 0.2045382857322693,
"learning_rate": 0.00028211098360831407,
"loss": 2.2576,
"step": 9050
},
{
"epoch": 1.0016510731975785,
"grad_norm": 0.2171631157398224,
"learning_rate": 0.00028185151871132894,
"loss": 2.3136,
"step": 9100
},
{
"epoch": 1.00715465052284,
"grad_norm": 0.19541560113430023,
"learning_rate": 0.00028159030677605017,
"loss": 2.2547,
"step": 9150
},
{
"epoch": 1.0126582278481013,
"grad_norm": 0.19888882339000702,
"learning_rate": 0.0002813273512635472,
"loss": 2.2275,
"step": 9200
},
{
"epoch": 1.0181618051733627,
"grad_norm": 0.24578270316123962,
"learning_rate": 0.0002810626556579918,
"loss": 2.2775,
"step": 9250
},
{
"epoch": 1.0236653824986242,
"grad_norm": 0.19902436435222626,
"learning_rate": 0.00028079622346661217,
"loss": 2.2353,
"step": 9300
},
{
"epoch": 1.0291689598238856,
"grad_norm": 0.19234856963157654,
"learning_rate": 0.00028052805821964633,
"loss": 2.2769,
"step": 9350
},
{
"epoch": 1.034672537149147,
"grad_norm": 0.19912928342819214,
"learning_rate": 0.0002802581634702952,
"loss": 2.2625,
"step": 9400
},
{
"epoch": 1.0401761144744084,
"grad_norm": 0.1918436884880066,
"learning_rate": 0.00027998654279467604,
"loss": 2.2335,
"step": 9450
},
{
"epoch": 1.0456796917996698,
"grad_norm": 0.20422010123729706,
"learning_rate": 0.0002797131997917743,
"loss": 2.3022,
"step": 9500
},
{
"epoch": 1.0511832691249312,
"grad_norm": 0.19621697068214417,
"learning_rate": 0.00027943813808339657,
"loss": 2.2378,
"step": 9550
},
{
"epoch": 1.0566868464501926,
"grad_norm": 0.18403683602809906,
"learning_rate": 0.0002791613613141223,
"loss": 2.2502,
"step": 9600
},
{
"epoch": 1.062190423775454,
"grad_norm": 0.2140798568725586,
"learning_rate": 0.0002788828731512556,
"loss": 2.2722,
"step": 9650
},
{
"epoch": 1.0676940011007154,
"grad_norm": 0.19459135830402374,
"learning_rate": 0.0002786026772847767,
"loss": 2.2361,
"step": 9700
},
{
"epoch": 1.0731975784259769,
"grad_norm": 0.20383061468601227,
"learning_rate": 0.00027832077742729277,
"loss": 2.2471,
"step": 9750
},
{
"epoch": 1.0787011557512383,
"grad_norm": 0.1901775300502777,
"learning_rate": 0.0002780371773139891,
"loss": 2.2968,
"step": 9800
},
{
"epoch": 1.0842047330764997,
"grad_norm": 0.20100219547748566,
"learning_rate": 0.0002777518807025793,
"loss": 2.2568,
"step": 9850
},
{
"epoch": 1.089708310401761,
"grad_norm": 0.2127520740032196,
"learning_rate": 0.00027746489137325586,
"loss": 2.28,
"step": 9900
},
{
"epoch": 1.0952118877270225,
"grad_norm": 0.19059176743030548,
"learning_rate": 0.00027717621312863965,
"loss": 2.284,
"step": 9950
},
{
"epoch": 1.100715465052284,
"grad_norm": 0.20845113694667816,
"learning_rate": 0.0002768858497937299,
"loss": 2.2409,
"step": 10000
},
{
"epoch": 1.1062190423775453,
"grad_norm": 0.18969906866550446,
"learning_rate": 0.00027659380521585325,
"loss": 2.2725,
"step": 10050
},
{
"epoch": 1.1117226197028067,
"grad_norm": 0.19071218371391296,
"learning_rate": 0.0002763000832646129,
"loss": 2.2624,
"step": 10100
},
{
"epoch": 1.1172261970280681,
"grad_norm": 0.21723783016204834,
"learning_rate": 0.0002760046878318373,
"loss": 2.2453,
"step": 10150
},
{
"epoch": 1.1227297743533298,
"grad_norm": 0.19336852431297302,
"learning_rate": 0.00027570762283152874,
"loss": 2.2322,
"step": 10200
},
{
"epoch": 1.1282333516785912,
"grad_norm": 0.19500704109668732,
"learning_rate": 0.0002754088921998112,
"loss": 2.1858,
"step": 10250
},
{
"epoch": 1.1337369290038526,
"grad_norm": 0.19920021295547485,
"learning_rate": 0.0002751084998948784,
"loss": 2.2286,
"step": 10300
},
{
"epoch": 1.139240506329114,
"grad_norm": 0.18863485753536224,
"learning_rate": 0.00027480644989694126,
"loss": 2.2623,
"step": 10350
},
{
"epoch": 1.1447440836543754,
"grad_norm": 0.18571630120277405,
"learning_rate": 0.0002745027462081753,
"loss": 2.2616,
"step": 10400
},
{
"epoch": 1.1502476609796368,
"grad_norm": 0.187313050031662,
"learning_rate": 0.00027419739285266745,
"loss": 2.2608,
"step": 10450
},
{
"epoch": 1.1557512383048982,
"grad_norm": 0.19139795005321503,
"learning_rate": 0.0002738903938763628,
"loss": 2.2204,
"step": 10500
},
{
"epoch": 1.1612548156301596,
"grad_norm": 0.188013955950737,
"learning_rate": 0.0002735817533470109,
"loss": 2.2596,
"step": 10550
},
{
"epoch": 1.166758392955421,
"grad_norm": 0.1884177178144455,
"learning_rate": 0.0002732714753541122,
"loss": 2.2618,
"step": 10600
},
{
"epoch": 1.1722619702806825,
"grad_norm": 0.19187124073505402,
"learning_rate": 0.00027295956400886335,
"loss": 2.2075,
"step": 10650
},
{
"epoch": 1.1777655476059439,
"grad_norm": 0.19815993309020996,
"learning_rate": 0.0002726460234441031,
"loss": 2.2677,
"step": 10700
},
{
"epoch": 1.1832691249312053,
"grad_norm": 0.18218094110488892,
"learning_rate": 0.00027233085781425744,
"loss": 2.2697,
"step": 10750
},
{
"epoch": 1.1887727022564667,
"grad_norm": 0.1923886239528656,
"learning_rate": 0.0002720140712952845,
"loss": 2.2127,
"step": 10800
},
{
"epoch": 1.194276279581728,
"grad_norm": 0.18907909095287323,
"learning_rate": 0.0002716956680846193,
"loss": 2.1989,
"step": 10850
},
{
"epoch": 1.1997798569069895,
"grad_norm": 0.1902381181716919,
"learning_rate": 0.000271375652401118,
"loss": 2.2211,
"step": 10900
},
{
"epoch": 1.205283434232251,
"grad_norm": 0.20184637606143951,
"learning_rate": 0.0002710540284850023,
"loss": 2.2451,
"step": 10950
},
{
"epoch": 1.2107870115575123,
"grad_norm": 0.19903656840324402,
"learning_rate": 0.00027073080059780285,
"loss": 2.2233,
"step": 11000
},
{
"epoch": 1.2162905888827737,
"grad_norm": 0.19023925065994263,
"learning_rate": 0.000270405973022303,
"loss": 2.2438,
"step": 11050
},
{
"epoch": 1.2217941662080352,
"grad_norm": 0.17722088098526,
"learning_rate": 0.0002700795500624822,
"loss": 2.259,
"step": 11100
},
{
"epoch": 1.2272977435332966,
"grad_norm": 0.19715473055839539,
"learning_rate": 0.0002697515360434587,
"loss": 2.2509,
"step": 11150
},
{
"epoch": 1.232801320858558,
"grad_norm": 0.18065761029720306,
"learning_rate": 0.00026942193531143225,
"loss": 2.2006,
"step": 11200
},
{
"epoch": 1.2383048981838194,
"grad_norm": 0.18757817149162292,
"learning_rate": 0.00026909075223362683,
"loss": 2.2215,
"step": 11250
},
{
"epoch": 1.243808475509081,
"grad_norm": 0.18816334009170532,
"learning_rate": 0.00026875799119823243,
"loss": 2.2314,
"step": 11300
},
{
"epoch": 1.2493120528343424,
"grad_norm": 0.19772210717201233,
"learning_rate": 0.0002684236566143471,
"loss": 2.2256,
"step": 11350
},
{
"epoch": 1.2548156301596038,
"grad_norm": 0.19510945677757263,
"learning_rate": 0.0002680877529119184,
"loss": 2.2427,
"step": 11400
},
{
"epoch": 1.2603192074848653,
"grad_norm": 0.19960781931877136,
"learning_rate": 0.0002677502845416849,
"loss": 2.2324,
"step": 11450
},
{
"epoch": 1.2658227848101267,
"grad_norm": 0.1882307380437851,
"learning_rate": 0.0002674112559751169,
"loss": 2.2432,
"step": 11500
},
{
"epoch": 1.271326362135388,
"grad_norm": 0.18959587812423706,
"learning_rate": 0.00026707067170435767,
"loss": 2.19,
"step": 11550
},
{
"epoch": 1.2768299394606495,
"grad_norm": 0.18768323957920074,
"learning_rate": 0.0002667285362421634,
"loss": 2.2294,
"step": 11600
},
{
"epoch": 1.282333516785911,
"grad_norm": 0.20077955722808838,
"learning_rate": 0.00026638485412184355,
"loss": 2.2518,
"step": 11650
},
{
"epoch": 1.2878370941111723,
"grad_norm": 0.1956523358821869,
"learning_rate": 0.00026603962989720105,
"loss": 2.2662,
"step": 11700
},
{
"epoch": 1.2933406714364337,
"grad_norm": 0.19226029515266418,
"learning_rate": 0.0002656928681424718,
"loss": 2.2133,
"step": 11750
},
{
"epoch": 1.2988442487616951,
"grad_norm": 0.16987648606300354,
"learning_rate": 0.0002653445734522639,
"loss": 2.2226,
"step": 11800
},
{
"epoch": 1.3043478260869565,
"grad_norm": 0.17524655163288116,
"learning_rate": 0.00026499475044149703,
"loss": 2.2555,
"step": 11850
},
{
"epoch": 1.309851403412218,
"grad_norm": 0.19188162684440613,
"learning_rate": 0.00026464340374534104,
"loss": 2.2234,
"step": 11900
},
{
"epoch": 1.3153549807374794,
"grad_norm": 0.17676417529582977,
"learning_rate": 0.00026429053801915487,
"loss": 2.2188,
"step": 11950
},
{
"epoch": 1.3208585580627408,
"grad_norm": 0.20071938633918762,
"learning_rate": 0.0002639361579384245,
"loss": 2.234,
"step": 12000
},
{
"epoch": 1.3263621353880022,
"grad_norm": 0.19267459213733673,
"learning_rate": 0.0002635802681987012,
"loss": 2.235,
"step": 12050
},
{
"epoch": 1.3318657127132636,
"grad_norm": 0.1724405139684677,
"learning_rate": 0.00026322287351553944,
"loss": 2.2388,
"step": 12100
},
{
"epoch": 1.337369290038525,
"grad_norm": 0.1907806396484375,
"learning_rate": 0.000262863978624434,
"loss": 2.1937,
"step": 12150
},
{
"epoch": 1.3428728673637864,
"grad_norm": 0.18452796339988708,
"learning_rate": 0.0002625035882807578,
"loss": 2.2429,
"step": 12200
},
{
"epoch": 1.3483764446890478,
"grad_norm": 0.17565739154815674,
"learning_rate": 0.0002621417072596982,
"loss": 2.2367,
"step": 12250
},
{
"epoch": 1.3538800220143092,
"grad_norm": 0.1761862337589264,
"learning_rate": 0.00026177834035619445,
"loss": 2.2258,
"step": 12300
},
{
"epoch": 1.3593835993395706,
"grad_norm": 0.18616576492786407,
"learning_rate": 0.0002614134923848736,
"loss": 2.2207,
"step": 12350
},
{
"epoch": 1.364887176664832,
"grad_norm": 0.18103408813476562,
"learning_rate": 0.00026104716817998696,
"loss": 2.2346,
"step": 12400
},
{
"epoch": 1.3703907539900935,
"grad_norm": 0.18511448800563812,
"learning_rate": 0.00026067937259534595,
"loss": 2.2363,
"step": 12450
},
{
"epoch": 1.3758943313153549,
"grad_norm": 0.1737717241048813,
"learning_rate": 0.00026031011050425796,
"loss": 2.2199,
"step": 12500
},
{
"epoch": 1.3813979086406163,
"grad_norm": 0.18761631846427917,
"learning_rate": 0.00025993938679946154,
"loss": 2.228,
"step": 12550
},
{
"epoch": 1.3869014859658777,
"grad_norm": 0.18815304338932037,
"learning_rate": 0.0002595672063930617,
"loss": 2.2034,
"step": 12600
},
{
"epoch": 1.3924050632911391,
"grad_norm": 0.1868327558040619,
"learning_rate": 0.00025919357421646486,
"loss": 2.197,
"step": 12650
},
{
"epoch": 1.3979086406164005,
"grad_norm": 0.17544154822826385,
"learning_rate": 0.00025881849522031345,
"loss": 2.1956,
"step": 12700
},
{
"epoch": 1.4034122179416622,
"grad_norm": 0.18338458240032196,
"learning_rate": 0.00025844197437442033,
"loss": 2.2291,
"step": 12750
},
{
"epoch": 1.4089157952669236,
"grad_norm": 0.17127631604671478,
"learning_rate": 0.00025806401666770277,
"loss": 2.2205,
"step": 12800
},
{
"epoch": 1.414419372592185,
"grad_norm": 0.17331157624721527,
"learning_rate": 0.00025768462710811673,
"loss": 2.2356,
"step": 12850
},
{
"epoch": 1.4199229499174464,
"grad_norm": 0.16098715364933014,
"learning_rate": 0.00025730381072259026,
"loss": 2.2065,
"step": 12900
},
{
"epoch": 1.4254265272427078,
"grad_norm": 0.17098484933376312,
"learning_rate": 0.00025692157255695667,
"loss": 2.2323,
"step": 12950
},
{
"epoch": 1.4309301045679692,
"grad_norm": 0.15763573348522186,
"learning_rate": 0.00025653791767588823,
"loss": 2.2064,
"step": 13000
},
{
"epoch": 1.4364336818932306,
"grad_norm": 0.17587807774543762,
"learning_rate": 0.0002561528511628286,
"loss": 2.2085,
"step": 13050
},
{
"epoch": 1.441937259218492,
"grad_norm": 0.17770230770111084,
"learning_rate": 0.00025576637811992555,
"loss": 2.2204,
"step": 13100
},
{
"epoch": 1.4474408365437534,
"grad_norm": 0.18236953020095825,
"learning_rate": 0.0002553785036679636,
"loss": 2.2388,
"step": 13150
},
{
"epoch": 1.4529444138690149,
"grad_norm": 0.16505366563796997,
"learning_rate": 0.00025498923294629594,
"loss": 2.2345,
"step": 13200
},
{
"epoch": 1.4584479911942763,
"grad_norm": 0.1687227189540863,
"learning_rate": 0.0002545985711127763,
"loss": 2.1652,
"step": 13250
},
{
"epoch": 1.4639515685195377,
"grad_norm": 0.17849218845367432,
"learning_rate": 0.00025420652334369085,
"loss": 2.2115,
"step": 13300
},
{
"epoch": 1.469455145844799,
"grad_norm": 0.17708005011081696,
"learning_rate": 0.0002538130948336894,
"loss": 2.2017,
"step": 13350
},
{
"epoch": 1.4749587231700605,
"grad_norm": 0.17772626876831055,
"learning_rate": 0.0002534182907957165,
"loss": 2.1926,
"step": 13400
},
{
"epoch": 1.480462300495322,
"grad_norm": 0.16764992475509644,
"learning_rate": 0.00025302211646094277,
"loss": 2.1986,
"step": 13450
},
{
"epoch": 1.4859658778205833,
"grad_norm": 0.19117778539657593,
"learning_rate": 0.00025262457707869506,
"loss": 2.2118,
"step": 13500
},
{
"epoch": 1.4914694551458447,
"grad_norm": 0.17632248997688293,
"learning_rate": 0.00025222567791638743,
"loss": 2.1897,
"step": 13550
},
{
"epoch": 1.4969730324711064,
"grad_norm": 0.19731369614601135,
"learning_rate": 0.0002518254242594508,
"loss": 2.1945,
"step": 13600
},
{
"epoch": 1.5024766097963678,
"grad_norm": 0.16187229752540588,
"learning_rate": 0.0002514238214112635,
"loss": 2.1959,
"step": 13650
},
{
"epoch": 1.5079801871216292,
"grad_norm": 0.1828337460756302,
"learning_rate": 0.00025102087469308036,
"loss": 2.1994,
"step": 13700
},
{
"epoch": 1.5134837644468906,
"grad_norm": 0.17188578844070435,
"learning_rate": 0.0002506165894439628,
"loss": 2.2087,
"step": 13750
},
{
"epoch": 1.518987341772152,
"grad_norm": 0.17914508283138275,
"learning_rate": 0.00025021097102070786,
"loss": 2.2266,
"step": 13800
},
{
"epoch": 1.5244909190974134,
"grad_norm": 0.17482663691043854,
"learning_rate": 0.0002498040247977769,
"loss": 2.2431,
"step": 13850
},
{
"epoch": 1.5299944964226748,
"grad_norm": 0.17342698574066162,
"learning_rate": 0.00024939575616722505,
"loss": 2.2045,
"step": 13900
},
{
"epoch": 1.5354980737479362,
"grad_norm": 0.18258632719516754,
"learning_rate": 0.00024898617053862904,
"loss": 2.1843,
"step": 13950
},
{
"epoch": 1.5410016510731976,
"grad_norm": 0.16960634291172028,
"learning_rate": 0.00024857527333901625,
"loss": 2.2096,
"step": 14000
},
{
"epoch": 1.546505228398459,
"grad_norm": 0.17008773982524872,
"learning_rate": 0.00024816307001279214,
"loss": 2.1522,
"step": 14050
},
{
"epoch": 1.5520088057237205,
"grad_norm": 0.16970837116241455,
"learning_rate": 0.00024774956602166857,
"loss": 2.2055,
"step": 14100
},
{
"epoch": 1.5575123830489819,
"grad_norm": 0.17486636340618134,
"learning_rate": 0.0002473347668445912,
"loss": 2.1906,
"step": 14150
},
{
"epoch": 1.5630159603742433,
"grad_norm": 0.19200065732002258,
"learning_rate": 0.00024691867797766684,
"loss": 2.1839,
"step": 14200
},
{
"epoch": 1.5685195376995047,
"grad_norm": 0.19441814720630646,
"learning_rate": 0.00024650130493409104,
"loss": 2.2121,
"step": 14250
},
{
"epoch": 1.574023115024766,
"grad_norm": 0.16507557034492493,
"learning_rate": 0.0002460826532440745,
"loss": 2.2016,
"step": 14300
},
{
"epoch": 1.5795266923500275,
"grad_norm": 0.16732299327850342,
"learning_rate": 0.00024566272845477014,
"loss": 2.2362,
"step": 14350
},
{
"epoch": 1.585030269675289,
"grad_norm": 0.18136867880821228,
"learning_rate": 0.00024524153613019947,
"loss": 2.193,
"step": 14400
},
{
"epoch": 1.5905338470005503,
"grad_norm": 0.16409073770046234,
"learning_rate": 0.0002448190818511789,
"loss": 2.1589,
"step": 14450
},
{
"epoch": 1.5960374243258117,
"grad_norm": 0.17381897568702698,
"learning_rate": 0.00024439537121524583,
"loss": 2.1834,
"step": 14500
},
{
"epoch": 1.6015410016510732,
"grad_norm": 0.1628628373146057,
"learning_rate": 0.00024397040983658436,
"loss": 2.172,
"step": 14550
},
{
"epoch": 1.6070445789763346,
"grad_norm": 0.16793246567249298,
"learning_rate": 0.00024354420334595105,
"loss": 2.176,
"step": 14600
},
{
"epoch": 1.612548156301596,
"grad_norm": 0.17394863069057465,
"learning_rate": 0.0002431167573906002,
"loss": 2.199,
"step": 14650
},
{
"epoch": 1.6180517336268574,
"grad_norm": 0.1554850935935974,
"learning_rate": 0.00024268807763420914,
"loss": 2.2042,
"step": 14700
},
{
"epoch": 1.6235553109521188,
"grad_norm": 0.17003560066223145,
"learning_rate": 0.00024225816975680312,
"loss": 2.1717,
"step": 14750
},
{
"epoch": 1.6290588882773802,
"grad_norm": 0.17207174003124237,
"learning_rate": 0.0002418270394546799,
"loss": 2.1894,
"step": 14800
},
{
"epoch": 1.6345624656026416,
"grad_norm": 0.1507563292980194,
"learning_rate": 0.00024139469244033467,
"loss": 2.2101,
"step": 14850
},
{
"epoch": 1.640066042927903,
"grad_norm": 0.17063497006893158,
"learning_rate": 0.0002409611344423838,
"loss": 2.1854,
"step": 14900
},
{
"epoch": 1.6455696202531644,
"grad_norm": 0.17093567550182343,
"learning_rate": 0.00024052637120548954,
"loss": 2.2028,
"step": 14950
},
{
"epoch": 1.6510731975784259,
"grad_norm": 0.17256326973438263,
"learning_rate": 0.00024009040849028341,
"loss": 2.2102,
"step": 15000
},
{
"epoch": 1.6565767749036873,
"grad_norm": 0.17648595571517944,
"learning_rate": 0.00023965325207329016,
"loss": 2.2094,
"step": 15050
},
{
"epoch": 1.6620803522289487,
"grad_norm": 0.17140266299247742,
"learning_rate": 0.0002392149077468511,
"loss": 2.1936,
"step": 15100
},
{
"epoch": 1.66758392955421,
"grad_norm": 0.16277456283569336,
"learning_rate": 0.00023877538131904743,
"loss": 2.1705,
"step": 15150
},
{
"epoch": 1.6730875068794715,
"grad_norm": 0.16912780702114105,
"learning_rate": 0.00023833467861362323,
"loss": 2.1834,
"step": 15200
},
{
"epoch": 1.6785910842047331,
"grad_norm": 0.1684638261795044,
"learning_rate": 0.00023789280546990824,
"loss": 2.2044,
"step": 15250
},
{
"epoch": 1.6840946615299945,
"grad_norm": 0.16793768107891083,
"learning_rate": 0.00023744976774274065,
"loss": 2.188,
"step": 15300
},
{
"epoch": 1.689598238855256,
"grad_norm": 0.15743038058280945,
"learning_rate": 0.00023700557130238944,
"loss": 2.183,
"step": 15350
},
{
"epoch": 1.6951018161805174,
"grad_norm": 0.1788504421710968,
"learning_rate": 0.0002365602220344765,
"loss": 2.1923,
"step": 15400
},
{
"epoch": 1.7006053935057788,
"grad_norm": 0.1614784598350525,
"learning_rate": 0.00023611372583989883,
"loss": 2.196,
"step": 15450
},
{
"epoch": 1.7061089708310402,
"grad_norm": 0.16596226394176483,
"learning_rate": 0.0002356660886347502,
"loss": 2.1952,
"step": 15500
},
{
"epoch": 1.7116125481563016,
"grad_norm": 0.18096621334552765,
"learning_rate": 0.00023521731635024274,
"loss": 2.1858,
"step": 15550
},
{
"epoch": 1.717116125481563,
"grad_norm": 0.1885058581829071,
"learning_rate": 0.0002347674149326286,
"loss": 2.1894,
"step": 15600
},
{
"epoch": 1.7226197028068244,
"grad_norm": 0.15273383259773254,
"learning_rate": 0.00023431639034312096,
"loss": 2.1896,
"step": 15650
},
{
"epoch": 1.7281232801320858,
"grad_norm": 0.17447490990161896,
"learning_rate": 0.00023386424855781495,
"loss": 2.1896,
"step": 15700
},
{
"epoch": 1.7336268574573472,
"grad_norm": 0.16734521090984344,
"learning_rate": 0.00023341099556760866,
"loss": 2.1886,
"step": 15750
},
{
"epoch": 1.7391304347826086,
"grad_norm": 0.17426900565624237,
"learning_rate": 0.0002329566373781238,
"loss": 2.1838,
"step": 15800
},
{
"epoch": 1.7446340121078703,
"grad_norm": 0.1640431135892868,
"learning_rate": 0.00023250118000962583,
"loss": 2.2026,
"step": 15850
},
{
"epoch": 1.7501375894331317,
"grad_norm": 0.16575609147548676,
"learning_rate": 0.00023204462949694447,
"loss": 2.1721,
"step": 15900
},
{
"epoch": 1.755641166758393,
"grad_norm": 0.17576967179775238,
"learning_rate": 0.00023158699188939371,
"loss": 2.1517,
"step": 15950
},
{
"epoch": 1.7611447440836545,
"grad_norm": 0.1661233901977539,
"learning_rate": 0.00023112827325069154,
"loss": 2.1594,
"step": 16000
},
{
"epoch": 1.766648321408916,
"grad_norm": 0.17433176934719086,
"learning_rate": 0.00023066847965887954,
"loss": 2.2069,
"step": 16050
},
{
"epoch": 1.7721518987341773,
"grad_norm": 0.164288729429245,
"learning_rate": 0.0002302076172062427,
"loss": 2.1945,
"step": 16100
},
{
"epoch": 1.7776554760594387,
"grad_norm": 0.16619046032428741,
"learning_rate": 0.00022974569199922827,
"loss": 2.1678,
"step": 16150
},
{
"epoch": 1.7831590533847002,
"grad_norm": 0.1572977751493454,
"learning_rate": 0.00022928271015836514,
"loss": 2.1834,
"step": 16200
},
{
"epoch": 1.7886626307099616,
"grad_norm": 0.15041351318359375,
"learning_rate": 0.0002288186778181826,
"loss": 2.1372,
"step": 16250
},
{
"epoch": 1.794166208035223,
"grad_norm": 0.17850619554519653,
"learning_rate": 0.00022835360112712913,
"loss": 2.2064,
"step": 16300
},
{
"epoch": 1.7996697853604844,
"grad_norm": 0.1657785028219223,
"learning_rate": 0.00022788748624749083,
"loss": 2.1948,
"step": 16350
},
{
"epoch": 1.8051733626857458,
"grad_norm": 0.16514721512794495,
"learning_rate": 0.00022742033935530995,
"loss": 2.209,
"step": 16400
},
{
"epoch": 1.8106769400110072,
"grad_norm": 0.1726425439119339,
"learning_rate": 0.00022695216664030297,
"loss": 2.1992,
"step": 16450
},
{
"epoch": 1.8161805173362686,
"grad_norm": 0.1753402054309845,
"learning_rate": 0.00022648297430577835,
"loss": 2.1929,
"step": 16500
},
{
"epoch": 1.82168409466153,
"grad_norm": 0.18259774148464203,
"learning_rate": 0.00022601276856855482,
"loss": 2.2083,
"step": 16550
},
{
"epoch": 1.8271876719867914,
"grad_norm": 0.16030113399028778,
"learning_rate": 0.0002255415556588786,
"loss": 2.1571,
"step": 16600
},
{
"epoch": 1.8326912493120529,
"grad_norm": 0.1431453675031662,
"learning_rate": 0.000225069341820341,
"loss": 2.149,
"step": 16650
},
{
"epoch": 1.8381948266373143,
"grad_norm": 0.1664692461490631,
"learning_rate": 0.0002245961333097956,
"loss": 2.1864,
"step": 16700
},
{
"epoch": 1.8436984039625757,
"grad_norm": 0.17130005359649658,
"learning_rate": 0.00022412193639727564,
"loss": 2.18,
"step": 16750
},
{
"epoch": 1.849201981287837,
"grad_norm": 0.15278129279613495,
"learning_rate": 0.00022364675736591058,
"loss": 2.1648,
"step": 16800
},
{
"epoch": 1.8547055586130985,
"grad_norm": 0.176371768116951,
"learning_rate": 0.000223170602511843,
"loss": 2.1582,
"step": 16850
},
{
"epoch": 1.86020913593836,
"grad_norm": 0.15902136266231537,
"learning_rate": 0.0002226934781441452,
"loss": 2.183,
"step": 16900
},
{
"epoch": 1.8657127132636213,
"grad_norm": 0.14821839332580566,
"learning_rate": 0.00022221539058473553,
"loss": 2.172,
"step": 16950
},
{
"epoch": 1.8712162905888827,
"grad_norm": 0.15611177682876587,
"learning_rate": 0.00022173634616829481,
"loss": 2.1715,
"step": 17000
},
{
"epoch": 1.8767198679141441,
"grad_norm": 0.1633993536233902,
"learning_rate": 0.00022125635124218207,
"loss": 2.1961,
"step": 17050
},
{
"epoch": 1.8822234452394055,
"grad_norm": 0.16087235510349274,
"learning_rate": 0.00022077541216635076,
"loss": 2.1963,
"step": 17100
},
{
"epoch": 1.887727022564667,
"grad_norm": 0.16733358800411224,
"learning_rate": 0.00022029353531326432,
"loss": 2.1605,
"step": 17150
},
{
"epoch": 1.8932305998899284,
"grad_norm": 0.16267576813697815,
"learning_rate": 0.00021981072706781185,
"loss": 2.1554,
"step": 17200
},
{
"epoch": 1.8987341772151898,
"grad_norm": 0.16028568148612976,
"learning_rate": 0.0002193269938272232,
"loss": 2.1793,
"step": 17250
},
{
"epoch": 1.9042377545404512,
"grad_norm": 0.15961337089538574,
"learning_rate": 0.00021884234200098473,
"loss": 2.1673,
"step": 17300
},
{
"epoch": 1.9097413318657126,
"grad_norm": 0.1637711226940155,
"learning_rate": 0.00021835677801075397,
"loss": 2.1676,
"step": 17350
},
{
"epoch": 1.915244909190974,
"grad_norm": 0.16490189731121063,
"learning_rate": 0.00021787030829027468,
"loss": 2.2119,
"step": 17400
},
{
"epoch": 1.9207484865162354,
"grad_norm": 0.17354600131511688,
"learning_rate": 0.00021738293928529154,
"loss": 2.2001,
"step": 17450
},
{
"epoch": 1.9262520638414968,
"grad_norm": 0.16645126044750214,
"learning_rate": 0.00021689467745346483,
"loss": 2.1915,
"step": 17500
},
{
"epoch": 1.9317556411667582,
"grad_norm": 0.16187366843223572,
"learning_rate": 0.0002164055292642849,
"loss": 2.1829,
"step": 17550
},
{
"epoch": 1.9372592184920197,
"grad_norm": 0.17859427630901337,
"learning_rate": 0.00021591550119898616,
"loss": 2.1203,
"step": 17600
},
{
"epoch": 1.9427627958172813,
"grad_norm": 0.15773439407348633,
"learning_rate": 0.00021542459975046164,
"loss": 2.1535,
"step": 17650
},
{
"epoch": 1.9482663731425427,
"grad_norm": 0.1696433126926422,
"learning_rate": 0.00021493283142317668,
"loss": 2.1271,
"step": 17700
},
{
"epoch": 1.953769950467804,
"grad_norm": 0.16615457832813263,
"learning_rate": 0.00021444020273308267,
"loss": 2.1688,
"step": 17750
},
{
"epoch": 1.9592735277930655,
"grad_norm": 0.16561925411224365,
"learning_rate": 0.00021394672020753102,
"loss": 2.1228,
"step": 17800
},
{
"epoch": 1.964777105118327,
"grad_norm": 0.1655699759721756,
"learning_rate": 0.00021345239038518638,
"loss": 2.1614,
"step": 17850
},
{
"epoch": 1.9702806824435883,
"grad_norm": 0.16106578707695007,
"learning_rate": 0.0002129572198159401,
"loss": 2.1471,
"step": 17900
},
{
"epoch": 1.9757842597688497,
"grad_norm": 0.1635727882385254,
"learning_rate": 0.00021246121506082367,
"loss": 2.184,
"step": 17950
},
{
"epoch": 1.9812878370941112,
"grad_norm": 0.15699493885040283,
"learning_rate": 0.0002119643826919213,
"loss": 2.1602,
"step": 18000
},
{
"epoch": 1.9867914144193726,
"grad_norm": 0.15784280002117157,
"learning_rate": 0.00021146672929228327,
"loss": 2.1781,
"step": 18050
},
{
"epoch": 1.992294991744634,
"grad_norm": 0.15157456696033478,
"learning_rate": 0.00021096826145583852,
"loss": 2.1532,
"step": 18100
},
{
"epoch": 1.9977985690698954,
"grad_norm": 0.15633012354373932,
"learning_rate": 0.0002104689857873074,
"loss": 2.1676,
"step": 18150
},
{
"epoch": 2.003302146395157,
"grad_norm": 0.16586118936538696,
"learning_rate": 0.00020996890890211386,
"loss": 2.1335,
"step": 18200
},
{
"epoch": 2.0088057237204184,
"grad_norm": 0.1573173552751541,
"learning_rate": 0.00020946803742629826,
"loss": 2.11,
"step": 18250
},
{
"epoch": 2.01430930104568,
"grad_norm": 0.18040034174919128,
"learning_rate": 0.00020896637799642915,
"loss": 2.1128,
"step": 18300
},
{
"epoch": 2.0198128783709413,
"grad_norm": 0.1664518266916275,
"learning_rate": 0.00020846393725951556,
"loss": 2.1384,
"step": 18350
},
{
"epoch": 2.0253164556962027,
"grad_norm": 0.17078042030334473,
"learning_rate": 0.0002079607218729188,
"loss": 2.0909,
"step": 18400
},
{
"epoch": 2.030820033021464,
"grad_norm": 0.15759077668190002,
"learning_rate": 0.00020745673850426446,
"loss": 2.1015,
"step": 18450
},
{
"epoch": 2.0363236103467255,
"grad_norm": 0.17311988770961761,
"learning_rate": 0.00020695199383135386,
"loss": 2.1436,
"step": 18500
},
{
"epoch": 2.041827187671987,
"grad_norm": 0.1612362265586853,
"learning_rate": 0.00020644649454207553,
"loss": 2.0868,
"step": 18550
},
{
"epoch": 2.0473307649972483,
"grad_norm": 0.16166241466999054,
"learning_rate": 0.0002059402473343169,
"loss": 2.0683,
"step": 18600
},
{
"epoch": 2.0528343423225097,
"grad_norm": 0.17205455899238586,
"learning_rate": 0.00020543325891587516,
"loss": 2.0618,
"step": 18650
},
{
"epoch": 2.058337919647771,
"grad_norm": 0.17534597218036652,
"learning_rate": 0.00020492553600436872,
"loss": 2.1173,
"step": 18700
},
{
"epoch": 2.0638414969730325,
"grad_norm": 0.15081949532032013,
"learning_rate": 0.00020441708532714797,
"loss": 2.0991,
"step": 18750
},
{
"epoch": 2.069345074298294,
"grad_norm": 0.15837198495864868,
"learning_rate": 0.00020390791362120625,
"loss": 2.1332,
"step": 18800
},
{
"epoch": 2.0748486516235554,
"grad_norm": 0.1598348468542099,
"learning_rate": 0.00020339802763309054,
"loss": 2.1164,
"step": 18850
},
{
"epoch": 2.0803522289488168,
"grad_norm": 0.1536133736371994,
"learning_rate": 0.0002028874341188122,
"loss": 2.1169,
"step": 18900
},
{
"epoch": 2.085855806274078,
"grad_norm": 0.1654282808303833,
"learning_rate": 0.00020237613984375711,
"loss": 2.1253,
"step": 18950
},
{
"epoch": 2.0913593835993396,
"grad_norm": 0.16015826165676117,
"learning_rate": 0.00020186415158259652,
"loss": 2.1196,
"step": 19000
},
{
"epoch": 2.096862960924601,
"grad_norm": 0.16205786168575287,
"learning_rate": 0.00020135147611919686,
"loss": 2.0872,
"step": 19050
},
{
"epoch": 2.1023665382498624,
"grad_norm": 0.16814254224300385,
"learning_rate": 0.00020083812024653013,
"loss": 2.1569,
"step": 19100
},
{
"epoch": 2.107870115575124,
"grad_norm": 0.15433181822299957,
"learning_rate": 0.00020032409076658364,
"loss": 2.0967,
"step": 19150
},
{
"epoch": 2.1133736929003852,
"grad_norm": 0.16261817514896393,
"learning_rate": 0.00019980939449027016,
"loss": 2.0938,
"step": 19200
},
{
"epoch": 2.1188772702256466,
"grad_norm": 0.17877838015556335,
"learning_rate": 0.00019929403823733748,
"loss": 2.1399,
"step": 19250
},
{
"epoch": 2.124380847550908,
"grad_norm": 0.178861603140831,
"learning_rate": 0.00019877802883627808,
"loss": 2.1362,
"step": 19300
},
{
"epoch": 2.1298844248761695,
"grad_norm": 0.1621016561985016,
"learning_rate": 0.00019826137312423873,
"loss": 2.1089,
"step": 19350
},
{
"epoch": 2.135388002201431,
"grad_norm": 0.17772269248962402,
"learning_rate": 0.00019774407794692986,
"loss": 2.1278,
"step": 19400
},
{
"epoch": 2.1408915795266923,
"grad_norm": 0.16339927911758423,
"learning_rate": 0.00019722615015853483,
"loss": 2.1244,
"step": 19450
},
{
"epoch": 2.1463951568519537,
"grad_norm": 0.16879414021968842,
"learning_rate": 0.0001967075966216191,
"loss": 2.106,
"step": 19500
},
{
"epoch": 2.151898734177215,
"grad_norm": 0.15002557635307312,
"learning_rate": 0.00019618842420703923,
"loss": 2.1007,
"step": 19550
},
{
"epoch": 2.1574023115024765,
"grad_norm": 0.17130738496780396,
"learning_rate": 0.0001956686397938522,
"loss": 2.1104,
"step": 19600
},
{
"epoch": 2.162905888827738,
"grad_norm": 0.16796913743019104,
"learning_rate": 0.00019514825026922377,
"loss": 2.1278,
"step": 19650
},
{
"epoch": 2.1684094661529993,
"grad_norm": 0.19111864268779755,
"learning_rate": 0.0001946272625283375,
"loss": 2.1402,
"step": 19700
},
{
"epoch": 2.1739130434782608,
"grad_norm": 0.17543087899684906,
"learning_rate": 0.00019410568347430344,
"loss": 2.0984,
"step": 19750
},
{
"epoch": 2.179416620803522,
"grad_norm": 0.15795323252677917,
"learning_rate": 0.00019358352001806642,
"loss": 2.1125,
"step": 19800
},
{
"epoch": 2.1849201981287836,
"grad_norm": 0.15581749379634857,
"learning_rate": 0.0001930607790783147,
"loss": 2.119,
"step": 19850
},
{
"epoch": 2.190423775454045,
"grad_norm": 0.17086704075336456,
"learning_rate": 0.00019253746758138822,
"loss": 2.1029,
"step": 19900
},
{
"epoch": 2.1959273527793064,
"grad_norm": 0.15870486199855804,
"learning_rate": 0.00019201359246118682,
"loss": 2.1133,
"step": 19950
},
{
"epoch": 2.201430930104568,
"grad_norm": 0.16840116679668427,
"learning_rate": 0.0001914891606590784,
"loss": 2.1196,
"step": 20000
},
{
"epoch": 2.206934507429829,
"grad_norm": 0.1643972396850586,
"learning_rate": 0.00019096417912380686,
"loss": 2.1173,
"step": 20050
},
{
"epoch": 2.2124380847550906,
"grad_norm": 0.1598280966281891,
"learning_rate": 0.0001904386548114001,
"loss": 2.1338,
"step": 20100
},
{
"epoch": 2.217941662080352,
"grad_norm": 0.16909152269363403,
"learning_rate": 0.0001899125946850779,
"loss": 2.0987,
"step": 20150
},
{
"epoch": 2.2234452394056134,
"grad_norm": 0.15597747266292572,
"learning_rate": 0.00018938600571515962,
"loss": 2.1295,
"step": 20200
},
{
"epoch": 2.228948816730875,
"grad_norm": 0.16759732365608215,
"learning_rate": 0.00018885889487897162,
"loss": 2.1374,
"step": 20250
},
{
"epoch": 2.2344523940561363,
"grad_norm": 0.15526509284973145,
"learning_rate": 0.00018833126916075526,
"loss": 2.1287,
"step": 20300
},
{
"epoch": 2.2399559713813977,
"grad_norm": 0.16562257707118988,
"learning_rate": 0.00018780313555157393,
"loss": 2.126,
"step": 20350
},
{
"epoch": 2.2454595487066595,
"grad_norm": 0.17471948266029358,
"learning_rate": 0.00018727450104922075,
"loss": 2.1175,
"step": 20400
},
{
"epoch": 2.2509631260319205,
"grad_norm": 0.16391079127788544,
"learning_rate": 0.00018674537265812548,
"loss": 2.0965,
"step": 20450
},
{
"epoch": 2.2564667033571824,
"grad_norm": 0.1693831831216812,
"learning_rate": 0.00018621575738926217,
"loss": 2.1124,
"step": 20500
},
{
"epoch": 2.2619702806824438,
"grad_norm": 0.16163313388824463,
"learning_rate": 0.00018568566226005587,
"loss": 2.1034,
"step": 20550
},
{
"epoch": 2.267473858007705,
"grad_norm": 0.1676999181509018,
"learning_rate": 0.00018515509429428988,
"loss": 2.0991,
"step": 20600
},
{
"epoch": 2.2729774353329666,
"grad_norm": 0.16531667113304138,
"learning_rate": 0.00018462406052201252,
"loss": 2.1183,
"step": 20650
},
{
"epoch": 2.278481012658228,
"grad_norm": 0.16208726167678833,
"learning_rate": 0.00018409256797944412,
"loss": 2.0981,
"step": 20700
},
{
"epoch": 2.2839845899834894,
"grad_norm": 0.1707569658756256,
"learning_rate": 0.0001835606237088837,
"loss": 2.0926,
"step": 20750
},
{
"epoch": 2.289488167308751,
"grad_norm": 0.16191871464252472,
"learning_rate": 0.00018302823475861582,
"loss": 2.0895,
"step": 20800
},
{
"epoch": 2.2949917446340122,
"grad_norm": 0.154753640294075,
"learning_rate": 0.0001824954081828168,
"loss": 2.0972,
"step": 20850
},
{
"epoch": 2.3004953219592736,
"grad_norm": 0.16868236660957336,
"learning_rate": 0.00018196215104146187,
"loss": 2.1009,
"step": 20900
},
{
"epoch": 2.305998899284535,
"grad_norm": 0.1602269858121872,
"learning_rate": 0.00018142847040023097,
"loss": 2.1063,
"step": 20950
},
{
"epoch": 2.3115024766097965,
"grad_norm": 0.17220260202884674,
"learning_rate": 0.00018089437333041562,
"loss": 2.1049,
"step": 21000
},
{
"epoch": 2.317006053935058,
"grad_norm": 0.14858907461166382,
"learning_rate": 0.0001803598669088249,
"loss": 2.1129,
"step": 21050
},
{
"epoch": 2.3225096312603193,
"grad_norm": 0.17586296796798706,
"learning_rate": 0.0001798249582176919,
"loss": 2.0961,
"step": 21100
},
{
"epoch": 2.3280132085855807,
"grad_norm": 0.14931316673755646,
"learning_rate": 0.00017928965434457983,
"loss": 2.1108,
"step": 21150
},
{
"epoch": 2.333516785910842,
"grad_norm": 0.17148137092590332,
"learning_rate": 0.00017875396238228798,
"loss": 2.1029,
"step": 21200
},
{
"epoch": 2.3390203632361035,
"grad_norm": 0.15888847410678864,
"learning_rate": 0.00017821788942875793,
"loss": 2.1258,
"step": 21250
},
{
"epoch": 2.344523940561365,
"grad_norm": 0.1749711036682129,
"learning_rate": 0.0001776814425869793,
"loss": 2.1176,
"step": 21300
},
{
"epoch": 2.3500275178866263,
"grad_norm": 0.16511616110801697,
"learning_rate": 0.0001771446289648959,
"loss": 2.0992,
"step": 21350
},
{
"epoch": 2.3555310952118877,
"grad_norm": 0.16869331896305084,
"learning_rate": 0.00017660745567531124,
"loss": 2.1274,
"step": 21400
},
{
"epoch": 2.361034672537149,
"grad_norm": 0.16024701297283173,
"learning_rate": 0.00017606992983579457,
"loss": 2.1379,
"step": 21450
},
{
"epoch": 2.3665382498624106,
"grad_norm": 0.1646272987127304,
"learning_rate": 0.00017553205856858637,
"loss": 2.1365,
"step": 21500
},
{
"epoch": 2.372041827187672,
"grad_norm": 0.1765459179878235,
"learning_rate": 0.00017499384900050404,
"loss": 2.1125,
"step": 21550
},
{
"epoch": 2.3775454045129334,
"grad_norm": 0.15617190301418304,
"learning_rate": 0.0001744553082628475,
"loss": 2.0985,
"step": 21600
},
{
"epoch": 2.383048981838195,
"grad_norm": 0.14687784016132355,
"learning_rate": 0.00017391644349130467,
"loss": 2.0771,
"step": 21650
},
{
"epoch": 2.388552559163456,
"grad_norm": 0.16840553283691406,
"learning_rate": 0.00017337726182585698,
"loss": 2.0909,
"step": 21700
},
{
"epoch": 2.3940561364887176,
"grad_norm": 0.16857478022575378,
"learning_rate": 0.00017283777041068472,
"loss": 2.0968,
"step": 21750
},
{
"epoch": 2.399559713813979,
"grad_norm": 0.1799498200416565,
"learning_rate": 0.00017229797639407233,
"loss": 2.1023,
"step": 21800
},
{
"epoch": 2.4050632911392404,
"grad_norm": 0.15700192749500275,
"learning_rate": 0.00017175788692831367,
"loss": 2.0986,
"step": 21850
},
{
"epoch": 2.410566868464502,
"grad_norm": 0.16778048872947693,
"learning_rate": 0.00017121750916961742,
"loss": 2.0914,
"step": 21900
},
{
"epoch": 2.4160704457897633,
"grad_norm": 0.15119898319244385,
"learning_rate": 0.0001706768502780122,
"loss": 2.0857,
"step": 21950
},
{
"epoch": 2.4215740231150247,
"grad_norm": 0.17003406584262848,
"learning_rate": 0.0001701359174172515,
"loss": 2.1366,
"step": 22000
},
{
"epoch": 2.427077600440286,
"grad_norm": 0.17182987928390503,
"learning_rate": 0.000169594717754719,
"loss": 2.1126,
"step": 22050
},
{
"epoch": 2.4325811777655475,
"grad_norm": 0.16026781499385834,
"learning_rate": 0.00016905325846133354,
"loss": 2.076,
"step": 22100
},
{
"epoch": 2.438084755090809,
"grad_norm": 0.16470612585544586,
"learning_rate": 0.00016851154671145403,
"loss": 2.0907,
"step": 22150
},
{
"epoch": 2.4435883324160703,
"grad_norm": 0.14621587097644806,
"learning_rate": 0.0001679695896827844,
"loss": 2.1073,
"step": 22200
},
{
"epoch": 2.4490919097413317,
"grad_norm": 0.1563323438167572,
"learning_rate": 0.00016742739455627862,
"loss": 2.0757,
"step": 22250
},
{
"epoch": 2.454595487066593,
"grad_norm": 0.17763760685920715,
"learning_rate": 0.0001668849685160455,
"loss": 2.0705,
"step": 22300
},
{
"epoch": 2.4600990643918546,
"grad_norm": 0.16651757061481476,
"learning_rate": 0.00016634231874925335,
"loss": 2.0881,
"step": 22350
},
{
"epoch": 2.465602641717116,
"grad_norm": 0.16752833127975464,
"learning_rate": 0.0001657994524460349,
"loss": 2.0639,
"step": 22400
},
{
"epoch": 2.4711062190423774,
"grad_norm": 0.14574387669563293,
"learning_rate": 0.000165256376799392,
"loss": 2.087,
"step": 22450
},
{
"epoch": 2.476609796367639,
"grad_norm": 0.15595164895057678,
"learning_rate": 0.00016471309900510043,
"loss": 2.0965,
"step": 22500
},
{
"epoch": 2.4821133736929,
"grad_norm": 0.17218752205371857,
"learning_rate": 0.0001641696262616142,
"loss": 2.0943,
"step": 22550
},
{
"epoch": 2.487616951018162,
"grad_norm": 0.1563566029071808,
"learning_rate": 0.00016362596576997057,
"loss": 2.0946,
"step": 22600
},
{
"epoch": 2.4931205283434235,
"grad_norm": 0.16190436482429504,
"learning_rate": 0.00016308212473369453,
"loss": 2.0924,
"step": 22650
},
{
"epoch": 2.498624105668685,
"grad_norm": 0.16890835762023926,
"learning_rate": 0.00016253811035870313,
"loss": 2.0954,
"step": 22700
},
{
"epoch": 2.5041276829939463,
"grad_norm": 0.16371208429336548,
"learning_rate": 0.00016199392985321017,
"loss": 2.0954,
"step": 22750
},
{
"epoch": 2.5096312603192077,
"grad_norm": 0.15898428857326508,
"learning_rate": 0.00016144959042763084,
"loss": 2.0766,
"step": 22800
},
{
"epoch": 2.515134837644469,
"grad_norm": 0.15516646206378937,
"learning_rate": 0.00016090509929448594,
"loss": 2.0891,
"step": 22850
},
{
"epoch": 2.5206384149697305,
"grad_norm": 0.15007217228412628,
"learning_rate": 0.00016036046366830635,
"loss": 2.078,
"step": 22900
},
{
"epoch": 2.526141992294992,
"grad_norm": 0.17989104986190796,
"learning_rate": 0.00015981569076553765,
"loss": 2.0888,
"step": 22950
},
{
"epoch": 2.5316455696202533,
"grad_norm": 0.16754071414470673,
"learning_rate": 0.00015927078780444403,
"loss": 2.0882,
"step": 23000
},
{
"epoch": 2.5371491469455147,
"grad_norm": 0.15897449851036072,
"learning_rate": 0.0001587257620050134,
"loss": 2.0933,
"step": 23050
},
{
"epoch": 2.542652724270776,
"grad_norm": 0.16594423353672028,
"learning_rate": 0.00015818062058886082,
"loss": 2.0968,
"step": 23100
},
{
"epoch": 2.5481563015960376,
"grad_norm": 0.1700371652841568,
"learning_rate": 0.00015763537077913365,
"loss": 2.096,
"step": 23150
},
{
"epoch": 2.553659878921299,
"grad_norm": 0.16013766825199127,
"learning_rate": 0.00015709001980041527,
"loss": 2.1134,
"step": 23200
},
{
"epoch": 2.5591634562465604,
"grad_norm": 0.1601455807685852,
"learning_rate": 0.00015654457487862961,
"loss": 2.0983,
"step": 23250
},
{
"epoch": 2.564667033571822,
"grad_norm": 0.16373537480831146,
"learning_rate": 0.00015599904324094534,
"loss": 2.0866,
"step": 23300
},
{
"epoch": 2.570170610897083,
"grad_norm": 0.15919940173625946,
"learning_rate": 0.0001554534321156801,
"loss": 2.1147,
"step": 23350
},
{
"epoch": 2.5756741882223446,
"grad_norm": 0.1523171216249466,
"learning_rate": 0.0001549077487322048,
"loss": 2.0887,
"step": 23400
},
{
"epoch": 2.581177765547606,
"grad_norm": 0.16425256431102753,
"learning_rate": 0.00015436200032084774,
"loss": 2.0918,
"step": 23450
},
{
"epoch": 2.5866813428728674,
"grad_norm": 0.16046775877475739,
"learning_rate": 0.0001538161941127988,
"loss": 2.0816,
"step": 23500
},
{
"epoch": 2.592184920198129,
"grad_norm": 0.1673419028520584,
"learning_rate": 0.00015327033734001375,
"loss": 2.1186,
"step": 23550
},
{
"epoch": 2.5976884975233903,
"grad_norm": 0.15573158860206604,
"learning_rate": 0.00015272443723511825,
"loss": 2.0721,
"step": 23600
},
{
"epoch": 2.6031920748486517,
"grad_norm": 0.16536468267440796,
"learning_rate": 0.00015217850103131227,
"loss": 2.0991,
"step": 23650
},
{
"epoch": 2.608695652173913,
"grad_norm": 0.15281961858272552,
"learning_rate": 0.00015163253596227393,
"loss": 2.0713,
"step": 23700
},
{
"epoch": 2.6141992294991745,
"grad_norm": 0.1582518070936203,
"learning_rate": 0.0001510865492620638,
"loss": 2.0923,
"step": 23750
},
{
"epoch": 2.619702806824436,
"grad_norm": 0.1581014096736908,
"learning_rate": 0.0001505405481650293,
"loss": 2.0895,
"step": 23800
},
{
"epoch": 2.6252063841496973,
"grad_norm": 0.16346633434295654,
"learning_rate": 0.00014999453990570839,
"loss": 2.1011,
"step": 23850
},
{
"epoch": 2.6307099614749587,
"grad_norm": 0.16536889970302582,
"learning_rate": 0.00014944853171873387,
"loss": 2.113,
"step": 23900
},
{
"epoch": 2.63621353880022,
"grad_norm": 0.170758455991745,
"learning_rate": 0.00014890253083873784,
"loss": 2.1124,
"step": 23950
},
{
"epoch": 2.6417171161254815,
"grad_norm": 0.15669451653957367,
"learning_rate": 0.00014835654450025542,
"loss": 2.0894,
"step": 24000
},
{
"epoch": 2.647220693450743,
"grad_norm": 0.16937313973903656,
"learning_rate": 0.00014781057993762902,
"loss": 2.127,
"step": 24050
},
{
"epoch": 2.6527242707760044,
"grad_norm": 0.15754981338977814,
"learning_rate": 0.0001472646443849126,
"loss": 2.0959,
"step": 24100
},
{
"epoch": 2.6582278481012658,
"grad_norm": 0.16420376300811768,
"learning_rate": 0.00014671874507577574,
"loss": 2.0742,
"step": 24150
},
{
"epoch": 2.663731425426527,
"grad_norm": 0.16400571167469025,
"learning_rate": 0.00014617288924340775,
"loss": 2.0992,
"step": 24200
},
{
"epoch": 2.6692350027517886,
"grad_norm": 0.1658066064119339,
"learning_rate": 0.00014562708412042193,
"loss": 2.0925,
"step": 24250
},
{
"epoch": 2.67473858007705,
"grad_norm": 0.16872134804725647,
"learning_rate": 0.00014508133693875953,
"loss": 2.0899,
"step": 24300
},
{
"epoch": 2.6802421574023114,
"grad_norm": 0.17002440989017487,
"learning_rate": 0.00014453565492959422,
"loss": 2.1252,
"step": 24350
},
{
"epoch": 2.685745734727573,
"grad_norm": 0.16417178511619568,
"learning_rate": 0.00014399004532323614,
"loss": 2.112,
"step": 24400
},
{
"epoch": 2.6912493120528342,
"grad_norm": 0.172217458486557,
"learning_rate": 0.000143444515349036,
"loss": 2.0784,
"step": 24450
},
{
"epoch": 2.6967528893780957,
"grad_norm": 0.16687516868114471,
"learning_rate": 0.0001428990722352894,
"loss": 2.0775,
"step": 24500
},
{
"epoch": 2.702256466703357,
"grad_norm": 0.162861630320549,
"learning_rate": 0.00014235372320914108,
"loss": 2.0928,
"step": 24550
},
{
"epoch": 2.7077600440286185,
"grad_norm": 0.1640874743461609,
"learning_rate": 0.00014180847549648913,
"loss": 2.1071,
"step": 24600
},
{
"epoch": 2.71326362135388,
"grad_norm": 0.1654275506734848,
"learning_rate": 0.0001412633363218891,
"loss": 2.0615,
"step": 24650
},
{
"epoch": 2.7187671986791413,
"grad_norm": 0.14277833700180054,
"learning_rate": 0.00014071831290845857,
"loss": 2.1078,
"step": 24700
},
{
"epoch": 2.7242707760044027,
"grad_norm": 0.1705639362335205,
"learning_rate": 0.00014017341247778123,
"loss": 2.0858,
"step": 24750
},
{
"epoch": 2.729774353329664,
"grad_norm": 0.16390259563922882,
"learning_rate": 0.0001396286422498112,
"loss": 2.0508,
"step": 24800
},
{
"epoch": 2.7352779306549255,
"grad_norm": 0.1526353508234024,
"learning_rate": 0.00013908400944277738,
"loss": 2.0784,
"step": 24850
},
{
"epoch": 2.740781507980187,
"grad_norm": 0.1581793874502182,
"learning_rate": 0.0001385395212730879,
"loss": 2.0596,
"step": 24900
},
{
"epoch": 2.7462850853054483,
"grad_norm": 0.1692277193069458,
"learning_rate": 0.00013799518495523445,
"loss": 2.0838,
"step": 24950
},
{
"epoch": 2.7517886626307098,
"grad_norm": 0.1614445298910141,
"learning_rate": 0.0001374510077016967,
"loss": 2.0928,
"step": 25000
},
{
"epoch": 2.757292239955971,
"grad_norm": 0.15747620165348053,
"learning_rate": 0.00013690699672284657,
"loss": 2.0686,
"step": 25050
},
{
"epoch": 2.7627958172812326,
"grad_norm": 0.17565257847309113,
"learning_rate": 0.00013636315922685294,
"loss": 2.0971,
"step": 25100
},
{
"epoch": 2.768299394606494,
"grad_norm": 0.16367433965206146,
"learning_rate": 0.0001358195024195861,
"loss": 2.1052,
"step": 25150
},
{
"epoch": 2.7738029719317554,
"grad_norm": 0.1564227044582367,
"learning_rate": 0.00013527603350452206,
"loss": 2.0902,
"step": 25200
},
{
"epoch": 2.779306549257017,
"grad_norm": 0.16555672883987427,
"learning_rate": 0.0001347327596826473,
"loss": 2.0683,
"step": 25250
},
{
"epoch": 2.7848101265822782,
"grad_norm": 0.16134041547775269,
"learning_rate": 0.00013418968815236346,
"loss": 2.0627,
"step": 25300
},
{
"epoch": 2.7903137039075396,
"grad_norm": 0.17097671329975128,
"learning_rate": 0.0001336468261093915,
"loss": 2.0724,
"step": 25350
},
{
"epoch": 2.795817281232801,
"grad_norm": 0.1656900942325592,
"learning_rate": 0.00013310418074667698,
"loss": 2.1266,
"step": 25400
},
{
"epoch": 2.801320858558063,
"grad_norm": 0.1745171695947647,
"learning_rate": 0.00013256175925429415,
"loss": 2.0983,
"step": 25450
},
{
"epoch": 2.8068244358833243,
"grad_norm": 0.15649184584617615,
"learning_rate": 0.0001320195688193512,
"loss": 2.0672,
"step": 25500
},
{
"epoch": 2.8123280132085857,
"grad_norm": 0.16284936666488647,
"learning_rate": 0.00013147761662589473,
"loss": 2.0865,
"step": 25550
},
{
"epoch": 2.817831590533847,
"grad_norm": 0.15236727893352509,
"learning_rate": 0.00013093590985481456,
"loss": 2.0595,
"step": 25600
},
{
"epoch": 2.8233351678591085,
"grad_norm": 0.166220560669899,
"learning_rate": 0.00013039445568374872,
"loss": 2.0478,
"step": 25650
},
{
"epoch": 2.82883874518437,
"grad_norm": 0.16826435923576355,
"learning_rate": 0.0001298532612869883,
"loss": 2.0756,
"step": 25700
},
{
"epoch": 2.8343423225096314,
"grad_norm": 0.16447168588638306,
"learning_rate": 0.0001293123338353823,
"loss": 2.0742,
"step": 25750
},
{
"epoch": 2.8398458998348928,
"grad_norm": 0.157373309135437,
"learning_rate": 0.00012877168049624268,
"loss": 2.0728,
"step": 25800
},
{
"epoch": 2.845349477160154,
"grad_norm": 0.1642802357673645,
"learning_rate": 0.00012823130843324948,
"loss": 2.0643,
"step": 25850
},
{
"epoch": 2.8508530544854156,
"grad_norm": 0.15527865290641785,
"learning_rate": 0.00012769122480635583,
"loss": 2.1088,
"step": 25900
},
{
"epoch": 2.856356631810677,
"grad_norm": 0.175029918551445,
"learning_rate": 0.00012715143677169297,
"loss": 2.0764,
"step": 25950
},
{
"epoch": 2.8618602091359384,
"grad_norm": 0.17664563655853271,
"learning_rate": 0.0001266119514814755,
"loss": 2.0527,
"step": 26000
},
{
"epoch": 2.8673637864612,
"grad_norm": 0.16839756071567535,
"learning_rate": 0.00012607277608390677,
"loss": 2.0708,
"step": 26050
},
{
"epoch": 2.8728673637864612,
"grad_norm": 0.162934809923172,
"learning_rate": 0.00012553391772308407,
"loss": 2.0984,
"step": 26100
},
{
"epoch": 2.8783709411117226,
"grad_norm": 0.1836494356393814,
"learning_rate": 0.0001249953835389037,
"loss": 2.1118,
"step": 26150
},
{
"epoch": 2.883874518436984,
"grad_norm": 0.15794949233531952,
"learning_rate": 0.00012445718066696687,
"loss": 2.0541,
"step": 26200
},
{
"epoch": 2.8893780957622455,
"grad_norm": 0.16560156643390656,
"learning_rate": 0.00012391931623848476,
"loss": 2.1111,
"step": 26250
},
{
"epoch": 2.894881673087507,
"grad_norm": 0.16835126280784607,
"learning_rate": 0.0001233817973801842,
"loss": 2.0635,
"step": 26300
},
{
"epoch": 2.9003852504127683,
"grad_norm": 0.16151325404644012,
"learning_rate": 0.0001228446312142131,
"loss": 2.0949,
"step": 26350
},
{
"epoch": 2.9058888277380297,
"grad_norm": 0.16751372814178467,
"learning_rate": 0.00012230782485804633,
"loss": 2.0896,
"step": 26400
},
{
"epoch": 2.911392405063291,
"grad_norm": 0.18384750187397003,
"learning_rate": 0.00012177138542439122,
"loss": 2.047,
"step": 26450
},
{
"epoch": 2.9168959823885525,
"grad_norm": 0.15618419647216797,
"learning_rate": 0.00012123532002109328,
"loss": 2.0787,
"step": 26500
},
{
"epoch": 2.922399559713814,
"grad_norm": 0.15803013741970062,
"learning_rate": 0.00012069963575104223,
"loss": 2.0735,
"step": 26550
},
{
"epoch": 2.9279031370390753,
"grad_norm": 0.1603049784898758,
"learning_rate": 0.00012016433971207757,
"loss": 2.0802,
"step": 26600
},
{
"epoch": 2.9334067143643368,
"grad_norm": 0.1752861887216568,
"learning_rate": 0.0001196294389968949,
"loss": 2.0969,
"step": 26650
},
{
"epoch": 2.938910291689598,
"grad_norm": 0.16382624208927155,
"learning_rate": 0.00011909494069295175,
"loss": 2.0933,
"step": 26700
},
{
"epoch": 2.9444138690148596,
"grad_norm": 0.16802163422107697,
"learning_rate": 0.00011856085188237357,
"loss": 2.1042,
"step": 26750
},
{
"epoch": 2.949917446340121,
"grad_norm": 0.15907612442970276,
"learning_rate": 0.00011802717964186006,
"loss": 2.0575,
"step": 26800
},
{
"epoch": 2.9554210236653824,
"grad_norm": 0.16981548070907593,
"learning_rate": 0.00011749393104259143,
"loss": 2.0951,
"step": 26850
},
{
"epoch": 2.960924600990644,
"grad_norm": 0.15169958770275116,
"learning_rate": 0.0001169611131501345,
"loss": 2.0778,
"step": 26900
},
{
"epoch": 2.966428178315905,
"grad_norm": 0.1601976752281189,
"learning_rate": 0.00011642873302434926,
"loss": 2.0621,
"step": 26950
},
{
"epoch": 2.9719317556411666,
"grad_norm": 0.16251477599143982,
"learning_rate": 0.00011589679771929529,
"loss": 2.08,
"step": 27000
},
{
"epoch": 2.977435332966428,
"grad_norm": 0.15681184828281403,
"learning_rate": 0.00011536531428313828,
"loss": 2.0518,
"step": 27050
},
{
"epoch": 2.9829389102916894,
"grad_norm": 0.17260442674160004,
"learning_rate": 0.00011483428975805664,
"loss": 2.0899,
"step": 27100
},
{
"epoch": 2.988442487616951,
"grad_norm": 0.16616973280906677,
"learning_rate": 0.00011430373118014808,
"loss": 2.0787,
"step": 27150
},
{
"epoch": 2.9939460649422127,
"grad_norm": 0.1666904091835022,
"learning_rate": 0.00011377364557933658,
"loss": 2.0555,
"step": 27200
},
{
"epoch": 2.999449642267474,
"grad_norm": 0.15540987253189087,
"learning_rate": 0.00011324403997927922,
"loss": 2.0676,
"step": 27250
},
{
"epoch": 3.004953219592735,
"grad_norm": 0.16125087440013885,
"learning_rate": 0.00011271492139727294,
"loss": 2.0144,
"step": 27300
},
{
"epoch": 3.0104567969179965,
"grad_norm": 0.17550891637802124,
"learning_rate": 0.00011218629684416168,
"loss": 2.0102,
"step": 27350
},
{
"epoch": 3.015960374243258,
"grad_norm": 0.16927887499332428,
"learning_rate": 0.00011165817332424356,
"loss": 1.9816,
"step": 27400
},
{
"epoch": 3.0214639515685193,
"grad_norm": 0.17522330582141876,
"learning_rate": 0.00011113055783517793,
"loss": 2.0066,
"step": 27450
},
{
"epoch": 3.026967528893781,
"grad_norm": 0.160496324300766,
"learning_rate": 0.00011060345736789265,
"loss": 2.0003,
"step": 27500
},
{
"epoch": 3.0324711062190426,
"grad_norm": 0.17563533782958984,
"learning_rate": 0.00011007687890649163,
"loss": 2.0272,
"step": 27550
},
{
"epoch": 3.037974683544304,
"grad_norm": 0.17000171542167664,
"learning_rate": 0.00010955082942816215,
"loss": 2.0447,
"step": 27600
},
{
"epoch": 3.0434782608695654,
"grad_norm": 0.17253416776657104,
"learning_rate": 0.00010902531590308236,
"loss": 1.9968,
"step": 27650
},
{
"epoch": 3.048981838194827,
"grad_norm": 0.16862879693508148,
"learning_rate": 0.00010850034529432919,
"loss": 1.9952,
"step": 27700
},
{
"epoch": 3.0544854155200882,
"grad_norm": 0.166275292634964,
"learning_rate": 0.00010797592455778562,
"loss": 2.0258,
"step": 27750
},
{
"epoch": 3.0599889928453496,
"grad_norm": 0.1734013706445694,
"learning_rate": 0.00010745206064204903,
"loss": 2.0047,
"step": 27800
},
{
"epoch": 3.065492570170611,
"grad_norm": 0.1760258674621582,
"learning_rate": 0.00010692876048833889,
"loss": 1.997,
"step": 27850
},
{
"epoch": 3.0709961474958725,
"grad_norm": 0.1616799384355545,
"learning_rate": 0.00010640603103040464,
"loss": 1.9817,
"step": 27900
},
{
"epoch": 3.076499724821134,
"grad_norm": 0.1775207221508026,
"learning_rate": 0.00010588387919443416,
"loss": 2.0391,
"step": 27950
},
{
"epoch": 3.0820033021463953,
"grad_norm": 0.17136207222938538,
"learning_rate": 0.00010536231189896175,
"loss": 1.999,
"step": 28000
},
{
"epoch": 3.0875068794716567,
"grad_norm": 0.1694386601448059,
"learning_rate": 0.00010484133605477644,
"loss": 2.0153,
"step": 28050
},
{
"epoch": 3.093010456796918,
"grad_norm": 0.17021538317203522,
"learning_rate": 0.00010432095856483057,
"loss": 2.0411,
"step": 28100
},
{
"epoch": 3.0985140341221795,
"grad_norm": 0.17548739910125732,
"learning_rate": 0.00010380118632414833,
"loss": 2.0416,
"step": 28150
},
{
"epoch": 3.104017611447441,
"grad_norm": 0.1785077303647995,
"learning_rate": 0.0001032820262197342,
"loss": 2.032,
"step": 28200
},
{
"epoch": 3.1095211887727023,
"grad_norm": 0.17226973176002502,
"learning_rate": 0.00010276348513048193,
"loss": 1.986,
"step": 28250
},
{
"epoch": 3.1150247660979637,
"grad_norm": 0.18131954967975616,
"learning_rate": 0.00010224556992708328,
"loss": 2.0149,
"step": 28300
},
{
"epoch": 3.120528343423225,
"grad_norm": 0.16980765759944916,
"learning_rate": 0.00010172828747193689,
"loss": 2.0309,
"step": 28350
},
{
"epoch": 3.1260319207484866,
"grad_norm": 0.16887415945529938,
"learning_rate": 0.00010121164461905759,
"loss": 1.9814,
"step": 28400
},
{
"epoch": 3.131535498073748,
"grad_norm": 0.18381917476654053,
"learning_rate": 0.00010069564821398536,
"loss": 2.0228,
"step": 28450
},
{
"epoch": 3.1370390753990094,
"grad_norm": 0.18095500767230988,
"learning_rate": 0.0001001803050936948,
"loss": 2.0197,
"step": 28500
},
{
"epoch": 3.142542652724271,
"grad_norm": 0.15869130194187164,
"learning_rate": 9.966562208650441e-05,
"loss": 2.0114,
"step": 28550
},
{
"epoch": 3.148046230049532,
"grad_norm": 0.17333059012889862,
"learning_rate": 9.915160601198616e-05,
"loss": 2.0025,
"step": 28600
},
{
"epoch": 3.1535498073747936,
"grad_norm": 0.161941796541214,
"learning_rate": 9.863826368087516e-05,
"loss": 2.0273,
"step": 28650
},
{
"epoch": 3.159053384700055,
"grad_norm": 0.16833285987377167,
"learning_rate": 9.812560189497935e-05,
"loss": 2.0108,
"step": 28700
},
{
"epoch": 3.1645569620253164,
"grad_norm": 0.15908978879451752,
"learning_rate": 9.76136274470895e-05,
"loss": 1.9852,
"step": 28750
},
{
"epoch": 3.170060539350578,
"grad_norm": 0.1724586933851242,
"learning_rate": 9.710234712088896e-05,
"loss": 2.0297,
"step": 28800
},
{
"epoch": 3.1755641166758393,
"grad_norm": 0.17113400995731354,
"learning_rate": 9.659176769086419e-05,
"loss": 1.9969,
"step": 28850
},
{
"epoch": 3.1810676940011007,
"grad_norm": 0.16423186659812927,
"learning_rate": 9.608189592221446e-05,
"loss": 2.0106,
"step": 28900
},
{
"epoch": 3.186571271326362,
"grad_norm": 0.16140978038311005,
"learning_rate": 9.557273857076271e-05,
"loss": 2.0339,
"step": 28950
},
{
"epoch": 3.1920748486516235,
"grad_norm": 0.1718360334634781,
"learning_rate": 9.506430238286566e-05,
"loss": 2.04,
"step": 29000
},
{
"epoch": 3.197578425976885,
"grad_norm": 0.164507195353508,
"learning_rate": 9.455659409532474e-05,
"loss": 2.0168,
"step": 29050
},
{
"epoch": 3.2030820033021463,
"grad_norm": 0.17989708483219147,
"learning_rate": 9.40496204352966e-05,
"loss": 2.0273,
"step": 29100
},
{
"epoch": 3.2085855806274077,
"grad_norm": 0.17319265007972717,
"learning_rate": 9.354338812020402e-05,
"loss": 1.9911,
"step": 29150
},
{
"epoch": 3.214089157952669,
"grad_norm": 0.17870402336120605,
"learning_rate": 9.303790385764694e-05,
"loss": 2.0103,
"step": 29200
},
{
"epoch": 3.2195927352779306,
"grad_norm": 0.1797982007265091,
"learning_rate": 9.253317434531357e-05,
"loss": 2.0169,
"step": 29250
},
{
"epoch": 3.225096312603192,
"grad_norm": 0.15193064510822296,
"learning_rate": 9.202920627089171e-05,
"loss": 2.0116,
"step": 29300
},
{
"epoch": 3.2305998899284534,
"grad_norm": 0.16679364442825317,
"learning_rate": 9.152600631197996e-05,
"loss": 2.0362,
"step": 29350
},
{
"epoch": 3.236103467253715,
"grad_norm": 0.17053711414337158,
"learning_rate": 9.102358113599942e-05,
"loss": 1.9829,
"step": 29400
},
{
"epoch": 3.241607044578976,
"grad_norm": 0.16923174262046814,
"learning_rate": 9.052193740010536e-05,
"loss": 2.0011,
"step": 29450
},
{
"epoch": 3.2471106219042376,
"grad_norm": 0.18323257565498352,
"learning_rate": 9.002108175109877e-05,
"loss": 1.9784,
"step": 29500
},
{
"epoch": 3.252614199229499,
"grad_norm": 0.16737329959869385,
"learning_rate": 8.952102082533855e-05,
"loss": 2.0276,
"step": 29550
},
{
"epoch": 3.2581177765547604,
"grad_norm": 0.17038094997406006,
"learning_rate": 8.902176124865348e-05,
"loss": 2.0315,
"step": 29600
},
{
"epoch": 3.263621353880022,
"grad_norm": 0.16778729856014252,
"learning_rate": 8.852330963625444e-05,
"loss": 2.0139,
"step": 29650
},
{
"epoch": 3.2691249312052832,
"grad_norm": 0.18061307072639465,
"learning_rate": 8.802567259264674e-05,
"loss": 2.002,
"step": 29700
},
{
"epoch": 3.2746285085305447,
"grad_norm": 0.17191821336746216,
"learning_rate": 8.752885671154253e-05,
"loss": 2.0218,
"step": 29750
},
{
"epoch": 3.280132085855806,
"grad_norm": 0.17957885563373566,
"learning_rate": 8.703286857577378e-05,
"loss": 2.0217,
"step": 29800
},
{
"epoch": 3.2856356631810675,
"grad_norm": 0.16803069412708282,
"learning_rate": 8.653771475720442e-05,
"loss": 2.0185,
"step": 29850
},
{
"epoch": 3.291139240506329,
"grad_norm": 0.1788654327392578,
"learning_rate": 8.604340181664395e-05,
"loss": 1.9973,
"step": 29900
},
{
"epoch": 3.2966428178315903,
"grad_norm": 0.17996376752853394,
"learning_rate": 8.554993630376006e-05,
"loss": 2.0161,
"step": 29950
},
{
"epoch": 3.3021463951568517,
"grad_norm": 0.18701910972595215,
"learning_rate": 8.50573247569921e-05,
"loss": 2.0156,
"step": 30000
},
{
"epoch": 3.3076499724821136,
"grad_norm": 0.1713022142648697,
"learning_rate": 8.45655737034641e-05,
"loss": 2.0144,
"step": 30050
},
{
"epoch": 3.313153549807375,
"grad_norm": 0.17666327953338623,
"learning_rate": 8.407468965889878e-05,
"loss": 2.0116,
"step": 30100
},
{
"epoch": 3.3186571271326364,
"grad_norm": 0.18481209874153137,
"learning_rate": 8.358467912753079e-05,
"loss": 2.0149,
"step": 30150
},
{
"epoch": 3.324160704457898,
"grad_norm": 0.17895914614200592,
"learning_rate": 8.309554860202082e-05,
"loss": 1.9913,
"step": 30200
},
{
"epoch": 3.329664281783159,
"grad_norm": 0.18427938222885132,
"learning_rate": 8.260730456336925e-05,
"loss": 1.9913,
"step": 30250
},
{
"epoch": 3.3351678591084206,
"grad_norm": 0.1806970238685608,
"learning_rate": 8.211995348083067e-05,
"loss": 2.0177,
"step": 30300
},
{
"epoch": 3.340671436433682,
"grad_norm": 0.1735043078660965,
"learning_rate": 8.163350181182787e-05,
"loss": 2.0275,
"step": 30350
},
{
"epoch": 3.3461750137589434,
"grad_norm": 0.17593735456466675,
"learning_rate": 8.11479560018664e-05,
"loss": 2.0525,
"step": 30400
},
{
"epoch": 3.351678591084205,
"grad_norm": 0.1755988448858261,
"learning_rate": 8.066332248444908e-05,
"loss": 2.0182,
"step": 30450
},
{
"epoch": 3.3571821684094663,
"grad_norm": 0.15805041790008545,
"learning_rate": 8.017960768099091e-05,
"loss": 1.9748,
"step": 30500
},
{
"epoch": 3.3626857457347277,
"grad_norm": 0.18487168848514557,
"learning_rate": 7.969681800073386e-05,
"loss": 2.021,
"step": 30550
},
{
"epoch": 3.368189323059989,
"grad_norm": 0.17120800912380219,
"learning_rate": 7.921495984066194e-05,
"loss": 2.0113,
"step": 30600
},
{
"epoch": 3.3736929003852505,
"grad_norm": 0.1721143126487732,
"learning_rate": 7.87340395854164e-05,
"loss": 1.9936,
"step": 30650
},
{
"epoch": 3.379196477710512,
"grad_norm": 0.17595061659812927,
"learning_rate": 7.825406360721139e-05,
"loss": 2.0044,
"step": 30700
},
{
"epoch": 3.3847000550357733,
"grad_norm": 0.16751642525196075,
"learning_rate": 7.777503826574925e-05,
"loss": 1.9988,
"step": 30750
},
{
"epoch": 3.3902036323610347,
"grad_norm": 0.17864260077476501,
"learning_rate": 7.729696990813634e-05,
"loss": 2.0189,
"step": 30800
},
{
"epoch": 3.395707209686296,
"grad_norm": 0.1791946142911911,
"learning_rate": 7.681986486879898e-05,
"loss": 2.0038,
"step": 30850
},
{
"epoch": 3.4012107870115575,
"grad_norm": 0.1704343855381012,
"learning_rate": 7.634372946939945e-05,
"loss": 1.9625,
"step": 30900
},
{
"epoch": 3.406714364336819,
"grad_norm": 0.17794691026210785,
"learning_rate": 7.586857001875235e-05,
"loss": 2.0017,
"step": 30950
},
{
"epoch": 3.4122179416620804,
"grad_norm": 0.17922665178775787,
"learning_rate": 7.539439281274071e-05,
"loss": 2.054,
"step": 31000
},
{
"epoch": 3.4177215189873418,
"grad_norm": 0.174763485789299,
"learning_rate": 7.492120413423296e-05,
"loss": 2.0185,
"step": 31050
},
{
"epoch": 3.423225096312603,
"grad_norm": 0.1687181144952774,
"learning_rate": 7.444901025299941e-05,
"loss": 2.0463,
"step": 31100
},
{
"epoch": 3.4287286736378646,
"grad_norm": 0.17793361842632294,
"learning_rate": 7.397781742562941e-05,
"loss": 2.0296,
"step": 31150
},
{
"epoch": 3.434232250963126,
"grad_norm": 0.16858656704425812,
"learning_rate": 7.350763189544803e-05,
"loss": 1.9916,
"step": 31200
},
{
"epoch": 3.4397358282883874,
"grad_norm": 0.16184011101722717,
"learning_rate": 7.303845989243384e-05,
"loss": 2.0314,
"step": 31250
},
{
"epoch": 3.445239405613649,
"grad_norm": 0.1733245998620987,
"learning_rate": 7.257030763313607e-05,
"loss": 2.0125,
"step": 31300
},
{
"epoch": 3.4507429829389102,
"grad_norm": 0.18471761047840118,
"learning_rate": 7.210318132059231e-05,
"loss": 1.9691,
"step": 31350
},
{
"epoch": 3.4562465602641717,
"grad_norm": 0.17260605096817017,
"learning_rate": 7.16370871442462e-05,
"loss": 2.0113,
"step": 31400
},
{
"epoch": 3.461750137589433,
"grad_norm": 0.1758509874343872,
"learning_rate": 7.117203127986569e-05,
"loss": 2.0239,
"step": 31450
},
{
"epoch": 3.4672537149146945,
"grad_norm": 0.18456335365772247,
"learning_rate": 7.070801988946098e-05,
"loss": 1.999,
"step": 31500
},
{
"epoch": 3.472757292239956,
"grad_norm": 0.17995281517505646,
"learning_rate": 7.024505912120293e-05,
"loss": 2.0141,
"step": 31550
},
{
"epoch": 3.4782608695652173,
"grad_norm": 0.16747882962226868,
"learning_rate": 6.978315510934164e-05,
"loss": 2.0073,
"step": 31600
},
{
"epoch": 3.4837644468904787,
"grad_norm": 0.1658480316400528,
"learning_rate": 6.932231397412522e-05,
"loss": 2.0079,
"step": 31650
},
{
"epoch": 3.48926802421574,
"grad_norm": 0.17987202107906342,
"learning_rate": 6.886254182171836e-05,
"loss": 2.0439,
"step": 31700
},
{
"epoch": 3.4947716015410015,
"grad_norm": 0.1672779768705368,
"learning_rate": 6.840384474412201e-05,
"loss": 1.9878,
"step": 31750
},
{
"epoch": 3.500275178866263,
"grad_norm": 0.1716667115688324,
"learning_rate": 6.794622881909207e-05,
"loss": 2.0248,
"step": 31800
},
{
"epoch": 3.5057787561915243,
"grad_norm": 0.17986662685871124,
"learning_rate": 6.748970011005924e-05,
"loss": 1.99,
"step": 31850
},
{
"epoch": 3.5112823335167858,
"grad_norm": 0.1808827668428421,
"learning_rate": 6.703426466604858e-05,
"loss": 1.9692,
"step": 31900
},
{
"epoch": 3.516785910842047,
"grad_norm": 0.17262905836105347,
"learning_rate": 6.65799285215993e-05,
"loss": 1.9636,
"step": 31950
},
{
"epoch": 3.522289488167309,
"grad_norm": 0.16857630014419556,
"learning_rate": 6.612669769668489e-05,
"loss": 1.9923,
"step": 32000
},
{
"epoch": 3.5277930654925704,
"grad_norm": 0.17498353123664856,
"learning_rate": 6.567457819663327e-05,
"loss": 2.0248,
"step": 32050
},
{
"epoch": 3.533296642817832,
"grad_norm": 0.16948480904102325,
"learning_rate": 6.522357601204719e-05,
"loss": 1.9966,
"step": 32100
},
{
"epoch": 3.5388002201430933,
"grad_norm": 0.1695474237203598,
"learning_rate": 6.477369711872502e-05,
"loss": 1.994,
"step": 32150
},
{
"epoch": 3.5443037974683547,
"grad_norm": 0.1761193871498108,
"learning_rate": 6.432494747758146e-05,
"loss": 1.9827,
"step": 32200
},
{
"epoch": 3.549807374793616,
"grad_norm": 0.17206500470638275,
"learning_rate": 6.387733303456858e-05,
"loss": 1.9944,
"step": 32250
},
{
"epoch": 3.5553109521188775,
"grad_norm": 0.19551701843738556,
"learning_rate": 6.3430859720597e-05,
"loss": 2.0118,
"step": 32300
},
{
"epoch": 3.560814529444139,
"grad_norm": 0.17827004194259644,
"learning_rate": 6.298553345145728e-05,
"loss": 1.9953,
"step": 32350
},
{
"epoch": 3.5663181067694003,
"grad_norm": 0.17359542846679688,
"learning_rate": 6.254136012774166e-05,
"loss": 2.0098,
"step": 32400
},
{
"epoch": 3.5718216840946617,
"grad_norm": 0.16225001215934753,
"learning_rate": 6.209834563476578e-05,
"loss": 2.008,
"step": 32450
},
{
"epoch": 3.577325261419923,
"grad_norm": 0.17888249456882477,
"learning_rate": 6.165649584249079e-05,
"loss": 1.9962,
"step": 32500
},
{
"epoch": 3.5828288387451845,
"grad_norm": 0.16299229860305786,
"learning_rate": 6.121581660544532e-05,
"loss": 2.0387,
"step": 32550
},
{
"epoch": 3.588332416070446,
"grad_norm": 0.16753186285495758,
"learning_rate": 6.077631376264832e-05,
"loss": 1.9662,
"step": 32600
},
{
"epoch": 3.5938359933957074,
"grad_norm": 0.18166500329971313,
"learning_rate": 6.033799313753129e-05,
"loss": 1.997,
"step": 32650
},
{
"epoch": 3.5993395707209688,
"grad_norm": 0.16691668331623077,
"learning_rate": 5.990086053786139e-05,
"loss": 2.0072,
"step": 32700
},
{
"epoch": 3.60484314804623,
"grad_norm": 0.18671706318855286,
"learning_rate": 5.946492175566438e-05,
"loss": 2.0217,
"step": 32750
},
{
"epoch": 3.6103467253714916,
"grad_norm": 0.16831424832344055,
"learning_rate": 5.9030182567147905e-05,
"loss": 2.0024,
"step": 32800
},
{
"epoch": 3.615850302696753,
"grad_norm": 0.17562079429626465,
"learning_rate": 5.859664873262483e-05,
"loss": 2.0351,
"step": 32850
},
{
"epoch": 3.6213538800220144,
"grad_norm": 0.18190255761146545,
"learning_rate": 5.816432599643719e-05,
"loss": 2.0117,
"step": 32900
},
{
"epoch": 3.626857457347276,
"grad_norm": 0.1830035150051117,
"learning_rate": 5.773322008687972e-05,
"loss": 2.012,
"step": 32950
},
{
"epoch": 3.6323610346725372,
"grad_norm": 0.17998354136943817,
"learning_rate": 5.730333671612435e-05,
"loss": 1.9977,
"step": 33000
},
{
"epoch": 3.6378646119977986,
"grad_norm": 0.1839103251695633,
"learning_rate": 5.687468158014421e-05,
"loss": 2.003,
"step": 33050
},
{
"epoch": 3.64336818932306,
"grad_norm": 0.18631495535373688,
"learning_rate": 5.6447260358638285e-05,
"loss": 2.0072,
"step": 33100
},
{
"epoch": 3.6488717666483215,
"grad_norm": 0.19609171152114868,
"learning_rate": 5.602107871495615e-05,
"loss": 1.9912,
"step": 33150
},
{
"epoch": 3.654375343973583,
"grad_norm": 0.16935598850250244,
"learning_rate": 5.5596142296022954e-05,
"loss": 2.0097,
"step": 33200
},
{
"epoch": 3.6598789212988443,
"grad_norm": 0.16973347961902618,
"learning_rate": 5.5172456732264435e-05,
"loss": 2.0151,
"step": 33250
},
{
"epoch": 3.6653824986241057,
"grad_norm": 0.17289844155311584,
"learning_rate": 5.475002763753257e-05,
"loss": 2.0393,
"step": 33300
},
{
"epoch": 3.670886075949367,
"grad_norm": 0.1864277720451355,
"learning_rate": 5.432886060903105e-05,
"loss": 2.0069,
"step": 33350
},
{
"epoch": 3.6763896532746285,
"grad_norm": 0.17381645739078522,
"learning_rate": 5.390896122724111e-05,
"loss": 2.0323,
"step": 33400
},
{
"epoch": 3.68189323059989,
"grad_norm": 0.18182097375392914,
"learning_rate": 5.349033505584767e-05,
"loss": 2.0344,
"step": 33450
},
{
"epoch": 3.6873968079251513,
"grad_norm": 0.17326848208904266,
"learning_rate": 5.3072987641665373e-05,
"loss": 1.9888,
"step": 33500
},
{
"epoch": 3.6929003852504128,
"grad_norm": 0.17132443189620972,
"learning_rate": 5.2656924514565515e-05,
"loss": 1.9741,
"step": 33550
},
{
"epoch": 3.698403962575674,
"grad_norm": 0.17236609756946564,
"learning_rate": 5.224215118740243e-05,
"loss": 1.9967,
"step": 33600
},
{
"epoch": 3.7039075399009356,
"grad_norm": 0.16628102958202362,
"learning_rate": 5.182867315594046e-05,
"loss": 2.0036,
"step": 33650
},
{
"epoch": 3.709411117226197,
"grad_norm": 0.17410129308700562,
"learning_rate": 5.141649589878134e-05,
"loss": 2.0173,
"step": 33700
},
{
"epoch": 3.7149146945514584,
"grad_norm": 0.1739521622657776,
"learning_rate": 5.100562487729148e-05,
"loss": 1.9977,
"step": 33750
},
{
"epoch": 3.72041827187672,
"grad_norm": 0.17988687753677368,
"learning_rate": 5.059606553552956e-05,
"loss": 1.9964,
"step": 33800
},
{
"epoch": 3.725921849201981,
"grad_norm": 0.17351609468460083,
"learning_rate": 5.018782330017448e-05,
"loss": 2.0078,
"step": 33850
},
{
"epoch": 3.7314254265272426,
"grad_norm": 0.176718607544899,
"learning_rate": 4.9780903580453435e-05,
"loss": 2.0036,
"step": 33900
},
{
"epoch": 3.736929003852504,
"grad_norm": 0.17504329979419708,
"learning_rate": 4.937531176807023e-05,
"loss": 1.9837,
"step": 33950
},
{
"epoch": 3.7424325811777654,
"grad_norm": 0.179831862449646,
"learning_rate": 4.897105323713374e-05,
"loss": 1.999,
"step": 34000
},
{
"epoch": 3.747936158503027,
"grad_norm": 0.17752991616725922,
"learning_rate": 4.8568133344086986e-05,
"loss": 2.0194,
"step": 34050
},
{
"epoch": 3.7534397358282883,
"grad_norm": 0.17888140678405762,
"learning_rate": 4.8166557427635774e-05,
"loss": 2.0108,
"step": 34100
},
{
"epoch": 3.7589433131535497,
"grad_norm": 0.17585836350917816,
"learning_rate": 4.776633080867834e-05,
"loss": 2.0421,
"step": 34150
},
{
"epoch": 3.764446890478811,
"grad_norm": 0.16860149800777435,
"learning_rate": 4.736745879023457e-05,
"loss": 2.0126,
"step": 34200
},
{
"epoch": 3.7699504678040725,
"grad_norm": 0.17601364850997925,
"learning_rate": 4.6969946657375865e-05,
"loss": 2.0127,
"step": 34250
},
{
"epoch": 3.775454045129334,
"grad_norm": 0.19177961349487305,
"learning_rate": 4.657379967715511e-05,
"loss": 2.0234,
"step": 34300
},
{
"epoch": 3.7809576224545953,
"grad_norm": 0.19012008607387543,
"learning_rate": 4.617902309853685e-05,
"loss": 1.976,
"step": 34350
},
{
"epoch": 3.7864611997798567,
"grad_norm": 0.18448469042778015,
"learning_rate": 4.578562215232766e-05,
"loss": 1.9897,
"step": 34400
},
{
"epoch": 3.791964777105118,
"grad_norm": 0.18167300522327423,
"learning_rate": 4.539360205110701e-05,
"loss": 2.0088,
"step": 34450
},
{
"epoch": 3.7974683544303796,
"grad_norm": 0.1806318610906601,
"learning_rate": 4.500296798915813e-05,
"loss": 1.9949,
"step": 34500
},
{
"epoch": 3.802971931755641,
"grad_norm": 0.17676065862178802,
"learning_rate": 4.461372514239911e-05,
"loss": 1.9966,
"step": 34550
},
{
"epoch": 3.8084755090809024,
"grad_norm": 0.16621781885623932,
"learning_rate": 4.422587866831446e-05,
"loss": 1.9952,
"step": 34600
},
{
"epoch": 3.813979086406164,
"grad_norm": 0.16662217676639557,
"learning_rate": 4.3839433705886646e-05,
"loss": 2.0032,
"step": 34650
},
{
"epoch": 3.819482663731425,
"grad_norm": 0.15986546874046326,
"learning_rate": 4.3454395375527966e-05,
"loss": 1.9974,
"step": 34700
},
{
"epoch": 3.8249862410566866,
"grad_norm": 0.18636077642440796,
"learning_rate": 4.307076877901293e-05,
"loss": 1.9959,
"step": 34750
},
{
"epoch": 3.830489818381948,
"grad_norm": 0.17939697206020355,
"learning_rate": 4.268855899941039e-05,
"loss": 1.9923,
"step": 34800
},
{
"epoch": 3.8359933957072094,
"grad_norm": 0.1743779331445694,
"learning_rate": 4.2307771101016366e-05,
"loss": 2.0055,
"step": 34850
},
{
"epoch": 3.841496973032471,
"grad_norm": 0.1767256259918213,
"learning_rate": 4.192841012928691e-05,
"loss": 1.9834,
"step": 34900
},
{
"epoch": 3.8470005503577323,
"grad_norm": 0.1775825172662735,
"learning_rate": 4.1550481110771224e-05,
"loss": 2.01,
"step": 34950
},
{
"epoch": 3.852504127682994,
"grad_norm": 0.1638861745595932,
"learning_rate": 4.1173989053045024e-05,
"loss": 2.0001,
"step": 35000
},
{
"epoch": 3.8580077050082555,
"grad_norm": 0.1761082410812378,
"learning_rate": 4.079893894464429e-05,
"loss": 2.0041,
"step": 35050
},
{
"epoch": 3.863511282333517,
"grad_norm": 0.17102594673633575,
"learning_rate": 4.042533575499914e-05,
"loss": 2.003,
"step": 35100
},
{
"epoch": 3.8690148596587783,
"grad_norm": 0.19267229735851288,
"learning_rate": 4.005318443436779e-05,
"loss": 1.973,
"step": 35150
},
{
"epoch": 3.8745184369840397,
"grad_norm": 0.16570790112018585,
"learning_rate": 3.968248991377137e-05,
"loss": 1.9834,
"step": 35200
},
{
"epoch": 3.880022014309301,
"grad_norm": 0.16816405951976776,
"learning_rate": 3.9313257104928094e-05,
"loss": 2.0011,
"step": 35250
},
{
"epoch": 3.8855255916345626,
"grad_norm": 0.16395322978496552,
"learning_rate": 3.8945490900188616e-05,
"loss": 1.9771,
"step": 35300
},
{
"epoch": 3.891029168959824,
"grad_norm": 0.17911794781684875,
"learning_rate": 3.857919617247097e-05,
"loss": 2.0,
"step": 35350
},
{
"epoch": 3.8965327462850854,
"grad_norm": 0.18344633281230927,
"learning_rate": 3.821437777519607e-05,
"loss": 1.9952,
"step": 35400
},
{
"epoch": 3.902036323610347,
"grad_norm": 0.1782706081867218,
"learning_rate": 3.785104054222334e-05,
"loss": 1.9952,
"step": 35450
},
{
"epoch": 3.907539900935608,
"grad_norm": 0.17840951681137085,
"learning_rate": 3.74891892877868e-05,
"loss": 1.9994,
"step": 35500
},
{
"epoch": 3.9130434782608696,
"grad_norm": 0.19014611840248108,
"learning_rate": 3.7128828806431046e-05,
"loss": 2.0023,
"step": 35550
},
{
"epoch": 3.918547055586131,
"grad_norm": 0.17002084851264954,
"learning_rate": 3.6769963872947996e-05,
"loss": 2.0068,
"step": 35600
},
{
"epoch": 3.9240506329113924,
"grad_norm": 0.16855107247829437,
"learning_rate": 3.641259924231345e-05,
"loss": 1.9847,
"step": 35650
},
{
"epoch": 3.929554210236654,
"grad_norm": 0.190132737159729,
"learning_rate": 3.605673964962414e-05,
"loss": 1.9955,
"step": 35700
},
{
"epoch": 3.9350577875619153,
"grad_norm": 0.1558249294757843,
"learning_rate": 3.570238981003498e-05,
"loss": 1.9861,
"step": 35750
},
{
"epoch": 3.9405613648871767,
"grad_norm": 0.18199962377548218,
"learning_rate": 3.53495544186966e-05,
"loss": 2.0136,
"step": 35800
},
{
"epoch": 3.946064942212438,
"grad_norm": 0.17813271284103394,
"learning_rate": 3.499823815069306e-05,
"loss": 2.0012,
"step": 35850
},
{
"epoch": 3.9515685195376995,
"grad_norm": 0.18256065249443054,
"learning_rate": 3.464844566098008e-05,
"loss": 1.9892,
"step": 35900
},
{
"epoch": 3.957072096862961,
"grad_norm": 0.1784873753786087,
"learning_rate": 3.4300181584323126e-05,
"loss": 1.9983,
"step": 35950
},
{
"epoch": 3.9625756741882223,
"grad_norm": 0.18530665338039398,
"learning_rate": 3.3953450535236226e-05,
"loss": 2.0357,
"step": 36000
},
{
"epoch": 3.9680792515134837,
"grad_norm": 0.17063435912132263,
"learning_rate": 3.360825710792068e-05,
"loss": 2.0029,
"step": 36050
},
{
"epoch": 3.973582828838745,
"grad_norm": 0.1718268245458603,
"learning_rate": 3.326460587620427e-05,
"loss": 1.9762,
"step": 36100
},
{
"epoch": 3.9790864061640066,
"grad_norm": 0.15901614725589752,
"learning_rate": 3.292250139348057e-05,
"loss": 1.9952,
"step": 36150
},
{
"epoch": 3.984589983489268,
"grad_norm": 0.18968409299850464,
"learning_rate": 3.258194819264871e-05,
"loss": 2.0127,
"step": 36200
},
{
"epoch": 3.9900935608145294,
"grad_norm": 0.18861430883407593,
"learning_rate": 3.2242950786053296e-05,
"loss": 1.9981,
"step": 36250
},
{
"epoch": 3.995597138139791,
"grad_norm": 0.1702040284872055,
"learning_rate": 3.1905513665424445e-05,
"loss": 2.0023,
"step": 36300
},
{
"epoch": 4.001100715465053,
"grad_norm": 0.18254053592681885,
"learning_rate": 3.156964130181859e-05,
"loss": 2.0038,
"step": 36350
},
{
"epoch": 4.006604292790314,
"grad_norm": 0.18867474794387817,
"learning_rate": 3.1235338145558925e-05,
"loss": 1.9139,
"step": 36400
},
{
"epoch": 4.0121078701155755,
"grad_norm": 0.1872694045305252,
"learning_rate": 3.090260862617671e-05,
"loss": 1.9484,
"step": 36450
},
{
"epoch": 4.017611447440837,
"grad_norm": 0.17595593631267548,
"learning_rate": 3.057145715235236e-05,
"loss": 1.9543,
"step": 36500
},
{
"epoch": 4.023115024766098,
"grad_norm": 0.17811767756938934,
"learning_rate": 3.0241888111857145e-05,
"loss": 1.9268,
"step": 36550
},
{
"epoch": 4.02861860209136,
"grad_norm": 0.17598308622837067,
"learning_rate": 2.9913905871495037e-05,
"loss": 1.9439,
"step": 36600
},
{
"epoch": 4.034122179416621,
"grad_norm": 0.19237573444843292,
"learning_rate": 2.9587514777044874e-05,
"loss": 1.9044,
"step": 36650
},
{
"epoch": 4.0396257567418825,
"grad_norm": 0.1873626410961151,
"learning_rate": 2.926271915320259e-05,
"loss": 1.943,
"step": 36700
},
{
"epoch": 4.045129334067144,
"grad_norm": 0.2001618891954422,
"learning_rate": 2.893952330352423e-05,
"loss": 1.9311,
"step": 36750
},
{
"epoch": 4.050632911392405,
"grad_norm": 0.1838996410369873,
"learning_rate": 2.86179315103687e-05,
"loss": 1.9373,
"step": 36800
},
{
"epoch": 4.056136488717667,
"grad_norm": 0.16863055527210236,
"learning_rate": 2.8297948034841074e-05,
"loss": 1.9289,
"step": 36850
},
{
"epoch": 4.061640066042928,
"grad_norm": 0.19162894785404205,
"learning_rate": 2.797957711673619e-05,
"loss": 1.9293,
"step": 36900
},
{
"epoch": 4.06714364336819,
"grad_norm": 0.1854933500289917,
"learning_rate": 2.7662822974482423e-05,
"loss": 1.9336,
"step": 36950
},
{
"epoch": 4.072647220693451,
"grad_norm": 0.18568897247314453,
"learning_rate": 2.7347689805085733e-05,
"loss": 1.9431,
"step": 37000
},
{
"epoch": 4.078150798018712,
"grad_norm": 0.1947336494922638,
"learning_rate": 2.7034181784074205e-05,
"loss": 1.9345,
"step": 37050
},
{
"epoch": 4.083654375343974,
"grad_norm": 0.18774710595607758,
"learning_rate": 2.672230306544254e-05,
"loss": 1.9364,
"step": 37100
},
{
"epoch": 4.089157952669235,
"grad_norm": 0.1984746754169464,
"learning_rate": 2.6412057781597223e-05,
"loss": 1.954,
"step": 37150
},
{
"epoch": 4.094661529994497,
"grad_norm": 0.1908876746892929,
"learning_rate": 2.61034500433016e-05,
"loss": 1.9422,
"step": 37200
},
{
"epoch": 4.100165107319758,
"grad_norm": 0.18860432505607605,
"learning_rate": 2.579648393962147e-05,
"loss": 1.9649,
"step": 37250
},
{
"epoch": 4.105668684645019,
"grad_norm": 0.18678942322731018,
"learning_rate": 2.5491163537870922e-05,
"loss": 1.9544,
"step": 37300
},
{
"epoch": 4.111172261970281,
"grad_norm": 0.19560185074806213,
"learning_rate": 2.5187492883558435e-05,
"loss": 1.9396,
"step": 37350
},
{
"epoch": 4.116675839295542,
"grad_norm": 0.1825692355632782,
"learning_rate": 2.488547600033317e-05,
"loss": 1.949,
"step": 37400
},
{
"epoch": 4.122179416620804,
"grad_norm": 0.1951771229505539,
"learning_rate": 2.4585116889931873e-05,
"loss": 1.9482,
"step": 37450
},
{
"epoch": 4.127682993946065,
"grad_norm": 0.1859281212091446,
"learning_rate": 2.4286419532125668e-05,
"loss": 1.9512,
"step": 37500
},
{
"epoch": 4.1331865712713265,
"grad_norm": 0.19174982607364655,
"learning_rate": 2.3989387884667365e-05,
"loss": 1.9358,
"step": 37550
},
{
"epoch": 4.138690148596588,
"grad_norm": 0.187012180685997,
"learning_rate": 2.369402588323908e-05,
"loss": 1.9114,
"step": 37600
},
{
"epoch": 4.144193725921849,
"grad_norm": 0.20616699755191803,
"learning_rate": 2.340033744140004e-05,
"loss": 1.8965,
"step": 37650
},
{
"epoch": 4.149697303247111,
"grad_norm": 0.17636051774024963,
"learning_rate": 2.3108326450534725e-05,
"loss": 1.9549,
"step": 37700
},
{
"epoch": 4.155200880572372,
"grad_norm": 0.19219453632831573,
"learning_rate": 2.281799677980136e-05,
"loss": 1.8911,
"step": 37750
},
{
"epoch": 4.1607044578976335,
"grad_norm": 0.19742678105831146,
"learning_rate": 2.2529352276080513e-05,
"loss": 1.9562,
"step": 37800
},
{
"epoch": 4.166208035222895,
"grad_norm": 0.18049471080303192,
"learning_rate": 2.2242396763924292e-05,
"loss": 1.9169,
"step": 37850
},
{
"epoch": 4.171711612548156,
"grad_norm": 0.2108810693025589,
"learning_rate": 2.1957134045505588e-05,
"loss": 1.9363,
"step": 37900
},
{
"epoch": 4.177215189873418,
"grad_norm": 0.20711076259613037,
"learning_rate": 2.1673567900567667e-05,
"loss": 1.9685,
"step": 37950
},
{
"epoch": 4.182718767198679,
"grad_norm": 0.18670059740543365,
"learning_rate": 2.139170208637415e-05,
"loss": 1.9537,
"step": 38000
},
{
"epoch": 4.188222344523941,
"grad_norm": 0.17974555492401123,
"learning_rate": 2.1111540337659227e-05,
"loss": 1.9476,
"step": 38050
},
{
"epoch": 4.193725921849202,
"grad_norm": 0.18271256983280182,
"learning_rate": 2.083308636657811e-05,
"loss": 1.9535,
"step": 38100
},
{
"epoch": 4.199229499174463,
"grad_norm": 0.19260470569133759,
"learning_rate": 2.0556343862657855e-05,
"loss": 1.9477,
"step": 38150
},
{
"epoch": 4.204733076499725,
"grad_norm": 0.19317923486232758,
"learning_rate": 2.0281316492748595e-05,
"loss": 1.9518,
"step": 38200
},
{
"epoch": 4.210236653824986,
"grad_norm": 0.1854427605867386,
"learning_rate": 2.0008007900974738e-05,
"loss": 1.9245,
"step": 38250
},
{
"epoch": 4.215740231150248,
"grad_norm": 0.1781974732875824,
"learning_rate": 1.9736421708686934e-05,
"loss": 1.9387,
"step": 38300
},
{
"epoch": 4.221243808475509,
"grad_norm": 0.18503354489803314,
"learning_rate": 1.946656151441389e-05,
"loss": 1.9683,
"step": 38350
},
{
"epoch": 4.2267473858007705,
"grad_norm": 0.19728437066078186,
"learning_rate": 1.9198430893814798e-05,
"loss": 1.9546,
"step": 38400
},
{
"epoch": 4.232250963126032,
"grad_norm": 0.1957421451807022,
"learning_rate": 1.89320333996319e-05,
"loss": 1.9526,
"step": 38450
},
{
"epoch": 4.237754540451293,
"grad_norm": 0.18944865465164185,
"learning_rate": 1.866737256164349e-05,
"loss": 1.9467,
"step": 38500
},
{
"epoch": 4.243258117776555,
"grad_norm": 0.1760573536157608,
"learning_rate": 1.8404451886617005e-05,
"loss": 1.9142,
"step": 38550
},
{
"epoch": 4.248761695101816,
"grad_norm": 0.19618487358093262,
"learning_rate": 1.8143274858262702e-05,
"loss": 1.9491,
"step": 38600
},
{
"epoch": 4.2542652724270775,
"grad_norm": 0.19029423594474792,
"learning_rate": 1.7883844937187495e-05,
"loss": 1.9553,
"step": 38650
},
{
"epoch": 4.259768849752339,
"grad_norm": 0.19688698649406433,
"learning_rate": 1.762616556084891e-05,
"loss": 1.9698,
"step": 38700
},
{
"epoch": 4.2652724270776,
"grad_norm": 0.16535942256450653,
"learning_rate": 1.737024014350983e-05,
"loss": 1.9395,
"step": 38750
},
{
"epoch": 4.270776004402862,
"grad_norm": 0.19671253859996796,
"learning_rate": 1.7116072076193e-05,
"loss": 1.9343,
"step": 38800
},
{
"epoch": 4.276279581728123,
"grad_norm": 0.18672049045562744,
"learning_rate": 1.6863664726636278e-05,
"loss": 1.939,
"step": 38850
},
{
"epoch": 4.281783159053385,
"grad_norm": 0.19714199006557465,
"learning_rate": 1.6613021439247914e-05,
"loss": 1.9455,
"step": 38900
},
{
"epoch": 4.287286736378646,
"grad_norm": 0.2018548846244812,
"learning_rate": 1.6364145535062172e-05,
"loss": 1.9169,
"step": 38950
},
{
"epoch": 4.292790313703907,
"grad_norm": 0.19057820737361908,
"learning_rate": 1.611704031169555e-05,
"loss": 1.956,
"step": 39000
},
{
"epoch": 4.298293891029169,
"grad_norm": 0.18719470500946045,
"learning_rate": 1.5871709043302876e-05,
"loss": 1.9483,
"step": 39050
},
{
"epoch": 4.30379746835443,
"grad_norm": 0.19479648768901825,
"learning_rate": 1.5628154980533984e-05,
"loss": 1.9507,
"step": 39100
},
{
"epoch": 4.309301045679692,
"grad_norm": 0.2034020870923996,
"learning_rate": 1.538638135049071e-05,
"loss": 1.9343,
"step": 39150
},
{
"epoch": 4.314804623004953,
"grad_norm": 0.18177463114261627,
"learning_rate": 1.5146391356684029e-05,
"loss": 1.9481,
"step": 39200
},
{
"epoch": 4.3203082003302145,
"grad_norm": 0.1850568652153015,
"learning_rate": 1.4908188178991714e-05,
"loss": 1.9324,
"step": 39250
},
{
"epoch": 4.325811777655476,
"grad_norm": 0.1880549043416977,
"learning_rate": 1.4671774973616102e-05,
"loss": 1.9406,
"step": 39300
},
{
"epoch": 4.331315354980737,
"grad_norm": 0.18556420505046844,
"learning_rate": 1.4437154873042279e-05,
"loss": 1.9516,
"step": 39350
},
{
"epoch": 4.336818932305999,
"grad_norm": 0.19034932553768158,
"learning_rate": 1.420433098599672e-05,
"loss": 1.9477,
"step": 39400
},
{
"epoch": 4.34232250963126,
"grad_norm": 0.18247225880622864,
"learning_rate": 1.3973306397405909e-05,
"loss": 1.9552,
"step": 39450
},
{
"epoch": 4.3478260869565215,
"grad_norm": 0.19472351670265198,
"learning_rate": 1.3744084168355612e-05,
"loss": 1.9543,
"step": 39500
},
{
"epoch": 4.353329664281783,
"grad_norm": 0.18814486265182495,
"learning_rate": 1.3516667336050219e-05,
"loss": 1.956,
"step": 39550
},
{
"epoch": 4.358833241607044,
"grad_norm": 0.18639877438545227,
"learning_rate": 1.3291058913772517e-05,
"loss": 1.9262,
"step": 39600
},
{
"epoch": 4.364336818932306,
"grad_norm": 0.19503189623355865,
"learning_rate": 1.306726189084385e-05,
"loss": 1.9768,
"step": 39650
},
{
"epoch": 4.369840396257567,
"grad_norm": 0.1823161244392395,
"learning_rate": 1.2845279232584354e-05,
"loss": 1.9588,
"step": 39700
},
{
"epoch": 4.375343973582829,
"grad_norm": 0.19661200046539307,
"learning_rate": 1.2625113880273786e-05,
"loss": 1.9576,
"step": 39750
},
{
"epoch": 4.38084755090809,
"grad_norm": 0.17353294789791107,
"learning_rate": 1.2406768751112572e-05,
"loss": 1.9249,
"step": 39800
},
{
"epoch": 4.386351128233351,
"grad_norm": 0.1832091063261032,
"learning_rate": 1.2190246738183074e-05,
"loss": 1.9169,
"step": 39850
},
{
"epoch": 4.391854705558613,
"grad_norm": 0.18823185563087463,
"learning_rate": 1.197555071041122e-05,
"loss": 1.9368,
"step": 39900
},
{
"epoch": 4.397358282883874,
"grad_norm": 0.1941109150648117,
"learning_rate": 1.1762683512528642e-05,
"loss": 1.96,
"step": 39950
},
{
"epoch": 4.402861860209136,
"grad_norm": 0.1973351240158081,
"learning_rate": 1.155164796503486e-05,
"loss": 1.9304,
"step": 40000
}
],
"logging_steps": 50,
"max_steps": 45425,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.91746903834624e+17,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}