{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 4.402861860209136, "eval_steps": 500, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00550357732526142, "grad_norm": 5.828019142150879, "learning_rate": 6.47007042253521e-06, "loss": 8.235, "step": 50 }, { "epoch": 0.01100715465052284, "grad_norm": 15.358248710632324, "learning_rate": 1.3072183098591547e-05, "loss": 6.0857, "step": 100 }, { "epoch": 0.01651073197578426, "grad_norm": 7.99273681640625, "learning_rate": 1.9674295774647885e-05, "loss": 4.4315, "step": 150 }, { "epoch": 0.02201430930104568, "grad_norm": 2.8550803661346436, "learning_rate": 2.6276408450704222e-05, "loss": 3.681, "step": 200 }, { "epoch": 0.0275178866263071, "grad_norm": 2.3769114017486572, "learning_rate": 3.2878521126760565e-05, "loss": 3.4729, "step": 250 }, { "epoch": 0.03302146395156852, "grad_norm": 5.98811149597168, "learning_rate": 3.94806338028169e-05, "loss": 3.3829, "step": 300 }, { "epoch": 0.03852504127682994, "grad_norm": 3.616163492202759, "learning_rate": 4.608274647887324e-05, "loss": 3.4119, "step": 350 }, { "epoch": 0.04402861860209136, "grad_norm": 2.1746344566345215, "learning_rate": 5.268485915492957e-05, "loss": 3.3955, "step": 400 }, { "epoch": 0.04953219592735278, "grad_norm": 1.3189276456832886, "learning_rate": 5.928697183098591e-05, "loss": 3.3177, "step": 450 }, { "epoch": 0.0550357732526142, "grad_norm": 1.2459770441055298, "learning_rate": 6.588908450704225e-05, "loss": 3.3095, "step": 500 }, { "epoch": 0.06053935057787562, "grad_norm": 1.2855902910232544, "learning_rate": 7.249119718309858e-05, "loss": 3.2437, "step": 550 }, { "epoch": 0.06604292790313704, "grad_norm": 1.2871235609054565, "learning_rate": 7.909330985915493e-05, "loss": 3.2069, "step": 600 }, { "epoch": 0.07154650522839846, "grad_norm": 1.0846847295761108, "learning_rate": 8.569542253521127e-05, "loss": 3.2512, "step": 650 }, { "epoch": 0.07705008255365987, "grad_norm": 1.5505499839782715, "learning_rate": 9.22975352112676e-05, "loss": 3.1704, "step": 700 }, { "epoch": 0.0825536598789213, "grad_norm": 1.178614616394043, "learning_rate": 9.889964788732394e-05, "loss": 3.1747, "step": 750 }, { "epoch": 0.08805723720418272, "grad_norm": 1.1504727602005005, "learning_rate": 0.00010550176056338028, "loss": 3.1241, "step": 800 }, { "epoch": 0.09356081452944413, "grad_norm": 1.02865469455719, "learning_rate": 0.00011210387323943662, "loss": 3.0481, "step": 850 }, { "epoch": 0.09906439185470556, "grad_norm": 1.1366077661514282, "learning_rate": 0.00011870598591549295, "loss": 3.0201, "step": 900 }, { "epoch": 0.10456796917996698, "grad_norm": 0.9753648042678833, "learning_rate": 0.00012530809859154929, "loss": 3.0145, "step": 950 }, { "epoch": 0.1100715465052284, "grad_norm": 0.6859256625175476, "learning_rate": 0.00013191021126760563, "loss": 2.9671, "step": 1000 }, { "epoch": 0.11557512383048982, "grad_norm": 0.8368203043937683, "learning_rate": 0.00013851232394366197, "loss": 2.992, "step": 1050 }, { "epoch": 0.12107870115575124, "grad_norm": 1.1109174489974976, "learning_rate": 0.0001451144366197183, "loss": 2.967, "step": 1100 }, { "epoch": 0.12658227848101267, "grad_norm": 0.6072912812232971, "learning_rate": 0.00015171654929577465, "loss": 2.9501, "step": 1150 }, { "epoch": 0.13208585580627408, "grad_norm": 0.7659889459609985, "learning_rate": 0.00015831866197183099, "loss": 2.9026, "step": 1200 }, { "epoch": 0.1375894331315355, "grad_norm": 0.5841110348701477, "learning_rate": 0.0001649207746478873, "loss": 2.8972, "step": 1250 }, { "epoch": 0.14309301045679693, "grad_norm": 0.5997458100318909, "learning_rate": 0.00017152288732394364, "loss": 2.9189, "step": 1300 }, { "epoch": 0.14859658778205834, "grad_norm": 0.6082264184951782, "learning_rate": 0.00017812499999999998, "loss": 2.8877, "step": 1350 }, { "epoch": 0.15410016510731975, "grad_norm": 0.696685254573822, "learning_rate": 0.00018472711267605632, "loss": 2.8752, "step": 1400 }, { "epoch": 0.15960374243258119, "grad_norm": 0.6795832514762878, "learning_rate": 0.00019132922535211266, "loss": 2.851, "step": 1450 }, { "epoch": 0.1651073197578426, "grad_norm": 0.5588585734367371, "learning_rate": 0.000197931338028169, "loss": 2.8546, "step": 1500 }, { "epoch": 0.170610897083104, "grad_norm": 0.5267760157585144, "learning_rate": 0.00020453345070422534, "loss": 2.8049, "step": 1550 }, { "epoch": 0.17611447440836545, "grad_norm": 0.591826856136322, "learning_rate": 0.00021113556338028168, "loss": 2.8071, "step": 1600 }, { "epoch": 0.18161805173362686, "grad_norm": 0.5463298559188843, "learning_rate": 0.00021773767605633802, "loss": 2.7667, "step": 1650 }, { "epoch": 0.18712162905888827, "grad_norm": 0.5745858550071716, "learning_rate": 0.00022433978873239433, "loss": 2.7968, "step": 1700 }, { "epoch": 0.1926252063841497, "grad_norm": 0.6239858865737915, "learning_rate": 0.00023094190140845067, "loss": 2.8206, "step": 1750 }, { "epoch": 0.19812878370941112, "grad_norm": 0.46496015787124634, "learning_rate": 0.000237544014084507, "loss": 2.7944, "step": 1800 }, { "epoch": 0.20363236103467253, "grad_norm": 0.5664075016975403, "learning_rate": 0.00024414612676056335, "loss": 2.7837, "step": 1850 }, { "epoch": 0.20913593835993396, "grad_norm": 0.537627100944519, "learning_rate": 0.0002507482394366197, "loss": 2.7179, "step": 1900 }, { "epoch": 0.21463951568519538, "grad_norm": 0.544585645198822, "learning_rate": 0.00025735035211267603, "loss": 2.7552, "step": 1950 }, { "epoch": 0.2201430930104568, "grad_norm": 0.5067969560623169, "learning_rate": 0.0002639524647887324, "loss": 2.7309, "step": 2000 }, { "epoch": 0.22564667033571822, "grad_norm": 0.462003618478775, "learning_rate": 0.0002705545774647887, "loss": 2.716, "step": 2050 }, { "epoch": 0.23115024766097964, "grad_norm": 0.44174736738204956, "learning_rate": 0.00027715669014084505, "loss": 2.7319, "step": 2100 }, { "epoch": 0.23665382498624105, "grad_norm": 0.5010894536972046, "learning_rate": 0.0002837588028169014, "loss": 2.7199, "step": 2150 }, { "epoch": 0.24215740231150248, "grad_norm": 0.44113022089004517, "learning_rate": 0.00029036091549295773, "loss": 2.7146, "step": 2200 }, { "epoch": 0.2476609796367639, "grad_norm": 0.4972345232963562, "learning_rate": 0.0002969630281690141, "loss": 2.6471, "step": 2250 }, { "epoch": 0.25316455696202533, "grad_norm": 0.5614003539085388, "learning_rate": 0.0002999997102213327, "loss": 2.687, "step": 2300 }, { "epoch": 0.25866813428728674, "grad_norm": 0.5223066806793213, "learning_rate": 0.0002999976432184194, "loss": 2.6279, "step": 2350 }, { "epoch": 0.26417171161254815, "grad_norm": 0.4381965100765228, "learning_rate": 0.0002999935887402823, "loss": 2.6889, "step": 2400 }, { "epoch": 0.26967528893780957, "grad_norm": 0.3969985842704773, "learning_rate": 0.00029998754684064345, "loss": 2.6565, "step": 2450 }, { "epoch": 0.275178866263071, "grad_norm": 0.44786953926086426, "learning_rate": 0.00029997951759955823, "loss": 2.6736, "step": 2500 }, { "epoch": 0.28068244358833244, "grad_norm": 0.3951723873615265, "learning_rate": 0.0002999695011234145, "loss": 2.6173, "step": 2550 }, { "epoch": 0.28618602091359385, "grad_norm": 0.4258750081062317, "learning_rate": 0.00029995749754493093, "loss": 2.5992, "step": 2600 }, { "epoch": 0.29168959823885526, "grad_norm": 0.41892707347869873, "learning_rate": 0.0002999435070231555, "loss": 2.6194, "step": 2650 }, { "epoch": 0.2971931755641167, "grad_norm": 0.4064221978187561, "learning_rate": 0.0002999275297434632, "loss": 2.6265, "step": 2700 }, { "epoch": 0.3026967528893781, "grad_norm": 0.3920956254005432, "learning_rate": 0.00029990956591755365, "loss": 2.6035, "step": 2750 }, { "epoch": 0.3082003302146395, "grad_norm": 0.366318941116333, "learning_rate": 0.0002998896157834484, "loss": 2.6352, "step": 2800 }, { "epoch": 0.31370390753990096, "grad_norm": 0.40867865085601807, "learning_rate": 0.0002998676796054875, "loss": 2.5674, "step": 2850 }, { "epoch": 0.31920748486516237, "grad_norm": 0.3685750365257263, "learning_rate": 0.00029984375767432627, "loss": 2.5748, "step": 2900 }, { "epoch": 0.3247110621904238, "grad_norm": 0.38134968280792236, "learning_rate": 0.0002998178503069314, "loss": 2.5703, "step": 2950 }, { "epoch": 0.3302146395156852, "grad_norm": 0.36195963621139526, "learning_rate": 0.00029978995784657643, "loss": 2.5367, "step": 3000 }, { "epoch": 0.3357182168409466, "grad_norm": 0.3696858584880829, "learning_rate": 0.0002997600806628379, "loss": 2.5693, "step": 3050 }, { "epoch": 0.341221794166208, "grad_norm": 0.3404608964920044, "learning_rate": 0.00029972821915158964, "loss": 2.5454, "step": 3100 }, { "epoch": 0.3467253714914695, "grad_norm": 0.33220136165618896, "learning_rate": 0.0002996943737349981, "loss": 2.5675, "step": 3150 }, { "epoch": 0.3522289488167309, "grad_norm": 0.3744208812713623, "learning_rate": 0.00029965854486151643, "loss": 2.5453, "step": 3200 }, { "epoch": 0.3577325261419923, "grad_norm": 0.36920756101608276, "learning_rate": 0.0002996207330058788, "loss": 2.5314, "step": 3250 }, { "epoch": 0.3632361034672537, "grad_norm": 0.36301830410957336, "learning_rate": 0.00029958093866909403, "loss": 2.5361, "step": 3300 }, { "epoch": 0.3687396807925151, "grad_norm": 0.3135142922401428, "learning_rate": 0.00029953916237843853, "loss": 2.5222, "step": 3350 }, { "epoch": 0.37424325811777653, "grad_norm": 0.3634016513824463, "learning_rate": 0.00029949540468745015, "loss": 2.5179, "step": 3400 }, { "epoch": 0.379746835443038, "grad_norm": 0.370914101600647, "learning_rate": 0.00029944966617592017, "loss": 2.5077, "step": 3450 }, { "epoch": 0.3852504127682994, "grad_norm": 0.32538744807243347, "learning_rate": 0.0002994019474498858, "loss": 2.4873, "step": 3500 }, { "epoch": 0.3907539900935608, "grad_norm": 0.3077157139778137, "learning_rate": 0.0002993522491416221, "loss": 2.5149, "step": 3550 }, { "epoch": 0.39625756741882223, "grad_norm": 0.32142704725265503, "learning_rate": 0.000299300571909634, "loss": 2.4989, "step": 3600 }, { "epoch": 0.40176114474408364, "grad_norm": 0.3485497832298279, "learning_rate": 0.00029924691643864684, "loss": 2.5114, "step": 3650 }, { "epoch": 0.40726472206934505, "grad_norm": 0.34483280777931213, "learning_rate": 0.0002991912834395981, "loss": 2.4863, "step": 3700 }, { "epoch": 0.4127682993946065, "grad_norm": 0.30801650881767273, "learning_rate": 0.00029913367364962733, "loss": 2.4935, "step": 3750 }, { "epoch": 0.41827187671986793, "grad_norm": 0.3099140524864197, "learning_rate": 0.00029907408783206674, "loss": 2.5122, "step": 3800 }, { "epoch": 0.42377545404512934, "grad_norm": 0.33073538541793823, "learning_rate": 0.0002990125267764309, "loss": 2.4613, "step": 3850 }, { "epoch": 0.42927903137039075, "grad_norm": 0.3025052845478058, "learning_rate": 0.00029894899129840653, "loss": 2.4855, "step": 3900 }, { "epoch": 0.43478260869565216, "grad_norm": 0.3200172483921051, "learning_rate": 0.00029888348223984143, "loss": 2.4542, "step": 3950 }, { "epoch": 0.4402861860209136, "grad_norm": 0.31750544905662537, "learning_rate": 0.0002988160004687335, "loss": 2.4714, "step": 4000 }, { "epoch": 0.44578976334617504, "grad_norm": 0.28732484579086304, "learning_rate": 0.00029874654687921895, "loss": 2.4595, "step": 4050 }, { "epoch": 0.45129334067143645, "grad_norm": 0.33312809467315674, "learning_rate": 0.0002986751223915609, "loss": 2.4797, "step": 4100 }, { "epoch": 0.45679691799669786, "grad_norm": 0.3109307289123535, "learning_rate": 0.00029860172795213695, "loss": 2.4506, "step": 4150 }, { "epoch": 0.46230049532195927, "grad_norm": 0.31127694249153137, "learning_rate": 0.0002985263645334266, "loss": 2.4544, "step": 4200 }, { "epoch": 0.4678040726472207, "grad_norm": 0.34132882952690125, "learning_rate": 0.0002984490331339982, "loss": 2.455, "step": 4250 }, { "epoch": 0.4733076499724821, "grad_norm": 0.3107958436012268, "learning_rate": 0.00029836973477849634, "loss": 2.4733, "step": 4300 }, { "epoch": 0.47881122729774356, "grad_norm": 0.27321889996528625, "learning_rate": 0.00029828847051762753, "loss": 2.4381, "step": 4350 }, { "epoch": 0.48431480462300497, "grad_norm": 0.3114171326160431, "learning_rate": 0.0002982052414281467, "loss": 2.4612, "step": 4400 }, { "epoch": 0.4898183819482664, "grad_norm": 0.2926501929759979, "learning_rate": 0.00029812004861284294, "loss": 2.4364, "step": 4450 }, { "epoch": 0.4953219592735278, "grad_norm": 0.2824021279811859, "learning_rate": 0.00029803289320052466, "loss": 2.4453, "step": 4500 }, { "epoch": 0.5008255365987893, "grad_norm": 0.2710123062133789, "learning_rate": 0.00029794377634600485, "loss": 2.4207, "step": 4550 }, { "epoch": 0.5063291139240507, "grad_norm": 0.28232479095458984, "learning_rate": 0.0002978526992300856, "loss": 2.4332, "step": 4600 }, { "epoch": 0.5118326912493121, "grad_norm": 0.2661692798137665, "learning_rate": 0.0002977596630595427, "loss": 2.4469, "step": 4650 }, { "epoch": 0.5173362685745735, "grad_norm": 0.317094087600708, "learning_rate": 0.0002976646690671094, "loss": 2.4486, "step": 4700 }, { "epoch": 0.5228398458998349, "grad_norm": 0.268915593624115, "learning_rate": 0.0002975677185114602, "loss": 2.4447, "step": 4750 }, { "epoch": 0.5283434232250963, "grad_norm": 0.2714874744415283, "learning_rate": 0.000297468812677194, "loss": 2.4065, "step": 4800 }, { "epoch": 0.5338470005503577, "grad_norm": 0.2872399389743805, "learning_rate": 0.0002973679528748175, "loss": 2.3943, "step": 4850 }, { "epoch": 0.5393505778756191, "grad_norm": 0.24350598454475403, "learning_rate": 0.00029726514044072736, "loss": 2.3888, "step": 4900 }, { "epoch": 0.5448541552008805, "grad_norm": 0.2634688913822174, "learning_rate": 0.00029716037673719275, "loss": 2.3953, "step": 4950 }, { "epoch": 0.550357732526142, "grad_norm": 0.2842876613140106, "learning_rate": 0.0002970536631523373, "loss": 2.403, "step": 5000 }, { "epoch": 0.5558613098514034, "grad_norm": 0.30130457878112793, "learning_rate": 0.00029694500110012055, "loss": 2.4393, "step": 5050 }, { "epoch": 0.5613648871766649, "grad_norm": 0.2594560980796814, "learning_rate": 0.00029683439202031936, "loss": 2.3827, "step": 5100 }, { "epoch": 0.5668684645019263, "grad_norm": 0.25417500734329224, "learning_rate": 0.0002967218373785088, "loss": 2.3789, "step": 5150 }, { "epoch": 0.5723720418271877, "grad_norm": 0.2533874213695526, "learning_rate": 0.0002966073386660428, "loss": 2.4211, "step": 5200 }, { "epoch": 0.5778756191524491, "grad_norm": 0.2343342900276184, "learning_rate": 0.0002964908974000341, "loss": 2.4332, "step": 5250 }, { "epoch": 0.5833791964777105, "grad_norm": 0.25453534722328186, "learning_rate": 0.0002963725151233345, "loss": 2.3984, "step": 5300 }, { "epoch": 0.5888827738029719, "grad_norm": 0.2670257091522217, "learning_rate": 0.00029625219340451435, "loss": 2.4126, "step": 5350 }, { "epoch": 0.5943863511282333, "grad_norm": 0.2728422284126282, "learning_rate": 0.00029612993383784154, "loss": 2.4152, "step": 5400 }, { "epoch": 0.5998899284534948, "grad_norm": 0.27964428067207336, "learning_rate": 0.0002960057380432606, "loss": 2.3675, "step": 5450 }, { "epoch": 0.6053935057787562, "grad_norm": 0.27354755997657776, "learning_rate": 0.00029587960766637103, "loss": 2.4109, "step": 5500 }, { "epoch": 0.6108970831040176, "grad_norm": 0.26175597310066223, "learning_rate": 0.000295751544378406, "loss": 2.3953, "step": 5550 }, { "epoch": 0.616400660429279, "grad_norm": 0.2590219974517822, "learning_rate": 0.0002956215498762093, "loss": 2.3699, "step": 5600 }, { "epoch": 0.6219042377545404, "grad_norm": 0.2495882660150528, "learning_rate": 0.0002954896258822139, "loss": 2.3949, "step": 5650 }, { "epoch": 0.6274078150798019, "grad_norm": 0.26052239537239075, "learning_rate": 0.0002953557741444183, "loss": 2.3642, "step": 5700 }, { "epoch": 0.6329113924050633, "grad_norm": 0.2627600431442261, "learning_rate": 0.0002952199964363638, "loss": 2.3895, "step": 5750 }, { "epoch": 0.6384149697303247, "grad_norm": 0.2417898327112198, "learning_rate": 0.00029508229455711086, "loss": 2.4087, "step": 5800 }, { "epoch": 0.6439185470555862, "grad_norm": 0.24891215562820435, "learning_rate": 0.00029494267033121525, "loss": 2.3642, "step": 5850 }, { "epoch": 0.6494221243808476, "grad_norm": 0.2453078180551529, "learning_rate": 0.0002948011256087041, "loss": 2.3761, "step": 5900 }, { "epoch": 0.654925701706109, "grad_norm": 0.24607980251312256, "learning_rate": 0.0002946576622650509, "loss": 2.3596, "step": 5950 }, { "epoch": 0.6604292790313704, "grad_norm": 0.2540590763092041, "learning_rate": 0.0002945122822011513, "loss": 2.3666, "step": 6000 }, { "epoch": 0.6659328563566318, "grad_norm": 0.25918814539909363, "learning_rate": 0.00029436498734329727, "loss": 2.3494, "step": 6050 }, { "epoch": 0.6714364336818932, "grad_norm": 0.27554330229759216, "learning_rate": 0.0002942157796431521, "loss": 2.3793, "step": 6100 }, { "epoch": 0.6769400110071546, "grad_norm": 0.22879928350448608, "learning_rate": 0.00029406466107772416, "loss": 2.3898, "step": 6150 }, { "epoch": 0.682443588332416, "grad_norm": 0.2851306200027466, "learning_rate": 0.00029391163364934095, "loss": 2.3493, "step": 6200 }, { "epoch": 0.6879471656576774, "grad_norm": 0.255014568567276, "learning_rate": 0.0002937566993856225, "loss": 2.3959, "step": 6250 }, { "epoch": 0.693450742982939, "grad_norm": 0.23992781341075897, "learning_rate": 0.00029359986033945454, "loss": 2.3588, "step": 6300 }, { "epoch": 0.6989543203082004, "grad_norm": 0.24902793765068054, "learning_rate": 0.000293441118588961, "loss": 2.3908, "step": 6350 }, { "epoch": 0.7044578976334618, "grad_norm": 0.2294873744249344, "learning_rate": 0.0002932804762374771, "loss": 2.398, "step": 6400 }, { "epoch": 0.7099614749587232, "grad_norm": 0.23560309410095215, "learning_rate": 0.00029311793541352075, "loss": 2.3481, "step": 6450 }, { "epoch": 0.7154650522839846, "grad_norm": 0.27493488788604736, "learning_rate": 0.00029295349827076497, "loss": 2.3248, "step": 6500 }, { "epoch": 0.720968629609246, "grad_norm": 0.22731854021549225, "learning_rate": 0.0002927871669880089, "loss": 2.3444, "step": 6550 }, { "epoch": 0.7264722069345074, "grad_norm": 0.22198539972305298, "learning_rate": 0.0002926189437691492, "loss": 2.3148, "step": 6600 }, { "epoch": 0.7319757842597688, "grad_norm": 0.25049686431884766, "learning_rate": 0.0002924488308431508, "loss": 2.3134, "step": 6650 }, { "epoch": 0.7374793615850302, "grad_norm": 0.2354484349489212, "learning_rate": 0.0002922768304640172, "loss": 2.348, "step": 6700 }, { "epoch": 0.7429829389102917, "grad_norm": 0.23955915868282318, "learning_rate": 0.00029210294491076094, "loss": 2.3172, "step": 6750 }, { "epoch": 0.7484865162355531, "grad_norm": 0.2476130574941635, "learning_rate": 0.000291927176487373, "loss": 2.3842, "step": 6800 }, { "epoch": 0.7539900935608145, "grad_norm": 0.2247840166091919, "learning_rate": 0.00029174952752279254, "loss": 2.3302, "step": 6850 }, { "epoch": 0.759493670886076, "grad_norm": 0.22530816495418549, "learning_rate": 0.000291570000370876, "loss": 2.3707, "step": 6900 }, { "epoch": 0.7649972482113374, "grad_norm": 0.2242722362279892, "learning_rate": 0.00029138859741036587, "loss": 2.312, "step": 6950 }, { "epoch": 0.7705008255365988, "grad_norm": 0.23376357555389404, "learning_rate": 0.0002912053210448592, "loss": 2.3334, "step": 7000 }, { "epoch": 0.7760044028618602, "grad_norm": 0.23131632804870605, "learning_rate": 0.0002910201737027757, "loss": 2.3293, "step": 7050 }, { "epoch": 0.7815079801871216, "grad_norm": 0.21440596878528595, "learning_rate": 0.0002908331578373256, "loss": 2.3593, "step": 7100 }, { "epoch": 0.787011557512383, "grad_norm": 0.225584477186203, "learning_rate": 0.00029064427592647715, "loss": 2.3464, "step": 7150 }, { "epoch": 0.7925151348376445, "grad_norm": 0.22819140553474426, "learning_rate": 0.0002904535304729238, "loss": 2.3145, "step": 7200 }, { "epoch": 0.7980187121629059, "grad_norm": 0.2503032982349396, "learning_rate": 0.00029026092400405115, "loss": 2.3376, "step": 7250 }, { "epoch": 0.8035222894881673, "grad_norm": 0.2061757743358612, "learning_rate": 0.00029006645907190295, "loss": 2.3252, "step": 7300 }, { "epoch": 0.8090258668134287, "grad_norm": 0.2099982351064682, "learning_rate": 0.00028987013825314804, "loss": 2.35, "step": 7350 }, { "epoch": 0.8145294441386901, "grad_norm": 0.212050199508667, "learning_rate": 0.00028967196414904573, "loss": 2.3304, "step": 7400 }, { "epoch": 0.8200330214639515, "grad_norm": 0.22283074259757996, "learning_rate": 0.0002894719393854114, "loss": 2.3389, "step": 7450 }, { "epoch": 0.825536598789213, "grad_norm": 0.2060365378856659, "learning_rate": 0.0002892700666125817, "loss": 2.3215, "step": 7500 }, { "epoch": 0.8310401761144744, "grad_norm": 0.22510799765586853, "learning_rate": 0.0002890663485053797, "loss": 2.3295, "step": 7550 }, { "epoch": 0.8365437534397359, "grad_norm": 0.21860332787036896, "learning_rate": 0.00028886078776307905, "loss": 2.3286, "step": 7600 }, { "epoch": 0.8420473307649973, "grad_norm": 0.20771068334579468, "learning_rate": 0.00028865338710936826, "loss": 2.3143, "step": 7650 }, { "epoch": 0.8475509080902587, "grad_norm": 0.23098550736904144, "learning_rate": 0.0002884441492923151, "loss": 2.3583, "step": 7700 }, { "epoch": 0.8530544854155201, "grad_norm": 0.204328715801239, "learning_rate": 0.00028823307708432963, "loss": 2.3425, "step": 7750 }, { "epoch": 0.8585580627407815, "grad_norm": 0.20610789954662323, "learning_rate": 0.0002880201732821275, "loss": 2.3269, "step": 7800 }, { "epoch": 0.8640616400660429, "grad_norm": 0.22584037482738495, "learning_rate": 0.0002878054407066935, "loss": 2.3061, "step": 7850 }, { "epoch": 0.8695652173913043, "grad_norm": 0.22317476570606232, "learning_rate": 0.0002875888822032433, "loss": 2.3047, "step": 7900 }, { "epoch": 0.8750687947165657, "grad_norm": 0.2132745087146759, "learning_rate": 0.00028737050064118645, "loss": 2.2925, "step": 7950 }, { "epoch": 0.8805723720418271, "grad_norm": 0.21960894763469696, "learning_rate": 0.00028715029891408805, "loss": 2.3365, "step": 8000 }, { "epoch": 0.8860759493670886, "grad_norm": 0.2113264799118042, "learning_rate": 0.0002869282799396305, "loss": 2.2814, "step": 8050 }, { "epoch": 0.8915795266923501, "grad_norm": 0.19959582388401031, "learning_rate": 0.00028670444665957465, "loss": 2.2787, "step": 8100 }, { "epoch": 0.8970831040176115, "grad_norm": 0.24770890176296234, "learning_rate": 0.00028647880203972115, "loss": 2.3312, "step": 8150 }, { "epoch": 0.9025866813428729, "grad_norm": 0.2229924350976944, "learning_rate": 0.0002862513490698709, "loss": 2.2825, "step": 8200 }, { "epoch": 0.9080902586681343, "grad_norm": 0.23118580877780914, "learning_rate": 0.0002860220907637856, "loss": 2.2973, "step": 8250 }, { "epoch": 0.9135938359933957, "grad_norm": 0.1978590488433838, "learning_rate": 0.0002857910301591475, "loss": 2.3106, "step": 8300 }, { "epoch": 0.9190974133186571, "grad_norm": 0.20972274243831635, "learning_rate": 0.0002855581703175198, "loss": 2.3489, "step": 8350 }, { "epoch": 0.9246009906439185, "grad_norm": 0.20369485020637512, "learning_rate": 0.0002853235143243052, "loss": 2.3135, "step": 8400 }, { "epoch": 0.93010456796918, "grad_norm": 0.19085292518138885, "learning_rate": 0.00028508706528870576, "loss": 2.2969, "step": 8450 }, { "epoch": 0.9356081452944414, "grad_norm": 0.21080902218818665, "learning_rate": 0.0002848488263436814, "loss": 2.2979, "step": 8500 }, { "epoch": 0.9411117226197028, "grad_norm": 0.1939728856086731, "learning_rate": 0.00028460880064590835, "loss": 2.31, "step": 8550 }, { "epoch": 0.9466152999449642, "grad_norm": 0.23909969627857208, "learning_rate": 0.0002843669913757375, "loss": 2.2585, "step": 8600 }, { "epoch": 0.9521188772702256, "grad_norm": 0.22809971868991852, "learning_rate": 0.000284123401737152, "loss": 2.3058, "step": 8650 }, { "epoch": 0.9576224545954871, "grad_norm": 0.2066497504711151, "learning_rate": 0.00028387803495772513, "loss": 2.2736, "step": 8700 }, { "epoch": 0.9631260319207485, "grad_norm": 0.1921154260635376, "learning_rate": 0.00028363089428857716, "loss": 2.2813, "step": 8750 }, { "epoch": 0.9686296092460099, "grad_norm": 0.19423851370811462, "learning_rate": 0.00028338198300433264, "loss": 2.314, "step": 8800 }, { "epoch": 0.9741331865712713, "grad_norm": 0.224751815199852, "learning_rate": 0.0002831313044030768, "loss": 2.3335, "step": 8850 }, { "epoch": 0.9796367638965328, "grad_norm": 0.21634523570537567, "learning_rate": 0.00028287886180631183, "loss": 2.3021, "step": 8900 }, { "epoch": 0.9851403412217942, "grad_norm": 0.19320085644721985, "learning_rate": 0.00028262465855891306, "loss": 2.2635, "step": 8950 }, { "epoch": 0.9906439185470556, "grad_norm": 0.20169596374034882, "learning_rate": 0.00028236869802908453, "loss": 2.3013, "step": 9000 }, { "epoch": 0.996147495872317, "grad_norm": 0.2045382857322693, "learning_rate": 0.00028211098360831407, "loss": 2.2576, "step": 9050 }, { "epoch": 1.0016510731975785, "grad_norm": 0.2171631157398224, "learning_rate": 0.00028185151871132894, "loss": 2.3136, "step": 9100 }, { "epoch": 1.00715465052284, "grad_norm": 0.19541560113430023, "learning_rate": 0.00028159030677605017, "loss": 2.2547, "step": 9150 }, { "epoch": 1.0126582278481013, "grad_norm": 0.19888882339000702, "learning_rate": 0.0002813273512635472, "loss": 2.2275, "step": 9200 }, { "epoch": 1.0181618051733627, "grad_norm": 0.24578270316123962, "learning_rate": 0.0002810626556579918, "loss": 2.2775, "step": 9250 }, { "epoch": 1.0236653824986242, "grad_norm": 0.19902436435222626, "learning_rate": 0.00028079622346661217, "loss": 2.2353, "step": 9300 }, { "epoch": 1.0291689598238856, "grad_norm": 0.19234856963157654, "learning_rate": 0.00028052805821964633, "loss": 2.2769, "step": 9350 }, { "epoch": 1.034672537149147, "grad_norm": 0.19912928342819214, "learning_rate": 0.0002802581634702952, "loss": 2.2625, "step": 9400 }, { "epoch": 1.0401761144744084, "grad_norm": 0.1918436884880066, "learning_rate": 0.00027998654279467604, "loss": 2.2335, "step": 9450 }, { "epoch": 1.0456796917996698, "grad_norm": 0.20422010123729706, "learning_rate": 0.0002797131997917743, "loss": 2.3022, "step": 9500 }, { "epoch": 1.0511832691249312, "grad_norm": 0.19621697068214417, "learning_rate": 0.00027943813808339657, "loss": 2.2378, "step": 9550 }, { "epoch": 1.0566868464501926, "grad_norm": 0.18403683602809906, "learning_rate": 0.0002791613613141223, "loss": 2.2502, "step": 9600 }, { "epoch": 1.062190423775454, "grad_norm": 0.2140798568725586, "learning_rate": 0.0002788828731512556, "loss": 2.2722, "step": 9650 }, { "epoch": 1.0676940011007154, "grad_norm": 0.19459135830402374, "learning_rate": 0.0002786026772847767, "loss": 2.2361, "step": 9700 }, { "epoch": 1.0731975784259769, "grad_norm": 0.20383061468601227, "learning_rate": 0.00027832077742729277, "loss": 2.2471, "step": 9750 }, { "epoch": 1.0787011557512383, "grad_norm": 0.1901775300502777, "learning_rate": 0.0002780371773139891, "loss": 2.2968, "step": 9800 }, { "epoch": 1.0842047330764997, "grad_norm": 0.20100219547748566, "learning_rate": 0.0002777518807025793, "loss": 2.2568, "step": 9850 }, { "epoch": 1.089708310401761, "grad_norm": 0.2127520740032196, "learning_rate": 0.00027746489137325586, "loss": 2.28, "step": 9900 }, { "epoch": 1.0952118877270225, "grad_norm": 0.19059176743030548, "learning_rate": 0.00027717621312863965, "loss": 2.284, "step": 9950 }, { "epoch": 1.100715465052284, "grad_norm": 0.20845113694667816, "learning_rate": 0.0002768858497937299, "loss": 2.2409, "step": 10000 }, { "epoch": 1.1062190423775453, "grad_norm": 0.18969906866550446, "learning_rate": 0.00027659380521585325, "loss": 2.2725, "step": 10050 }, { "epoch": 1.1117226197028067, "grad_norm": 0.19071218371391296, "learning_rate": 0.0002763000832646129, "loss": 2.2624, "step": 10100 }, { "epoch": 1.1172261970280681, "grad_norm": 0.21723783016204834, "learning_rate": 0.0002760046878318373, "loss": 2.2453, "step": 10150 }, { "epoch": 1.1227297743533298, "grad_norm": 0.19336852431297302, "learning_rate": 0.00027570762283152874, "loss": 2.2322, "step": 10200 }, { "epoch": 1.1282333516785912, "grad_norm": 0.19500704109668732, "learning_rate": 0.0002754088921998112, "loss": 2.1858, "step": 10250 }, { "epoch": 1.1337369290038526, "grad_norm": 0.19920021295547485, "learning_rate": 0.0002751084998948784, "loss": 2.2286, "step": 10300 }, { "epoch": 1.139240506329114, "grad_norm": 0.18863485753536224, "learning_rate": 0.00027480644989694126, "loss": 2.2623, "step": 10350 }, { "epoch": 1.1447440836543754, "grad_norm": 0.18571630120277405, "learning_rate": 0.0002745027462081753, "loss": 2.2616, "step": 10400 }, { "epoch": 1.1502476609796368, "grad_norm": 0.187313050031662, "learning_rate": 0.00027419739285266745, "loss": 2.2608, "step": 10450 }, { "epoch": 1.1557512383048982, "grad_norm": 0.19139795005321503, "learning_rate": 0.0002738903938763628, "loss": 2.2204, "step": 10500 }, { "epoch": 1.1612548156301596, "grad_norm": 0.188013955950737, "learning_rate": 0.0002735817533470109, "loss": 2.2596, "step": 10550 }, { "epoch": 1.166758392955421, "grad_norm": 0.1884177178144455, "learning_rate": 0.0002732714753541122, "loss": 2.2618, "step": 10600 }, { "epoch": 1.1722619702806825, "grad_norm": 0.19187124073505402, "learning_rate": 0.00027295956400886335, "loss": 2.2075, "step": 10650 }, { "epoch": 1.1777655476059439, "grad_norm": 0.19815993309020996, "learning_rate": 0.0002726460234441031, "loss": 2.2677, "step": 10700 }, { "epoch": 1.1832691249312053, "grad_norm": 0.18218094110488892, "learning_rate": 0.00027233085781425744, "loss": 2.2697, "step": 10750 }, { "epoch": 1.1887727022564667, "grad_norm": 0.1923886239528656, "learning_rate": 0.0002720140712952845, "loss": 2.2127, "step": 10800 }, { "epoch": 1.194276279581728, "grad_norm": 0.18907909095287323, "learning_rate": 0.0002716956680846193, "loss": 2.1989, "step": 10850 }, { "epoch": 1.1997798569069895, "grad_norm": 0.1902381181716919, "learning_rate": 0.000271375652401118, "loss": 2.2211, "step": 10900 }, { "epoch": 1.205283434232251, "grad_norm": 0.20184637606143951, "learning_rate": 0.0002710540284850023, "loss": 2.2451, "step": 10950 }, { "epoch": 1.2107870115575123, "grad_norm": 0.19903656840324402, "learning_rate": 0.00027073080059780285, "loss": 2.2233, "step": 11000 }, { "epoch": 1.2162905888827737, "grad_norm": 0.19023925065994263, "learning_rate": 0.000270405973022303, "loss": 2.2438, "step": 11050 }, { "epoch": 1.2217941662080352, "grad_norm": 0.17722088098526, "learning_rate": 0.0002700795500624822, "loss": 2.259, "step": 11100 }, { "epoch": 1.2272977435332966, "grad_norm": 0.19715473055839539, "learning_rate": 0.0002697515360434587, "loss": 2.2509, "step": 11150 }, { "epoch": 1.232801320858558, "grad_norm": 0.18065761029720306, "learning_rate": 0.00026942193531143225, "loss": 2.2006, "step": 11200 }, { "epoch": 1.2383048981838194, "grad_norm": 0.18757817149162292, "learning_rate": 0.00026909075223362683, "loss": 2.2215, "step": 11250 }, { "epoch": 1.243808475509081, "grad_norm": 0.18816334009170532, "learning_rate": 0.00026875799119823243, "loss": 2.2314, "step": 11300 }, { "epoch": 1.2493120528343424, "grad_norm": 0.19772210717201233, "learning_rate": 0.0002684236566143471, "loss": 2.2256, "step": 11350 }, { "epoch": 1.2548156301596038, "grad_norm": 0.19510945677757263, "learning_rate": 0.0002680877529119184, "loss": 2.2427, "step": 11400 }, { "epoch": 1.2603192074848653, "grad_norm": 0.19960781931877136, "learning_rate": 0.0002677502845416849, "loss": 2.2324, "step": 11450 }, { "epoch": 1.2658227848101267, "grad_norm": 0.1882307380437851, "learning_rate": 0.0002674112559751169, "loss": 2.2432, "step": 11500 }, { "epoch": 1.271326362135388, "grad_norm": 0.18959587812423706, "learning_rate": 0.00026707067170435767, "loss": 2.19, "step": 11550 }, { "epoch": 1.2768299394606495, "grad_norm": 0.18768323957920074, "learning_rate": 0.0002667285362421634, "loss": 2.2294, "step": 11600 }, { "epoch": 1.282333516785911, "grad_norm": 0.20077955722808838, "learning_rate": 0.00026638485412184355, "loss": 2.2518, "step": 11650 }, { "epoch": 1.2878370941111723, "grad_norm": 0.1956523358821869, "learning_rate": 0.00026603962989720105, "loss": 2.2662, "step": 11700 }, { "epoch": 1.2933406714364337, "grad_norm": 0.19226029515266418, "learning_rate": 0.0002656928681424718, "loss": 2.2133, "step": 11750 }, { "epoch": 1.2988442487616951, "grad_norm": 0.16987648606300354, "learning_rate": 0.0002653445734522639, "loss": 2.2226, "step": 11800 }, { "epoch": 1.3043478260869565, "grad_norm": 0.17524655163288116, "learning_rate": 0.00026499475044149703, "loss": 2.2555, "step": 11850 }, { "epoch": 1.309851403412218, "grad_norm": 0.19188162684440613, "learning_rate": 0.00026464340374534104, "loss": 2.2234, "step": 11900 }, { "epoch": 1.3153549807374794, "grad_norm": 0.17676417529582977, "learning_rate": 0.00026429053801915487, "loss": 2.2188, "step": 11950 }, { "epoch": 1.3208585580627408, "grad_norm": 0.20071938633918762, "learning_rate": 0.0002639361579384245, "loss": 2.234, "step": 12000 }, { "epoch": 1.3263621353880022, "grad_norm": 0.19267459213733673, "learning_rate": 0.0002635802681987012, "loss": 2.235, "step": 12050 }, { "epoch": 1.3318657127132636, "grad_norm": 0.1724405139684677, "learning_rate": 0.00026322287351553944, "loss": 2.2388, "step": 12100 }, { "epoch": 1.337369290038525, "grad_norm": 0.1907806396484375, "learning_rate": 0.000262863978624434, "loss": 2.1937, "step": 12150 }, { "epoch": 1.3428728673637864, "grad_norm": 0.18452796339988708, "learning_rate": 0.0002625035882807578, "loss": 2.2429, "step": 12200 }, { "epoch": 1.3483764446890478, "grad_norm": 0.17565739154815674, "learning_rate": 0.0002621417072596982, "loss": 2.2367, "step": 12250 }, { "epoch": 1.3538800220143092, "grad_norm": 0.1761862337589264, "learning_rate": 0.00026177834035619445, "loss": 2.2258, "step": 12300 }, { "epoch": 1.3593835993395706, "grad_norm": 0.18616576492786407, "learning_rate": 0.0002614134923848736, "loss": 2.2207, "step": 12350 }, { "epoch": 1.364887176664832, "grad_norm": 0.18103408813476562, "learning_rate": 0.00026104716817998696, "loss": 2.2346, "step": 12400 }, { "epoch": 1.3703907539900935, "grad_norm": 0.18511448800563812, "learning_rate": 0.00026067937259534595, "loss": 2.2363, "step": 12450 }, { "epoch": 1.3758943313153549, "grad_norm": 0.1737717241048813, "learning_rate": 0.00026031011050425796, "loss": 2.2199, "step": 12500 }, { "epoch": 1.3813979086406163, "grad_norm": 0.18761631846427917, "learning_rate": 0.00025993938679946154, "loss": 2.228, "step": 12550 }, { "epoch": 1.3869014859658777, "grad_norm": 0.18815304338932037, "learning_rate": 0.0002595672063930617, "loss": 2.2034, "step": 12600 }, { "epoch": 1.3924050632911391, "grad_norm": 0.1868327558040619, "learning_rate": 0.00025919357421646486, "loss": 2.197, "step": 12650 }, { "epoch": 1.3979086406164005, "grad_norm": 0.17544154822826385, "learning_rate": 0.00025881849522031345, "loss": 2.1956, "step": 12700 }, { "epoch": 1.4034122179416622, "grad_norm": 0.18338458240032196, "learning_rate": 0.00025844197437442033, "loss": 2.2291, "step": 12750 }, { "epoch": 1.4089157952669236, "grad_norm": 0.17127631604671478, "learning_rate": 0.00025806401666770277, "loss": 2.2205, "step": 12800 }, { "epoch": 1.414419372592185, "grad_norm": 0.17331157624721527, "learning_rate": 0.00025768462710811673, "loss": 2.2356, "step": 12850 }, { "epoch": 1.4199229499174464, "grad_norm": 0.16098715364933014, "learning_rate": 0.00025730381072259026, "loss": 2.2065, "step": 12900 }, { "epoch": 1.4254265272427078, "grad_norm": 0.17098484933376312, "learning_rate": 0.00025692157255695667, "loss": 2.2323, "step": 12950 }, { "epoch": 1.4309301045679692, "grad_norm": 0.15763573348522186, "learning_rate": 0.00025653791767588823, "loss": 2.2064, "step": 13000 }, { "epoch": 1.4364336818932306, "grad_norm": 0.17587807774543762, "learning_rate": 0.0002561528511628286, "loss": 2.2085, "step": 13050 }, { "epoch": 1.441937259218492, "grad_norm": 0.17770230770111084, "learning_rate": 0.00025576637811992555, "loss": 2.2204, "step": 13100 }, { "epoch": 1.4474408365437534, "grad_norm": 0.18236953020095825, "learning_rate": 0.0002553785036679636, "loss": 2.2388, "step": 13150 }, { "epoch": 1.4529444138690149, "grad_norm": 0.16505366563796997, "learning_rate": 0.00025498923294629594, "loss": 2.2345, "step": 13200 }, { "epoch": 1.4584479911942763, "grad_norm": 0.1687227189540863, "learning_rate": 0.0002545985711127763, "loss": 2.1652, "step": 13250 }, { "epoch": 1.4639515685195377, "grad_norm": 0.17849218845367432, "learning_rate": 0.00025420652334369085, "loss": 2.2115, "step": 13300 }, { "epoch": 1.469455145844799, "grad_norm": 0.17708005011081696, "learning_rate": 0.0002538130948336894, "loss": 2.2017, "step": 13350 }, { "epoch": 1.4749587231700605, "grad_norm": 0.17772626876831055, "learning_rate": 0.0002534182907957165, "loss": 2.1926, "step": 13400 }, { "epoch": 1.480462300495322, "grad_norm": 0.16764992475509644, "learning_rate": 0.00025302211646094277, "loss": 2.1986, "step": 13450 }, { "epoch": 1.4859658778205833, "grad_norm": 0.19117778539657593, "learning_rate": 0.00025262457707869506, "loss": 2.2118, "step": 13500 }, { "epoch": 1.4914694551458447, "grad_norm": 0.17632248997688293, "learning_rate": 0.00025222567791638743, "loss": 2.1897, "step": 13550 }, { "epoch": 1.4969730324711064, "grad_norm": 0.19731369614601135, "learning_rate": 0.0002518254242594508, "loss": 2.1945, "step": 13600 }, { "epoch": 1.5024766097963678, "grad_norm": 0.16187229752540588, "learning_rate": 0.0002514238214112635, "loss": 2.1959, "step": 13650 }, { "epoch": 1.5079801871216292, "grad_norm": 0.1828337460756302, "learning_rate": 0.00025102087469308036, "loss": 2.1994, "step": 13700 }, { "epoch": 1.5134837644468906, "grad_norm": 0.17188578844070435, "learning_rate": 0.0002506165894439628, "loss": 2.2087, "step": 13750 }, { "epoch": 1.518987341772152, "grad_norm": 0.17914508283138275, "learning_rate": 0.00025021097102070786, "loss": 2.2266, "step": 13800 }, { "epoch": 1.5244909190974134, "grad_norm": 0.17482663691043854, "learning_rate": 0.0002498040247977769, "loss": 2.2431, "step": 13850 }, { "epoch": 1.5299944964226748, "grad_norm": 0.17342698574066162, "learning_rate": 0.00024939575616722505, "loss": 2.2045, "step": 13900 }, { "epoch": 1.5354980737479362, "grad_norm": 0.18258632719516754, "learning_rate": 0.00024898617053862904, "loss": 2.1843, "step": 13950 }, { "epoch": 1.5410016510731976, "grad_norm": 0.16960634291172028, "learning_rate": 0.00024857527333901625, "loss": 2.2096, "step": 14000 }, { "epoch": 1.546505228398459, "grad_norm": 0.17008773982524872, "learning_rate": 0.00024816307001279214, "loss": 2.1522, "step": 14050 }, { "epoch": 1.5520088057237205, "grad_norm": 0.16970837116241455, "learning_rate": 0.00024774956602166857, "loss": 2.2055, "step": 14100 }, { "epoch": 1.5575123830489819, "grad_norm": 0.17486636340618134, "learning_rate": 0.0002473347668445912, "loss": 2.1906, "step": 14150 }, { "epoch": 1.5630159603742433, "grad_norm": 0.19200065732002258, "learning_rate": 0.00024691867797766684, "loss": 2.1839, "step": 14200 }, { "epoch": 1.5685195376995047, "grad_norm": 0.19441814720630646, "learning_rate": 0.00024650130493409104, "loss": 2.2121, "step": 14250 }, { "epoch": 1.574023115024766, "grad_norm": 0.16507557034492493, "learning_rate": 0.0002460826532440745, "loss": 2.2016, "step": 14300 }, { "epoch": 1.5795266923500275, "grad_norm": 0.16732299327850342, "learning_rate": 0.00024566272845477014, "loss": 2.2362, "step": 14350 }, { "epoch": 1.585030269675289, "grad_norm": 0.18136867880821228, "learning_rate": 0.00024524153613019947, "loss": 2.193, "step": 14400 }, { "epoch": 1.5905338470005503, "grad_norm": 0.16409073770046234, "learning_rate": 0.0002448190818511789, "loss": 2.1589, "step": 14450 }, { "epoch": 1.5960374243258117, "grad_norm": 0.17381897568702698, "learning_rate": 0.00024439537121524583, "loss": 2.1834, "step": 14500 }, { "epoch": 1.6015410016510732, "grad_norm": 0.1628628373146057, "learning_rate": 0.00024397040983658436, "loss": 2.172, "step": 14550 }, { "epoch": 1.6070445789763346, "grad_norm": 0.16793246567249298, "learning_rate": 0.00024354420334595105, "loss": 2.176, "step": 14600 }, { "epoch": 1.612548156301596, "grad_norm": 0.17394863069057465, "learning_rate": 0.0002431167573906002, "loss": 2.199, "step": 14650 }, { "epoch": 1.6180517336268574, "grad_norm": 0.1554850935935974, "learning_rate": 0.00024268807763420914, "loss": 2.2042, "step": 14700 }, { "epoch": 1.6235553109521188, "grad_norm": 0.17003560066223145, "learning_rate": 0.00024225816975680312, "loss": 2.1717, "step": 14750 }, { "epoch": 1.6290588882773802, "grad_norm": 0.17207174003124237, "learning_rate": 0.0002418270394546799, "loss": 2.1894, "step": 14800 }, { "epoch": 1.6345624656026416, "grad_norm": 0.1507563292980194, "learning_rate": 0.00024139469244033467, "loss": 2.2101, "step": 14850 }, { "epoch": 1.640066042927903, "grad_norm": 0.17063497006893158, "learning_rate": 0.0002409611344423838, "loss": 2.1854, "step": 14900 }, { "epoch": 1.6455696202531644, "grad_norm": 0.17093567550182343, "learning_rate": 0.00024052637120548954, "loss": 2.2028, "step": 14950 }, { "epoch": 1.6510731975784259, "grad_norm": 0.17256326973438263, "learning_rate": 0.00024009040849028341, "loss": 2.2102, "step": 15000 }, { "epoch": 1.6565767749036873, "grad_norm": 0.17648595571517944, "learning_rate": 0.00023965325207329016, "loss": 2.2094, "step": 15050 }, { "epoch": 1.6620803522289487, "grad_norm": 0.17140266299247742, "learning_rate": 0.0002392149077468511, "loss": 2.1936, "step": 15100 }, { "epoch": 1.66758392955421, "grad_norm": 0.16277456283569336, "learning_rate": 0.00023877538131904743, "loss": 2.1705, "step": 15150 }, { "epoch": 1.6730875068794715, "grad_norm": 0.16912780702114105, "learning_rate": 0.00023833467861362323, "loss": 2.1834, "step": 15200 }, { "epoch": 1.6785910842047331, "grad_norm": 0.1684638261795044, "learning_rate": 0.00023789280546990824, "loss": 2.2044, "step": 15250 }, { "epoch": 1.6840946615299945, "grad_norm": 0.16793768107891083, "learning_rate": 0.00023744976774274065, "loss": 2.188, "step": 15300 }, { "epoch": 1.689598238855256, "grad_norm": 0.15743038058280945, "learning_rate": 0.00023700557130238944, "loss": 2.183, "step": 15350 }, { "epoch": 1.6951018161805174, "grad_norm": 0.1788504421710968, "learning_rate": 0.0002365602220344765, "loss": 2.1923, "step": 15400 }, { "epoch": 1.7006053935057788, "grad_norm": 0.1614784598350525, "learning_rate": 0.00023611372583989883, "loss": 2.196, "step": 15450 }, { "epoch": 1.7061089708310402, "grad_norm": 0.16596226394176483, "learning_rate": 0.0002356660886347502, "loss": 2.1952, "step": 15500 }, { "epoch": 1.7116125481563016, "grad_norm": 0.18096621334552765, "learning_rate": 0.00023521731635024274, "loss": 2.1858, "step": 15550 }, { "epoch": 1.717116125481563, "grad_norm": 0.1885058581829071, "learning_rate": 0.0002347674149326286, "loss": 2.1894, "step": 15600 }, { "epoch": 1.7226197028068244, "grad_norm": 0.15273383259773254, "learning_rate": 0.00023431639034312096, "loss": 2.1896, "step": 15650 }, { "epoch": 1.7281232801320858, "grad_norm": 0.17447490990161896, "learning_rate": 0.00023386424855781495, "loss": 2.1896, "step": 15700 }, { "epoch": 1.7336268574573472, "grad_norm": 0.16734521090984344, "learning_rate": 0.00023341099556760866, "loss": 2.1886, "step": 15750 }, { "epoch": 1.7391304347826086, "grad_norm": 0.17426900565624237, "learning_rate": 0.0002329566373781238, "loss": 2.1838, "step": 15800 }, { "epoch": 1.7446340121078703, "grad_norm": 0.1640431135892868, "learning_rate": 0.00023250118000962583, "loss": 2.2026, "step": 15850 }, { "epoch": 1.7501375894331317, "grad_norm": 0.16575609147548676, "learning_rate": 0.00023204462949694447, "loss": 2.1721, "step": 15900 }, { "epoch": 1.755641166758393, "grad_norm": 0.17576967179775238, "learning_rate": 0.00023158699188939371, "loss": 2.1517, "step": 15950 }, { "epoch": 1.7611447440836545, "grad_norm": 0.1661233901977539, "learning_rate": 0.00023112827325069154, "loss": 2.1594, "step": 16000 }, { "epoch": 1.766648321408916, "grad_norm": 0.17433176934719086, "learning_rate": 0.00023066847965887954, "loss": 2.2069, "step": 16050 }, { "epoch": 1.7721518987341773, "grad_norm": 0.164288729429245, "learning_rate": 0.0002302076172062427, "loss": 2.1945, "step": 16100 }, { "epoch": 1.7776554760594387, "grad_norm": 0.16619046032428741, "learning_rate": 0.00022974569199922827, "loss": 2.1678, "step": 16150 }, { "epoch": 1.7831590533847002, "grad_norm": 0.1572977751493454, "learning_rate": 0.00022928271015836514, "loss": 2.1834, "step": 16200 }, { "epoch": 1.7886626307099616, "grad_norm": 0.15041351318359375, "learning_rate": 0.0002288186778181826, "loss": 2.1372, "step": 16250 }, { "epoch": 1.794166208035223, "grad_norm": 0.17850619554519653, "learning_rate": 0.00022835360112712913, "loss": 2.2064, "step": 16300 }, { "epoch": 1.7996697853604844, "grad_norm": 0.1657785028219223, "learning_rate": 0.00022788748624749083, "loss": 2.1948, "step": 16350 }, { "epoch": 1.8051733626857458, "grad_norm": 0.16514721512794495, "learning_rate": 0.00022742033935530995, "loss": 2.209, "step": 16400 }, { "epoch": 1.8106769400110072, "grad_norm": 0.1726425439119339, "learning_rate": 0.00022695216664030297, "loss": 2.1992, "step": 16450 }, { "epoch": 1.8161805173362686, "grad_norm": 0.1753402054309845, "learning_rate": 0.00022648297430577835, "loss": 2.1929, "step": 16500 }, { "epoch": 1.82168409466153, "grad_norm": 0.18259774148464203, "learning_rate": 0.00022601276856855482, "loss": 2.2083, "step": 16550 }, { "epoch": 1.8271876719867914, "grad_norm": 0.16030113399028778, "learning_rate": 0.0002255415556588786, "loss": 2.1571, "step": 16600 }, { "epoch": 1.8326912493120529, "grad_norm": 0.1431453675031662, "learning_rate": 0.000225069341820341, "loss": 2.149, "step": 16650 }, { "epoch": 1.8381948266373143, "grad_norm": 0.1664692461490631, "learning_rate": 0.0002245961333097956, "loss": 2.1864, "step": 16700 }, { "epoch": 1.8436984039625757, "grad_norm": 0.17130005359649658, "learning_rate": 0.00022412193639727564, "loss": 2.18, "step": 16750 }, { "epoch": 1.849201981287837, "grad_norm": 0.15278129279613495, "learning_rate": 0.00022364675736591058, "loss": 2.1648, "step": 16800 }, { "epoch": 1.8547055586130985, "grad_norm": 0.176371768116951, "learning_rate": 0.000223170602511843, "loss": 2.1582, "step": 16850 }, { "epoch": 1.86020913593836, "grad_norm": 0.15902136266231537, "learning_rate": 0.0002226934781441452, "loss": 2.183, "step": 16900 }, { "epoch": 1.8657127132636213, "grad_norm": 0.14821839332580566, "learning_rate": 0.00022221539058473553, "loss": 2.172, "step": 16950 }, { "epoch": 1.8712162905888827, "grad_norm": 0.15611177682876587, "learning_rate": 0.00022173634616829481, "loss": 2.1715, "step": 17000 }, { "epoch": 1.8767198679141441, "grad_norm": 0.1633993536233902, "learning_rate": 0.00022125635124218207, "loss": 2.1961, "step": 17050 }, { "epoch": 1.8822234452394055, "grad_norm": 0.16087235510349274, "learning_rate": 0.00022077541216635076, "loss": 2.1963, "step": 17100 }, { "epoch": 1.887727022564667, "grad_norm": 0.16733358800411224, "learning_rate": 0.00022029353531326432, "loss": 2.1605, "step": 17150 }, { "epoch": 1.8932305998899284, "grad_norm": 0.16267576813697815, "learning_rate": 0.00021981072706781185, "loss": 2.1554, "step": 17200 }, { "epoch": 1.8987341772151898, "grad_norm": 0.16028568148612976, "learning_rate": 0.0002193269938272232, "loss": 2.1793, "step": 17250 }, { "epoch": 1.9042377545404512, "grad_norm": 0.15961337089538574, "learning_rate": 0.00021884234200098473, "loss": 2.1673, "step": 17300 }, { "epoch": 1.9097413318657126, "grad_norm": 0.1637711226940155, "learning_rate": 0.00021835677801075397, "loss": 2.1676, "step": 17350 }, { "epoch": 1.915244909190974, "grad_norm": 0.16490189731121063, "learning_rate": 0.00021787030829027468, "loss": 2.2119, "step": 17400 }, { "epoch": 1.9207484865162354, "grad_norm": 0.17354600131511688, "learning_rate": 0.00021738293928529154, "loss": 2.2001, "step": 17450 }, { "epoch": 1.9262520638414968, "grad_norm": 0.16645126044750214, "learning_rate": 0.00021689467745346483, "loss": 2.1915, "step": 17500 }, { "epoch": 1.9317556411667582, "grad_norm": 0.16187366843223572, "learning_rate": 0.0002164055292642849, "loss": 2.1829, "step": 17550 }, { "epoch": 1.9372592184920197, "grad_norm": 0.17859427630901337, "learning_rate": 0.00021591550119898616, "loss": 2.1203, "step": 17600 }, { "epoch": 1.9427627958172813, "grad_norm": 0.15773439407348633, "learning_rate": 0.00021542459975046164, "loss": 2.1535, "step": 17650 }, { "epoch": 1.9482663731425427, "grad_norm": 0.1696433126926422, "learning_rate": 0.00021493283142317668, "loss": 2.1271, "step": 17700 }, { "epoch": 1.953769950467804, "grad_norm": 0.16615457832813263, "learning_rate": 0.00021444020273308267, "loss": 2.1688, "step": 17750 }, { "epoch": 1.9592735277930655, "grad_norm": 0.16561925411224365, "learning_rate": 0.00021394672020753102, "loss": 2.1228, "step": 17800 }, { "epoch": 1.964777105118327, "grad_norm": 0.1655699759721756, "learning_rate": 0.00021345239038518638, "loss": 2.1614, "step": 17850 }, { "epoch": 1.9702806824435883, "grad_norm": 0.16106578707695007, "learning_rate": 0.0002129572198159401, "loss": 2.1471, "step": 17900 }, { "epoch": 1.9757842597688497, "grad_norm": 0.1635727882385254, "learning_rate": 0.00021246121506082367, "loss": 2.184, "step": 17950 }, { "epoch": 1.9812878370941112, "grad_norm": 0.15699493885040283, "learning_rate": 0.0002119643826919213, "loss": 2.1602, "step": 18000 }, { "epoch": 1.9867914144193726, "grad_norm": 0.15784280002117157, "learning_rate": 0.00021146672929228327, "loss": 2.1781, "step": 18050 }, { "epoch": 1.992294991744634, "grad_norm": 0.15157456696033478, "learning_rate": 0.00021096826145583852, "loss": 2.1532, "step": 18100 }, { "epoch": 1.9977985690698954, "grad_norm": 0.15633012354373932, "learning_rate": 0.0002104689857873074, "loss": 2.1676, "step": 18150 }, { "epoch": 2.003302146395157, "grad_norm": 0.16586118936538696, "learning_rate": 0.00020996890890211386, "loss": 2.1335, "step": 18200 }, { "epoch": 2.0088057237204184, "grad_norm": 0.1573173552751541, "learning_rate": 0.00020946803742629826, "loss": 2.11, "step": 18250 }, { "epoch": 2.01430930104568, "grad_norm": 0.18040034174919128, "learning_rate": 0.00020896637799642915, "loss": 2.1128, "step": 18300 }, { "epoch": 2.0198128783709413, "grad_norm": 0.1664518266916275, "learning_rate": 0.00020846393725951556, "loss": 2.1384, "step": 18350 }, { "epoch": 2.0253164556962027, "grad_norm": 0.17078042030334473, "learning_rate": 0.0002079607218729188, "loss": 2.0909, "step": 18400 }, { "epoch": 2.030820033021464, "grad_norm": 0.15759077668190002, "learning_rate": 0.00020745673850426446, "loss": 2.1015, "step": 18450 }, { "epoch": 2.0363236103467255, "grad_norm": 0.17311988770961761, "learning_rate": 0.00020695199383135386, "loss": 2.1436, "step": 18500 }, { "epoch": 2.041827187671987, "grad_norm": 0.1612362265586853, "learning_rate": 0.00020644649454207553, "loss": 2.0868, "step": 18550 }, { "epoch": 2.0473307649972483, "grad_norm": 0.16166241466999054, "learning_rate": 0.0002059402473343169, "loss": 2.0683, "step": 18600 }, { "epoch": 2.0528343423225097, "grad_norm": 0.17205455899238586, "learning_rate": 0.00020543325891587516, "loss": 2.0618, "step": 18650 }, { "epoch": 2.058337919647771, "grad_norm": 0.17534597218036652, "learning_rate": 0.00020492553600436872, "loss": 2.1173, "step": 18700 }, { "epoch": 2.0638414969730325, "grad_norm": 0.15081949532032013, "learning_rate": 0.00020441708532714797, "loss": 2.0991, "step": 18750 }, { "epoch": 2.069345074298294, "grad_norm": 0.15837198495864868, "learning_rate": 0.00020390791362120625, "loss": 2.1332, "step": 18800 }, { "epoch": 2.0748486516235554, "grad_norm": 0.1598348468542099, "learning_rate": 0.00020339802763309054, "loss": 2.1164, "step": 18850 }, { "epoch": 2.0803522289488168, "grad_norm": 0.1536133736371994, "learning_rate": 0.0002028874341188122, "loss": 2.1169, "step": 18900 }, { "epoch": 2.085855806274078, "grad_norm": 0.1654282808303833, "learning_rate": 0.00020237613984375711, "loss": 2.1253, "step": 18950 }, { "epoch": 2.0913593835993396, "grad_norm": 0.16015826165676117, "learning_rate": 0.00020186415158259652, "loss": 2.1196, "step": 19000 }, { "epoch": 2.096862960924601, "grad_norm": 0.16205786168575287, "learning_rate": 0.00020135147611919686, "loss": 2.0872, "step": 19050 }, { "epoch": 2.1023665382498624, "grad_norm": 0.16814254224300385, "learning_rate": 0.00020083812024653013, "loss": 2.1569, "step": 19100 }, { "epoch": 2.107870115575124, "grad_norm": 0.15433181822299957, "learning_rate": 0.00020032409076658364, "loss": 2.0967, "step": 19150 }, { "epoch": 2.1133736929003852, "grad_norm": 0.16261817514896393, "learning_rate": 0.00019980939449027016, "loss": 2.0938, "step": 19200 }, { "epoch": 2.1188772702256466, "grad_norm": 0.17877838015556335, "learning_rate": 0.00019929403823733748, "loss": 2.1399, "step": 19250 }, { "epoch": 2.124380847550908, "grad_norm": 0.178861603140831, "learning_rate": 0.00019877802883627808, "loss": 2.1362, "step": 19300 }, { "epoch": 2.1298844248761695, "grad_norm": 0.1621016561985016, "learning_rate": 0.00019826137312423873, "loss": 2.1089, "step": 19350 }, { "epoch": 2.135388002201431, "grad_norm": 0.17772269248962402, "learning_rate": 0.00019774407794692986, "loss": 2.1278, "step": 19400 }, { "epoch": 2.1408915795266923, "grad_norm": 0.16339927911758423, "learning_rate": 0.00019722615015853483, "loss": 2.1244, "step": 19450 }, { "epoch": 2.1463951568519537, "grad_norm": 0.16879414021968842, "learning_rate": 0.0001967075966216191, "loss": 2.106, "step": 19500 }, { "epoch": 2.151898734177215, "grad_norm": 0.15002557635307312, "learning_rate": 0.00019618842420703923, "loss": 2.1007, "step": 19550 }, { "epoch": 2.1574023115024765, "grad_norm": 0.17130738496780396, "learning_rate": 0.0001956686397938522, "loss": 2.1104, "step": 19600 }, { "epoch": 2.162905888827738, "grad_norm": 0.16796913743019104, "learning_rate": 0.00019514825026922377, "loss": 2.1278, "step": 19650 }, { "epoch": 2.1684094661529993, "grad_norm": 0.19111864268779755, "learning_rate": 0.0001946272625283375, "loss": 2.1402, "step": 19700 }, { "epoch": 2.1739130434782608, "grad_norm": 0.17543087899684906, "learning_rate": 0.00019410568347430344, "loss": 2.0984, "step": 19750 }, { "epoch": 2.179416620803522, "grad_norm": 0.15795323252677917, "learning_rate": 0.00019358352001806642, "loss": 2.1125, "step": 19800 }, { "epoch": 2.1849201981287836, "grad_norm": 0.15581749379634857, "learning_rate": 0.0001930607790783147, "loss": 2.119, "step": 19850 }, { "epoch": 2.190423775454045, "grad_norm": 0.17086704075336456, "learning_rate": 0.00019253746758138822, "loss": 2.1029, "step": 19900 }, { "epoch": 2.1959273527793064, "grad_norm": 0.15870486199855804, "learning_rate": 0.00019201359246118682, "loss": 2.1133, "step": 19950 }, { "epoch": 2.201430930104568, "grad_norm": 0.16840116679668427, "learning_rate": 0.0001914891606590784, "loss": 2.1196, "step": 20000 }, { "epoch": 2.206934507429829, "grad_norm": 0.1643972396850586, "learning_rate": 0.00019096417912380686, "loss": 2.1173, "step": 20050 }, { "epoch": 2.2124380847550906, "grad_norm": 0.1598280966281891, "learning_rate": 0.0001904386548114001, "loss": 2.1338, "step": 20100 }, { "epoch": 2.217941662080352, "grad_norm": 0.16909152269363403, "learning_rate": 0.0001899125946850779, "loss": 2.0987, "step": 20150 }, { "epoch": 2.2234452394056134, "grad_norm": 0.15597747266292572, "learning_rate": 0.00018938600571515962, "loss": 2.1295, "step": 20200 }, { "epoch": 2.228948816730875, "grad_norm": 0.16759732365608215, "learning_rate": 0.00018885889487897162, "loss": 2.1374, "step": 20250 }, { "epoch": 2.2344523940561363, "grad_norm": 0.15526509284973145, "learning_rate": 0.00018833126916075526, "loss": 2.1287, "step": 20300 }, { "epoch": 2.2399559713813977, "grad_norm": 0.16562257707118988, "learning_rate": 0.00018780313555157393, "loss": 2.126, "step": 20350 }, { "epoch": 2.2454595487066595, "grad_norm": 0.17471948266029358, "learning_rate": 0.00018727450104922075, "loss": 2.1175, "step": 20400 }, { "epoch": 2.2509631260319205, "grad_norm": 0.16391079127788544, "learning_rate": 0.00018674537265812548, "loss": 2.0965, "step": 20450 }, { "epoch": 2.2564667033571824, "grad_norm": 0.1693831831216812, "learning_rate": 0.00018621575738926217, "loss": 2.1124, "step": 20500 }, { "epoch": 2.2619702806824438, "grad_norm": 0.16163313388824463, "learning_rate": 0.00018568566226005587, "loss": 2.1034, "step": 20550 }, { "epoch": 2.267473858007705, "grad_norm": 0.1676999181509018, "learning_rate": 0.00018515509429428988, "loss": 2.0991, "step": 20600 }, { "epoch": 2.2729774353329666, "grad_norm": 0.16531667113304138, "learning_rate": 0.00018462406052201252, "loss": 2.1183, "step": 20650 }, { "epoch": 2.278481012658228, "grad_norm": 0.16208726167678833, "learning_rate": 0.00018409256797944412, "loss": 2.0981, "step": 20700 }, { "epoch": 2.2839845899834894, "grad_norm": 0.1707569658756256, "learning_rate": 0.0001835606237088837, "loss": 2.0926, "step": 20750 }, { "epoch": 2.289488167308751, "grad_norm": 0.16191871464252472, "learning_rate": 0.00018302823475861582, "loss": 2.0895, "step": 20800 }, { "epoch": 2.2949917446340122, "grad_norm": 0.154753640294075, "learning_rate": 0.0001824954081828168, "loss": 2.0972, "step": 20850 }, { "epoch": 2.3004953219592736, "grad_norm": 0.16868236660957336, "learning_rate": 0.00018196215104146187, "loss": 2.1009, "step": 20900 }, { "epoch": 2.305998899284535, "grad_norm": 0.1602269858121872, "learning_rate": 0.00018142847040023097, "loss": 2.1063, "step": 20950 }, { "epoch": 2.3115024766097965, "grad_norm": 0.17220260202884674, "learning_rate": 0.00018089437333041562, "loss": 2.1049, "step": 21000 }, { "epoch": 2.317006053935058, "grad_norm": 0.14858907461166382, "learning_rate": 0.0001803598669088249, "loss": 2.1129, "step": 21050 }, { "epoch": 2.3225096312603193, "grad_norm": 0.17586296796798706, "learning_rate": 0.0001798249582176919, "loss": 2.0961, "step": 21100 }, { "epoch": 2.3280132085855807, "grad_norm": 0.14931316673755646, "learning_rate": 0.00017928965434457983, "loss": 2.1108, "step": 21150 }, { "epoch": 2.333516785910842, "grad_norm": 0.17148137092590332, "learning_rate": 0.00017875396238228798, "loss": 2.1029, "step": 21200 }, { "epoch": 2.3390203632361035, "grad_norm": 0.15888847410678864, "learning_rate": 0.00017821788942875793, "loss": 2.1258, "step": 21250 }, { "epoch": 2.344523940561365, "grad_norm": 0.1749711036682129, "learning_rate": 0.0001776814425869793, "loss": 2.1176, "step": 21300 }, { "epoch": 2.3500275178866263, "grad_norm": 0.16511616110801697, "learning_rate": 0.0001771446289648959, "loss": 2.0992, "step": 21350 }, { "epoch": 2.3555310952118877, "grad_norm": 0.16869331896305084, "learning_rate": 0.00017660745567531124, "loss": 2.1274, "step": 21400 }, { "epoch": 2.361034672537149, "grad_norm": 0.16024701297283173, "learning_rate": 0.00017606992983579457, "loss": 2.1379, "step": 21450 }, { "epoch": 2.3665382498624106, "grad_norm": 0.1646272987127304, "learning_rate": 0.00017553205856858637, "loss": 2.1365, "step": 21500 }, { "epoch": 2.372041827187672, "grad_norm": 0.1765459179878235, "learning_rate": 0.00017499384900050404, "loss": 2.1125, "step": 21550 }, { "epoch": 2.3775454045129334, "grad_norm": 0.15617190301418304, "learning_rate": 0.0001744553082628475, "loss": 2.0985, "step": 21600 }, { "epoch": 2.383048981838195, "grad_norm": 0.14687784016132355, "learning_rate": 0.00017391644349130467, "loss": 2.0771, "step": 21650 }, { "epoch": 2.388552559163456, "grad_norm": 0.16840553283691406, "learning_rate": 0.00017337726182585698, "loss": 2.0909, "step": 21700 }, { "epoch": 2.3940561364887176, "grad_norm": 0.16857478022575378, "learning_rate": 0.00017283777041068472, "loss": 2.0968, "step": 21750 }, { "epoch": 2.399559713813979, "grad_norm": 0.1799498200416565, "learning_rate": 0.00017229797639407233, "loss": 2.1023, "step": 21800 }, { "epoch": 2.4050632911392404, "grad_norm": 0.15700192749500275, "learning_rate": 0.00017175788692831367, "loss": 2.0986, "step": 21850 }, { "epoch": 2.410566868464502, "grad_norm": 0.16778048872947693, "learning_rate": 0.00017121750916961742, "loss": 2.0914, "step": 21900 }, { "epoch": 2.4160704457897633, "grad_norm": 0.15119898319244385, "learning_rate": 0.0001706768502780122, "loss": 2.0857, "step": 21950 }, { "epoch": 2.4215740231150247, "grad_norm": 0.17003406584262848, "learning_rate": 0.0001701359174172515, "loss": 2.1366, "step": 22000 }, { "epoch": 2.427077600440286, "grad_norm": 0.17182987928390503, "learning_rate": 0.000169594717754719, "loss": 2.1126, "step": 22050 }, { "epoch": 2.4325811777655475, "grad_norm": 0.16026781499385834, "learning_rate": 0.00016905325846133354, "loss": 2.076, "step": 22100 }, { "epoch": 2.438084755090809, "grad_norm": 0.16470612585544586, "learning_rate": 0.00016851154671145403, "loss": 2.0907, "step": 22150 }, { "epoch": 2.4435883324160703, "grad_norm": 0.14621587097644806, "learning_rate": 0.0001679695896827844, "loss": 2.1073, "step": 22200 }, { "epoch": 2.4490919097413317, "grad_norm": 0.1563323438167572, "learning_rate": 0.00016742739455627862, "loss": 2.0757, "step": 22250 }, { "epoch": 2.454595487066593, "grad_norm": 0.17763760685920715, "learning_rate": 0.0001668849685160455, "loss": 2.0705, "step": 22300 }, { "epoch": 2.4600990643918546, "grad_norm": 0.16651757061481476, "learning_rate": 0.00016634231874925335, "loss": 2.0881, "step": 22350 }, { "epoch": 2.465602641717116, "grad_norm": 0.16752833127975464, "learning_rate": 0.0001657994524460349, "loss": 2.0639, "step": 22400 }, { "epoch": 2.4711062190423774, "grad_norm": 0.14574387669563293, "learning_rate": 0.000165256376799392, "loss": 2.087, "step": 22450 }, { "epoch": 2.476609796367639, "grad_norm": 0.15595164895057678, "learning_rate": 0.00016471309900510043, "loss": 2.0965, "step": 22500 }, { "epoch": 2.4821133736929, "grad_norm": 0.17218752205371857, "learning_rate": 0.0001641696262616142, "loss": 2.0943, "step": 22550 }, { "epoch": 2.487616951018162, "grad_norm": 0.1563566029071808, "learning_rate": 0.00016362596576997057, "loss": 2.0946, "step": 22600 }, { "epoch": 2.4931205283434235, "grad_norm": 0.16190436482429504, "learning_rate": 0.00016308212473369453, "loss": 2.0924, "step": 22650 }, { "epoch": 2.498624105668685, "grad_norm": 0.16890835762023926, "learning_rate": 0.00016253811035870313, "loss": 2.0954, "step": 22700 }, { "epoch": 2.5041276829939463, "grad_norm": 0.16371208429336548, "learning_rate": 0.00016199392985321017, "loss": 2.0954, "step": 22750 }, { "epoch": 2.5096312603192077, "grad_norm": 0.15898428857326508, "learning_rate": 0.00016144959042763084, "loss": 2.0766, "step": 22800 }, { "epoch": 2.515134837644469, "grad_norm": 0.15516646206378937, "learning_rate": 0.00016090509929448594, "loss": 2.0891, "step": 22850 }, { "epoch": 2.5206384149697305, "grad_norm": 0.15007217228412628, "learning_rate": 0.00016036046366830635, "loss": 2.078, "step": 22900 }, { "epoch": 2.526141992294992, "grad_norm": 0.17989104986190796, "learning_rate": 0.00015981569076553765, "loss": 2.0888, "step": 22950 }, { "epoch": 2.5316455696202533, "grad_norm": 0.16754071414470673, "learning_rate": 0.00015927078780444403, "loss": 2.0882, "step": 23000 }, { "epoch": 2.5371491469455147, "grad_norm": 0.15897449851036072, "learning_rate": 0.0001587257620050134, "loss": 2.0933, "step": 23050 }, { "epoch": 2.542652724270776, "grad_norm": 0.16594423353672028, "learning_rate": 0.00015818062058886082, "loss": 2.0968, "step": 23100 }, { "epoch": 2.5481563015960376, "grad_norm": 0.1700371652841568, "learning_rate": 0.00015763537077913365, "loss": 2.096, "step": 23150 }, { "epoch": 2.553659878921299, "grad_norm": 0.16013766825199127, "learning_rate": 0.00015709001980041527, "loss": 2.1134, "step": 23200 }, { "epoch": 2.5591634562465604, "grad_norm": 0.1601455807685852, "learning_rate": 0.00015654457487862961, "loss": 2.0983, "step": 23250 }, { "epoch": 2.564667033571822, "grad_norm": 0.16373537480831146, "learning_rate": 0.00015599904324094534, "loss": 2.0866, "step": 23300 }, { "epoch": 2.570170610897083, "grad_norm": 0.15919940173625946, "learning_rate": 0.0001554534321156801, "loss": 2.1147, "step": 23350 }, { "epoch": 2.5756741882223446, "grad_norm": 0.1523171216249466, "learning_rate": 0.0001549077487322048, "loss": 2.0887, "step": 23400 }, { "epoch": 2.581177765547606, "grad_norm": 0.16425256431102753, "learning_rate": 0.00015436200032084774, "loss": 2.0918, "step": 23450 }, { "epoch": 2.5866813428728674, "grad_norm": 0.16046775877475739, "learning_rate": 0.0001538161941127988, "loss": 2.0816, "step": 23500 }, { "epoch": 2.592184920198129, "grad_norm": 0.1673419028520584, "learning_rate": 0.00015327033734001375, "loss": 2.1186, "step": 23550 }, { "epoch": 2.5976884975233903, "grad_norm": 0.15573158860206604, "learning_rate": 0.00015272443723511825, "loss": 2.0721, "step": 23600 }, { "epoch": 2.6031920748486517, "grad_norm": 0.16536468267440796, "learning_rate": 0.00015217850103131227, "loss": 2.0991, "step": 23650 }, { "epoch": 2.608695652173913, "grad_norm": 0.15281961858272552, "learning_rate": 0.00015163253596227393, "loss": 2.0713, "step": 23700 }, { "epoch": 2.6141992294991745, "grad_norm": 0.1582518070936203, "learning_rate": 0.0001510865492620638, "loss": 2.0923, "step": 23750 }, { "epoch": 2.619702806824436, "grad_norm": 0.1581014096736908, "learning_rate": 0.0001505405481650293, "loss": 2.0895, "step": 23800 }, { "epoch": 2.6252063841496973, "grad_norm": 0.16346633434295654, "learning_rate": 0.00014999453990570839, "loss": 2.1011, "step": 23850 }, { "epoch": 2.6307099614749587, "grad_norm": 0.16536889970302582, "learning_rate": 0.00014944853171873387, "loss": 2.113, "step": 23900 }, { "epoch": 2.63621353880022, "grad_norm": 0.170758455991745, "learning_rate": 0.00014890253083873784, "loss": 2.1124, "step": 23950 }, { "epoch": 2.6417171161254815, "grad_norm": 0.15669451653957367, "learning_rate": 0.00014835654450025542, "loss": 2.0894, "step": 24000 }, { "epoch": 2.647220693450743, "grad_norm": 0.16937313973903656, "learning_rate": 0.00014781057993762902, "loss": 2.127, "step": 24050 }, { "epoch": 2.6527242707760044, "grad_norm": 0.15754981338977814, "learning_rate": 0.0001472646443849126, "loss": 2.0959, "step": 24100 }, { "epoch": 2.6582278481012658, "grad_norm": 0.16420376300811768, "learning_rate": 0.00014671874507577574, "loss": 2.0742, "step": 24150 }, { "epoch": 2.663731425426527, "grad_norm": 0.16400571167469025, "learning_rate": 0.00014617288924340775, "loss": 2.0992, "step": 24200 }, { "epoch": 2.6692350027517886, "grad_norm": 0.1658066064119339, "learning_rate": 0.00014562708412042193, "loss": 2.0925, "step": 24250 }, { "epoch": 2.67473858007705, "grad_norm": 0.16872134804725647, "learning_rate": 0.00014508133693875953, "loss": 2.0899, "step": 24300 }, { "epoch": 2.6802421574023114, "grad_norm": 0.17002440989017487, "learning_rate": 0.00014453565492959422, "loss": 2.1252, "step": 24350 }, { "epoch": 2.685745734727573, "grad_norm": 0.16417178511619568, "learning_rate": 0.00014399004532323614, "loss": 2.112, "step": 24400 }, { "epoch": 2.6912493120528342, "grad_norm": 0.172217458486557, "learning_rate": 0.000143444515349036, "loss": 2.0784, "step": 24450 }, { "epoch": 2.6967528893780957, "grad_norm": 0.16687516868114471, "learning_rate": 0.0001428990722352894, "loss": 2.0775, "step": 24500 }, { "epoch": 2.702256466703357, "grad_norm": 0.162861630320549, "learning_rate": 0.00014235372320914108, "loss": 2.0928, "step": 24550 }, { "epoch": 2.7077600440286185, "grad_norm": 0.1640874743461609, "learning_rate": 0.00014180847549648913, "loss": 2.1071, "step": 24600 }, { "epoch": 2.71326362135388, "grad_norm": 0.1654275506734848, "learning_rate": 0.0001412633363218891, "loss": 2.0615, "step": 24650 }, { "epoch": 2.7187671986791413, "grad_norm": 0.14277833700180054, "learning_rate": 0.00014071831290845857, "loss": 2.1078, "step": 24700 }, { "epoch": 2.7242707760044027, "grad_norm": 0.1705639362335205, "learning_rate": 0.00014017341247778123, "loss": 2.0858, "step": 24750 }, { "epoch": 2.729774353329664, "grad_norm": 0.16390259563922882, "learning_rate": 0.0001396286422498112, "loss": 2.0508, "step": 24800 }, { "epoch": 2.7352779306549255, "grad_norm": 0.1526353508234024, "learning_rate": 0.00013908400944277738, "loss": 2.0784, "step": 24850 }, { "epoch": 2.740781507980187, "grad_norm": 0.1581793874502182, "learning_rate": 0.0001385395212730879, "loss": 2.0596, "step": 24900 }, { "epoch": 2.7462850853054483, "grad_norm": 0.1692277193069458, "learning_rate": 0.00013799518495523445, "loss": 2.0838, "step": 24950 }, { "epoch": 2.7517886626307098, "grad_norm": 0.1614445298910141, "learning_rate": 0.0001374510077016967, "loss": 2.0928, "step": 25000 }, { "epoch": 2.757292239955971, "grad_norm": 0.15747620165348053, "learning_rate": 0.00013690699672284657, "loss": 2.0686, "step": 25050 }, { "epoch": 2.7627958172812326, "grad_norm": 0.17565257847309113, "learning_rate": 0.00013636315922685294, "loss": 2.0971, "step": 25100 }, { "epoch": 2.768299394606494, "grad_norm": 0.16367433965206146, "learning_rate": 0.0001358195024195861, "loss": 2.1052, "step": 25150 }, { "epoch": 2.7738029719317554, "grad_norm": 0.1564227044582367, "learning_rate": 0.00013527603350452206, "loss": 2.0902, "step": 25200 }, { "epoch": 2.779306549257017, "grad_norm": 0.16555672883987427, "learning_rate": 0.0001347327596826473, "loss": 2.0683, "step": 25250 }, { "epoch": 2.7848101265822782, "grad_norm": 0.16134041547775269, "learning_rate": 0.00013418968815236346, "loss": 2.0627, "step": 25300 }, { "epoch": 2.7903137039075396, "grad_norm": 0.17097671329975128, "learning_rate": 0.0001336468261093915, "loss": 2.0724, "step": 25350 }, { "epoch": 2.795817281232801, "grad_norm": 0.1656900942325592, "learning_rate": 0.00013310418074667698, "loss": 2.1266, "step": 25400 }, { "epoch": 2.801320858558063, "grad_norm": 0.1745171695947647, "learning_rate": 0.00013256175925429415, "loss": 2.0983, "step": 25450 }, { "epoch": 2.8068244358833243, "grad_norm": 0.15649184584617615, "learning_rate": 0.0001320195688193512, "loss": 2.0672, "step": 25500 }, { "epoch": 2.8123280132085857, "grad_norm": 0.16284936666488647, "learning_rate": 0.00013147761662589473, "loss": 2.0865, "step": 25550 }, { "epoch": 2.817831590533847, "grad_norm": 0.15236727893352509, "learning_rate": 0.00013093590985481456, "loss": 2.0595, "step": 25600 }, { "epoch": 2.8233351678591085, "grad_norm": 0.166220560669899, "learning_rate": 0.00013039445568374872, "loss": 2.0478, "step": 25650 }, { "epoch": 2.82883874518437, "grad_norm": 0.16826435923576355, "learning_rate": 0.0001298532612869883, "loss": 2.0756, "step": 25700 }, { "epoch": 2.8343423225096314, "grad_norm": 0.16447168588638306, "learning_rate": 0.0001293123338353823, "loss": 2.0742, "step": 25750 }, { "epoch": 2.8398458998348928, "grad_norm": 0.157373309135437, "learning_rate": 0.00012877168049624268, "loss": 2.0728, "step": 25800 }, { "epoch": 2.845349477160154, "grad_norm": 0.1642802357673645, "learning_rate": 0.00012823130843324948, "loss": 2.0643, "step": 25850 }, { "epoch": 2.8508530544854156, "grad_norm": 0.15527865290641785, "learning_rate": 0.00012769122480635583, "loss": 2.1088, "step": 25900 }, { "epoch": 2.856356631810677, "grad_norm": 0.175029918551445, "learning_rate": 0.00012715143677169297, "loss": 2.0764, "step": 25950 }, { "epoch": 2.8618602091359384, "grad_norm": 0.17664563655853271, "learning_rate": 0.0001266119514814755, "loss": 2.0527, "step": 26000 }, { "epoch": 2.8673637864612, "grad_norm": 0.16839756071567535, "learning_rate": 0.00012607277608390677, "loss": 2.0708, "step": 26050 }, { "epoch": 2.8728673637864612, "grad_norm": 0.162934809923172, "learning_rate": 0.00012553391772308407, "loss": 2.0984, "step": 26100 }, { "epoch": 2.8783709411117226, "grad_norm": 0.1836494356393814, "learning_rate": 0.0001249953835389037, "loss": 2.1118, "step": 26150 }, { "epoch": 2.883874518436984, "grad_norm": 0.15794949233531952, "learning_rate": 0.00012445718066696687, "loss": 2.0541, "step": 26200 }, { "epoch": 2.8893780957622455, "grad_norm": 0.16560156643390656, "learning_rate": 0.00012391931623848476, "loss": 2.1111, "step": 26250 }, { "epoch": 2.894881673087507, "grad_norm": 0.16835126280784607, "learning_rate": 0.0001233817973801842, "loss": 2.0635, "step": 26300 }, { "epoch": 2.9003852504127683, "grad_norm": 0.16151325404644012, "learning_rate": 0.0001228446312142131, "loss": 2.0949, "step": 26350 }, { "epoch": 2.9058888277380297, "grad_norm": 0.16751372814178467, "learning_rate": 0.00012230782485804633, "loss": 2.0896, "step": 26400 }, { "epoch": 2.911392405063291, "grad_norm": 0.18384750187397003, "learning_rate": 0.00012177138542439122, "loss": 2.047, "step": 26450 }, { "epoch": 2.9168959823885525, "grad_norm": 0.15618419647216797, "learning_rate": 0.00012123532002109328, "loss": 2.0787, "step": 26500 }, { "epoch": 2.922399559713814, "grad_norm": 0.15803013741970062, "learning_rate": 0.00012069963575104223, "loss": 2.0735, "step": 26550 }, { "epoch": 2.9279031370390753, "grad_norm": 0.1603049784898758, "learning_rate": 0.00012016433971207757, "loss": 2.0802, "step": 26600 }, { "epoch": 2.9334067143643368, "grad_norm": 0.1752861887216568, "learning_rate": 0.0001196294389968949, "loss": 2.0969, "step": 26650 }, { "epoch": 2.938910291689598, "grad_norm": 0.16382624208927155, "learning_rate": 0.00011909494069295175, "loss": 2.0933, "step": 26700 }, { "epoch": 2.9444138690148596, "grad_norm": 0.16802163422107697, "learning_rate": 0.00011856085188237357, "loss": 2.1042, "step": 26750 }, { "epoch": 2.949917446340121, "grad_norm": 0.15907612442970276, "learning_rate": 0.00011802717964186006, "loss": 2.0575, "step": 26800 }, { "epoch": 2.9554210236653824, "grad_norm": 0.16981548070907593, "learning_rate": 0.00011749393104259143, "loss": 2.0951, "step": 26850 }, { "epoch": 2.960924600990644, "grad_norm": 0.15169958770275116, "learning_rate": 0.0001169611131501345, "loss": 2.0778, "step": 26900 }, { "epoch": 2.966428178315905, "grad_norm": 0.1601976752281189, "learning_rate": 0.00011642873302434926, "loss": 2.0621, "step": 26950 }, { "epoch": 2.9719317556411666, "grad_norm": 0.16251477599143982, "learning_rate": 0.00011589679771929529, "loss": 2.08, "step": 27000 }, { "epoch": 2.977435332966428, "grad_norm": 0.15681184828281403, "learning_rate": 0.00011536531428313828, "loss": 2.0518, "step": 27050 }, { "epoch": 2.9829389102916894, "grad_norm": 0.17260442674160004, "learning_rate": 0.00011483428975805664, "loss": 2.0899, "step": 27100 }, { "epoch": 2.988442487616951, "grad_norm": 0.16616973280906677, "learning_rate": 0.00011430373118014808, "loss": 2.0787, "step": 27150 }, { "epoch": 2.9939460649422127, "grad_norm": 0.1666904091835022, "learning_rate": 0.00011377364557933658, "loss": 2.0555, "step": 27200 }, { "epoch": 2.999449642267474, "grad_norm": 0.15540987253189087, "learning_rate": 0.00011324403997927922, "loss": 2.0676, "step": 27250 }, { "epoch": 3.004953219592735, "grad_norm": 0.16125087440013885, "learning_rate": 0.00011271492139727294, "loss": 2.0144, "step": 27300 }, { "epoch": 3.0104567969179965, "grad_norm": 0.17550891637802124, "learning_rate": 0.00011218629684416168, "loss": 2.0102, "step": 27350 }, { "epoch": 3.015960374243258, "grad_norm": 0.16927887499332428, "learning_rate": 0.00011165817332424356, "loss": 1.9816, "step": 27400 }, { "epoch": 3.0214639515685193, "grad_norm": 0.17522330582141876, "learning_rate": 0.00011113055783517793, "loss": 2.0066, "step": 27450 }, { "epoch": 3.026967528893781, "grad_norm": 0.160496324300766, "learning_rate": 0.00011060345736789265, "loss": 2.0003, "step": 27500 }, { "epoch": 3.0324711062190426, "grad_norm": 0.17563533782958984, "learning_rate": 0.00011007687890649163, "loss": 2.0272, "step": 27550 }, { "epoch": 3.037974683544304, "grad_norm": 0.17000171542167664, "learning_rate": 0.00010955082942816215, "loss": 2.0447, "step": 27600 }, { "epoch": 3.0434782608695654, "grad_norm": 0.17253416776657104, "learning_rate": 0.00010902531590308236, "loss": 1.9968, "step": 27650 }, { "epoch": 3.048981838194827, "grad_norm": 0.16862879693508148, "learning_rate": 0.00010850034529432919, "loss": 1.9952, "step": 27700 }, { "epoch": 3.0544854155200882, "grad_norm": 0.166275292634964, "learning_rate": 0.00010797592455778562, "loss": 2.0258, "step": 27750 }, { "epoch": 3.0599889928453496, "grad_norm": 0.1734013706445694, "learning_rate": 0.00010745206064204903, "loss": 2.0047, "step": 27800 }, { "epoch": 3.065492570170611, "grad_norm": 0.1760258674621582, "learning_rate": 0.00010692876048833889, "loss": 1.997, "step": 27850 }, { "epoch": 3.0709961474958725, "grad_norm": 0.1616799384355545, "learning_rate": 0.00010640603103040464, "loss": 1.9817, "step": 27900 }, { "epoch": 3.076499724821134, "grad_norm": 0.1775207221508026, "learning_rate": 0.00010588387919443416, "loss": 2.0391, "step": 27950 }, { "epoch": 3.0820033021463953, "grad_norm": 0.17136207222938538, "learning_rate": 0.00010536231189896175, "loss": 1.999, "step": 28000 }, { "epoch": 3.0875068794716567, "grad_norm": 0.1694386601448059, "learning_rate": 0.00010484133605477644, "loss": 2.0153, "step": 28050 }, { "epoch": 3.093010456796918, "grad_norm": 0.17021538317203522, "learning_rate": 0.00010432095856483057, "loss": 2.0411, "step": 28100 }, { "epoch": 3.0985140341221795, "grad_norm": 0.17548739910125732, "learning_rate": 0.00010380118632414833, "loss": 2.0416, "step": 28150 }, { "epoch": 3.104017611447441, "grad_norm": 0.1785077303647995, "learning_rate": 0.0001032820262197342, "loss": 2.032, "step": 28200 }, { "epoch": 3.1095211887727023, "grad_norm": 0.17226973176002502, "learning_rate": 0.00010276348513048193, "loss": 1.986, "step": 28250 }, { "epoch": 3.1150247660979637, "grad_norm": 0.18131954967975616, "learning_rate": 0.00010224556992708328, "loss": 2.0149, "step": 28300 }, { "epoch": 3.120528343423225, "grad_norm": 0.16980765759944916, "learning_rate": 0.00010172828747193689, "loss": 2.0309, "step": 28350 }, { "epoch": 3.1260319207484866, "grad_norm": 0.16887415945529938, "learning_rate": 0.00010121164461905759, "loss": 1.9814, "step": 28400 }, { "epoch": 3.131535498073748, "grad_norm": 0.18381917476654053, "learning_rate": 0.00010069564821398536, "loss": 2.0228, "step": 28450 }, { "epoch": 3.1370390753990094, "grad_norm": 0.18095500767230988, "learning_rate": 0.0001001803050936948, "loss": 2.0197, "step": 28500 }, { "epoch": 3.142542652724271, "grad_norm": 0.15869130194187164, "learning_rate": 9.966562208650441e-05, "loss": 2.0114, "step": 28550 }, { "epoch": 3.148046230049532, "grad_norm": 0.17333059012889862, "learning_rate": 9.915160601198616e-05, "loss": 2.0025, "step": 28600 }, { "epoch": 3.1535498073747936, "grad_norm": 0.161941796541214, "learning_rate": 9.863826368087516e-05, "loss": 2.0273, "step": 28650 }, { "epoch": 3.159053384700055, "grad_norm": 0.16833285987377167, "learning_rate": 9.812560189497935e-05, "loss": 2.0108, "step": 28700 }, { "epoch": 3.1645569620253164, "grad_norm": 0.15908978879451752, "learning_rate": 9.76136274470895e-05, "loss": 1.9852, "step": 28750 }, { "epoch": 3.170060539350578, "grad_norm": 0.1724586933851242, "learning_rate": 9.710234712088896e-05, "loss": 2.0297, "step": 28800 }, { "epoch": 3.1755641166758393, "grad_norm": 0.17113400995731354, "learning_rate": 9.659176769086419e-05, "loss": 1.9969, "step": 28850 }, { "epoch": 3.1810676940011007, "grad_norm": 0.16423186659812927, "learning_rate": 9.608189592221446e-05, "loss": 2.0106, "step": 28900 }, { "epoch": 3.186571271326362, "grad_norm": 0.16140978038311005, "learning_rate": 9.557273857076271e-05, "loss": 2.0339, "step": 28950 }, { "epoch": 3.1920748486516235, "grad_norm": 0.1718360334634781, "learning_rate": 9.506430238286566e-05, "loss": 2.04, "step": 29000 }, { "epoch": 3.197578425976885, "grad_norm": 0.164507195353508, "learning_rate": 9.455659409532474e-05, "loss": 2.0168, "step": 29050 }, { "epoch": 3.2030820033021463, "grad_norm": 0.17989708483219147, "learning_rate": 9.40496204352966e-05, "loss": 2.0273, "step": 29100 }, { "epoch": 3.2085855806274077, "grad_norm": 0.17319265007972717, "learning_rate": 9.354338812020402e-05, "loss": 1.9911, "step": 29150 }, { "epoch": 3.214089157952669, "grad_norm": 0.17870402336120605, "learning_rate": 9.303790385764694e-05, "loss": 2.0103, "step": 29200 }, { "epoch": 3.2195927352779306, "grad_norm": 0.1797982007265091, "learning_rate": 9.253317434531357e-05, "loss": 2.0169, "step": 29250 }, { "epoch": 3.225096312603192, "grad_norm": 0.15193064510822296, "learning_rate": 9.202920627089171e-05, "loss": 2.0116, "step": 29300 }, { "epoch": 3.2305998899284534, "grad_norm": 0.16679364442825317, "learning_rate": 9.152600631197996e-05, "loss": 2.0362, "step": 29350 }, { "epoch": 3.236103467253715, "grad_norm": 0.17053711414337158, "learning_rate": 9.102358113599942e-05, "loss": 1.9829, "step": 29400 }, { "epoch": 3.241607044578976, "grad_norm": 0.16923174262046814, "learning_rate": 9.052193740010536e-05, "loss": 2.0011, "step": 29450 }, { "epoch": 3.2471106219042376, "grad_norm": 0.18323257565498352, "learning_rate": 9.002108175109877e-05, "loss": 1.9784, "step": 29500 }, { "epoch": 3.252614199229499, "grad_norm": 0.16737329959869385, "learning_rate": 8.952102082533855e-05, "loss": 2.0276, "step": 29550 }, { "epoch": 3.2581177765547604, "grad_norm": 0.17038094997406006, "learning_rate": 8.902176124865348e-05, "loss": 2.0315, "step": 29600 }, { "epoch": 3.263621353880022, "grad_norm": 0.16778729856014252, "learning_rate": 8.852330963625444e-05, "loss": 2.0139, "step": 29650 }, { "epoch": 3.2691249312052832, "grad_norm": 0.18061307072639465, "learning_rate": 8.802567259264674e-05, "loss": 2.002, "step": 29700 }, { "epoch": 3.2746285085305447, "grad_norm": 0.17191821336746216, "learning_rate": 8.752885671154253e-05, "loss": 2.0218, "step": 29750 }, { "epoch": 3.280132085855806, "grad_norm": 0.17957885563373566, "learning_rate": 8.703286857577378e-05, "loss": 2.0217, "step": 29800 }, { "epoch": 3.2856356631810675, "grad_norm": 0.16803069412708282, "learning_rate": 8.653771475720442e-05, "loss": 2.0185, "step": 29850 }, { "epoch": 3.291139240506329, "grad_norm": 0.1788654327392578, "learning_rate": 8.604340181664395e-05, "loss": 1.9973, "step": 29900 }, { "epoch": 3.2966428178315903, "grad_norm": 0.17996376752853394, "learning_rate": 8.554993630376006e-05, "loss": 2.0161, "step": 29950 }, { "epoch": 3.3021463951568517, "grad_norm": 0.18701910972595215, "learning_rate": 8.50573247569921e-05, "loss": 2.0156, "step": 30000 }, { "epoch": 3.3076499724821136, "grad_norm": 0.1713022142648697, "learning_rate": 8.45655737034641e-05, "loss": 2.0144, "step": 30050 }, { "epoch": 3.313153549807375, "grad_norm": 0.17666327953338623, "learning_rate": 8.407468965889878e-05, "loss": 2.0116, "step": 30100 }, { "epoch": 3.3186571271326364, "grad_norm": 0.18481209874153137, "learning_rate": 8.358467912753079e-05, "loss": 2.0149, "step": 30150 }, { "epoch": 3.324160704457898, "grad_norm": 0.17895914614200592, "learning_rate": 8.309554860202082e-05, "loss": 1.9913, "step": 30200 }, { "epoch": 3.329664281783159, "grad_norm": 0.18427938222885132, "learning_rate": 8.260730456336925e-05, "loss": 1.9913, "step": 30250 }, { "epoch": 3.3351678591084206, "grad_norm": 0.1806970238685608, "learning_rate": 8.211995348083067e-05, "loss": 2.0177, "step": 30300 }, { "epoch": 3.340671436433682, "grad_norm": 0.1735043078660965, "learning_rate": 8.163350181182787e-05, "loss": 2.0275, "step": 30350 }, { "epoch": 3.3461750137589434, "grad_norm": 0.17593735456466675, "learning_rate": 8.11479560018664e-05, "loss": 2.0525, "step": 30400 }, { "epoch": 3.351678591084205, "grad_norm": 0.1755988448858261, "learning_rate": 8.066332248444908e-05, "loss": 2.0182, "step": 30450 }, { "epoch": 3.3571821684094663, "grad_norm": 0.15805041790008545, "learning_rate": 8.017960768099091e-05, "loss": 1.9748, "step": 30500 }, { "epoch": 3.3626857457347277, "grad_norm": 0.18487168848514557, "learning_rate": 7.969681800073386e-05, "loss": 2.021, "step": 30550 }, { "epoch": 3.368189323059989, "grad_norm": 0.17120800912380219, "learning_rate": 7.921495984066194e-05, "loss": 2.0113, "step": 30600 }, { "epoch": 3.3736929003852505, "grad_norm": 0.1721143126487732, "learning_rate": 7.87340395854164e-05, "loss": 1.9936, "step": 30650 }, { "epoch": 3.379196477710512, "grad_norm": 0.17595061659812927, "learning_rate": 7.825406360721139e-05, "loss": 2.0044, "step": 30700 }, { "epoch": 3.3847000550357733, "grad_norm": 0.16751642525196075, "learning_rate": 7.777503826574925e-05, "loss": 1.9988, "step": 30750 }, { "epoch": 3.3902036323610347, "grad_norm": 0.17864260077476501, "learning_rate": 7.729696990813634e-05, "loss": 2.0189, "step": 30800 }, { "epoch": 3.395707209686296, "grad_norm": 0.1791946142911911, "learning_rate": 7.681986486879898e-05, "loss": 2.0038, "step": 30850 }, { "epoch": 3.4012107870115575, "grad_norm": 0.1704343855381012, "learning_rate": 7.634372946939945e-05, "loss": 1.9625, "step": 30900 }, { "epoch": 3.406714364336819, "grad_norm": 0.17794691026210785, "learning_rate": 7.586857001875235e-05, "loss": 2.0017, "step": 30950 }, { "epoch": 3.4122179416620804, "grad_norm": 0.17922665178775787, "learning_rate": 7.539439281274071e-05, "loss": 2.054, "step": 31000 }, { "epoch": 3.4177215189873418, "grad_norm": 0.174763485789299, "learning_rate": 7.492120413423296e-05, "loss": 2.0185, "step": 31050 }, { "epoch": 3.423225096312603, "grad_norm": 0.1687181144952774, "learning_rate": 7.444901025299941e-05, "loss": 2.0463, "step": 31100 }, { "epoch": 3.4287286736378646, "grad_norm": 0.17793361842632294, "learning_rate": 7.397781742562941e-05, "loss": 2.0296, "step": 31150 }, { "epoch": 3.434232250963126, "grad_norm": 0.16858656704425812, "learning_rate": 7.350763189544803e-05, "loss": 1.9916, "step": 31200 }, { "epoch": 3.4397358282883874, "grad_norm": 0.16184011101722717, "learning_rate": 7.303845989243384e-05, "loss": 2.0314, "step": 31250 }, { "epoch": 3.445239405613649, "grad_norm": 0.1733245998620987, "learning_rate": 7.257030763313607e-05, "loss": 2.0125, "step": 31300 }, { "epoch": 3.4507429829389102, "grad_norm": 0.18471761047840118, "learning_rate": 7.210318132059231e-05, "loss": 1.9691, "step": 31350 }, { "epoch": 3.4562465602641717, "grad_norm": 0.17260605096817017, "learning_rate": 7.16370871442462e-05, "loss": 2.0113, "step": 31400 }, { "epoch": 3.461750137589433, "grad_norm": 0.1758509874343872, "learning_rate": 7.117203127986569e-05, "loss": 2.0239, "step": 31450 }, { "epoch": 3.4672537149146945, "grad_norm": 0.18456335365772247, "learning_rate": 7.070801988946098e-05, "loss": 1.999, "step": 31500 }, { "epoch": 3.472757292239956, "grad_norm": 0.17995281517505646, "learning_rate": 7.024505912120293e-05, "loss": 2.0141, "step": 31550 }, { "epoch": 3.4782608695652173, "grad_norm": 0.16747882962226868, "learning_rate": 6.978315510934164e-05, "loss": 2.0073, "step": 31600 }, { "epoch": 3.4837644468904787, "grad_norm": 0.1658480316400528, "learning_rate": 6.932231397412522e-05, "loss": 2.0079, "step": 31650 }, { "epoch": 3.48926802421574, "grad_norm": 0.17987202107906342, "learning_rate": 6.886254182171836e-05, "loss": 2.0439, "step": 31700 }, { "epoch": 3.4947716015410015, "grad_norm": 0.1672779768705368, "learning_rate": 6.840384474412201e-05, "loss": 1.9878, "step": 31750 }, { "epoch": 3.500275178866263, "grad_norm": 0.1716667115688324, "learning_rate": 6.794622881909207e-05, "loss": 2.0248, "step": 31800 }, { "epoch": 3.5057787561915243, "grad_norm": 0.17986662685871124, "learning_rate": 6.748970011005924e-05, "loss": 1.99, "step": 31850 }, { "epoch": 3.5112823335167858, "grad_norm": 0.1808827668428421, "learning_rate": 6.703426466604858e-05, "loss": 1.9692, "step": 31900 }, { "epoch": 3.516785910842047, "grad_norm": 0.17262905836105347, "learning_rate": 6.65799285215993e-05, "loss": 1.9636, "step": 31950 }, { "epoch": 3.522289488167309, "grad_norm": 0.16857630014419556, "learning_rate": 6.612669769668489e-05, "loss": 1.9923, "step": 32000 }, { "epoch": 3.5277930654925704, "grad_norm": 0.17498353123664856, "learning_rate": 6.567457819663327e-05, "loss": 2.0248, "step": 32050 }, { "epoch": 3.533296642817832, "grad_norm": 0.16948480904102325, "learning_rate": 6.522357601204719e-05, "loss": 1.9966, "step": 32100 }, { "epoch": 3.5388002201430933, "grad_norm": 0.1695474237203598, "learning_rate": 6.477369711872502e-05, "loss": 1.994, "step": 32150 }, { "epoch": 3.5443037974683547, "grad_norm": 0.1761193871498108, "learning_rate": 6.432494747758146e-05, "loss": 1.9827, "step": 32200 }, { "epoch": 3.549807374793616, "grad_norm": 0.17206500470638275, "learning_rate": 6.387733303456858e-05, "loss": 1.9944, "step": 32250 }, { "epoch": 3.5553109521188775, "grad_norm": 0.19551701843738556, "learning_rate": 6.3430859720597e-05, "loss": 2.0118, "step": 32300 }, { "epoch": 3.560814529444139, "grad_norm": 0.17827004194259644, "learning_rate": 6.298553345145728e-05, "loss": 1.9953, "step": 32350 }, { "epoch": 3.5663181067694003, "grad_norm": 0.17359542846679688, "learning_rate": 6.254136012774166e-05, "loss": 2.0098, "step": 32400 }, { "epoch": 3.5718216840946617, "grad_norm": 0.16225001215934753, "learning_rate": 6.209834563476578e-05, "loss": 2.008, "step": 32450 }, { "epoch": 3.577325261419923, "grad_norm": 0.17888249456882477, "learning_rate": 6.165649584249079e-05, "loss": 1.9962, "step": 32500 }, { "epoch": 3.5828288387451845, "grad_norm": 0.16299229860305786, "learning_rate": 6.121581660544532e-05, "loss": 2.0387, "step": 32550 }, { "epoch": 3.588332416070446, "grad_norm": 0.16753186285495758, "learning_rate": 6.077631376264832e-05, "loss": 1.9662, "step": 32600 }, { "epoch": 3.5938359933957074, "grad_norm": 0.18166500329971313, "learning_rate": 6.033799313753129e-05, "loss": 1.997, "step": 32650 }, { "epoch": 3.5993395707209688, "grad_norm": 0.16691668331623077, "learning_rate": 5.990086053786139e-05, "loss": 2.0072, "step": 32700 }, { "epoch": 3.60484314804623, "grad_norm": 0.18671706318855286, "learning_rate": 5.946492175566438e-05, "loss": 2.0217, "step": 32750 }, { "epoch": 3.6103467253714916, "grad_norm": 0.16831424832344055, "learning_rate": 5.9030182567147905e-05, "loss": 2.0024, "step": 32800 }, { "epoch": 3.615850302696753, "grad_norm": 0.17562079429626465, "learning_rate": 5.859664873262483e-05, "loss": 2.0351, "step": 32850 }, { "epoch": 3.6213538800220144, "grad_norm": 0.18190255761146545, "learning_rate": 5.816432599643719e-05, "loss": 2.0117, "step": 32900 }, { "epoch": 3.626857457347276, "grad_norm": 0.1830035150051117, "learning_rate": 5.773322008687972e-05, "loss": 2.012, "step": 32950 }, { "epoch": 3.6323610346725372, "grad_norm": 0.17998354136943817, "learning_rate": 5.730333671612435e-05, "loss": 1.9977, "step": 33000 }, { "epoch": 3.6378646119977986, "grad_norm": 0.1839103251695633, "learning_rate": 5.687468158014421e-05, "loss": 2.003, "step": 33050 }, { "epoch": 3.64336818932306, "grad_norm": 0.18631495535373688, "learning_rate": 5.6447260358638285e-05, "loss": 2.0072, "step": 33100 }, { "epoch": 3.6488717666483215, "grad_norm": 0.19609171152114868, "learning_rate": 5.602107871495615e-05, "loss": 1.9912, "step": 33150 }, { "epoch": 3.654375343973583, "grad_norm": 0.16935598850250244, "learning_rate": 5.5596142296022954e-05, "loss": 2.0097, "step": 33200 }, { "epoch": 3.6598789212988443, "grad_norm": 0.16973347961902618, "learning_rate": 5.5172456732264435e-05, "loss": 2.0151, "step": 33250 }, { "epoch": 3.6653824986241057, "grad_norm": 0.17289844155311584, "learning_rate": 5.475002763753257e-05, "loss": 2.0393, "step": 33300 }, { "epoch": 3.670886075949367, "grad_norm": 0.1864277720451355, "learning_rate": 5.432886060903105e-05, "loss": 2.0069, "step": 33350 }, { "epoch": 3.6763896532746285, "grad_norm": 0.17381645739078522, "learning_rate": 5.390896122724111e-05, "loss": 2.0323, "step": 33400 }, { "epoch": 3.68189323059989, "grad_norm": 0.18182097375392914, "learning_rate": 5.349033505584767e-05, "loss": 2.0344, "step": 33450 }, { "epoch": 3.6873968079251513, "grad_norm": 0.17326848208904266, "learning_rate": 5.3072987641665373e-05, "loss": 1.9888, "step": 33500 }, { "epoch": 3.6929003852504128, "grad_norm": 0.17132443189620972, "learning_rate": 5.2656924514565515e-05, "loss": 1.9741, "step": 33550 }, { "epoch": 3.698403962575674, "grad_norm": 0.17236609756946564, "learning_rate": 5.224215118740243e-05, "loss": 1.9967, "step": 33600 }, { "epoch": 3.7039075399009356, "grad_norm": 0.16628102958202362, "learning_rate": 5.182867315594046e-05, "loss": 2.0036, "step": 33650 }, { "epoch": 3.709411117226197, "grad_norm": 0.17410129308700562, "learning_rate": 5.141649589878134e-05, "loss": 2.0173, "step": 33700 }, { "epoch": 3.7149146945514584, "grad_norm": 0.1739521622657776, "learning_rate": 5.100562487729148e-05, "loss": 1.9977, "step": 33750 }, { "epoch": 3.72041827187672, "grad_norm": 0.17988687753677368, "learning_rate": 5.059606553552956e-05, "loss": 1.9964, "step": 33800 }, { "epoch": 3.725921849201981, "grad_norm": 0.17351609468460083, "learning_rate": 5.018782330017448e-05, "loss": 2.0078, "step": 33850 }, { "epoch": 3.7314254265272426, "grad_norm": 0.176718607544899, "learning_rate": 4.9780903580453435e-05, "loss": 2.0036, "step": 33900 }, { "epoch": 3.736929003852504, "grad_norm": 0.17504329979419708, "learning_rate": 4.937531176807023e-05, "loss": 1.9837, "step": 33950 }, { "epoch": 3.7424325811777654, "grad_norm": 0.179831862449646, "learning_rate": 4.897105323713374e-05, "loss": 1.999, "step": 34000 }, { "epoch": 3.747936158503027, "grad_norm": 0.17752991616725922, "learning_rate": 4.8568133344086986e-05, "loss": 2.0194, "step": 34050 }, { "epoch": 3.7534397358282883, "grad_norm": 0.17888140678405762, "learning_rate": 4.8166557427635774e-05, "loss": 2.0108, "step": 34100 }, { "epoch": 3.7589433131535497, "grad_norm": 0.17585836350917816, "learning_rate": 4.776633080867834e-05, "loss": 2.0421, "step": 34150 }, { "epoch": 3.764446890478811, "grad_norm": 0.16860149800777435, "learning_rate": 4.736745879023457e-05, "loss": 2.0126, "step": 34200 }, { "epoch": 3.7699504678040725, "grad_norm": 0.17601364850997925, "learning_rate": 4.6969946657375865e-05, "loss": 2.0127, "step": 34250 }, { "epoch": 3.775454045129334, "grad_norm": 0.19177961349487305, "learning_rate": 4.657379967715511e-05, "loss": 2.0234, "step": 34300 }, { "epoch": 3.7809576224545953, "grad_norm": 0.19012008607387543, "learning_rate": 4.617902309853685e-05, "loss": 1.976, "step": 34350 }, { "epoch": 3.7864611997798567, "grad_norm": 0.18448469042778015, "learning_rate": 4.578562215232766e-05, "loss": 1.9897, "step": 34400 }, { "epoch": 3.791964777105118, "grad_norm": 0.18167300522327423, "learning_rate": 4.539360205110701e-05, "loss": 2.0088, "step": 34450 }, { "epoch": 3.7974683544303796, "grad_norm": 0.1806318610906601, "learning_rate": 4.500296798915813e-05, "loss": 1.9949, "step": 34500 }, { "epoch": 3.802971931755641, "grad_norm": 0.17676065862178802, "learning_rate": 4.461372514239911e-05, "loss": 1.9966, "step": 34550 }, { "epoch": 3.8084755090809024, "grad_norm": 0.16621781885623932, "learning_rate": 4.422587866831446e-05, "loss": 1.9952, "step": 34600 }, { "epoch": 3.813979086406164, "grad_norm": 0.16662217676639557, "learning_rate": 4.3839433705886646e-05, "loss": 2.0032, "step": 34650 }, { "epoch": 3.819482663731425, "grad_norm": 0.15986546874046326, "learning_rate": 4.3454395375527966e-05, "loss": 1.9974, "step": 34700 }, { "epoch": 3.8249862410566866, "grad_norm": 0.18636077642440796, "learning_rate": 4.307076877901293e-05, "loss": 1.9959, "step": 34750 }, { "epoch": 3.830489818381948, "grad_norm": 0.17939697206020355, "learning_rate": 4.268855899941039e-05, "loss": 1.9923, "step": 34800 }, { "epoch": 3.8359933957072094, "grad_norm": 0.1743779331445694, "learning_rate": 4.2307771101016366e-05, "loss": 2.0055, "step": 34850 }, { "epoch": 3.841496973032471, "grad_norm": 0.1767256259918213, "learning_rate": 4.192841012928691e-05, "loss": 1.9834, "step": 34900 }, { "epoch": 3.8470005503577323, "grad_norm": 0.1775825172662735, "learning_rate": 4.1550481110771224e-05, "loss": 2.01, "step": 34950 }, { "epoch": 3.852504127682994, "grad_norm": 0.1638861745595932, "learning_rate": 4.1173989053045024e-05, "loss": 2.0001, "step": 35000 }, { "epoch": 3.8580077050082555, "grad_norm": 0.1761082410812378, "learning_rate": 4.079893894464429e-05, "loss": 2.0041, "step": 35050 }, { "epoch": 3.863511282333517, "grad_norm": 0.17102594673633575, "learning_rate": 4.042533575499914e-05, "loss": 2.003, "step": 35100 }, { "epoch": 3.8690148596587783, "grad_norm": 0.19267229735851288, "learning_rate": 4.005318443436779e-05, "loss": 1.973, "step": 35150 }, { "epoch": 3.8745184369840397, "grad_norm": 0.16570790112018585, "learning_rate": 3.968248991377137e-05, "loss": 1.9834, "step": 35200 }, { "epoch": 3.880022014309301, "grad_norm": 0.16816405951976776, "learning_rate": 3.9313257104928094e-05, "loss": 2.0011, "step": 35250 }, { "epoch": 3.8855255916345626, "grad_norm": 0.16395322978496552, "learning_rate": 3.8945490900188616e-05, "loss": 1.9771, "step": 35300 }, { "epoch": 3.891029168959824, "grad_norm": 0.17911794781684875, "learning_rate": 3.857919617247097e-05, "loss": 2.0, "step": 35350 }, { "epoch": 3.8965327462850854, "grad_norm": 0.18344633281230927, "learning_rate": 3.821437777519607e-05, "loss": 1.9952, "step": 35400 }, { "epoch": 3.902036323610347, "grad_norm": 0.1782706081867218, "learning_rate": 3.785104054222334e-05, "loss": 1.9952, "step": 35450 }, { "epoch": 3.907539900935608, "grad_norm": 0.17840951681137085, "learning_rate": 3.74891892877868e-05, "loss": 1.9994, "step": 35500 }, { "epoch": 3.9130434782608696, "grad_norm": 0.19014611840248108, "learning_rate": 3.7128828806431046e-05, "loss": 2.0023, "step": 35550 }, { "epoch": 3.918547055586131, "grad_norm": 0.17002084851264954, "learning_rate": 3.6769963872947996e-05, "loss": 2.0068, "step": 35600 }, { "epoch": 3.9240506329113924, "grad_norm": 0.16855107247829437, "learning_rate": 3.641259924231345e-05, "loss": 1.9847, "step": 35650 }, { "epoch": 3.929554210236654, "grad_norm": 0.190132737159729, "learning_rate": 3.605673964962414e-05, "loss": 1.9955, "step": 35700 }, { "epoch": 3.9350577875619153, "grad_norm": 0.1558249294757843, "learning_rate": 3.570238981003498e-05, "loss": 1.9861, "step": 35750 }, { "epoch": 3.9405613648871767, "grad_norm": 0.18199962377548218, "learning_rate": 3.53495544186966e-05, "loss": 2.0136, "step": 35800 }, { "epoch": 3.946064942212438, "grad_norm": 0.17813271284103394, "learning_rate": 3.499823815069306e-05, "loss": 2.0012, "step": 35850 }, { "epoch": 3.9515685195376995, "grad_norm": 0.18256065249443054, "learning_rate": 3.464844566098008e-05, "loss": 1.9892, "step": 35900 }, { "epoch": 3.957072096862961, "grad_norm": 0.1784873753786087, "learning_rate": 3.4300181584323126e-05, "loss": 1.9983, "step": 35950 }, { "epoch": 3.9625756741882223, "grad_norm": 0.18530665338039398, "learning_rate": 3.3953450535236226e-05, "loss": 2.0357, "step": 36000 }, { "epoch": 3.9680792515134837, "grad_norm": 0.17063435912132263, "learning_rate": 3.360825710792068e-05, "loss": 2.0029, "step": 36050 }, { "epoch": 3.973582828838745, "grad_norm": 0.1718268245458603, "learning_rate": 3.326460587620427e-05, "loss": 1.9762, "step": 36100 }, { "epoch": 3.9790864061640066, "grad_norm": 0.15901614725589752, "learning_rate": 3.292250139348057e-05, "loss": 1.9952, "step": 36150 }, { "epoch": 3.984589983489268, "grad_norm": 0.18968409299850464, "learning_rate": 3.258194819264871e-05, "loss": 2.0127, "step": 36200 }, { "epoch": 3.9900935608145294, "grad_norm": 0.18861430883407593, "learning_rate": 3.2242950786053296e-05, "loss": 1.9981, "step": 36250 }, { "epoch": 3.995597138139791, "grad_norm": 0.1702040284872055, "learning_rate": 3.1905513665424445e-05, "loss": 2.0023, "step": 36300 }, { "epoch": 4.001100715465053, "grad_norm": 0.18254053592681885, "learning_rate": 3.156964130181859e-05, "loss": 2.0038, "step": 36350 }, { "epoch": 4.006604292790314, "grad_norm": 0.18867474794387817, "learning_rate": 3.1235338145558925e-05, "loss": 1.9139, "step": 36400 }, { "epoch": 4.0121078701155755, "grad_norm": 0.1872694045305252, "learning_rate": 3.090260862617671e-05, "loss": 1.9484, "step": 36450 }, { "epoch": 4.017611447440837, "grad_norm": 0.17595593631267548, "learning_rate": 3.057145715235236e-05, "loss": 1.9543, "step": 36500 }, { "epoch": 4.023115024766098, "grad_norm": 0.17811767756938934, "learning_rate": 3.0241888111857145e-05, "loss": 1.9268, "step": 36550 }, { "epoch": 4.02861860209136, "grad_norm": 0.17598308622837067, "learning_rate": 2.9913905871495037e-05, "loss": 1.9439, "step": 36600 }, { "epoch": 4.034122179416621, "grad_norm": 0.19237573444843292, "learning_rate": 2.9587514777044874e-05, "loss": 1.9044, "step": 36650 }, { "epoch": 4.0396257567418825, "grad_norm": 0.1873626410961151, "learning_rate": 2.926271915320259e-05, "loss": 1.943, "step": 36700 }, { "epoch": 4.045129334067144, "grad_norm": 0.2001618891954422, "learning_rate": 2.893952330352423e-05, "loss": 1.9311, "step": 36750 }, { "epoch": 4.050632911392405, "grad_norm": 0.1838996410369873, "learning_rate": 2.86179315103687e-05, "loss": 1.9373, "step": 36800 }, { "epoch": 4.056136488717667, "grad_norm": 0.16863055527210236, "learning_rate": 2.8297948034841074e-05, "loss": 1.9289, "step": 36850 }, { "epoch": 4.061640066042928, "grad_norm": 0.19162894785404205, "learning_rate": 2.797957711673619e-05, "loss": 1.9293, "step": 36900 }, { "epoch": 4.06714364336819, "grad_norm": 0.1854933500289917, "learning_rate": 2.7662822974482423e-05, "loss": 1.9336, "step": 36950 }, { "epoch": 4.072647220693451, "grad_norm": 0.18568897247314453, "learning_rate": 2.7347689805085733e-05, "loss": 1.9431, "step": 37000 }, { "epoch": 4.078150798018712, "grad_norm": 0.1947336494922638, "learning_rate": 2.7034181784074205e-05, "loss": 1.9345, "step": 37050 }, { "epoch": 4.083654375343974, "grad_norm": 0.18774710595607758, "learning_rate": 2.672230306544254e-05, "loss": 1.9364, "step": 37100 }, { "epoch": 4.089157952669235, "grad_norm": 0.1984746754169464, "learning_rate": 2.6412057781597223e-05, "loss": 1.954, "step": 37150 }, { "epoch": 4.094661529994497, "grad_norm": 0.1908876746892929, "learning_rate": 2.61034500433016e-05, "loss": 1.9422, "step": 37200 }, { "epoch": 4.100165107319758, "grad_norm": 0.18860432505607605, "learning_rate": 2.579648393962147e-05, "loss": 1.9649, "step": 37250 }, { "epoch": 4.105668684645019, "grad_norm": 0.18678942322731018, "learning_rate": 2.5491163537870922e-05, "loss": 1.9544, "step": 37300 }, { "epoch": 4.111172261970281, "grad_norm": 0.19560185074806213, "learning_rate": 2.5187492883558435e-05, "loss": 1.9396, "step": 37350 }, { "epoch": 4.116675839295542, "grad_norm": 0.1825692355632782, "learning_rate": 2.488547600033317e-05, "loss": 1.949, "step": 37400 }, { "epoch": 4.122179416620804, "grad_norm": 0.1951771229505539, "learning_rate": 2.4585116889931873e-05, "loss": 1.9482, "step": 37450 }, { "epoch": 4.127682993946065, "grad_norm": 0.1859281212091446, "learning_rate": 2.4286419532125668e-05, "loss": 1.9512, "step": 37500 }, { "epoch": 4.1331865712713265, "grad_norm": 0.19174982607364655, "learning_rate": 2.3989387884667365e-05, "loss": 1.9358, "step": 37550 }, { "epoch": 4.138690148596588, "grad_norm": 0.187012180685997, "learning_rate": 2.369402588323908e-05, "loss": 1.9114, "step": 37600 }, { "epoch": 4.144193725921849, "grad_norm": 0.20616699755191803, "learning_rate": 2.340033744140004e-05, "loss": 1.8965, "step": 37650 }, { "epoch": 4.149697303247111, "grad_norm": 0.17636051774024963, "learning_rate": 2.3108326450534725e-05, "loss": 1.9549, "step": 37700 }, { "epoch": 4.155200880572372, "grad_norm": 0.19219453632831573, "learning_rate": 2.281799677980136e-05, "loss": 1.8911, "step": 37750 }, { "epoch": 4.1607044578976335, "grad_norm": 0.19742678105831146, "learning_rate": 2.2529352276080513e-05, "loss": 1.9562, "step": 37800 }, { "epoch": 4.166208035222895, "grad_norm": 0.18049471080303192, "learning_rate": 2.2242396763924292e-05, "loss": 1.9169, "step": 37850 }, { "epoch": 4.171711612548156, "grad_norm": 0.2108810693025589, "learning_rate": 2.1957134045505588e-05, "loss": 1.9363, "step": 37900 }, { "epoch": 4.177215189873418, "grad_norm": 0.20711076259613037, "learning_rate": 2.1673567900567667e-05, "loss": 1.9685, "step": 37950 }, { "epoch": 4.182718767198679, "grad_norm": 0.18670059740543365, "learning_rate": 2.139170208637415e-05, "loss": 1.9537, "step": 38000 }, { "epoch": 4.188222344523941, "grad_norm": 0.17974555492401123, "learning_rate": 2.1111540337659227e-05, "loss": 1.9476, "step": 38050 }, { "epoch": 4.193725921849202, "grad_norm": 0.18271256983280182, "learning_rate": 2.083308636657811e-05, "loss": 1.9535, "step": 38100 }, { "epoch": 4.199229499174463, "grad_norm": 0.19260470569133759, "learning_rate": 2.0556343862657855e-05, "loss": 1.9477, "step": 38150 }, { "epoch": 4.204733076499725, "grad_norm": 0.19317923486232758, "learning_rate": 2.0281316492748595e-05, "loss": 1.9518, "step": 38200 }, { "epoch": 4.210236653824986, "grad_norm": 0.1854427605867386, "learning_rate": 2.0008007900974738e-05, "loss": 1.9245, "step": 38250 }, { "epoch": 4.215740231150248, "grad_norm": 0.1781974732875824, "learning_rate": 1.9736421708686934e-05, "loss": 1.9387, "step": 38300 }, { "epoch": 4.221243808475509, "grad_norm": 0.18503354489803314, "learning_rate": 1.946656151441389e-05, "loss": 1.9683, "step": 38350 }, { "epoch": 4.2267473858007705, "grad_norm": 0.19728437066078186, "learning_rate": 1.9198430893814798e-05, "loss": 1.9546, "step": 38400 }, { "epoch": 4.232250963126032, "grad_norm": 0.1957421451807022, "learning_rate": 1.89320333996319e-05, "loss": 1.9526, "step": 38450 }, { "epoch": 4.237754540451293, "grad_norm": 0.18944865465164185, "learning_rate": 1.866737256164349e-05, "loss": 1.9467, "step": 38500 }, { "epoch": 4.243258117776555, "grad_norm": 0.1760573536157608, "learning_rate": 1.8404451886617005e-05, "loss": 1.9142, "step": 38550 }, { "epoch": 4.248761695101816, "grad_norm": 0.19618487358093262, "learning_rate": 1.8143274858262702e-05, "loss": 1.9491, "step": 38600 }, { "epoch": 4.2542652724270775, "grad_norm": 0.19029423594474792, "learning_rate": 1.7883844937187495e-05, "loss": 1.9553, "step": 38650 }, { "epoch": 4.259768849752339, "grad_norm": 0.19688698649406433, "learning_rate": 1.762616556084891e-05, "loss": 1.9698, "step": 38700 }, { "epoch": 4.2652724270776, "grad_norm": 0.16535942256450653, "learning_rate": 1.737024014350983e-05, "loss": 1.9395, "step": 38750 }, { "epoch": 4.270776004402862, "grad_norm": 0.19671253859996796, "learning_rate": 1.7116072076193e-05, "loss": 1.9343, "step": 38800 }, { "epoch": 4.276279581728123, "grad_norm": 0.18672049045562744, "learning_rate": 1.6863664726636278e-05, "loss": 1.939, "step": 38850 }, { "epoch": 4.281783159053385, "grad_norm": 0.19714199006557465, "learning_rate": 1.6613021439247914e-05, "loss": 1.9455, "step": 38900 }, { "epoch": 4.287286736378646, "grad_norm": 0.2018548846244812, "learning_rate": 1.6364145535062172e-05, "loss": 1.9169, "step": 38950 }, { "epoch": 4.292790313703907, "grad_norm": 0.19057820737361908, "learning_rate": 1.611704031169555e-05, "loss": 1.956, "step": 39000 }, { "epoch": 4.298293891029169, "grad_norm": 0.18719470500946045, "learning_rate": 1.5871709043302876e-05, "loss": 1.9483, "step": 39050 }, { "epoch": 4.30379746835443, "grad_norm": 0.19479648768901825, "learning_rate": 1.5628154980533984e-05, "loss": 1.9507, "step": 39100 }, { "epoch": 4.309301045679692, "grad_norm": 0.2034020870923996, "learning_rate": 1.538638135049071e-05, "loss": 1.9343, "step": 39150 }, { "epoch": 4.314804623004953, "grad_norm": 0.18177463114261627, "learning_rate": 1.5146391356684029e-05, "loss": 1.9481, "step": 39200 }, { "epoch": 4.3203082003302145, "grad_norm": 0.1850568652153015, "learning_rate": 1.4908188178991714e-05, "loss": 1.9324, "step": 39250 }, { "epoch": 4.325811777655476, "grad_norm": 0.1880549043416977, "learning_rate": 1.4671774973616102e-05, "loss": 1.9406, "step": 39300 }, { "epoch": 4.331315354980737, "grad_norm": 0.18556420505046844, "learning_rate": 1.4437154873042279e-05, "loss": 1.9516, "step": 39350 }, { "epoch": 4.336818932305999, "grad_norm": 0.19034932553768158, "learning_rate": 1.420433098599672e-05, "loss": 1.9477, "step": 39400 }, { "epoch": 4.34232250963126, "grad_norm": 0.18247225880622864, "learning_rate": 1.3973306397405909e-05, "loss": 1.9552, "step": 39450 }, { "epoch": 4.3478260869565215, "grad_norm": 0.19472351670265198, "learning_rate": 1.3744084168355612e-05, "loss": 1.9543, "step": 39500 }, { "epoch": 4.353329664281783, "grad_norm": 0.18814486265182495, "learning_rate": 1.3516667336050219e-05, "loss": 1.956, "step": 39550 }, { "epoch": 4.358833241607044, "grad_norm": 0.18639877438545227, "learning_rate": 1.3291058913772517e-05, "loss": 1.9262, "step": 39600 }, { "epoch": 4.364336818932306, "grad_norm": 0.19503189623355865, "learning_rate": 1.306726189084385e-05, "loss": 1.9768, "step": 39650 }, { "epoch": 4.369840396257567, "grad_norm": 0.1823161244392395, "learning_rate": 1.2845279232584354e-05, "loss": 1.9588, "step": 39700 }, { "epoch": 4.375343973582829, "grad_norm": 0.19661200046539307, "learning_rate": 1.2625113880273786e-05, "loss": 1.9576, "step": 39750 }, { "epoch": 4.38084755090809, "grad_norm": 0.17353294789791107, "learning_rate": 1.2406768751112572e-05, "loss": 1.9249, "step": 39800 }, { "epoch": 4.386351128233351, "grad_norm": 0.1832091063261032, "learning_rate": 1.2190246738183074e-05, "loss": 1.9169, "step": 39850 }, { "epoch": 4.391854705558613, "grad_norm": 0.18823185563087463, "learning_rate": 1.197555071041122e-05, "loss": 1.9368, "step": 39900 }, { "epoch": 4.397358282883874, "grad_norm": 0.1941109150648117, "learning_rate": 1.1762683512528642e-05, "loss": 1.96, "step": 39950 }, { "epoch": 4.402861860209136, "grad_norm": 0.1973351240158081, "learning_rate": 1.155164796503486e-05, "loss": 1.9304, "step": 40000 } ], "logging_steps": 50, "max_steps": 45425, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.91746903834624e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }