{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.999764280663157, "eval_steps": 500, "global_step": 3181, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0003142924491239098, "grad_norm": 19.593584060668945, "learning_rate": 3.134796238244514e-06, "loss": 7.2211, "step": 1 }, { "epoch": 0.001571462245619549, "grad_norm": 17.929481506347656, "learning_rate": 1.567398119122257e-05, "loss": 7.2713, "step": 5 }, { "epoch": 0.003142924491239098, "grad_norm": 17.97577476501465, "learning_rate": 3.134796238244514e-05, "loss": 7.2643, "step": 10 }, { "epoch": 0.004714386736858647, "grad_norm": 9.52695369720459, "learning_rate": 4.7021943573667716e-05, "loss": 6.9744, "step": 15 }, { "epoch": 0.006285848982478196, "grad_norm": 4.145598888397217, "learning_rate": 6.269592476489028e-05, "loss": 6.5934, "step": 20 }, { "epoch": 0.007857311228097744, "grad_norm": 3.3066627979278564, "learning_rate": 7.836990595611286e-05, "loss": 6.1739, "step": 25 }, { "epoch": 0.009428773473717294, "grad_norm": 3.240945339202881, "learning_rate": 9.404388714733543e-05, "loss": 5.9143, "step": 30 }, { "epoch": 0.011000235719336842, "grad_norm": 1.6531422138214111, "learning_rate": 0.000109717868338558, "loss": 5.6534, "step": 35 }, { "epoch": 0.012571697964956392, "grad_norm": 0.9349701404571533, "learning_rate": 0.00012539184952978057, "loss": 5.471, "step": 40 }, { "epoch": 0.01414316021057594, "grad_norm": 0.565790057182312, "learning_rate": 0.00014106583072100311, "loss": 5.3034, "step": 45 }, { "epoch": 0.01571462245619549, "grad_norm": 0.4517938494682312, "learning_rate": 0.00015673981191222572, "loss": 5.1125, "step": 50 }, { "epoch": 0.01728608470181504, "grad_norm": 0.4741256833076477, "learning_rate": 0.0001724137931034483, "loss": 5.009, "step": 55 }, { "epoch": 0.01885754694743459, "grad_norm": 0.4407913386821747, "learning_rate": 0.00018808777429467086, "loss": 4.9626, "step": 60 }, { "epoch": 0.020429009193054137, "grad_norm": 0.6502472758293152, "learning_rate": 0.00020376175548589344, "loss": 4.837, "step": 65 }, { "epoch": 0.022000471438673685, "grad_norm": 0.6139042377471924, "learning_rate": 0.000219435736677116, "loss": 4.7188, "step": 70 }, { "epoch": 0.023571933684293236, "grad_norm": 0.5190150141716003, "learning_rate": 0.00023510971786833856, "loss": 4.5818, "step": 75 }, { "epoch": 0.025143395929912785, "grad_norm": 0.4606360197067261, "learning_rate": 0.00025078369905956113, "loss": 4.4429, "step": 80 }, { "epoch": 0.026714858175532333, "grad_norm": 0.38270899653434753, "learning_rate": 0.0002664576802507837, "loss": 4.4131, "step": 85 }, { "epoch": 0.02828632042115188, "grad_norm": 0.4721100330352783, "learning_rate": 0.00028213166144200623, "loss": 4.248, "step": 90 }, { "epoch": 0.02985778266677143, "grad_norm": 0.4452402889728546, "learning_rate": 0.00029780564263322886, "loss": 4.1921, "step": 95 }, { "epoch": 0.03142924491239098, "grad_norm": 0.4978654980659485, "learning_rate": 0.00031347962382445143, "loss": 4.0356, "step": 100 }, { "epoch": 0.033000707158010525, "grad_norm": 0.6152021288871765, "learning_rate": 0.000329153605015674, "loss": 3.8516, "step": 105 }, { "epoch": 0.03457216940363008, "grad_norm": 0.4162922203540802, "learning_rate": 0.0003448275862068966, "loss": 3.8018, "step": 110 }, { "epoch": 0.03614363164924963, "grad_norm": 0.44047296047210693, "learning_rate": 0.0003605015673981191, "loss": 3.5919, "step": 115 }, { "epoch": 0.03771509389486918, "grad_norm": 0.4196017384529114, "learning_rate": 0.00037617554858934173, "loss": 3.5899, "step": 120 }, { "epoch": 0.039286556140488725, "grad_norm": 0.4823077917098999, "learning_rate": 0.00039184952978056425, "loss": 3.4321, "step": 125 }, { "epoch": 0.04085801838610827, "grad_norm": 0.3624018132686615, "learning_rate": 0.0004075235109717869, "loss": 3.3484, "step": 130 }, { "epoch": 0.04242948063172782, "grad_norm": 0.3260553181171417, "learning_rate": 0.0004231974921630094, "loss": 3.2639, "step": 135 }, { "epoch": 0.04400094287734737, "grad_norm": 0.38659989833831787, "learning_rate": 0.000438871473354232, "loss": 3.1999, "step": 140 }, { "epoch": 0.04557240512296692, "grad_norm": 0.4590894281864166, "learning_rate": 0.00045454545454545455, "loss": 3.1469, "step": 145 }, { "epoch": 0.04714386736858647, "grad_norm": 0.39775681495666504, "learning_rate": 0.0004702194357366771, "loss": 3.0631, "step": 150 }, { "epoch": 0.04871532961420602, "grad_norm": 0.43203604221343994, "learning_rate": 0.0004858934169278997, "loss": 3.0505, "step": 155 }, { "epoch": 0.05028679185982557, "grad_norm": 0.3682110607624054, "learning_rate": 0.0005015673981191223, "loss": 3.0316, "step": 160 }, { "epoch": 0.05185825410544512, "grad_norm": 0.32574963569641113, "learning_rate": 0.0005172413793103448, "loss": 3.0132, "step": 165 }, { "epoch": 0.053429716351064666, "grad_norm": 0.4408009946346283, "learning_rate": 0.0005329153605015674, "loss": 3.0166, "step": 170 }, { "epoch": 0.055001178596684214, "grad_norm": 0.3008413016796112, "learning_rate": 0.00054858934169279, "loss": 2.8596, "step": 175 }, { "epoch": 0.05657264084230376, "grad_norm": 0.3133067786693573, "learning_rate": 0.0005642633228840125, "loss": 2.9631, "step": 180 }, { "epoch": 0.05814410308792331, "grad_norm": 0.4327049255371094, "learning_rate": 0.0005799373040752351, "loss": 2.8666, "step": 185 }, { "epoch": 0.05971556533354286, "grad_norm": 0.44304072856903076, "learning_rate": 0.0005956112852664577, "loss": 2.7938, "step": 190 }, { "epoch": 0.06128702757916241, "grad_norm": 0.4220844805240631, "learning_rate": 0.0006112852664576803, "loss": 2.8245, "step": 195 }, { "epoch": 0.06285848982478195, "grad_norm": 0.36100244522094727, "learning_rate": 0.0006269592476489029, "loss": 2.7925, "step": 200 }, { "epoch": 0.0644299520704015, "grad_norm": 0.4607883095741272, "learning_rate": 0.0006426332288401254, "loss": 2.7257, "step": 205 }, { "epoch": 0.06600141431602105, "grad_norm": 0.3919202387332916, "learning_rate": 0.000658307210031348, "loss": 2.6557, "step": 210 }, { "epoch": 0.06757287656164061, "grad_norm": 0.3630659282207489, "learning_rate": 0.0006739811912225705, "loss": 2.7476, "step": 215 }, { "epoch": 0.06914433880726016, "grad_norm": 0.3189098834991455, "learning_rate": 0.0006896551724137932, "loss": 2.6834, "step": 220 }, { "epoch": 0.07071580105287971, "grad_norm": 0.3652380108833313, "learning_rate": 0.0007053291536050157, "loss": 2.6362, "step": 225 }, { "epoch": 0.07228726329849926, "grad_norm": 0.35851606726646423, "learning_rate": 0.0007210031347962382, "loss": 2.637, "step": 230 }, { "epoch": 0.0738587255441188, "grad_norm": 0.7696281671524048, "learning_rate": 0.0007366771159874608, "loss": 2.5676, "step": 235 }, { "epoch": 0.07543018778973835, "grad_norm": 0.6321612000465393, "learning_rate": 0.0007523510971786835, "loss": 2.6147, "step": 240 }, { "epoch": 0.0770016500353579, "grad_norm": 0.386197030544281, "learning_rate": 0.000768025078369906, "loss": 2.5907, "step": 245 }, { "epoch": 0.07857311228097745, "grad_norm": 0.4827702045440674, "learning_rate": 0.0007836990595611285, "loss": 2.5142, "step": 250 }, { "epoch": 0.080144574526597, "grad_norm": 0.292161762714386, "learning_rate": 0.0007993730407523511, "loss": 2.4859, "step": 255 }, { "epoch": 0.08171603677221655, "grad_norm": 0.33461418747901917, "learning_rate": 0.0008150470219435738, "loss": 2.5654, "step": 260 }, { "epoch": 0.0832874990178361, "grad_norm": 0.5843150019645691, "learning_rate": 0.0008307210031347962, "loss": 2.5168, "step": 265 }, { "epoch": 0.08485896126345564, "grad_norm": 0.5514886975288391, "learning_rate": 0.0008463949843260188, "loss": 2.4514, "step": 270 }, { "epoch": 0.08643042350907519, "grad_norm": 0.32894566655158997, "learning_rate": 0.0008620689655172414, "loss": 2.4504, "step": 275 }, { "epoch": 0.08800188575469474, "grad_norm": 0.30089420080184937, "learning_rate": 0.000877742946708464, "loss": 2.5003, "step": 280 }, { "epoch": 0.08957334800031429, "grad_norm": 0.41669008135795593, "learning_rate": 0.0008934169278996865, "loss": 2.4293, "step": 285 }, { "epoch": 0.09114481024593384, "grad_norm": 0.25633400678634644, "learning_rate": 0.0009090909090909091, "loss": 2.4447, "step": 290 }, { "epoch": 0.09271627249155338, "grad_norm": 0.27528685331344604, "learning_rate": 0.0009247648902821318, "loss": 2.4672, "step": 295 }, { "epoch": 0.09428773473717295, "grad_norm": 0.3092777132987976, "learning_rate": 0.0009404388714733542, "loss": 2.3757, "step": 300 }, { "epoch": 0.0958591969827925, "grad_norm": 0.29513272643089294, "learning_rate": 0.0009561128526645768, "loss": 2.4152, "step": 305 }, { "epoch": 0.09743065922841204, "grad_norm": 0.43028154969215393, "learning_rate": 0.0009717868338557994, "loss": 2.4507, "step": 310 }, { "epoch": 0.09900212147403159, "grad_norm": 0.33603623509407043, "learning_rate": 0.000987460815047022, "loss": 2.3948, "step": 315 }, { "epoch": 0.10057358371965114, "grad_norm": 0.36047083139419556, "learning_rate": 0.0009999996987684534, "loss": 2.3747, "step": 320 }, { "epoch": 0.10214504596527069, "grad_norm": 0.47083067893981934, "learning_rate": 0.0009999891557024327, "loss": 2.4144, "step": 325 }, { "epoch": 0.10371650821089023, "grad_norm": 0.3539304733276367, "learning_rate": 0.0009999635514220413, "loss": 2.365, "step": 330 }, { "epoch": 0.10528797045650978, "grad_norm": 0.3079536557197571, "learning_rate": 0.0009999228866985584, "loss": 2.3087, "step": 335 }, { "epoch": 0.10685943270212933, "grad_norm": 0.3304389715194702, "learning_rate": 0.0009998671627569314, "loss": 2.2729, "step": 340 }, { "epoch": 0.10843089494774888, "grad_norm": 0.3577703535556793, "learning_rate": 0.0009997963812757367, "loss": 2.362, "step": 345 }, { "epoch": 0.11000235719336843, "grad_norm": 0.3626112937927246, "learning_rate": 0.000999710544387131, "loss": 2.3232, "step": 350 }, { "epoch": 0.11157381943898798, "grad_norm": 0.33722984790802, "learning_rate": 0.000999609654676786, "loss": 2.3837, "step": 355 }, { "epoch": 0.11314528168460752, "grad_norm": 0.6007051467895508, "learning_rate": 0.0009994937151838103, "loss": 2.2621, "step": 360 }, { "epoch": 0.11471674393022707, "grad_norm": 0.24785129725933075, "learning_rate": 0.0009993627294006592, "loss": 2.3242, "step": 365 }, { "epoch": 0.11628820617584662, "grad_norm": 0.3804435431957245, "learning_rate": 0.000999216701273028, "loss": 2.2901, "step": 370 }, { "epoch": 0.11785966842146617, "grad_norm": 0.2868124544620514, "learning_rate": 0.000999055635199734, "loss": 2.2492, "step": 375 }, { "epoch": 0.11943113066708572, "grad_norm": 0.45313236117362976, "learning_rate": 0.0009988795360325836, "loss": 2.3535, "step": 380 }, { "epoch": 0.12100259291270528, "grad_norm": 0.37607526779174805, "learning_rate": 0.0009986884090762266, "loss": 2.2953, "step": 385 }, { "epoch": 0.12257405515832483, "grad_norm": 0.27437832951545715, "learning_rate": 0.000998482260087996, "loss": 2.2992, "step": 390 }, { "epoch": 0.12414551740394437, "grad_norm": 1.2427669763565063, "learning_rate": 0.0009982610952777348, "loss": 2.23, "step": 395 }, { "epoch": 0.1257169796495639, "grad_norm": 0.2712255120277405, "learning_rate": 0.0009980249213076085, "loss": 2.2312, "step": 400 }, { "epoch": 0.12728844189518346, "grad_norm": 0.3325655162334442, "learning_rate": 0.0009977737452919052, "loss": 2.24, "step": 405 }, { "epoch": 0.128859904140803, "grad_norm": 0.3058350384235382, "learning_rate": 0.0009975075747968203, "loss": 2.2233, "step": 410 }, { "epoch": 0.13043136638642255, "grad_norm": 0.46767985820770264, "learning_rate": 0.00099722641784023, "loss": 2.2153, "step": 415 }, { "epoch": 0.1320028286320421, "grad_norm": 0.27317899465560913, "learning_rate": 0.0009969302828914477, "loss": 2.2048, "step": 420 }, { "epoch": 0.13357429087766168, "grad_norm": 0.49403005838394165, "learning_rate": 0.0009966191788709714, "loss": 2.1963, "step": 425 }, { "epoch": 0.13514575312328123, "grad_norm": 0.28752318024635315, "learning_rate": 0.0009962931151502128, "loss": 2.2472, "step": 430 }, { "epoch": 0.13671721536890077, "grad_norm": 0.2463676631450653, "learning_rate": 0.000995952101551216, "loss": 2.1985, "step": 435 }, { "epoch": 0.13828867761452032, "grad_norm": 0.24134370684623718, "learning_rate": 0.000995596148346362, "loss": 2.2209, "step": 440 }, { "epoch": 0.13986013986013987, "grad_norm": 0.28822633624076843, "learning_rate": 0.000995225266258058, "loss": 2.1771, "step": 445 }, { "epoch": 0.14143160210575942, "grad_norm": 0.35365989804267883, "learning_rate": 0.0009948394664584155, "loss": 2.1384, "step": 450 }, { "epoch": 0.14300306435137897, "grad_norm": 0.25629669427871704, "learning_rate": 0.0009944387605689139, "loss": 2.1782, "step": 455 }, { "epoch": 0.14457452659699851, "grad_norm": 0.3197285532951355, "learning_rate": 0.0009940231606600494, "loss": 2.1927, "step": 460 }, { "epoch": 0.14614598884261806, "grad_norm": 0.2365492731332779, "learning_rate": 0.0009935926792509723, "loss": 2.1824, "step": 465 }, { "epoch": 0.1477174510882376, "grad_norm": 0.3567672073841095, "learning_rate": 0.000993147329309109, "loss": 2.1744, "step": 470 }, { "epoch": 0.14928891333385716, "grad_norm": 0.28331291675567627, "learning_rate": 0.0009926871242497731, "loss": 2.1513, "step": 475 }, { "epoch": 0.1508603755794767, "grad_norm": 0.2945978343486786, "learning_rate": 0.0009922120779357587, "loss": 2.1584, "step": 480 }, { "epoch": 0.15243183782509626, "grad_norm": 0.2360486388206482, "learning_rate": 0.0009917222046769252, "loss": 2.1771, "step": 485 }, { "epoch": 0.1540033000707158, "grad_norm": 0.26957565546035767, "learning_rate": 0.0009912175192297648, "loss": 2.1623, "step": 490 }, { "epoch": 0.15557476231633535, "grad_norm": 0.2816142141819, "learning_rate": 0.0009906980367969589, "loss": 2.1792, "step": 495 }, { "epoch": 0.1571462245619549, "grad_norm": 0.23650062084197998, "learning_rate": 0.0009901637730269192, "loss": 2.1513, "step": 500 }, { "epoch": 0.15871768680757445, "grad_norm": 0.20437929034233093, "learning_rate": 0.0009896147440133173, "loss": 2.1128, "step": 505 }, { "epoch": 0.160289149053194, "grad_norm": 0.4117949903011322, "learning_rate": 0.0009890509662945992, "loss": 2.1282, "step": 510 }, { "epoch": 0.16186061129881354, "grad_norm": 0.2904440462589264, "learning_rate": 0.0009884724568534873, "loss": 2.1036, "step": 515 }, { "epoch": 0.1634320735444331, "grad_norm": 0.30711230635643005, "learning_rate": 0.000987879233116469, "loss": 2.1463, "step": 520 }, { "epoch": 0.16500353579005264, "grad_norm": 0.28438258171081543, "learning_rate": 0.0009872713129532717, "loss": 2.0971, "step": 525 }, { "epoch": 0.1665749980356722, "grad_norm": 0.2756672203540802, "learning_rate": 0.000986648714676324, "loss": 2.1291, "step": 530 }, { "epoch": 0.16814646028129174, "grad_norm": 0.23554356396198273, "learning_rate": 0.0009860114570402056, "loss": 2.0797, "step": 535 }, { "epoch": 0.16971792252691129, "grad_norm": 0.3189121186733246, "learning_rate": 0.0009853595592410798, "loss": 2.1122, "step": 540 }, { "epoch": 0.17128938477253083, "grad_norm": 0.2167348712682724, "learning_rate": 0.000984693040916118, "loss": 2.1103, "step": 545 }, { "epoch": 0.17286084701815038, "grad_norm": 0.3020518124103546, "learning_rate": 0.0009840119221429062, "loss": 2.1314, "step": 550 }, { "epoch": 0.17443230926376993, "grad_norm": 0.2934524118900299, "learning_rate": 0.0009833162234388414, "loss": 2.1118, "step": 555 }, { "epoch": 0.17600377150938948, "grad_norm": 0.4206439256668091, "learning_rate": 0.0009826059657605123, "loss": 2.0729, "step": 560 }, { "epoch": 0.17757523375500903, "grad_norm": 0.2568371593952179, "learning_rate": 0.0009818811705030695, "loss": 2.0941, "step": 565 }, { "epoch": 0.17914669600062857, "grad_norm": 0.3140527009963989, "learning_rate": 0.00098114185949958, "loss": 2.0994, "step": 570 }, { "epoch": 0.18071815824624812, "grad_norm": 0.2604500651359558, "learning_rate": 0.0009803880550203698, "loss": 2.1462, "step": 575 }, { "epoch": 0.18228962049186767, "grad_norm": 0.3531113266944885, "learning_rate": 0.0009796197797723532, "loss": 2.0743, "step": 580 }, { "epoch": 0.18386108273748722, "grad_norm": 0.24229241907596588, "learning_rate": 0.0009788370568983484, "loss": 2.0614, "step": 585 }, { "epoch": 0.18543254498310677, "grad_norm": 0.43031787872314453, "learning_rate": 0.000978039909976381, "loss": 2.1287, "step": 590 }, { "epoch": 0.18700400722872634, "grad_norm": 0.210410937666893, "learning_rate": 0.0009772283630189726, "loss": 2.0234, "step": 595 }, { "epoch": 0.1885754694743459, "grad_norm": 0.28944680094718933, "learning_rate": 0.0009764024404724194, "loss": 2.0951, "step": 600 }, { "epoch": 0.19014693171996544, "grad_norm": 0.2910195589065552, "learning_rate": 0.0009755621672160539, "loss": 2.077, "step": 605 }, { "epoch": 0.191718393965585, "grad_norm": 0.27437564730644226, "learning_rate": 0.0009747075685614961, "loss": 2.0639, "step": 610 }, { "epoch": 0.19328985621120454, "grad_norm": 0.26965805888175964, "learning_rate": 0.0009738386702518917, "loss": 2.0205, "step": 615 }, { "epoch": 0.19486131845682408, "grad_norm": 0.22133469581604004, "learning_rate": 0.0009729554984611357, "loss": 2.0871, "step": 620 }, { "epoch": 0.19643278070244363, "grad_norm": 0.26669520139694214, "learning_rate": 0.0009720580797930845, "loss": 2.1095, "step": 625 }, { "epoch": 0.19800424294806318, "grad_norm": 0.22317343950271606, "learning_rate": 0.0009711464412807542, "loss": 2.0449, "step": 630 }, { "epoch": 0.19957570519368273, "grad_norm": 0.4366435706615448, "learning_rate": 0.0009702206103855065, "loss": 2.0923, "step": 635 }, { "epoch": 0.20114716743930228, "grad_norm": 0.29953792691230774, "learning_rate": 0.0009692806149962215, "loss": 2.0278, "step": 640 }, { "epoch": 0.20271862968492182, "grad_norm": 0.3391275405883789, "learning_rate": 0.0009683264834284575, "loss": 2.0462, "step": 645 }, { "epoch": 0.20429009193054137, "grad_norm": 0.3684813976287842, "learning_rate": 0.0009673582444235978, "loss": 2.0406, "step": 650 }, { "epoch": 0.20586155417616092, "grad_norm": 0.3173034191131592, "learning_rate": 0.0009663759271479858, "loss": 1.9821, "step": 655 }, { "epoch": 0.20743301642178047, "grad_norm": 0.28240126371383667, "learning_rate": 0.0009653795611920448, "loss": 2.0112, "step": 660 }, { "epoch": 0.20900447866740002, "grad_norm": 0.31004390120506287, "learning_rate": 0.000964369176569389, "loss": 2.0155, "step": 665 }, { "epoch": 0.21057594091301957, "grad_norm": 0.29802826046943665, "learning_rate": 0.0009633448037159167, "loss": 2.0641, "step": 670 }, { "epoch": 0.21214740315863911, "grad_norm": 0.2916626036167145, "learning_rate": 0.0009623064734888958, "loss": 1.9948, "step": 675 }, { "epoch": 0.21371886540425866, "grad_norm": 0.24406270682811737, "learning_rate": 0.0009612542171660328, "loss": 1.9714, "step": 680 }, { "epoch": 0.2152903276498782, "grad_norm": 0.31989896297454834, "learning_rate": 0.0009601880664445312, "loss": 2.0075, "step": 685 }, { "epoch": 0.21686178989549776, "grad_norm": 0.16441141068935394, "learning_rate": 0.000959108053440137, "loss": 1.9964, "step": 690 }, { "epoch": 0.2184332521411173, "grad_norm": 0.27429190278053284, "learning_rate": 0.0009580142106861702, "loss": 2.013, "step": 695 }, { "epoch": 0.22000471438673685, "grad_norm": 0.21225886046886444, "learning_rate": 0.0009569065711325461, "loss": 2.0092, "step": 700 }, { "epoch": 0.2215761766323564, "grad_norm": 0.24281561374664307, "learning_rate": 0.0009557851681447816, "loss": 2.0172, "step": 705 }, { "epoch": 0.22314763887797595, "grad_norm": 0.20721524953842163, "learning_rate": 0.0009546500355029912, "loss": 2.0265, "step": 710 }, { "epoch": 0.2247191011235955, "grad_norm": 0.268388032913208, "learning_rate": 0.0009535012074008687, "loss": 2.0329, "step": 715 }, { "epoch": 0.22629056336921505, "grad_norm": 0.25884491205215454, "learning_rate": 0.0009523387184446569, "loss": 1.9673, "step": 720 }, { "epoch": 0.2278620256148346, "grad_norm": 0.26054859161376953, "learning_rate": 0.0009511626036521064, "loss": 1.9775, "step": 725 }, { "epoch": 0.22943348786045414, "grad_norm": 0.23918680846691132, "learning_rate": 0.0009499728984514194, "loss": 1.9753, "step": 730 }, { "epoch": 0.2310049501060737, "grad_norm": 0.18698784708976746, "learning_rate": 0.0009487696386801834, "loss": 2.0547, "step": 735 }, { "epoch": 0.23257641235169324, "grad_norm": 0.23608291149139404, "learning_rate": 0.0009475528605842913, "loss": 1.9701, "step": 740 }, { "epoch": 0.2341478745973128, "grad_norm": 0.20181405544281006, "learning_rate": 0.0009463226008168493, "loss": 1.9994, "step": 745 }, { "epoch": 0.23571933684293234, "grad_norm": 0.421236515045166, "learning_rate": 0.0009450788964370737, "loss": 1.9943, "step": 750 }, { "epoch": 0.23729079908855188, "grad_norm": 0.2977412939071655, "learning_rate": 0.0009438217849091732, "loss": 1.998, "step": 755 }, { "epoch": 0.23886226133417143, "grad_norm": 0.3429923355579376, "learning_rate": 0.0009425513041012219, "loss": 2.0098, "step": 760 }, { "epoch": 0.240433723579791, "grad_norm": 0.28545793890953064, "learning_rate": 0.0009412674922840173, "loss": 1.9684, "step": 765 }, { "epoch": 0.24200518582541056, "grad_norm": 0.1793077141046524, "learning_rate": 0.0009399703881299281, "loss": 2.0074, "step": 770 }, { "epoch": 0.2435766480710301, "grad_norm": 0.280935674905777, "learning_rate": 0.0009386600307117292, "loss": 2.0081, "step": 775 }, { "epoch": 0.24514811031664965, "grad_norm": 0.30971866846084595, "learning_rate": 0.0009373364595014245, "loss": 1.9568, "step": 780 }, { "epoch": 0.2467195725622692, "grad_norm": 0.26543349027633667, "learning_rate": 0.0009359997143690587, "loss": 1.9926, "step": 785 }, { "epoch": 0.24829103480788875, "grad_norm": 0.27692025899887085, "learning_rate": 0.0009346498355815143, "loss": 1.9465, "step": 790 }, { "epoch": 0.2498624970535083, "grad_norm": 0.2559816539287567, "learning_rate": 0.0009332868638013016, "loss": 1.9473, "step": 795 }, { "epoch": 0.2514339592991278, "grad_norm": 0.330941766500473, "learning_rate": 0.0009319108400853309, "loss": 1.9892, "step": 800 }, { "epoch": 0.25300542154474737, "grad_norm": 0.33171382546424866, "learning_rate": 0.0009305218058836777, "loss": 1.9819, "step": 805 }, { "epoch": 0.2545768837903669, "grad_norm": 0.49775874614715576, "learning_rate": 0.0009291198030383335, "loss": 1.9979, "step": 810 }, { "epoch": 0.25614834603598646, "grad_norm": 0.4397469758987427, "learning_rate": 0.0009277048737819445, "loss": 1.9923, "step": 815 }, { "epoch": 0.257719808281606, "grad_norm": 0.38636040687561035, "learning_rate": 0.0009262770607365412, "loss": 1.989, "step": 820 }, { "epoch": 0.25929127052722556, "grad_norm": 0.19002190232276917, "learning_rate": 0.0009248364069122531, "loss": 1.9449, "step": 825 }, { "epoch": 0.2608627327728451, "grad_norm": 0.21801802515983582, "learning_rate": 0.0009233829557060136, "loss": 1.9449, "step": 830 }, { "epoch": 0.26243419501846466, "grad_norm": 0.2232774794101715, "learning_rate": 0.0009219167509002526, "loss": 2.0174, "step": 835 }, { "epoch": 0.2640056572640842, "grad_norm": 0.21366550028324127, "learning_rate": 0.0009204378366615778, "loss": 1.9458, "step": 840 }, { "epoch": 0.26557711950970375, "grad_norm": 0.22979532182216644, "learning_rate": 0.0009189462575394443, "loss": 1.968, "step": 845 }, { "epoch": 0.26714858175532336, "grad_norm": 0.3232949376106262, "learning_rate": 0.0009174420584648122, "loss": 1.9344, "step": 850 }, { "epoch": 0.2687200440009429, "grad_norm": 0.3545362055301666, "learning_rate": 0.0009159252847487938, "loss": 1.9558, "step": 855 }, { "epoch": 0.27029150624656245, "grad_norm": 0.25105616450309753, "learning_rate": 0.0009143959820812882, "loss": 1.9726, "step": 860 }, { "epoch": 0.271862968492182, "grad_norm": 0.37967053055763245, "learning_rate": 0.0009128541965296051, "loss": 1.9593, "step": 865 }, { "epoch": 0.27343443073780155, "grad_norm": 0.47272658348083496, "learning_rate": 0.0009112999745370774, "loss": 1.9306, "step": 870 }, { "epoch": 0.2750058929834211, "grad_norm": 0.40520063042640686, "learning_rate": 0.0009097333629216616, "loss": 2.0052, "step": 875 }, { "epoch": 0.27657735522904064, "grad_norm": 0.16213001310825348, "learning_rate": 0.000908154408874528, "loss": 1.9685, "step": 880 }, { "epoch": 0.2781488174746602, "grad_norm": 0.2853317856788635, "learning_rate": 0.000906563159958639, "loss": 1.972, "step": 885 }, { "epoch": 0.27972027972027974, "grad_norm": 0.3446250259876251, "learning_rate": 0.0009049596641073161, "loss": 1.9114, "step": 890 }, { "epoch": 0.2812917419658993, "grad_norm": 0.19377753138542175, "learning_rate": 0.0009033439696227966, "loss": 1.9119, "step": 895 }, { "epoch": 0.28286320421151884, "grad_norm": 0.249973326921463, "learning_rate": 0.0009017161251747779, "loss": 1.9657, "step": 900 }, { "epoch": 0.2844346664571384, "grad_norm": 0.26847654581069946, "learning_rate": 0.0009000761797989521, "loss": 1.9543, "step": 905 }, { "epoch": 0.28600612870275793, "grad_norm": 0.23439465463161469, "learning_rate": 0.0008984241828955281, "loss": 1.9104, "step": 910 }, { "epoch": 0.2875775909483775, "grad_norm": 0.37913578748703003, "learning_rate": 0.0008967601842277444, "loss": 1.9149, "step": 915 }, { "epoch": 0.28914905319399703, "grad_norm": 0.2542668581008911, "learning_rate": 0.0008950842339203695, "loss": 1.9682, "step": 920 }, { "epoch": 0.2907205154396166, "grad_norm": 0.3627631962299347, "learning_rate": 0.0008933963824581919, "loss": 1.9878, "step": 925 }, { "epoch": 0.2922919776852361, "grad_norm": 0.25258249044418335, "learning_rate": 0.0008916966806844996, "loss": 1.9425, "step": 930 }, { "epoch": 0.2938634399308557, "grad_norm": 0.502772867679596, "learning_rate": 0.0008899851797995489, "loss": 1.956, "step": 935 }, { "epoch": 0.2954349021764752, "grad_norm": 0.3263178765773773, "learning_rate": 0.0008882619313590213, "loss": 1.9896, "step": 940 }, { "epoch": 0.29700636442209477, "grad_norm": 0.24707098305225372, "learning_rate": 0.0008865269872724708, "loss": 1.901, "step": 945 }, { "epoch": 0.2985778266677143, "grad_norm": 0.26257702708244324, "learning_rate": 0.000884780399801761, "loss": 1.957, "step": 950 }, { "epoch": 0.30014928891333387, "grad_norm": 0.2532920241355896, "learning_rate": 0.000883022221559489, "loss": 1.9059, "step": 955 }, { "epoch": 0.3017207511589534, "grad_norm": 0.26241347193717957, "learning_rate": 0.0008812525055074028, "loss": 1.938, "step": 960 }, { "epoch": 0.30329221340457296, "grad_norm": 0.28568172454833984, "learning_rate": 0.000879471304954804, "loss": 1.8915, "step": 965 }, { "epoch": 0.3048636756501925, "grad_norm": 0.23444630205631256, "learning_rate": 0.0008776786735569431, "loss": 1.9095, "step": 970 }, { "epoch": 0.30643513789581206, "grad_norm": 0.22640594840049744, "learning_rate": 0.0008758746653134029, "loss": 1.8879, "step": 975 }, { "epoch": 0.3080066001414316, "grad_norm": 0.5037649869918823, "learning_rate": 0.0008740593345664716, "loss": 1.889, "step": 980 }, { "epoch": 0.30957806238705116, "grad_norm": 0.18424461781978607, "learning_rate": 0.0008722327359995064, "loss": 1.8806, "step": 985 }, { "epoch": 0.3111495246326707, "grad_norm": 0.2199254333972931, "learning_rate": 0.0008703949246352858, "loss": 1.8767, "step": 990 }, { "epoch": 0.31272098687829025, "grad_norm": 0.26489025354385376, "learning_rate": 0.0008685459558343523, "loss": 1.8609, "step": 995 }, { "epoch": 0.3142924491239098, "grad_norm": 0.164434015750885, "learning_rate": 0.000866685885293345, "loss": 1.8941, "step": 1000 }, { "epoch": 0.31586391136952935, "grad_norm": 0.21282632648944855, "learning_rate": 0.0008648147690433212, "loss": 1.8818, "step": 1005 }, { "epoch": 0.3174353736151489, "grad_norm": 0.21168777346611023, "learning_rate": 0.0008629326634480697, "loss": 1.917, "step": 1010 }, { "epoch": 0.31900683586076845, "grad_norm": 0.2421833872795105, "learning_rate": 0.0008610396252024113, "loss": 1.9548, "step": 1015 }, { "epoch": 0.320578298106388, "grad_norm": 0.17949452996253967, "learning_rate": 0.0008591357113304927, "loss": 1.907, "step": 1020 }, { "epoch": 0.32214976035200754, "grad_norm": 0.25564640760421753, "learning_rate": 0.0008572209791840677, "loss": 1.9133, "step": 1025 }, { "epoch": 0.3237212225976271, "grad_norm": 0.21499499678611755, "learning_rate": 0.0008552954864407698, "loss": 1.8724, "step": 1030 }, { "epoch": 0.32529268484324664, "grad_norm": 0.286145955324173, "learning_rate": 0.000853359291102375, "loss": 1.9174, "step": 1035 }, { "epoch": 0.3268641470888662, "grad_norm": 0.216496080160141, "learning_rate": 0.0008514124514930544, "loss": 1.9212, "step": 1040 }, { "epoch": 0.32843560933448573, "grad_norm": 0.20780375599861145, "learning_rate": 0.0008494550262576173, "loss": 1.9042, "step": 1045 }, { "epoch": 0.3300070715801053, "grad_norm": 0.19447444379329681, "learning_rate": 0.0008474870743597448, "loss": 1.9154, "step": 1050 }, { "epoch": 0.33157853382572483, "grad_norm": 0.2113378793001175, "learning_rate": 0.0008455086550802132, "loss": 1.9357, "step": 1055 }, { "epoch": 0.3331499960713444, "grad_norm": 0.2649160921573639, "learning_rate": 0.0008435198280151091, "loss": 1.9016, "step": 1060 }, { "epoch": 0.3347214583169639, "grad_norm": 0.37894386053085327, "learning_rate": 0.0008415206530740331, "loss": 1.9078, "step": 1065 }, { "epoch": 0.3362929205625835, "grad_norm": 0.19303838908672333, "learning_rate": 0.0008395111904782964, "loss": 1.9018, "step": 1070 }, { "epoch": 0.337864382808203, "grad_norm": 0.25691303610801697, "learning_rate": 0.0008374915007591053, "loss": 1.9319, "step": 1075 }, { "epoch": 0.33943584505382257, "grad_norm": 0.3956037759780884, "learning_rate": 0.0008354616447557392, "loss": 1.9003, "step": 1080 }, { "epoch": 0.3410073072994421, "grad_norm": 0.29999446868896484, "learning_rate": 0.000833421683613717, "loss": 1.9377, "step": 1085 }, { "epoch": 0.34257876954506167, "grad_norm": 0.23759868741035461, "learning_rate": 0.0008313716787829558, "loss": 1.8833, "step": 1090 }, { "epoch": 0.3441502317906812, "grad_norm": 0.33085566759109497, "learning_rate": 0.000829311692015919, "loss": 1.936, "step": 1095 }, { "epoch": 0.34572169403630076, "grad_norm": 0.32732662558555603, "learning_rate": 0.0008272417853657571, "loss": 1.8785, "step": 1100 }, { "epoch": 0.3472931562819203, "grad_norm": 0.23105382919311523, "learning_rate": 0.0008251620211844383, "loss": 1.9553, "step": 1105 }, { "epoch": 0.34886461852753986, "grad_norm": 0.34907397627830505, "learning_rate": 0.0008230724621208693, "loss": 1.8749, "step": 1110 }, { "epoch": 0.3504360807731594, "grad_norm": 0.2972997725009918, "learning_rate": 0.0008209731711190099, "loss": 1.8532, "step": 1115 }, { "epoch": 0.35200754301877896, "grad_norm": 0.21005867421627045, "learning_rate": 0.0008188642114159747, "loss": 1.8804, "step": 1120 }, { "epoch": 0.3535790052643985, "grad_norm": 0.1943141371011734, "learning_rate": 0.0008167456465401299, "loss": 1.8385, "step": 1125 }, { "epoch": 0.35515046751001805, "grad_norm": 0.2459149956703186, "learning_rate": 0.0008146175403091795, "loss": 1.8629, "step": 1130 }, { "epoch": 0.3567219297556376, "grad_norm": 0.2223600149154663, "learning_rate": 0.0008124799568282418, "loss": 1.8939, "step": 1135 }, { "epoch": 0.35829339200125715, "grad_norm": 0.3591344654560089, "learning_rate": 0.0008103329604879195, "loss": 1.835, "step": 1140 }, { "epoch": 0.3598648542468767, "grad_norm": 0.33568888902664185, "learning_rate": 0.0008081766159623596, "loss": 1.9111, "step": 1145 }, { "epoch": 0.36143631649249625, "grad_norm": 0.19357386231422424, "learning_rate": 0.0008060109882073055, "loss": 1.8748, "step": 1150 }, { "epoch": 0.3630077787381158, "grad_norm": 0.20973102748394012, "learning_rate": 0.0008038361424581392, "loss": 1.8755, "step": 1155 }, { "epoch": 0.36457924098373534, "grad_norm": 0.19260567426681519, "learning_rate": 0.000801652144227918, "loss": 1.8516, "step": 1160 }, { "epoch": 0.3661507032293549, "grad_norm": 0.21391819417476654, "learning_rate": 0.0007994590593054001, "loss": 1.8751, "step": 1165 }, { "epoch": 0.36772216547497444, "grad_norm": 0.23929333686828613, "learning_rate": 0.0007972569537530623, "loss": 1.8371, "step": 1170 }, { "epoch": 0.369293627720594, "grad_norm": 0.2769240438938141, "learning_rate": 0.0007950458939051108, "loss": 1.8872, "step": 1175 }, { "epoch": 0.37086508996621353, "grad_norm": 0.22298012673854828, "learning_rate": 0.000792825946365483, "loss": 1.8764, "step": 1180 }, { "epoch": 0.37243655221183314, "grad_norm": 0.1722257286310196, "learning_rate": 0.000790597178005841, "loss": 1.802, "step": 1185 }, { "epoch": 0.3740080144574527, "grad_norm": 0.29332205653190613, "learning_rate": 0.0007883596559635567, "loss": 1.8884, "step": 1190 }, { "epoch": 0.37557947670307223, "grad_norm": 0.2480962872505188, "learning_rate": 0.0007861134476396901, "loss": 1.8287, "step": 1195 }, { "epoch": 0.3771509389486918, "grad_norm": 0.22237135469913483, "learning_rate": 0.0007838586206969593, "loss": 1.8414, "step": 1200 }, { "epoch": 0.37872240119431133, "grad_norm": 0.2007599025964737, "learning_rate": 0.0007815952430577014, "loss": 1.9258, "step": 1205 }, { "epoch": 0.3802938634399309, "grad_norm": 0.2816416323184967, "learning_rate": 0.0007793233829018263, "loss": 1.8544, "step": 1210 }, { "epoch": 0.3818653256855504, "grad_norm": 0.1743992269039154, "learning_rate": 0.0007770431086647642, "loss": 1.8735, "step": 1215 }, { "epoch": 0.38343678793117, "grad_norm": 0.20147810876369476, "learning_rate": 0.000774754489035403, "loss": 1.8541, "step": 1220 }, { "epoch": 0.3850082501767895, "grad_norm": 0.19584012031555176, "learning_rate": 0.0007724575929540197, "loss": 1.821, "step": 1225 }, { "epoch": 0.38657971242240907, "grad_norm": 0.32170602679252625, "learning_rate": 0.0007701524896102037, "loss": 1.8277, "step": 1230 }, { "epoch": 0.3881511746680286, "grad_norm": 0.16368649899959564, "learning_rate": 0.0007678392484407717, "loss": 1.8317, "step": 1235 }, { "epoch": 0.38972263691364817, "grad_norm": 0.17887534201145172, "learning_rate": 0.0007655179391276775, "loss": 1.8234, "step": 1240 }, { "epoch": 0.3912940991592677, "grad_norm": 0.2335747331380844, "learning_rate": 0.0007631886315959121, "loss": 1.8789, "step": 1245 }, { "epoch": 0.39286556140488726, "grad_norm": 0.16466860473155975, "learning_rate": 0.0007608513960113975, "loss": 1.8777, "step": 1250 }, { "epoch": 0.3944370236505068, "grad_norm": 0.1995929628610611, "learning_rate": 0.000758506302778873, "loss": 1.8159, "step": 1255 }, { "epoch": 0.39600848589612636, "grad_norm": 0.28415942192077637, "learning_rate": 0.0007561534225397744, "loss": 1.8151, "step": 1260 }, { "epoch": 0.3975799481417459, "grad_norm": 0.15427584946155548, "learning_rate": 0.0007537928261701064, "loss": 1.8402, "step": 1265 }, { "epoch": 0.39915141038736546, "grad_norm": 0.211939737200737, "learning_rate": 0.0007514245847783069, "loss": 1.8554, "step": 1270 }, { "epoch": 0.400722872632985, "grad_norm": 0.1703938990831375, "learning_rate": 0.0007490487697031061, "loss": 1.8369, "step": 1275 }, { "epoch": 0.40229433487860455, "grad_norm": 0.19989508390426636, "learning_rate": 0.0007466654525113761, "loss": 1.8075, "step": 1280 }, { "epoch": 0.4038657971242241, "grad_norm": 0.3463574945926666, "learning_rate": 0.0007442747049959765, "loss": 1.8606, "step": 1285 }, { "epoch": 0.40543725936984365, "grad_norm": 0.2496725171804428, "learning_rate": 0.0007418765991735908, "loss": 1.8269, "step": 1290 }, { "epoch": 0.4070087216154632, "grad_norm": 0.29229408502578735, "learning_rate": 0.0007394712072825576, "loss": 1.8034, "step": 1295 }, { "epoch": 0.40858018386108275, "grad_norm": 0.2583361566066742, "learning_rate": 0.0007370586017806941, "loss": 1.8845, "step": 1300 }, { "epoch": 0.4101516461067023, "grad_norm": 0.168927401304245, "learning_rate": 0.0007346388553431141, "loss": 1.8509, "step": 1305 }, { "epoch": 0.41172310835232184, "grad_norm": 0.24876132607460022, "learning_rate": 0.0007322120408600379, "loss": 1.9044, "step": 1310 }, { "epoch": 0.4132945705979414, "grad_norm": 0.16044846177101135, "learning_rate": 0.0007297782314345972, "loss": 1.8406, "step": 1315 }, { "epoch": 0.41486603284356094, "grad_norm": 0.18333998322486877, "learning_rate": 0.0007273375003806335, "loss": 1.8755, "step": 1320 }, { "epoch": 0.4164374950891805, "grad_norm": 0.2609647512435913, "learning_rate": 0.0007248899212204883, "loss": 1.8146, "step": 1325 }, { "epoch": 0.41800895733480004, "grad_norm": 0.20326748490333557, "learning_rate": 0.0007224355676827897, "loss": 1.8173, "step": 1330 }, { "epoch": 0.4195804195804196, "grad_norm": 0.32762306928634644, "learning_rate": 0.0007199745137002305, "loss": 1.892, "step": 1335 }, { "epoch": 0.42115188182603913, "grad_norm": 0.25635045766830444, "learning_rate": 0.000717506833407342, "loss": 1.8104, "step": 1340 }, { "epoch": 0.4227233440716587, "grad_norm": 0.29682666063308716, "learning_rate": 0.0007150326011382603, "loss": 1.8018, "step": 1345 }, { "epoch": 0.42429480631727823, "grad_norm": 0.22666814923286438, "learning_rate": 0.0007125518914244868, "loss": 1.8358, "step": 1350 }, { "epoch": 0.4258662685628978, "grad_norm": 0.26551052927970886, "learning_rate": 0.000710064778992644, "loss": 1.8332, "step": 1355 }, { "epoch": 0.4274377308085173, "grad_norm": 0.16310109198093414, "learning_rate": 0.0007075713387622236, "loss": 1.7835, "step": 1360 }, { "epoch": 0.4290091930541369, "grad_norm": 0.19358539581298828, "learning_rate": 0.0007050716458433305, "loss": 1.7956, "step": 1365 }, { "epoch": 0.4305806552997564, "grad_norm": 0.16724836826324463, "learning_rate": 0.0007025657755344197, "loss": 1.874, "step": 1370 }, { "epoch": 0.43215211754537597, "grad_norm": 0.2080153524875641, "learning_rate": 0.0007000538033200279, "loss": 1.8706, "step": 1375 }, { "epoch": 0.4337235797909955, "grad_norm": 0.1837550699710846, "learning_rate": 0.0006975358048685004, "loss": 1.8576, "step": 1380 }, { "epoch": 0.43529504203661507, "grad_norm": 0.26517677307128906, "learning_rate": 0.0006950118560297112, "loss": 1.8121, "step": 1385 }, { "epoch": 0.4368665042822346, "grad_norm": 0.2400379180908203, "learning_rate": 0.0006924820328327785, "loss": 1.7908, "step": 1390 }, { "epoch": 0.43843796652785416, "grad_norm": 0.1548314243555069, "learning_rate": 0.0006899464114837739, "loss": 1.8439, "step": 1395 }, { "epoch": 0.4400094287734737, "grad_norm": 0.1790788173675537, "learning_rate": 0.0006874050683634273, "loss": 1.8511, "step": 1400 }, { "epoch": 0.44158089101909326, "grad_norm": 0.2156343162059784, "learning_rate": 0.0006848580800248262, "loss": 1.8442, "step": 1405 }, { "epoch": 0.4431523532647128, "grad_norm": 0.24280230700969696, "learning_rate": 0.0006823055231911093, "loss": 1.8091, "step": 1410 }, { "epoch": 0.44472381551033235, "grad_norm": 0.18132703006267548, "learning_rate": 0.0006797474747531558, "loss": 1.8596, "step": 1415 }, { "epoch": 0.4462952777559519, "grad_norm": 0.2499558925628662, "learning_rate": 0.0006771840117672684, "loss": 1.7982, "step": 1420 }, { "epoch": 0.44786674000157145, "grad_norm": 0.14542344212532043, "learning_rate": 0.0006746152114528531, "loss": 1.8436, "step": 1425 }, { "epoch": 0.449438202247191, "grad_norm": 0.23537123203277588, "learning_rate": 0.0006720411511900927, "loss": 1.8342, "step": 1430 }, { "epoch": 0.45100966449281055, "grad_norm": 0.2291416972875595, "learning_rate": 0.000669461908517616, "loss": 1.815, "step": 1435 }, { "epoch": 0.4525811267384301, "grad_norm": 0.17745240032672882, "learning_rate": 0.0006668775611301611, "loss": 1.807, "step": 1440 }, { "epoch": 0.45415258898404964, "grad_norm": 0.2866009771823883, "learning_rate": 0.0006642881868762368, "loss": 1.8313, "step": 1445 }, { "epoch": 0.4557240512296692, "grad_norm": 0.23607899248600006, "learning_rate": 0.0006616938637557761, "loss": 1.8308, "step": 1450 }, { "epoch": 0.45729551347528874, "grad_norm": 0.25843170285224915, "learning_rate": 0.0006590946699177875, "loss": 1.8076, "step": 1455 }, { "epoch": 0.4588669757209083, "grad_norm": 0.2201550155878067, "learning_rate": 0.0006564906836580004, "loss": 1.7999, "step": 1460 }, { "epoch": 0.46043843796652784, "grad_norm": 0.2689286172389984, "learning_rate": 0.0006538819834165061, "loss": 1.8498, "step": 1465 }, { "epoch": 0.4620099002121474, "grad_norm": 0.19224600493907928, "learning_rate": 0.0006512686477753966, "loss": 1.8525, "step": 1470 }, { "epoch": 0.46358136245776693, "grad_norm": 0.18337461352348328, "learning_rate": 0.0006486507554563953, "loss": 1.8378, "step": 1475 }, { "epoch": 0.4651528247033865, "grad_norm": 0.22880521416664124, "learning_rate": 0.0006460283853184879, "loss": 1.8124, "step": 1480 }, { "epoch": 0.46672428694900603, "grad_norm": 0.3217616081237793, "learning_rate": 0.0006434016163555452, "loss": 1.8509, "step": 1485 }, { "epoch": 0.4682957491946256, "grad_norm": 0.1924976408481598, "learning_rate": 0.000640770527693944, "loss": 1.8106, "step": 1490 }, { "epoch": 0.4698672114402451, "grad_norm": 0.24654747545719147, "learning_rate": 0.0006381351985901842, "loss": 1.79, "step": 1495 }, { "epoch": 0.4714386736858647, "grad_norm": 0.18487554788589478, "learning_rate": 0.0006354957084285007, "loss": 1.7589, "step": 1500 }, { "epoch": 0.4730101359314842, "grad_norm": 0.19454938173294067, "learning_rate": 0.0006328521367184721, "loss": 1.7756, "step": 1505 }, { "epoch": 0.47458159817710377, "grad_norm": 0.2697226107120514, "learning_rate": 0.000630204563092626, "loss": 1.8154, "step": 1510 }, { "epoch": 0.4761530604227233, "grad_norm": 0.2261214405298233, "learning_rate": 0.0006275530673040401, "loss": 1.7989, "step": 1515 }, { "epoch": 0.47772452266834287, "grad_norm": 0.29914212226867676, "learning_rate": 0.0006248977292239395, "loss": 1.7528, "step": 1520 }, { "epoch": 0.47929598491396247, "grad_norm": 0.21949878334999084, "learning_rate": 0.0006222386288392914, "loss": 1.7694, "step": 1525 }, { "epoch": 0.480867447159582, "grad_norm": 0.22492919862270355, "learning_rate": 0.0006195758462503947, "loss": 1.8911, "step": 1530 }, { "epoch": 0.48243890940520157, "grad_norm": 0.1728420853614807, "learning_rate": 0.0006169094616684678, "loss": 1.7795, "step": 1535 }, { "epoch": 0.4840103716508211, "grad_norm": 0.1900889128446579, "learning_rate": 0.0006142395554132324, "loss": 1.8095, "step": 1540 }, { "epoch": 0.48558183389644066, "grad_norm": 0.2331051081418991, "learning_rate": 0.0006115662079104937, "loss": 1.8101, "step": 1545 }, { "epoch": 0.4871532961420602, "grad_norm": 0.28242024779319763, "learning_rate": 0.000608889499689718, "loss": 1.8349, "step": 1550 }, { "epoch": 0.48872475838767976, "grad_norm": 0.2434227615594864, "learning_rate": 0.0006062095113816069, "loss": 1.7964, "step": 1555 }, { "epoch": 0.4902962206332993, "grad_norm": 0.2203354686498642, "learning_rate": 0.0006035263237156676, "loss": 1.7928, "step": 1560 }, { "epoch": 0.49186768287891885, "grad_norm": 0.2670794427394867, "learning_rate": 0.0006008400175177827, "loss": 1.8127, "step": 1565 }, { "epoch": 0.4934391451245384, "grad_norm": 0.3301917016506195, "learning_rate": 0.0005981506737077743, "loss": 1.8027, "step": 1570 }, { "epoch": 0.49501060737015795, "grad_norm": 0.20628106594085693, "learning_rate": 0.0005954583732969666, "loss": 1.7938, "step": 1575 }, { "epoch": 0.4965820696157775, "grad_norm": 0.16303130984306335, "learning_rate": 0.000592763197385746, "loss": 1.8486, "step": 1580 }, { "epoch": 0.49815353186139705, "grad_norm": 0.15457695722579956, "learning_rate": 0.0005900652271611175, "loss": 1.773, "step": 1585 }, { "epoch": 0.4997249941070166, "grad_norm": 0.1589762419462204, "learning_rate": 0.0005873645438942595, "loss": 1.7507, "step": 1590 }, { "epoch": 0.5012964563526361, "grad_norm": 0.19667518138885498, "learning_rate": 0.000584661228938076, "loss": 1.8197, "step": 1595 }, { "epoch": 0.5028679185982556, "grad_norm": 0.40122613310813904, "learning_rate": 0.000581955363724745, "loss": 1.8094, "step": 1600 }, { "epoch": 0.5044393808438752, "grad_norm": 0.19897930324077606, "learning_rate": 0.0005792470297632666, "loss": 1.7865, "step": 1605 }, { "epoch": 0.5060108430894947, "grad_norm": 0.3172074854373932, "learning_rate": 0.0005765363086370068, "loss": 1.7962, "step": 1610 }, { "epoch": 0.5075823053351143, "grad_norm": 0.20447920262813568, "learning_rate": 0.0005738232820012407, "loss": 1.7691, "step": 1615 }, { "epoch": 0.5091537675807338, "grad_norm": 0.16863569617271423, "learning_rate": 0.0005711080315806921, "loss": 1.8213, "step": 1620 }, { "epoch": 0.5107252298263534, "grad_norm": 0.29164841771125793, "learning_rate": 0.0005683906391670727, "loss": 1.7405, "step": 1625 }, { "epoch": 0.5122966920719729, "grad_norm": 0.1419832408428192, "learning_rate": 0.0005656711866166167, "loss": 1.8176, "step": 1630 }, { "epoch": 0.5138681543175925, "grad_norm": 0.1718250811100006, "learning_rate": 0.0005629497558476167, "loss": 1.8431, "step": 1635 }, { "epoch": 0.515439616563212, "grad_norm": 0.14950774610042572, "learning_rate": 0.0005602264288379551, "loss": 1.8238, "step": 1640 }, { "epoch": 0.5170110788088316, "grad_norm": 0.204507514834404, "learning_rate": 0.0005575012876226347, "loss": 1.831, "step": 1645 }, { "epoch": 0.5185825410544511, "grad_norm": 0.16311819851398468, "learning_rate": 0.0005547744142913084, "loss": 1.7805, "step": 1650 }, { "epoch": 0.5201540033000707, "grad_norm": 0.17227725684642792, "learning_rate": 0.0005520458909858048, "loss": 1.8119, "step": 1655 }, { "epoch": 0.5217254655456902, "grad_norm": 0.22001691162586212, "learning_rate": 0.0005493157998976559, "loss": 1.7803, "step": 1660 }, { "epoch": 0.5232969277913098, "grad_norm": 0.3537323474884033, "learning_rate": 0.0005465842232656194, "loss": 1.8142, "step": 1665 }, { "epoch": 0.5248683900369293, "grad_norm": 0.7048746347427368, "learning_rate": 0.0005438512433732023, "loss": 1.7881, "step": 1670 }, { "epoch": 0.5264398522825489, "grad_norm": 0.24534624814987183, "learning_rate": 0.0005411169425461822, "loss": 1.8228, "step": 1675 }, { "epoch": 0.5280113145281684, "grad_norm": 0.16271162033081055, "learning_rate": 0.0005383814031501272, "loss": 1.8155, "step": 1680 }, { "epoch": 0.529582776773788, "grad_norm": 0.3233349323272705, "learning_rate": 0.0005356447075879153, "loss": 1.8067, "step": 1685 }, { "epoch": 0.5311542390194075, "grad_norm": 0.20079372823238373, "learning_rate": 0.0005329069382972513, "loss": 1.8035, "step": 1690 }, { "epoch": 0.5327257012650272, "grad_norm": 0.17062976956367493, "learning_rate": 0.0005301681777481846, "loss": 1.7636, "step": 1695 }, { "epoch": 0.5342971635106467, "grad_norm": 0.24763086438179016, "learning_rate": 0.0005274285084406234, "loss": 1.8238, "step": 1700 }, { "epoch": 0.5358686257562663, "grad_norm": 0.18531553447246552, "learning_rate": 0.0005246880129018515, "loss": 1.7826, "step": 1705 }, { "epoch": 0.5374400880018858, "grad_norm": 0.13745392858982086, "learning_rate": 0.0005219467736840409, "loss": 1.7463, "step": 1710 }, { "epoch": 0.5390115502475054, "grad_norm": 0.2211742103099823, "learning_rate": 0.0005192048733617654, "loss": 1.775, "step": 1715 }, { "epoch": 0.5405830124931249, "grad_norm": 0.1478564590215683, "learning_rate": 0.0005164623945295136, "loss": 1.7752, "step": 1720 }, { "epoch": 0.5421544747387445, "grad_norm": 0.19443662464618683, "learning_rate": 0.0005137194197992001, "loss": 1.7751, "step": 1725 }, { "epoch": 0.543725936984364, "grad_norm": 0.27985528111457825, "learning_rate": 0.0005109760317976782, "loss": 1.7892, "step": 1730 }, { "epoch": 0.5452973992299835, "grad_norm": 0.3474476635456085, "learning_rate": 0.0005082323131642496, "loss": 1.7856, "step": 1735 }, { "epoch": 0.5468688614756031, "grad_norm": 0.19026850163936615, "learning_rate": 0.0005054883465481761, "loss": 1.7514, "step": 1740 }, { "epoch": 0.5484403237212226, "grad_norm": 0.15917906165122986, "learning_rate": 0.0005027442146061889, "loss": 1.8218, "step": 1745 }, { "epoch": 0.5500117859668422, "grad_norm": 0.1620151400566101, "learning_rate": 0.0005, "loss": 1.7814, "step": 1750 }, { "epoch": 0.5515832482124617, "grad_norm": 0.19481198489665985, "learning_rate": 0.0004972557853938111, "loss": 1.8251, "step": 1755 }, { "epoch": 0.5531547104580813, "grad_norm": 0.20308136940002441, "learning_rate": 0.000494511653451824, "loss": 1.8222, "step": 1760 }, { "epoch": 0.5547261727037008, "grad_norm": 0.1648699939250946, "learning_rate": 0.0004917676868357503, "loss": 1.81, "step": 1765 }, { "epoch": 0.5562976349493204, "grad_norm": 0.2805931568145752, "learning_rate": 0.0004890239682023217, "loss": 1.7743, "step": 1770 }, { "epoch": 0.5578690971949399, "grad_norm": 0.22301295399665833, "learning_rate": 0.00048628058020080007, "loss": 1.7481, "step": 1775 }, { "epoch": 0.5594405594405595, "grad_norm": 0.18746043741703033, "learning_rate": 0.0004835376054704866, "loss": 1.7545, "step": 1780 }, { "epoch": 0.561012021686179, "grad_norm": 0.22121259570121765, "learning_rate": 0.00048079512663823474, "loss": 1.8134, "step": 1785 }, { "epoch": 0.5625834839317986, "grad_norm": 0.22781746089458466, "learning_rate": 0.0004780532263159592, "loss": 1.8077, "step": 1790 }, { "epoch": 0.5641549461774181, "grad_norm": 0.26571694016456604, "learning_rate": 0.00047531198709814857, "loss": 1.8487, "step": 1795 }, { "epoch": 0.5657264084230377, "grad_norm": 0.1296350359916687, "learning_rate": 0.00047257149155937667, "loss": 1.7883, "step": 1800 }, { "epoch": 0.5672978706686572, "grad_norm": 0.18556547164916992, "learning_rate": 0.00046983182225181555, "loss": 1.7306, "step": 1805 }, { "epoch": 0.5688693329142768, "grad_norm": 0.1794668436050415, "learning_rate": 0.00046709306170274867, "loss": 1.8121, "step": 1810 }, { "epoch": 0.5704407951598963, "grad_norm": 0.14622707664966583, "learning_rate": 0.0004643552924120847, "loss": 1.7723, "step": 1815 }, { "epoch": 0.5720122574055159, "grad_norm": 0.2089068442583084, "learning_rate": 0.00046161859684987303, "loss": 1.792, "step": 1820 }, { "epoch": 0.5735837196511354, "grad_norm": 0.18128232657909393, "learning_rate": 0.000458883057453818, "loss": 1.7802, "step": 1825 }, { "epoch": 0.575155181896755, "grad_norm": 0.3529801666736603, "learning_rate": 0.00045614875662679797, "loss": 1.7848, "step": 1830 }, { "epoch": 0.5767266441423745, "grad_norm": 0.1691349595785141, "learning_rate": 0.00045341577673438073, "loss": 1.7563, "step": 1835 }, { "epoch": 0.5782981063879941, "grad_norm": 0.2139570415019989, "learning_rate": 0.00045068420010234417, "loss": 1.7557, "step": 1840 }, { "epoch": 0.5798695686336136, "grad_norm": 0.3797873556613922, "learning_rate": 0.00044795410901419527, "loss": 1.8131, "step": 1845 }, { "epoch": 0.5814410308792332, "grad_norm": 0.20959897339344025, "learning_rate": 0.00044522558570869177, "loss": 1.7783, "step": 1850 }, { "epoch": 0.5830124931248527, "grad_norm": 0.21909403800964355, "learning_rate": 0.0004424987123773653, "loss": 1.7801, "step": 1855 }, { "epoch": 0.5845839553704723, "grad_norm": 0.18797878921031952, "learning_rate": 0.0004397735711620451, "loss": 1.7918, "step": 1860 }, { "epoch": 0.5861554176160918, "grad_norm": 0.15942728519439697, "learning_rate": 0.0004370502441523834, "loss": 1.7746, "step": 1865 }, { "epoch": 0.5877268798617113, "grad_norm": 0.14537079632282257, "learning_rate": 0.0004343288133833835, "loss": 1.7475, "step": 1870 }, { "epoch": 0.5892983421073309, "grad_norm": 0.2374078631401062, "learning_rate": 0.0004316093608329275, "loss": 1.802, "step": 1875 }, { "epoch": 0.5908698043529504, "grad_norm": 0.16053801774978638, "learning_rate": 0.000428891968419308, "loss": 1.7559, "step": 1880 }, { "epoch": 0.59244126659857, "grad_norm": 0.15984462201595306, "learning_rate": 0.00042617671799875947, "loss": 1.7853, "step": 1885 }, { "epoch": 0.5940127288441895, "grad_norm": 0.1509033590555191, "learning_rate": 0.00042346369136299334, "loss": 1.78, "step": 1890 }, { "epoch": 0.5955841910898091, "grad_norm": 0.21747443079948425, "learning_rate": 0.0004207529702367335, "loss": 1.7661, "step": 1895 }, { "epoch": 0.5971556533354286, "grad_norm": 0.23470966517925262, "learning_rate": 0.00041804463627525504, "loss": 1.7534, "step": 1900 }, { "epoch": 0.5987271155810482, "grad_norm": 0.1689888834953308, "learning_rate": 0.00041533877106192407, "loss": 1.7705, "step": 1905 }, { "epoch": 0.6002985778266677, "grad_norm": 0.1531875878572464, "learning_rate": 0.0004126354561057404, "loss": 1.7506, "step": 1910 }, { "epoch": 0.6018700400722873, "grad_norm": 0.22668935358524323, "learning_rate": 0.00040993477283888266, "loss": 1.781, "step": 1915 }, { "epoch": 0.6034415023179068, "grad_norm": 0.15343786776065826, "learning_rate": 0.0004072368026142541, "loss": 1.7676, "step": 1920 }, { "epoch": 0.6050129645635264, "grad_norm": 0.19431781768798828, "learning_rate": 0.0004045416267030335, "loss": 1.7892, "step": 1925 }, { "epoch": 0.6065844268091459, "grad_norm": 0.1378578096628189, "learning_rate": 0.00040184932629222574, "loss": 1.7712, "step": 1930 }, { "epoch": 0.6081558890547655, "grad_norm": 0.19014595448970795, "learning_rate": 0.0003991599824822174, "loss": 1.8178, "step": 1935 }, { "epoch": 0.609727351300385, "grad_norm": 0.21545611321926117, "learning_rate": 0.00039647367628433246, "loss": 1.8122, "step": 1940 }, { "epoch": 0.6112988135460046, "grad_norm": 0.12300246208906174, "learning_rate": 0.0003937904886183933, "loss": 1.7544, "step": 1945 }, { "epoch": 0.6128702757916241, "grad_norm": 0.18749244511127472, "learning_rate": 0.00039111050031028193, "loss": 1.7679, "step": 1950 }, { "epoch": 0.6144417380372437, "grad_norm": 0.20615063607692719, "learning_rate": 0.00038843379208950617, "loss": 1.7784, "step": 1955 }, { "epoch": 0.6160132002828632, "grad_norm": 0.1621728390455246, "learning_rate": 0.0003857604445867677, "loss": 1.7688, "step": 1960 }, { "epoch": 0.6175846625284828, "grad_norm": 0.2304105907678604, "learning_rate": 0.00038309053833153234, "loss": 1.8177, "step": 1965 }, { "epoch": 0.6191561247741023, "grad_norm": 0.19086576998233795, "learning_rate": 0.0003804241537496055, "loss": 1.7566, "step": 1970 }, { "epoch": 0.6207275870197219, "grad_norm": 0.21728602051734924, "learning_rate": 0.00037776137116070867, "loss": 1.7514, "step": 1975 }, { "epoch": 0.6222990492653414, "grad_norm": 0.1872587502002716, "learning_rate": 0.0003751022707760605, "loss": 1.8045, "step": 1980 }, { "epoch": 0.623870511510961, "grad_norm": 0.21256718039512634, "learning_rate": 0.00037244693269596, "loss": 1.7916, "step": 1985 }, { "epoch": 0.6254419737565805, "grad_norm": 0.2095334529876709, "learning_rate": 0.00036979543690737407, "loss": 1.7581, "step": 1990 }, { "epoch": 0.6270134360022, "grad_norm": 0.22193801403045654, "learning_rate": 0.00036714786328152804, "loss": 1.779, "step": 1995 }, { "epoch": 0.6285848982478196, "grad_norm": 0.16215133666992188, "learning_rate": 0.00036450429157149934, "loss": 1.7565, "step": 2000 }, { "epoch": 0.6301563604934391, "grad_norm": 0.17526470124721527, "learning_rate": 0.00036186480140981583, "loss": 1.789, "step": 2005 }, { "epoch": 0.6317278227390587, "grad_norm": 0.3180091679096222, "learning_rate": 0.00035922947230605605, "loss": 1.7617, "step": 2010 }, { "epoch": 0.6332992849846782, "grad_norm": 0.11836399137973785, "learning_rate": 0.00035659838364445503, "loss": 1.8111, "step": 2015 }, { "epoch": 0.6348707472302978, "grad_norm": 0.15318524837493896, "learning_rate": 0.0003539716146815122, "loss": 1.7409, "step": 2020 }, { "epoch": 0.6364422094759173, "grad_norm": 0.13768509030342102, "learning_rate": 0.0003513492445436048, "loss": 1.7733, "step": 2025 }, { "epoch": 0.6380136717215369, "grad_norm": 0.16969747841358185, "learning_rate": 0.0003487313522246036, "loss": 1.8085, "step": 2030 }, { "epoch": 0.6395851339671564, "grad_norm": 0.15154893696308136, "learning_rate": 0.00034611801658349393, "loss": 1.7053, "step": 2035 }, { "epoch": 0.641156596212776, "grad_norm": 0.16899384558200836, "learning_rate": 0.0003435093163419998, "loss": 1.8229, "step": 2040 }, { "epoch": 0.6427280584583955, "grad_norm": 0.23929907381534576, "learning_rate": 0.00034090533008221234, "loss": 1.7719, "step": 2045 }, { "epoch": 0.6442995207040151, "grad_norm": 0.12470386922359467, "learning_rate": 0.00033830613624422377, "loss": 1.8131, "step": 2050 }, { "epoch": 0.6458709829496346, "grad_norm": 0.15960881114006042, "learning_rate": 0.00033571181312376335, "loss": 1.7428, "step": 2055 }, { "epoch": 0.6474424451952542, "grad_norm": 0.2688332796096802, "learning_rate": 0.00033312243886983906, "loss": 1.7652, "step": 2060 }, { "epoch": 0.6490139074408737, "grad_norm": 0.20620514452457428, "learning_rate": 0.00033053809148238423, "loss": 1.7579, "step": 2065 }, { "epoch": 0.6505853696864933, "grad_norm": 0.18862473964691162, "learning_rate": 0.0003279588488099073, "loss": 1.7364, "step": 2070 }, { "epoch": 0.6521568319321128, "grad_norm": 0.155025452375412, "learning_rate": 0.0003253847885471469, "loss": 1.7118, "step": 2075 }, { "epoch": 0.6537282941777324, "grad_norm": 0.27940723299980164, "learning_rate": 0.0003228159882327317, "loss": 1.6794, "step": 2080 }, { "epoch": 0.6552997564233519, "grad_norm": 0.19810332357883453, "learning_rate": 0.0003202525252468443, "loss": 1.7879, "step": 2085 }, { "epoch": 0.6568712186689715, "grad_norm": 0.1640649139881134, "learning_rate": 0.00031769447680889064, "loss": 1.7904, "step": 2090 }, { "epoch": 0.658442680914591, "grad_norm": 0.2099301815032959, "learning_rate": 0.00031514191997517385, "loss": 1.7922, "step": 2095 }, { "epoch": 0.6600141431602106, "grad_norm": 0.19281212985515594, "learning_rate": 0.0003125949316365728, "loss": 1.7957, "step": 2100 }, { "epoch": 0.6615856054058301, "grad_norm": 0.20853348076343536, "learning_rate": 0.00031005358851622633, "loss": 1.7531, "step": 2105 }, { "epoch": 0.6631570676514497, "grad_norm": 0.16896933317184448, "learning_rate": 0.00030751796716722157, "loss": 1.7632, "step": 2110 }, { "epoch": 0.6647285298970692, "grad_norm": 0.6465707421302795, "learning_rate": 0.0003049881439702888, "loss": 1.7804, "step": 2115 }, { "epoch": 0.6662999921426888, "grad_norm": 0.1943897157907486, "learning_rate": 0.00030246419513149967, "loss": 1.7897, "step": 2120 }, { "epoch": 0.6678714543883083, "grad_norm": 0.20024192333221436, "learning_rate": 0.00029994619667997216, "loss": 1.734, "step": 2125 }, { "epoch": 0.6694429166339279, "grad_norm": 0.18751543760299683, "learning_rate": 0.0002974342244655804, "loss": 1.7113, "step": 2130 }, { "epoch": 0.6710143788795474, "grad_norm": 0.1718306988477707, "learning_rate": 0.0002949283541566694, "loss": 1.6794, "step": 2135 }, { "epoch": 0.672585841125167, "grad_norm": 0.18411104381084442, "learning_rate": 0.0002924286612377764, "loss": 1.7223, "step": 2140 }, { "epoch": 0.6741573033707865, "grad_norm": 0.14647985994815826, "learning_rate": 0.0002899352210073562, "loss": 1.7483, "step": 2145 }, { "epoch": 0.675728765616406, "grad_norm": 0.27714547514915466, "learning_rate": 0.0002874481085755133, "loss": 1.7302, "step": 2150 }, { "epoch": 0.6773002278620256, "grad_norm": 0.32130590081214905, "learning_rate": 0.0002849673988617399, "loss": 1.812, "step": 2155 }, { "epoch": 0.6788716901076451, "grad_norm": 0.17637619376182556, "learning_rate": 0.000282493166592658, "loss": 1.729, "step": 2160 }, { "epoch": 0.6804431523532647, "grad_norm": 0.19782552123069763, "learning_rate": 0.0002800254862997695, "loss": 1.7661, "step": 2165 }, { "epoch": 0.6820146145988842, "grad_norm": 0.21681202948093414, "learning_rate": 0.0002775644323172105, "loss": 1.7431, "step": 2170 }, { "epoch": 0.6835860768445038, "grad_norm": 0.21317927539348602, "learning_rate": 0.0002751100787795118, "loss": 1.7573, "step": 2175 }, { "epoch": 0.6851575390901233, "grad_norm": 0.2029709368944168, "learning_rate": 0.0002726624996193665, "loss": 1.776, "step": 2180 }, { "epoch": 0.6867290013357429, "grad_norm": 0.15296722948551178, "learning_rate": 0.0002702217685654028, "loss": 1.7741, "step": 2185 }, { "epoch": 0.6883004635813624, "grad_norm": 0.14518578350543976, "learning_rate": 0.00026778795913996224, "loss": 1.7665, "step": 2190 }, { "epoch": 0.689871925826982, "grad_norm": 0.17168502509593964, "learning_rate": 0.0002653611446568861, "loss": 1.7154, "step": 2195 }, { "epoch": 0.6914433880726015, "grad_norm": 0.1342601329088211, "learning_rate": 0.00026294139821930593, "loss": 1.7174, "step": 2200 }, { "epoch": 0.6930148503182211, "grad_norm": 0.1369861364364624, "learning_rate": 0.00026052879271744263, "loss": 1.753, "step": 2205 }, { "epoch": 0.6945863125638406, "grad_norm": 0.1410820037126541, "learning_rate": 0.00025812340082640936, "loss": 1.6835, "step": 2210 }, { "epoch": 0.6961577748094602, "grad_norm": 0.16122332215309143, "learning_rate": 0.00025572529500402365, "loss": 1.7404, "step": 2215 }, { "epoch": 0.6977292370550797, "grad_norm": 0.19913320243358612, "learning_rate": 0.00025333454748862396, "loss": 1.7498, "step": 2220 }, { "epoch": 0.6993006993006993, "grad_norm": 0.14058250188827515, "learning_rate": 0.0002509512302968941, "loss": 1.7571, "step": 2225 }, { "epoch": 0.7008721615463188, "grad_norm": 0.2978239059448242, "learning_rate": 0.0002485754152216931, "loss": 1.7602, "step": 2230 }, { "epoch": 0.7024436237919384, "grad_norm": 0.13628768920898438, "learning_rate": 0.0002462071738298936, "loss": 1.7331, "step": 2235 }, { "epoch": 0.7040150860375579, "grad_norm": 0.16833730041980743, "learning_rate": 0.00024384657746022564, "loss": 1.7697, "step": 2240 }, { "epoch": 0.7055865482831775, "grad_norm": 0.19926880300045013, "learning_rate": 0.00024149369722112717, "loss": 1.7079, "step": 2245 }, { "epoch": 0.707158010528797, "grad_norm": 0.13357147574424744, "learning_rate": 0.00023914860398860255, "loss": 1.702, "step": 2250 }, { "epoch": 0.7087294727744166, "grad_norm": 0.1692400723695755, "learning_rate": 0.00023681136840408786, "loss": 1.7342, "step": 2255 }, { "epoch": 0.7103009350200361, "grad_norm": 0.1352614313364029, "learning_rate": 0.00023448206087232267, "loss": 1.7437, "step": 2260 }, { "epoch": 0.7118723972656557, "grad_norm": 0.17008154094219208, "learning_rate": 0.00023216075155922845, "loss": 1.6892, "step": 2265 }, { "epoch": 0.7134438595112752, "grad_norm": 0.17110054194927216, "learning_rate": 0.0002298475103897964, "loss": 1.7326, "step": 2270 }, { "epoch": 0.7150153217568948, "grad_norm": 0.16124136745929718, "learning_rate": 0.0002275424070459803, "loss": 1.766, "step": 2275 }, { "epoch": 0.7165867840025143, "grad_norm": 0.14922770857810974, "learning_rate": 0.000225245510964597, "loss": 1.7667, "step": 2280 }, { "epoch": 0.7181582462481338, "grad_norm": 0.17472444474697113, "learning_rate": 0.000222956891335236, "loss": 1.7224, "step": 2285 }, { "epoch": 0.7197297084937534, "grad_norm": 0.14927974343299866, "learning_rate": 0.00022067661709817383, "loss": 1.7444, "step": 2290 }, { "epoch": 0.7213011707393729, "grad_norm": 0.1595926582813263, "learning_rate": 0.00021840475694229888, "loss": 1.7501, "step": 2295 }, { "epoch": 0.7228726329849925, "grad_norm": 0.1755470633506775, "learning_rate": 0.00021614137930304068, "loss": 1.7742, "step": 2300 }, { "epoch": 0.724444095230612, "grad_norm": 0.15455584228038788, "learning_rate": 0.00021388655236030985, "loss": 1.7152, "step": 2305 }, { "epoch": 0.7260155574762316, "grad_norm": 0.13549718260765076, "learning_rate": 0.00021164034403644338, "loss": 1.7603, "step": 2310 }, { "epoch": 0.7275870197218511, "grad_norm": 0.20018717646598816, "learning_rate": 0.00020940282199415915, "loss": 1.7403, "step": 2315 }, { "epoch": 0.7291584819674707, "grad_norm": 0.14150027930736542, "learning_rate": 0.00020717405363451696, "loss": 1.7578, "step": 2320 }, { "epoch": 0.7307299442130902, "grad_norm": 0.17683018743991852, "learning_rate": 0.00020495410609488912, "loss": 1.7105, "step": 2325 }, { "epoch": 0.7323014064587098, "grad_norm": 0.1376308798789978, "learning_rate": 0.00020274304624693778, "loss": 1.6991, "step": 2330 }, { "epoch": 0.7338728687043293, "grad_norm": 0.2307780683040619, "learning_rate": 0.0002005409406946, "loss": 1.7478, "step": 2335 }, { "epoch": 0.7354443309499489, "grad_norm": 0.22559094429016113, "learning_rate": 0.00019834785577208192, "loss": 1.7321, "step": 2340 }, { "epoch": 0.7370157931955684, "grad_norm": 0.2081470936536789, "learning_rate": 0.00019616385754186078, "loss": 1.7659, "step": 2345 }, { "epoch": 0.738587255441188, "grad_norm": 0.17917020618915558, "learning_rate": 0.00019398901179269474, "loss": 1.7489, "step": 2350 }, { "epoch": 0.7401587176868075, "grad_norm": 0.1390395164489746, "learning_rate": 0.00019182338403764038, "loss": 1.7142, "step": 2355 }, { "epoch": 0.7417301799324271, "grad_norm": 0.13578888773918152, "learning_rate": 0.00018966703951208048, "loss": 1.7468, "step": 2360 }, { "epoch": 0.7433016421780466, "grad_norm": 0.1478307545185089, "learning_rate": 0.00018752004317175832, "loss": 1.7042, "step": 2365 }, { "epoch": 0.7448731044236663, "grad_norm": 0.1503387987613678, "learning_rate": 0.00018538245969082056, "loss": 1.7176, "step": 2370 }, { "epoch": 0.7464445666692858, "grad_norm": 0.15558657050132751, "learning_rate": 0.00018325435345986995, "loss": 1.7821, "step": 2375 }, { "epoch": 0.7480160289149054, "grad_norm": 0.14257632195949554, "learning_rate": 0.0001811357885840254, "loss": 1.7148, "step": 2380 }, { "epoch": 0.7495874911605249, "grad_norm": 0.13766352832317352, "learning_rate": 0.00017902682888099026, "loss": 1.7506, "step": 2385 }, { "epoch": 0.7511589534061445, "grad_norm": 0.16231706738471985, "learning_rate": 0.00017692753787913057, "loss": 1.7785, "step": 2390 }, { "epoch": 0.752730415651764, "grad_norm": 0.15460623800754547, "learning_rate": 0.00017483797881556173, "loss": 1.7757, "step": 2395 }, { "epoch": 0.7543018778973836, "grad_norm": 0.15915700793266296, "learning_rate": 0.000172758214634243, "loss": 1.6892, "step": 2400 }, { "epoch": 0.7558733401430031, "grad_norm": 0.14042919874191284, "learning_rate": 0.0001706883079840812, "loss": 1.7892, "step": 2405 }, { "epoch": 0.7574448023886227, "grad_norm": 0.23500895500183105, "learning_rate": 0.00016862832121704435, "loss": 1.7211, "step": 2410 }, { "epoch": 0.7590162646342422, "grad_norm": 0.20954306423664093, "learning_rate": 0.00016657831638628297, "loss": 1.7364, "step": 2415 }, { "epoch": 0.7605877268798618, "grad_norm": 0.18037594854831696, "learning_rate": 0.00016453835524426086, "loss": 1.7445, "step": 2420 }, { "epoch": 0.7621591891254813, "grad_norm": 0.1708739697933197, "learning_rate": 0.00016250849924089484, "loss": 1.7493, "step": 2425 }, { "epoch": 0.7637306513711009, "grad_norm": 0.16356390714645386, "learning_rate": 0.00016048880952170374, "loss": 1.7218, "step": 2430 }, { "epoch": 0.7653021136167204, "grad_norm": 0.1173071339726448, "learning_rate": 0.00015847934692596688, "loss": 1.7069, "step": 2435 }, { "epoch": 0.76687357586234, "grad_norm": 0.15243308246135712, "learning_rate": 0.00015648017198489106, "loss": 1.7909, "step": 2440 }, { "epoch": 0.7684450381079595, "grad_norm": 0.12692369520664215, "learning_rate": 0.00015449134491978683, "loss": 1.7751, "step": 2445 }, { "epoch": 0.770016500353579, "grad_norm": 0.13145235180854797, "learning_rate": 0.00015251292564025527, "loss": 1.76, "step": 2450 }, { "epoch": 0.7715879625991986, "grad_norm": 0.12512874603271484, "learning_rate": 0.00015054497374238275, "loss": 1.7219, "step": 2455 }, { "epoch": 0.7731594248448181, "grad_norm": 0.1528131067752838, "learning_rate": 0.0001485875485069456, "loss": 1.7519, "step": 2460 }, { "epoch": 0.7747308870904377, "grad_norm": 0.213288813829422, "learning_rate": 0.00014664070889762492, "loss": 1.7176, "step": 2465 }, { "epoch": 0.7763023493360572, "grad_norm": 0.12732981145381927, "learning_rate": 0.00014470451355923025, "loss": 1.7407, "step": 2470 }, { "epoch": 0.7778738115816768, "grad_norm": 0.13688194751739502, "learning_rate": 0.00014277902081593252, "loss": 1.7018, "step": 2475 }, { "epoch": 0.7794452738272963, "grad_norm": 0.1379719078540802, "learning_rate": 0.00014086428866950744, "loss": 1.7401, "step": 2480 }, { "epoch": 0.7810167360729159, "grad_norm": 0.15917198359966278, "learning_rate": 0.00013896037479758878, "loss": 1.7188, "step": 2485 }, { "epoch": 0.7825881983185354, "grad_norm": 0.16691961884498596, "learning_rate": 0.00013706733655193055, "loss": 1.6855, "step": 2490 }, { "epoch": 0.784159660564155, "grad_norm": 0.11223277449607849, "learning_rate": 0.0001351852309566788, "loss": 1.7897, "step": 2495 }, { "epoch": 0.7857311228097745, "grad_norm": 0.13406723737716675, "learning_rate": 0.00013331411470665505, "loss": 1.7386, "step": 2500 }, { "epoch": 0.7873025850553941, "grad_norm": 0.2705506980419159, "learning_rate": 0.0001314540441656476, "loss": 1.6985, "step": 2505 }, { "epoch": 0.7888740473010136, "grad_norm": 0.22507880628108978, "learning_rate": 0.00012960507536471428, "loss": 1.721, "step": 2510 }, { "epoch": 0.7904455095466332, "grad_norm": 0.1900377720594406, "learning_rate": 0.0001277672640004936, "loss": 1.7351, "step": 2515 }, { "epoch": 0.7920169717922527, "grad_norm": 0.17875802516937256, "learning_rate": 0.0001259406654335285, "loss": 1.7385, "step": 2520 }, { "epoch": 0.7935884340378723, "grad_norm": 0.23769760131835938, "learning_rate": 0.0001241253346865972, "loss": 1.7105, "step": 2525 }, { "epoch": 0.7951598962834918, "grad_norm": 0.15305249392986298, "learning_rate": 0.000122321326443057, "loss": 1.7535, "step": 2530 }, { "epoch": 0.7967313585291114, "grad_norm": 0.13712617754936218, "learning_rate": 0.00012052869504519603, "loss": 1.6869, "step": 2535 }, { "epoch": 0.7983028207747309, "grad_norm": 0.14712879061698914, "learning_rate": 0.0001187474944925972, "loss": 1.6889, "step": 2540 }, { "epoch": 0.7998742830203505, "grad_norm": 0.20566821098327637, "learning_rate": 0.00011697777844051105, "loss": 1.73, "step": 2545 }, { "epoch": 0.80144574526597, "grad_norm": 0.16050195693969727, "learning_rate": 0.00011521960019823913, "loss": 1.705, "step": 2550 }, { "epoch": 0.8030172075115896, "grad_norm": 0.2171618640422821, "learning_rate": 0.00011347301272752913, "loss": 1.7078, "step": 2555 }, { "epoch": 0.8045886697572091, "grad_norm": 0.19696617126464844, "learning_rate": 0.00011173806864097885, "loss": 1.7577, "step": 2560 }, { "epoch": 0.8061601320028287, "grad_norm": 0.15522879362106323, "learning_rate": 0.00011001482020045128, "loss": 1.7271, "step": 2565 }, { "epoch": 0.8077315942484482, "grad_norm": 0.13360817730426788, "learning_rate": 0.00010830331931550047, "loss": 1.7681, "step": 2570 }, { "epoch": 0.8093030564940678, "grad_norm": 0.17120471596717834, "learning_rate": 0.0001066036175418082, "loss": 1.7188, "step": 2575 }, { "epoch": 0.8108745187396873, "grad_norm": 0.13265547156333923, "learning_rate": 0.00010491576607963066, "loss": 1.7485, "step": 2580 }, { "epoch": 0.8124459809853068, "grad_norm": 0.14022652804851532, "learning_rate": 0.0001032398157722556, "loss": 1.6629, "step": 2585 }, { "epoch": 0.8140174432309264, "grad_norm": 0.14860029518604279, "learning_rate": 0.0001015758171044719, "loss": 1.6937, "step": 2590 }, { "epoch": 0.815588905476546, "grad_norm": 0.1722240149974823, "learning_rate": 9.992382020104807e-05, "loss": 1.7502, "step": 2595 }, { "epoch": 0.8171603677221655, "grad_norm": 0.1788044422864914, "learning_rate": 9.828387482522216e-05, "loss": 1.6794, "step": 2600 }, { "epoch": 0.818731829967785, "grad_norm": 0.17370399832725525, "learning_rate": 9.66560303772035e-05, "loss": 1.6838, "step": 2605 }, { "epoch": 0.8203032922134046, "grad_norm": 0.12996020913124084, "learning_rate": 9.504033589268401e-05, "loss": 1.7152, "step": 2610 }, { "epoch": 0.8218747544590241, "grad_norm": 0.14151506125926971, "learning_rate": 9.343684004136121e-05, "loss": 1.7185, "step": 2615 }, { "epoch": 0.8234462167046437, "grad_norm": 0.15836787223815918, "learning_rate": 9.184559112547208e-05, "loss": 1.7237, "step": 2620 }, { "epoch": 0.8250176789502632, "grad_norm": 0.1398027092218399, "learning_rate": 9.026663707833843e-05, "loss": 1.7814, "step": 2625 }, { "epoch": 0.8265891411958828, "grad_norm": 0.14515486359596252, "learning_rate": 8.870002546292256e-05, "loss": 1.6791, "step": 2630 }, { "epoch": 0.8281606034415023, "grad_norm": 0.16488778591156006, "learning_rate": 8.714580347039492e-05, "loss": 1.75, "step": 2635 }, { "epoch": 0.8297320656871219, "grad_norm": 0.1568734496831894, "learning_rate": 8.560401791871186e-05, "loss": 1.7421, "step": 2640 }, { "epoch": 0.8313035279327414, "grad_norm": 0.15567830204963684, "learning_rate": 8.407471525120625e-05, "loss": 1.7411, "step": 2645 }, { "epoch": 0.832874990178361, "grad_norm": 0.16733458638191223, "learning_rate": 8.255794153518798e-05, "loss": 1.7286, "step": 2650 }, { "epoch": 0.8344464524239805, "grad_norm": 0.1331174075603485, "learning_rate": 8.10537424605558e-05, "loss": 1.6844, "step": 2655 }, { "epoch": 0.8360179146696001, "grad_norm": 0.18669481575489044, "learning_rate": 7.95621633384223e-05, "loss": 1.7329, "step": 2660 }, { "epoch": 0.8375893769152196, "grad_norm": 0.1392640769481659, "learning_rate": 7.808324909974745e-05, "loss": 1.7276, "step": 2665 }, { "epoch": 0.8391608391608392, "grad_norm": 0.15594840049743652, "learning_rate": 7.661704429398653e-05, "loss": 1.6907, "step": 2670 }, { "epoch": 0.8407323014064587, "grad_norm": 0.15163709223270416, "learning_rate": 7.516359308774695e-05, "loss": 1.7359, "step": 2675 }, { "epoch": 0.8423037636520783, "grad_norm": 0.14341649413108826, "learning_rate": 7.37229392634588e-05, "loss": 1.7746, "step": 2680 }, { "epoch": 0.8438752258976978, "grad_norm": 0.1520870178937912, "learning_rate": 7.229512621805562e-05, "loss": 1.7143, "step": 2685 }, { "epoch": 0.8454466881433174, "grad_norm": 0.1629820466041565, "learning_rate": 7.08801969616667e-05, "loss": 1.818, "step": 2690 }, { "epoch": 0.8470181503889369, "grad_norm": 0.14012764394283295, "learning_rate": 6.947819411632222e-05, "loss": 1.7606, "step": 2695 }, { "epoch": 0.8485896126345565, "grad_norm": 0.18674196302890778, "learning_rate": 6.808915991466902e-05, "loss": 1.7707, "step": 2700 }, { "epoch": 0.850161074880176, "grad_norm": 0.15248626470565796, "learning_rate": 6.671313619869857e-05, "loss": 1.7617, "step": 2705 }, { "epoch": 0.8517325371257956, "grad_norm": 0.185978963971138, "learning_rate": 6.535016441848573e-05, "loss": 1.6929, "step": 2710 }, { "epoch": 0.8533039993714151, "grad_norm": 0.14573198556900024, "learning_rate": 6.400028563094152e-05, "loss": 1.725, "step": 2715 }, { "epoch": 0.8548754616170346, "grad_norm": 0.18840889632701874, "learning_rate": 6.266354049857543e-05, "loss": 1.7846, "step": 2720 }, { "epoch": 0.8564469238626542, "grad_norm": 0.13628707826137543, "learning_rate": 6.13399692882709e-05, "loss": 1.7298, "step": 2725 }, { "epoch": 0.8580183861082737, "grad_norm": 0.14333242177963257, "learning_rate": 6.002961187007194e-05, "loss": 1.7341, "step": 2730 }, { "epoch": 0.8595898483538933, "grad_norm": 0.13763877749443054, "learning_rate": 5.873250771598265e-05, "loss": 1.7192, "step": 2735 }, { "epoch": 0.8611613105995128, "grad_norm": 0.14965565502643585, "learning_rate": 5.7448695898778106e-05, "loss": 1.7346, "step": 2740 }, { "epoch": 0.8627327728451324, "grad_norm": 0.11032088100910187, "learning_rate": 5.617821509082671e-05, "loss": 1.6652, "step": 2745 }, { "epoch": 0.8643042350907519, "grad_norm": 0.17685818672180176, "learning_rate": 5.49211035629264e-05, "loss": 1.8149, "step": 2750 }, { "epoch": 0.8658756973363715, "grad_norm": 0.14297960698604584, "learning_rate": 5.3677399183150674e-05, "loss": 1.7199, "step": 2755 }, { "epoch": 0.867447159581991, "grad_norm": 0.1297394186258316, "learning_rate": 5.244713941570889e-05, "loss": 1.7095, "step": 2760 }, { "epoch": 0.8690186218276106, "grad_norm": 0.13748817145824432, "learning_rate": 5.123036131981668e-05, "loss": 1.7151, "step": 2765 }, { "epoch": 0.8705900840732301, "grad_norm": 0.11178262531757355, "learning_rate": 5.002710154858065e-05, "loss": 1.7202, "step": 2770 }, { "epoch": 0.8721615463188497, "grad_norm": 0.12300541251897812, "learning_rate": 4.883739634789375e-05, "loss": 1.7699, "step": 2775 }, { "epoch": 0.8737330085644692, "grad_norm": 0.12739528715610504, "learning_rate": 4.7661281555343164e-05, "loss": 1.7716, "step": 2780 }, { "epoch": 0.8753044708100888, "grad_norm": 0.12120873481035233, "learning_rate": 4.649879259913137e-05, "loss": 1.7218, "step": 2785 }, { "epoch": 0.8768759330557083, "grad_norm": 0.1548471450805664, "learning_rate": 4.534996449700879e-05, "loss": 1.7433, "step": 2790 }, { "epoch": 0.8784473953013279, "grad_norm": 0.16176049411296844, "learning_rate": 4.421483185521835e-05, "loss": 1.728, "step": 2795 }, { "epoch": 0.8800188575469474, "grad_norm": 0.19101834297180176, "learning_rate": 4.309342886745399e-05, "loss": 1.7275, "step": 2800 }, { "epoch": 0.881590319792567, "grad_norm": 0.14591899514198303, "learning_rate": 4.198578931382979e-05, "loss": 1.7068, "step": 2805 }, { "epoch": 0.8831617820381865, "grad_norm": 0.16475893557071686, "learning_rate": 4.0891946559863055e-05, "loss": 1.7479, "step": 2810 }, { "epoch": 0.8847332442838061, "grad_norm": 0.18267378211021423, "learning_rate": 3.981193355546869e-05, "loss": 1.7445, "step": 2815 }, { "epoch": 0.8863047065294256, "grad_norm": 0.10409973561763763, "learning_rate": 3.874578283396718e-05, "loss": 1.7277, "step": 2820 }, { "epoch": 0.8878761687750452, "grad_norm": 0.13992512226104736, "learning_rate": 3.769352651110419e-05, "loss": 1.7217, "step": 2825 }, { "epoch": 0.8894476310206647, "grad_norm": 0.1338614523410797, "learning_rate": 3.6655196284083314e-05, "loss": 1.7598, "step": 2830 }, { "epoch": 0.8910190932662843, "grad_norm": 0.15006397664546967, "learning_rate": 3.563082343061108e-05, "loss": 1.7017, "step": 2835 }, { "epoch": 0.8925905555119038, "grad_norm": 0.14285215735435486, "learning_rate": 3.4620438807955125e-05, "loss": 1.7068, "step": 2840 }, { "epoch": 0.8941620177575234, "grad_norm": 0.11642735451459885, "learning_rate": 3.3624072852014354e-05, "loss": 1.7363, "step": 2845 }, { "epoch": 0.8957334800031429, "grad_norm": 0.1364676058292389, "learning_rate": 3.2641755576402255e-05, "loss": 1.79, "step": 2850 }, { "epoch": 0.8973049422487624, "grad_norm": 0.13236261904239655, "learning_rate": 3.16735165715426e-05, "loss": 1.7463, "step": 2855 }, { "epoch": 0.898876404494382, "grad_norm": 0.16500313580036163, "learning_rate": 3.071938500377852e-05, "loss": 1.7057, "step": 2860 }, { "epoch": 0.9004478667400015, "grad_norm": 0.15041331946849823, "learning_rate": 2.9779389614493558e-05, "loss": 1.8108, "step": 2865 }, { "epoch": 0.9020193289856211, "grad_norm": 0.1336473524570465, "learning_rate": 2.8853558719245833e-05, "loss": 1.7186, "step": 2870 }, { "epoch": 0.9035907912312406, "grad_norm": 0.1339365541934967, "learning_rate": 2.794192020691544e-05, "loss": 1.7605, "step": 2875 }, { "epoch": 0.9051622534768602, "grad_norm": 0.09974883496761322, "learning_rate": 2.704450153886423e-05, "loss": 1.7394, "step": 2880 }, { "epoch": 0.9067337157224797, "grad_norm": 0.1279599815607071, "learning_rate": 2.6161329748108253e-05, "loss": 1.7415, "step": 2885 }, { "epoch": 0.9083051779680993, "grad_norm": 0.1555188149213791, "learning_rate": 2.5292431438503905e-05, "loss": 1.8056, "step": 2890 }, { "epoch": 0.9098766402137188, "grad_norm": 0.19093474745750427, "learning_rate": 2.4437832783946234e-05, "loss": 1.7738, "step": 2895 }, { "epoch": 0.9114481024593384, "grad_norm": 0.09890997409820557, "learning_rate": 2.3597559527580692e-05, "loss": 1.6863, "step": 2900 }, { "epoch": 0.9130195647049579, "grad_norm": 0.14242489635944366, "learning_rate": 2.2771636981027467e-05, "loss": 1.6858, "step": 2905 }, { "epoch": 0.9145910269505775, "grad_norm": 0.15494489669799805, "learning_rate": 2.1960090023619205e-05, "loss": 1.7568, "step": 2910 }, { "epoch": 0.916162489196197, "grad_norm": 0.1707945168018341, "learning_rate": 2.1162943101651622e-05, "loss": 1.7361, "step": 2915 }, { "epoch": 0.9177339514418166, "grad_norm": 0.17627616226673126, "learning_rate": 2.038022022764685e-05, "loss": 1.791, "step": 2920 }, { "epoch": 0.9193054136874361, "grad_norm": 0.15799571573734283, "learning_rate": 1.9611944979630204e-05, "loss": 1.7233, "step": 2925 }, { "epoch": 0.9208768759330557, "grad_norm": 0.11753001809120178, "learning_rate": 1.8858140500420005e-05, "loss": 1.7659, "step": 2930 }, { "epoch": 0.9224483381786752, "grad_norm": 0.11653709411621094, "learning_rate": 1.8118829496930557e-05, "loss": 1.7428, "step": 2935 }, { "epoch": 0.9240198004242948, "grad_norm": 0.1365276426076889, "learning_rate": 1.739403423948782e-05, "loss": 1.728, "step": 2940 }, { "epoch": 0.9255912626699143, "grad_norm": 0.1241711974143982, "learning_rate": 1.668377656115877e-05, "loss": 1.7144, "step": 2945 }, { "epoch": 0.9271627249155339, "grad_norm": 0.15429584681987762, "learning_rate": 1.5988077857093775e-05, "loss": 1.6854, "step": 2950 }, { "epoch": 0.9287341871611534, "grad_norm": 0.11500924080610275, "learning_rate": 1.5306959083882078e-05, "loss": 1.753, "step": 2955 }, { "epoch": 0.930305649406773, "grad_norm": 0.12633784115314484, "learning_rate": 1.4640440758920293e-05, "loss": 1.7387, "step": 2960 }, { "epoch": 0.9318771116523925, "grad_norm": 0.13106156885623932, "learning_rate": 1.3988542959794625e-05, "loss": 1.7124, "step": 2965 }, { "epoch": 0.9334485738980121, "grad_norm": 0.10034507513046265, "learning_rate": 1.3351285323676022e-05, "loss": 1.7571, "step": 2970 }, { "epoch": 0.9350200361436316, "grad_norm": 0.1519390344619751, "learning_rate": 1.2728687046728526e-05, "loss": 1.6967, "step": 2975 }, { "epoch": 0.9365914983892512, "grad_norm": 0.15289808809757233, "learning_rate": 1.2120766883531087e-05, "loss": 1.7167, "step": 2980 }, { "epoch": 0.9381629606348707, "grad_norm": 0.11695173382759094, "learning_rate": 1.152754314651283e-05, "loss": 1.7096, "step": 2985 }, { "epoch": 0.9397344228804902, "grad_norm": 0.11231189966201782, "learning_rate": 1.0949033705400902e-05, "loss": 1.7092, "step": 2990 }, { "epoch": 0.9413058851261098, "grad_norm": 0.12071159482002258, "learning_rate": 1.0385255986682718e-05, "loss": 1.7096, "step": 2995 }, { "epoch": 0.9428773473717293, "grad_norm": 0.12497507780790329, "learning_rate": 9.836226973080786e-06, "loss": 1.7723, "step": 3000 }, { "epoch": 0.9444488096173489, "grad_norm": 0.11592131853103638, "learning_rate": 9.30196320304122e-06, "loss": 1.7458, "step": 3005 }, { "epoch": 0.9460202718629684, "grad_norm": 0.16098622977733612, "learning_rate": 8.782480770235246e-06, "loss": 1.7568, "step": 3010 }, { "epoch": 0.947591734108588, "grad_norm": 0.1251290738582611, "learning_rate": 8.277795323074933e-06, "loss": 1.7218, "step": 3015 }, { "epoch": 0.9491631963542075, "grad_norm": 0.15294887125492096, "learning_rate": 7.787922064241393e-06, "loss": 1.7038, "step": 3020 }, { "epoch": 0.9507346585998271, "grad_norm": 0.12903992831707, "learning_rate": 7.312875750227044e-06, "loss": 1.7287, "step": 3025 }, { "epoch": 0.9523061208454466, "grad_norm": 0.17125993967056274, "learning_rate": 6.852670690890961e-06, "loss": 1.7112, "step": 3030 }, { "epoch": 0.9538775830910662, "grad_norm": 0.13478563725948334, "learning_rate": 6.40732074902789e-06, "loss": 1.7588, "step": 3035 }, { "epoch": 0.9554490453366857, "grad_norm": 0.14833632111549377, "learning_rate": 5.97683933995069e-06, "loss": 1.7438, "step": 3040 }, { "epoch": 0.9570205075823053, "grad_norm": 0.16726098954677582, "learning_rate": 5.561239431086218e-06, "loss": 1.7639, "step": 3045 }, { "epoch": 0.9585919698279249, "grad_norm": 0.16357098519802094, "learning_rate": 5.160533541584578e-06, "loss": 1.6912, "step": 3050 }, { "epoch": 0.9601634320735445, "grad_norm": 0.12722498178482056, "learning_rate": 4.774733741942205e-06, "loss": 1.7576, "step": 3055 }, { "epoch": 0.961734894319164, "grad_norm": 0.16135632991790771, "learning_rate": 4.403851653638158e-06, "loss": 1.7702, "step": 3060 }, { "epoch": 0.9633063565647836, "grad_norm": 0.156170055270195, "learning_rate": 4.0478984487838935e-06, "loss": 1.7429, "step": 3065 }, { "epoch": 0.9648778188104031, "grad_norm": 0.1406138390302658, "learning_rate": 3.706884849787151e-06, "loss": 1.7209, "step": 3070 }, { "epoch": 0.9664492810560227, "grad_norm": 0.1389617770910263, "learning_rate": 3.3808211290284885e-06, "loss": 1.7126, "step": 3075 }, { "epoch": 0.9680207433016422, "grad_norm": 0.1216338723897934, "learning_rate": 3.0697171085521946e-06, "loss": 1.7186, "step": 3080 }, { "epoch": 0.9695922055472618, "grad_norm": 0.13428504765033722, "learning_rate": 2.7735821597701382e-06, "loss": 1.7334, "step": 3085 }, { "epoch": 0.9711636677928813, "grad_norm": 0.1258799433708191, "learning_rate": 2.49242520317966e-06, "loss": 1.7303, "step": 3090 }, { "epoch": 0.9727351300385009, "grad_norm": 0.144920215010643, "learning_rate": 2.2262547080948992e-06, "loss": 1.7595, "step": 3095 }, { "epoch": 0.9743065922841204, "grad_norm": 0.12858329713344574, "learning_rate": 1.975078692391552e-06, "loss": 1.7411, "step": 3100 }, { "epoch": 0.97587805452974, "grad_norm": 0.1087680459022522, "learning_rate": 1.7389047222652888e-06, "loss": 1.7469, "step": 3105 }, { "epoch": 0.9774495167753595, "grad_norm": 0.12177930027246475, "learning_rate": 1.5177399120039904e-06, "loss": 1.6998, "step": 3110 }, { "epoch": 0.9790209790209791, "grad_norm": 0.11609877645969391, "learning_rate": 1.3115909237734204e-06, "loss": 1.7057, "step": 3115 }, { "epoch": 0.9805924412665986, "grad_norm": 0.13257728517055511, "learning_rate": 1.1204639674164962e-06, "loss": 1.7442, "step": 3120 }, { "epoch": 0.9821639035122182, "grad_norm": 0.12050619721412659, "learning_rate": 9.44364800266162e-07, "loss": 1.7151, "step": 3125 }, { "epoch": 0.9837353657578377, "grad_norm": 0.12853524088859558, "learning_rate": 7.832987269720815e-07, "loss": 1.7062, "step": 3130 }, { "epoch": 0.9853068280034573, "grad_norm": 0.11707114428281784, "learning_rate": 6.372705993408223e-07, "loss": 1.7023, "step": 3135 }, { "epoch": 0.9868782902490768, "grad_norm": 0.12321787327528, "learning_rate": 5.062848161896394e-07, "loss": 1.7308, "step": 3140 }, { "epoch": 0.9884497524946964, "grad_norm": 0.13473524153232574, "learning_rate": 3.903453232140808e-07, "loss": 1.7428, "step": 3145 }, { "epoch": 0.9900212147403159, "grad_norm": 0.12741941213607788, "learning_rate": 2.894556128689163e-07, "loss": 1.6881, "step": 3150 }, { "epoch": 0.9915926769859355, "grad_norm": 0.1763051599264145, "learning_rate": 2.03618724263277e-07, "loss": 1.7244, "step": 3155 }, { "epoch": 0.993164139231555, "grad_norm": 0.11934591829776764, "learning_rate": 1.3283724306867306e-07, "loss": 1.7025, "step": 3160 }, { "epoch": 0.9947356014771745, "grad_norm": 0.12859931588172913, "learning_rate": 7.711330144161144e-08, "loss": 1.7378, "step": 3165 }, { "epoch": 0.9963070637227941, "grad_norm": 0.14746056497097015, "learning_rate": 3.644857795886969e-08, "loss": 1.7293, "step": 3170 }, { "epoch": 0.9978785259684136, "grad_norm": 0.16102327406406403, "learning_rate": 1.0844297567258466e-08, "loss": 1.6806, "step": 3175 }, { "epoch": 0.9994499882140332, "grad_norm": 0.15685699880123138, "learning_rate": 3.012315465955595e-10, "loss": 1.8048, "step": 3180 }, { "epoch": 0.999764280663157, "eval_loss": 1.7209471464157104, "eval_runtime": 333.4066, "eval_samples_per_second": 31.946, "eval_steps_per_second": 1.998, "step": 3181 }, { "epoch": 0.999764280663157, "step": 3181, "total_flos": 2.656972328528773e+17, "train_loss": 2.0235598598475426, "train_runtime": 14726.7001, "train_samples_per_second": 13.827, "train_steps_per_second": 0.216 } ], "logging_steps": 5, "max_steps": 3181, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.656972328528773e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }