| { | |
| "best_metric": 3.101016044616699, | |
| "best_model_checkpoint": "./output/models/parallel-gpt2-medium-wikitext/checkpoint-17500", | |
| "epoch": 4.998736842105263, | |
| "eval_steps": 500, | |
| "global_step": 17810, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.028070175438596492, | |
| "grad_norm": 2.028989553451538, | |
| "learning_rate": 5.614823133071308e-06, | |
| "loss": 9.2256, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.056140350877192984, | |
| "grad_norm": 1.8919556140899658, | |
| "learning_rate": 1.1229646266142617e-05, | |
| "loss": 7.8733, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.08421052631578947, | |
| "grad_norm": 1.433174967765808, | |
| "learning_rate": 1.6844469399213926e-05, | |
| "loss": 7.0853, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.11228070175438597, | |
| "grad_norm": 1.4522157907485962, | |
| "learning_rate": 2.2459292532285233e-05, | |
| "loss": 6.6702, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.14035087719298245, | |
| "grad_norm": 1.322840929031372, | |
| "learning_rate": 2.8074115665356544e-05, | |
| "loss": 6.4455, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.14035087719298245, | |
| "eval_accuracy": 0.17660504282313108, | |
| "eval_bleu": 0.02573169351336114, | |
| "eval_loss": 6.331261157989502, | |
| "eval_perplexity": 561.8647474275982, | |
| "eval_runtime": 20.1474, | |
| "eval_samples_per_second": 56.633, | |
| "eval_steps_per_second": 3.574, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.16842105263157894, | |
| "grad_norm": 1.5128698348999023, | |
| "learning_rate": 3.368893879842785e-05, | |
| "loss": 6.2694, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.19649122807017544, | |
| "grad_norm": 1.4778157472610474, | |
| "learning_rate": 3.930376193149916e-05, | |
| "loss": 6.1217, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.22456140350877193, | |
| "grad_norm": 1.1647950410842896, | |
| "learning_rate": 4.4918585064570466e-05, | |
| "loss": 5.9897, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.25263157894736843, | |
| "grad_norm": 1.3264423608779907, | |
| "learning_rate": 5.053340819764178e-05, | |
| "loss": 5.8659, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.2807017543859649, | |
| "grad_norm": 1.521269679069519, | |
| "learning_rate": 5.614823133071309e-05, | |
| "loss": 5.7254, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.2807017543859649, | |
| "eval_accuracy": 0.2136290386834618, | |
| "eval_bleu": 0.045427591918014404, | |
| "eval_loss": 5.623491287231445, | |
| "eval_perplexity": 276.85427569402634, | |
| "eval_runtime": 20.1737, | |
| "eval_samples_per_second": 56.559, | |
| "eval_steps_per_second": 3.569, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3087719298245614, | |
| "grad_norm": 1.1612671613693237, | |
| "learning_rate": 6.176305446378439e-05, | |
| "loss": 5.5759, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.3368421052631579, | |
| "grad_norm": 1.1074801683425903, | |
| "learning_rate": 6.73778775968557e-05, | |
| "loss": 5.4609, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.3649122807017544, | |
| "grad_norm": 1.0632935762405396, | |
| "learning_rate": 7.299270072992701e-05, | |
| "loss": 5.3257, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.3929824561403509, | |
| "grad_norm": 1.1046158075332642, | |
| "learning_rate": 7.860752386299832e-05, | |
| "loss": 5.1946, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.42105263157894735, | |
| "grad_norm": 1.1495596170425415, | |
| "learning_rate": 8.422234699606962e-05, | |
| "loss": 5.1084, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.42105263157894735, | |
| "eval_accuracy": 0.2575530545053601, | |
| "eval_bleu": 0.06490299322624599, | |
| "eval_loss": 4.982165813446045, | |
| "eval_perplexity": 145.78979346372705, | |
| "eval_runtime": 20.1855, | |
| "eval_samples_per_second": 56.526, | |
| "eval_steps_per_second": 3.567, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.44912280701754387, | |
| "grad_norm": 0.9580332040786743, | |
| "learning_rate": 8.983717012914093e-05, | |
| "loss": 4.9932, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.47719298245614034, | |
| "grad_norm": 0.9515564441680908, | |
| "learning_rate": 9.545199326221224e-05, | |
| "loss": 4.8819, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.5052631578947369, | |
| "grad_norm": 0.9855053424835205, | |
| "learning_rate": 9.98814648449685e-05, | |
| "loss": 4.7868, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.5333333333333333, | |
| "grad_norm": 1.0349925756454468, | |
| "learning_rate": 9.925759560796058e-05, | |
| "loss": 4.6832, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.5614035087719298, | |
| "grad_norm": 1.0200501680374146, | |
| "learning_rate": 9.863372637095265e-05, | |
| "loss": 4.5994, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.5614035087719298, | |
| "eval_accuracy": 0.29288759924026103, | |
| "eval_bleu": 0.07406906637840037, | |
| "eval_loss": 4.5052409172058105, | |
| "eval_perplexity": 90.49014205518364, | |
| "eval_runtime": 19.5875, | |
| "eval_samples_per_second": 58.251, | |
| "eval_steps_per_second": 3.676, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.5894736842105263, | |
| "grad_norm": 0.9353361129760742, | |
| "learning_rate": 9.800985713394473e-05, | |
| "loss": 4.5176, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.6175438596491228, | |
| "grad_norm": 0.9589385390281677, | |
| "learning_rate": 9.73859878969368e-05, | |
| "loss": 4.4355, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.6456140350877193, | |
| "grad_norm": 1.0331965684890747, | |
| "learning_rate": 9.676211865992888e-05, | |
| "loss": 4.3647, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.6736842105263158, | |
| "grad_norm": 0.9204744100570679, | |
| "learning_rate": 9.613824942292095e-05, | |
| "loss": 4.2935, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.7017543859649122, | |
| "grad_norm": 0.9139348268508911, | |
| "learning_rate": 9.551438018591303e-05, | |
| "loss": 4.2338, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.7017543859649122, | |
| "eval_accuracy": 0.32728917628977, | |
| "eval_bleu": 0.09369678160356684, | |
| "eval_loss": 4.137840747833252, | |
| "eval_perplexity": 62.66736062058415, | |
| "eval_runtime": 20.1595, | |
| "eval_samples_per_second": 56.598, | |
| "eval_steps_per_second": 3.572, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.7298245614035088, | |
| "grad_norm": 0.9106621146202087, | |
| "learning_rate": 9.489051094890511e-05, | |
| "loss": 4.1694, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.7578947368421053, | |
| "grad_norm": 0.9036598801612854, | |
| "learning_rate": 9.426664171189718e-05, | |
| "loss": 4.1386, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.7859649122807018, | |
| "grad_norm": 0.8367689847946167, | |
| "learning_rate": 9.364277247488927e-05, | |
| "loss": 4.0731, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.8140350877192982, | |
| "grad_norm": 0.9330971837043762, | |
| "learning_rate": 9.301890323788135e-05, | |
| "loss": 4.0342, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "grad_norm": 0.847179114818573, | |
| "learning_rate": 9.239503400087343e-05, | |
| "loss": 3.9975, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "eval_accuracy": 0.3465208187155545, | |
| "eval_bleu": 0.10305263495616372, | |
| "eval_loss": 3.928612470626831, | |
| "eval_perplexity": 50.83639172374704, | |
| "eval_runtime": 20.1917, | |
| "eval_samples_per_second": 56.508, | |
| "eval_steps_per_second": 3.566, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.8701754385964913, | |
| "grad_norm": 0.8944078087806702, | |
| "learning_rate": 9.17711647638655e-05, | |
| "loss": 3.9805, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.8982456140350877, | |
| "grad_norm": 0.8461546301841736, | |
| "learning_rate": 9.114729552685758e-05, | |
| "loss": 3.9468, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.9263157894736842, | |
| "grad_norm": 0.8034185767173767, | |
| "learning_rate": 9.052342628984965e-05, | |
| "loss": 3.9049, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.9543859649122807, | |
| "grad_norm": 0.8450652360916138, | |
| "learning_rate": 8.989955705284174e-05, | |
| "loss": 3.8727, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.9824561403508771, | |
| "grad_norm": 0.7972449064254761, | |
| "learning_rate": 8.92756878158338e-05, | |
| "loss": 3.8648, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.9824561403508771, | |
| "eval_accuracy": 0.35825016727450926, | |
| "eval_bleu": 0.11656022916295831, | |
| "eval_loss": 3.7925570011138916, | |
| "eval_perplexity": 44.36970874398872, | |
| "eval_runtime": 20.1635, | |
| "eval_samples_per_second": 56.587, | |
| "eval_steps_per_second": 3.571, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.0103859649122806, | |
| "grad_norm": 0.766300618648529, | |
| "learning_rate": 8.865181857882589e-05, | |
| "loss": 3.8193, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.0384561403508772, | |
| "grad_norm": 0.8034301996231079, | |
| "learning_rate": 8.802794934181796e-05, | |
| "loss": 3.7801, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.0665263157894738, | |
| "grad_norm": 0.8166210651397705, | |
| "learning_rate": 8.740408010481004e-05, | |
| "loss": 3.7505, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.0945964912280701, | |
| "grad_norm": 0.7962071895599365, | |
| "learning_rate": 8.678021086780212e-05, | |
| "loss": 3.7398, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.1226666666666667, | |
| "grad_norm": 0.7868255972862244, | |
| "learning_rate": 8.615634163079419e-05, | |
| "loss": 3.7164, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.1226666666666667, | |
| "eval_accuracy": 0.3667368320049895, | |
| "eval_bleu": 0.1225879386890452, | |
| "eval_loss": 3.6986546516418457, | |
| "eval_perplexity": 40.39292523319135, | |
| "eval_runtime": 20.1969, | |
| "eval_samples_per_second": 56.494, | |
| "eval_steps_per_second": 3.565, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.150736842105263, | |
| "grad_norm": 0.7974035739898682, | |
| "learning_rate": 8.553247239378627e-05, | |
| "loss": 3.7099, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.1788070175438596, | |
| "grad_norm": 0.7970170974731445, | |
| "learning_rate": 8.490860315677834e-05, | |
| "loss": 3.6871, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.2068771929824562, | |
| "grad_norm": 0.7822019457817078, | |
| "learning_rate": 8.428473391977042e-05, | |
| "loss": 3.6628, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.2349473684210526, | |
| "grad_norm": 0.7597511410713196, | |
| "learning_rate": 8.36608646827625e-05, | |
| "loss": 3.6631, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.2630175438596491, | |
| "grad_norm": 0.7736372947692871, | |
| "learning_rate": 8.303699544575458e-05, | |
| "loss": 3.6639, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.2630175438596491, | |
| "eval_accuracy": 0.37342095861787133, | |
| "eval_bleu": 0.12824268044271067, | |
| "eval_loss": 3.622089147567749, | |
| "eval_perplexity": 37.415653048434656, | |
| "eval_runtime": 20.1594, | |
| "eval_samples_per_second": 56.599, | |
| "eval_steps_per_second": 3.572, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.2910877192982455, | |
| "grad_norm": 0.7472370266914368, | |
| "learning_rate": 8.241312620874664e-05, | |
| "loss": 3.6404, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.319157894736842, | |
| "grad_norm": 0.771800696849823, | |
| "learning_rate": 8.178925697173873e-05, | |
| "loss": 3.6235, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.3472280701754387, | |
| "grad_norm": 0.7624268531799316, | |
| "learning_rate": 8.11653877347308e-05, | |
| "loss": 3.6123, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.375298245614035, | |
| "grad_norm": 0.7671138644218445, | |
| "learning_rate": 8.054151849772288e-05, | |
| "loss": 3.592, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.4033684210526316, | |
| "grad_norm": 0.7753613591194153, | |
| "learning_rate": 7.991764926071496e-05, | |
| "loss": 3.582, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.4033684210526316, | |
| "eval_accuracy": 0.3795687787375893, | |
| "eval_bleu": 0.1276585909940598, | |
| "eval_loss": 3.5575244426727295, | |
| "eval_perplexity": 35.07625629269647, | |
| "eval_runtime": 20.1451, | |
| "eval_samples_per_second": 56.639, | |
| "eval_steps_per_second": 3.574, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.431438596491228, | |
| "grad_norm": 0.7279810309410095, | |
| "learning_rate": 7.929378002370703e-05, | |
| "loss": 3.5846, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.4595087719298245, | |
| "grad_norm": 0.728085458278656, | |
| "learning_rate": 7.866991078669911e-05, | |
| "loss": 3.5677, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.4875789473684211, | |
| "grad_norm": 0.7683869004249573, | |
| "learning_rate": 7.804604154969118e-05, | |
| "loss": 3.5452, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.5156491228070177, | |
| "grad_norm": 0.7563459277153015, | |
| "learning_rate": 7.742217231268327e-05, | |
| "loss": 3.5557, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.543719298245614, | |
| "grad_norm": 0.7477959394454956, | |
| "learning_rate": 7.679830307567533e-05, | |
| "loss": 3.5315, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.543719298245614, | |
| "eval_accuracy": 0.38395261312340273, | |
| "eval_bleu": 0.13124123866011556, | |
| "eval_loss": 3.5063869953155518, | |
| "eval_perplexity": 33.32763708661619, | |
| "eval_runtime": 20.2862, | |
| "eval_samples_per_second": 56.245, | |
| "eval_steps_per_second": 3.549, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.5717894736842104, | |
| "grad_norm": 0.732016921043396, | |
| "learning_rate": 7.617443383866742e-05, | |
| "loss": 3.5185, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.599859649122807, | |
| "grad_norm": 0.7203788757324219, | |
| "learning_rate": 7.555056460165949e-05, | |
| "loss": 3.5216, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.6279298245614036, | |
| "grad_norm": 0.7561742663383484, | |
| "learning_rate": 7.492669536465158e-05, | |
| "loss": 3.5057, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.6560000000000001, | |
| "grad_norm": 0.7377160787582397, | |
| "learning_rate": 7.430282612764365e-05, | |
| "loss": 3.5058, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.6840701754385965, | |
| "grad_norm": 0.8002681136131287, | |
| "learning_rate": 7.367895689063573e-05, | |
| "loss": 3.5025, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.6840701754385965, | |
| "eval_accuracy": 0.3880588703466202, | |
| "eval_bleu": 0.13662311841214886, | |
| "eval_loss": 3.4594311714172363, | |
| "eval_perplexity": 31.798883255478383, | |
| "eval_runtime": 20.2931, | |
| "eval_samples_per_second": 56.226, | |
| "eval_steps_per_second": 3.548, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.7121403508771929, | |
| "grad_norm": 0.7451775074005127, | |
| "learning_rate": 7.30550876536278e-05, | |
| "loss": 3.4714, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.7402105263157894, | |
| "grad_norm": 0.7121933102607727, | |
| "learning_rate": 7.243121841661989e-05, | |
| "loss": 3.474, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.768280701754386, | |
| "grad_norm": 0.7507015466690063, | |
| "learning_rate": 7.180734917961195e-05, | |
| "loss": 3.4577, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.7963508771929826, | |
| "grad_norm": 0.7061077356338501, | |
| "learning_rate": 7.118347994260404e-05, | |
| "loss": 3.463, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.824421052631579, | |
| "grad_norm": 0.7301498055458069, | |
| "learning_rate": 7.055961070559612e-05, | |
| "loss": 3.4462, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.824421052631579, | |
| "eval_accuracy": 0.3919063982392698, | |
| "eval_bleu": 0.13096345408557636, | |
| "eval_loss": 3.420844554901123, | |
| "eval_perplexity": 30.5952434755694, | |
| "eval_runtime": 20.2154, | |
| "eval_samples_per_second": 56.442, | |
| "eval_steps_per_second": 3.562, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.8524912280701753, | |
| "grad_norm": 0.713631272315979, | |
| "learning_rate": 6.993574146858819e-05, | |
| "loss": 3.4245, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.8805614035087719, | |
| "grad_norm": 0.7331680059432983, | |
| "learning_rate": 6.931187223158027e-05, | |
| "loss": 3.4289, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.9086315789473685, | |
| "grad_norm": 0.7402526140213013, | |
| "learning_rate": 6.868800299457234e-05, | |
| "loss": 3.434, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.936701754385965, | |
| "grad_norm": 0.7350389361381531, | |
| "learning_rate": 6.806413375756442e-05, | |
| "loss": 3.413, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.9647719298245614, | |
| "grad_norm": 0.7257172465324402, | |
| "learning_rate": 6.744026452055649e-05, | |
| "loss": 3.4167, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.9647719298245614, | |
| "eval_accuracy": 0.39561171067207085, | |
| "eval_bleu": 0.1355256129038967, | |
| "eval_loss": 3.3863015174865723, | |
| "eval_perplexity": 29.556435911658944, | |
| "eval_runtime": 20.2056, | |
| "eval_samples_per_second": 56.469, | |
| "eval_steps_per_second": 3.563, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.9928421052631577, | |
| "grad_norm": 0.7187970876693726, | |
| "learning_rate": 6.681639528354857e-05, | |
| "loss": 3.4086, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.020771929824561, | |
| "grad_norm": 0.7202402353286743, | |
| "learning_rate": 6.619252604654064e-05, | |
| "loss": 3.3372, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.048842105263158, | |
| "grad_norm": 0.7403327822685242, | |
| "learning_rate": 6.556865680953273e-05, | |
| "loss": 3.3079, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 2.0769122807017544, | |
| "grad_norm": 0.7321439385414124, | |
| "learning_rate": 6.49447875725248e-05, | |
| "loss": 3.3128, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.104982456140351, | |
| "grad_norm": 0.7240020632743835, | |
| "learning_rate": 6.432091833551688e-05, | |
| "loss": 3.2967, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.104982456140351, | |
| "eval_accuracy": 0.39890408423952856, | |
| "eval_bleu": 0.1316612565684532, | |
| "eval_loss": 3.3547873497009277, | |
| "eval_perplexity": 28.639513342041703, | |
| "eval_runtime": 20.2147, | |
| "eval_samples_per_second": 56.444, | |
| "eval_steps_per_second": 3.562, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.1330526315789475, | |
| "grad_norm": 0.7483673095703125, | |
| "learning_rate": 6.369704909850896e-05, | |
| "loss": 3.2857, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 2.1611228070175437, | |
| "grad_norm": 0.7482850551605225, | |
| "learning_rate": 6.307317986150103e-05, | |
| "loss": 3.3038, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 2.1891929824561402, | |
| "grad_norm": 0.7295832633972168, | |
| "learning_rate": 6.244931062449311e-05, | |
| "loss": 3.284, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.217263157894737, | |
| "grad_norm": 0.7610414028167725, | |
| "learning_rate": 6.182544138748518e-05, | |
| "loss": 3.2862, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 2.2453333333333334, | |
| "grad_norm": 0.7399555444717407, | |
| "learning_rate": 6.120157215047726e-05, | |
| "loss": 3.2909, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.2453333333333334, | |
| "eval_accuracy": 0.4014682461149906, | |
| "eval_bleu": 0.13811168690281525, | |
| "eval_loss": 3.329040050506592, | |
| "eval_perplexity": 27.911535174787044, | |
| "eval_runtime": 20.1568, | |
| "eval_samples_per_second": 56.606, | |
| "eval_steps_per_second": 3.572, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.27340350877193, | |
| "grad_norm": 0.7306208610534668, | |
| "learning_rate": 6.057770291346934e-05, | |
| "loss": 3.2853, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 2.301473684210526, | |
| "grad_norm": 0.7453253269195557, | |
| "learning_rate": 5.9953833676461415e-05, | |
| "loss": 3.2744, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 2.3295438596491227, | |
| "grad_norm": 0.7369946837425232, | |
| "learning_rate": 5.932996443945349e-05, | |
| "loss": 3.2604, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 2.3576140350877193, | |
| "grad_norm": 0.7455064654350281, | |
| "learning_rate": 5.870609520244557e-05, | |
| "loss": 3.2622, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 2.385684210526316, | |
| "grad_norm": 0.7593878507614136, | |
| "learning_rate": 5.808222596543764e-05, | |
| "loss": 3.2593, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.385684210526316, | |
| "eval_accuracy": 0.40385506702546087, | |
| "eval_bleu": 0.14220030200690656, | |
| "eval_loss": 3.3044025897979736, | |
| "eval_perplexity": 27.232267893585384, | |
| "eval_runtime": 20.1746, | |
| "eval_samples_per_second": 56.556, | |
| "eval_steps_per_second": 3.569, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.4137543859649124, | |
| "grad_norm": 0.7376691102981567, | |
| "learning_rate": 5.745835672842972e-05, | |
| "loss": 3.2582, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 2.441824561403509, | |
| "grad_norm": 0.7452288269996643, | |
| "learning_rate": 5.683448749142181e-05, | |
| "loss": 3.2491, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 2.469894736842105, | |
| "grad_norm": 0.749320924282074, | |
| "learning_rate": 5.6210618254413884e-05, | |
| "loss": 3.24, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 2.4979649122807017, | |
| "grad_norm": 0.7482838034629822, | |
| "learning_rate": 5.558674901740596e-05, | |
| "loss": 3.2374, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 2.5260350877192983, | |
| "grad_norm": 0.7401648163795471, | |
| "learning_rate": 5.4962879780398035e-05, | |
| "loss": 3.2408, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.5260350877192983, | |
| "eval_accuracy": 0.4061330845419506, | |
| "eval_bleu": 0.14119139857697727, | |
| "eval_loss": 3.282592296600342, | |
| "eval_perplexity": 26.644754356923094, | |
| "eval_runtime": 20.1938, | |
| "eval_samples_per_second": 56.503, | |
| "eval_steps_per_second": 3.565, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.554105263157895, | |
| "grad_norm": 0.7236805558204651, | |
| "learning_rate": 5.433901054339011e-05, | |
| "loss": 3.2434, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 2.582175438596491, | |
| "grad_norm": 0.7386437058448792, | |
| "learning_rate": 5.371514130638219e-05, | |
| "loss": 3.2445, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 2.6102456140350876, | |
| "grad_norm": 0.7296267747879028, | |
| "learning_rate": 5.309127206937426e-05, | |
| "loss": 3.2257, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 2.638315789473684, | |
| "grad_norm": 0.7237492203712463, | |
| "learning_rate": 5.246740283236634e-05, | |
| "loss": 3.2336, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 2.6663859649122807, | |
| "grad_norm": 0.7474483847618103, | |
| "learning_rate": 5.1843533595358414e-05, | |
| "loss": 3.2278, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 2.6663859649122807, | |
| "eval_accuracy": 0.40903479395464354, | |
| "eval_bleu": 0.14356756322892422, | |
| "eval_loss": 3.2591938972473145, | |
| "eval_perplexity": 26.028547000163957, | |
| "eval_runtime": 20.1915, | |
| "eval_samples_per_second": 56.509, | |
| "eval_steps_per_second": 3.566, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 2.6944561403508773, | |
| "grad_norm": 0.7365037798881531, | |
| "learning_rate": 5.12196643583505e-05, | |
| "loss": 3.2235, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 2.722526315789474, | |
| "grad_norm": 0.7330195307731628, | |
| "learning_rate": 5.059579512134257e-05, | |
| "loss": 3.2179, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 2.75059649122807, | |
| "grad_norm": 0.7449477314949036, | |
| "learning_rate": 4.997192588433465e-05, | |
| "loss": 3.2178, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 2.7786666666666666, | |
| "grad_norm": 0.7243569493293762, | |
| "learning_rate": 4.9348056647326725e-05, | |
| "loss": 3.2126, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 2.806736842105263, | |
| "grad_norm": 0.7603092789649963, | |
| "learning_rate": 4.87241874103188e-05, | |
| "loss": 3.2172, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.806736842105263, | |
| "eval_accuracy": 0.4104792232637077, | |
| "eval_bleu": 0.1411554473599086, | |
| "eval_loss": 3.2415478229522705, | |
| "eval_perplexity": 25.573274029997943, | |
| "eval_runtime": 19.9283, | |
| "eval_samples_per_second": 57.255, | |
| "eval_steps_per_second": 3.613, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.8348070175438598, | |
| "grad_norm": 0.7319552302360535, | |
| "learning_rate": 4.8100318173310876e-05, | |
| "loss": 3.2138, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 2.862877192982456, | |
| "grad_norm": 0.7321156859397888, | |
| "learning_rate": 4.747644893630295e-05, | |
| "loss": 3.2114, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 2.8909473684210525, | |
| "grad_norm": 0.7428148984909058, | |
| "learning_rate": 4.685257969929503e-05, | |
| "loss": 3.1985, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 2.919017543859649, | |
| "grad_norm": 0.7498809099197388, | |
| "learning_rate": 4.6228710462287104e-05, | |
| "loss": 3.1861, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 2.9470877192982456, | |
| "grad_norm": 0.7468050718307495, | |
| "learning_rate": 4.560484122527918e-05, | |
| "loss": 3.2145, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.9470877192982456, | |
| "eval_accuracy": 0.41254477431006226, | |
| "eval_bleu": 0.14022835224132563, | |
| "eval_loss": 3.222651243209839, | |
| "eval_perplexity": 25.094563855273414, | |
| "eval_runtime": 20.1985, | |
| "eval_samples_per_second": 56.489, | |
| "eval_steps_per_second": 3.565, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.9751578947368422, | |
| "grad_norm": 0.7379462122917175, | |
| "learning_rate": 4.498097198827126e-05, | |
| "loss": 3.1799, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 3.0030877192982457, | |
| "grad_norm": 0.7374199032783508, | |
| "learning_rate": 4.435710275126334e-05, | |
| "loss": 3.1681, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 3.0311578947368423, | |
| "grad_norm": 0.7500795722007751, | |
| "learning_rate": 4.3733233514255414e-05, | |
| "loss": 3.096, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 3.0592280701754384, | |
| "grad_norm": 0.7644433975219727, | |
| "learning_rate": 4.3109364277247496e-05, | |
| "loss": 3.0842, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 3.087298245614035, | |
| "grad_norm": 0.7695332169532776, | |
| "learning_rate": 4.248549504023957e-05, | |
| "loss": 3.0749, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.087298245614035, | |
| "eval_accuracy": 0.41432503771708207, | |
| "eval_bleu": 0.1413248900872156, | |
| "eval_loss": 3.209906578063965, | |
| "eval_perplexity": 24.776771422507743, | |
| "eval_runtime": 20.1941, | |
| "eval_samples_per_second": 56.502, | |
| "eval_steps_per_second": 3.565, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.1153684210526316, | |
| "grad_norm": 0.7983810901641846, | |
| "learning_rate": 4.186162580323165e-05, | |
| "loss": 3.0886, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 3.143438596491228, | |
| "grad_norm": 0.7757524847984314, | |
| "learning_rate": 4.1237756566223724e-05, | |
| "loss": 3.0873, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 3.1715087719298247, | |
| "grad_norm": 0.8084014654159546, | |
| "learning_rate": 4.06138873292158e-05, | |
| "loss": 3.0882, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 3.199578947368421, | |
| "grad_norm": 0.7784711718559265, | |
| "learning_rate": 3.9990018092207875e-05, | |
| "loss": 3.0807, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 3.2276491228070174, | |
| "grad_norm": 0.7808659076690674, | |
| "learning_rate": 3.936614885519995e-05, | |
| "loss": 3.0777, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 3.2276491228070174, | |
| "eval_accuracy": 0.4159836469355567, | |
| "eval_bleu": 0.14196972146127593, | |
| "eval_loss": 3.1977927684783936, | |
| "eval_perplexity": 24.478440938871064, | |
| "eval_runtime": 20.2549, | |
| "eval_samples_per_second": 56.332, | |
| "eval_steps_per_second": 3.555, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 3.255719298245614, | |
| "grad_norm": 0.7880111336708069, | |
| "learning_rate": 3.874227961819203e-05, | |
| "loss": 3.0907, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 3.2837894736842106, | |
| "grad_norm": 0.7764289975166321, | |
| "learning_rate": 3.81184103811841e-05, | |
| "loss": 3.0829, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 3.311859649122807, | |
| "grad_norm": 0.7774072885513306, | |
| "learning_rate": 3.749454114417618e-05, | |
| "loss": 3.084, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 3.3399298245614037, | |
| "grad_norm": 0.7633748650550842, | |
| "learning_rate": 3.687067190716826e-05, | |
| "loss": 3.0832, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 3.368, | |
| "grad_norm": 0.7821294069290161, | |
| "learning_rate": 3.624680267016034e-05, | |
| "loss": 3.0743, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 3.368, | |
| "eval_accuracy": 0.41744863751592426, | |
| "eval_bleu": 0.1437616042304935, | |
| "eval_loss": 3.185511589050293, | |
| "eval_perplexity": 24.1796552890801, | |
| "eval_runtime": 20.1722, | |
| "eval_samples_per_second": 56.563, | |
| "eval_steps_per_second": 3.569, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 3.3960701754385965, | |
| "grad_norm": 0.7699295878410339, | |
| "learning_rate": 3.562293343315241e-05, | |
| "loss": 3.0603, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 3.424140350877193, | |
| "grad_norm": 0.7893390655517578, | |
| "learning_rate": 3.499906419614449e-05, | |
| "loss": 3.064, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 3.4522105263157896, | |
| "grad_norm": 0.7758604884147644, | |
| "learning_rate": 3.4375194959136565e-05, | |
| "loss": 3.0791, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 3.4802807017543858, | |
| "grad_norm": 0.7830843925476074, | |
| "learning_rate": 3.375132572212865e-05, | |
| "loss": 3.0691, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 3.5083508771929823, | |
| "grad_norm": 0.7860715389251709, | |
| "learning_rate": 3.312745648512072e-05, | |
| "loss": 3.0679, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 3.5083508771929823, | |
| "eval_accuracy": 0.4182539539753076, | |
| "eval_bleu": 0.13973142885022957, | |
| "eval_loss": 3.1735119819641113, | |
| "eval_perplexity": 23.891242805090002, | |
| "eval_runtime": 20.2985, | |
| "eval_samples_per_second": 56.211, | |
| "eval_steps_per_second": 3.547, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 3.536421052631579, | |
| "grad_norm": 0.7781902551651001, | |
| "learning_rate": 3.25035872481128e-05, | |
| "loss": 3.0577, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 3.5644912280701755, | |
| "grad_norm": 0.7784900665283203, | |
| "learning_rate": 3.1879718011104875e-05, | |
| "loss": 3.0775, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 3.592561403508772, | |
| "grad_norm": 0.8009820580482483, | |
| "learning_rate": 3.125584877409695e-05, | |
| "loss": 3.0532, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 3.6206315789473686, | |
| "grad_norm": 0.7675787210464478, | |
| "learning_rate": 3.0631979537089026e-05, | |
| "loss": 3.0597, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 3.648701754385965, | |
| "grad_norm": 0.7848743796348572, | |
| "learning_rate": 3.0008110300081106e-05, | |
| "loss": 3.0635, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 3.648701754385965, | |
| "eval_accuracy": 0.4199536857363891, | |
| "eval_bleu": 0.14234657853553717, | |
| "eval_loss": 3.1599371433258057, | |
| "eval_perplexity": 23.56911440636153, | |
| "eval_runtime": 20.1704, | |
| "eval_samples_per_second": 56.568, | |
| "eval_steps_per_second": 3.57, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 3.6767719298245614, | |
| "grad_norm": 0.7748121023178101, | |
| "learning_rate": 2.938424106307318e-05, | |
| "loss": 3.0489, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 3.704842105263158, | |
| "grad_norm": 0.7777696251869202, | |
| "learning_rate": 2.8760371826065257e-05, | |
| "loss": 3.0603, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 3.7329122807017545, | |
| "grad_norm": 0.7850595116615295, | |
| "learning_rate": 2.8136502589057333e-05, | |
| "loss": 3.0361, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 3.7609824561403506, | |
| "grad_norm": 0.7980267405509949, | |
| "learning_rate": 2.7512633352049412e-05, | |
| "loss": 3.0565, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 3.7890526315789472, | |
| "grad_norm": 0.8132106065750122, | |
| "learning_rate": 2.6888764115041488e-05, | |
| "loss": 3.0262, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 3.7890526315789472, | |
| "eval_accuracy": 0.4210717048635117, | |
| "eval_bleu": 0.14320922241168227, | |
| "eval_loss": 3.148859977722168, | |
| "eval_perplexity": 23.309476106050013, | |
| "eval_runtime": 20.2248, | |
| "eval_samples_per_second": 56.416, | |
| "eval_steps_per_second": 3.56, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 3.817122807017544, | |
| "grad_norm": 0.8002068400382996, | |
| "learning_rate": 2.6264894878033564e-05, | |
| "loss": 3.0504, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 3.8451929824561404, | |
| "grad_norm": 0.7945306301116943, | |
| "learning_rate": 2.564102564102564e-05, | |
| "loss": 3.0412, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 3.873263157894737, | |
| "grad_norm": 0.7735077738761902, | |
| "learning_rate": 2.5017156404017715e-05, | |
| "loss": 3.0193, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 3.9013333333333335, | |
| "grad_norm": 0.7823268175125122, | |
| "learning_rate": 2.4393287167009795e-05, | |
| "loss": 3.0408, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 3.9294035087719297, | |
| "grad_norm": 0.7943819165229797, | |
| "learning_rate": 2.376941793000187e-05, | |
| "loss": 3.0382, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 3.9294035087719297, | |
| "eval_accuracy": 0.4223199453755559, | |
| "eval_bleu": 0.1460635167412373, | |
| "eval_loss": 3.139704465866089, | |
| "eval_perplexity": 23.097039886291473, | |
| "eval_runtime": 20.2485, | |
| "eval_samples_per_second": 56.35, | |
| "eval_steps_per_second": 3.556, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 3.9574736842105263, | |
| "grad_norm": 0.7909451723098755, | |
| "learning_rate": 2.314554869299395e-05, | |
| "loss": 3.032, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 3.985543859649123, | |
| "grad_norm": 0.7830550670623779, | |
| "learning_rate": 2.2521679455986026e-05, | |
| "loss": 3.0383, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 4.013473684210527, | |
| "grad_norm": 0.7939472794532776, | |
| "learning_rate": 2.1897810218978105e-05, | |
| "loss": 2.9744, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 4.041543859649122, | |
| "grad_norm": 0.8031191229820251, | |
| "learning_rate": 2.127394098197018e-05, | |
| "loss": 2.9426, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 4.069614035087719, | |
| "grad_norm": 0.8071146607398987, | |
| "learning_rate": 2.0650071744962256e-05, | |
| "loss": 2.9525, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 4.069614035087719, | |
| "eval_accuracy": 0.4232880385660912, | |
| "eval_bleu": 0.1457336973656826, | |
| "eval_loss": 3.1334879398345947, | |
| "eval_perplexity": 22.95390190803781, | |
| "eval_runtime": 20.1672, | |
| "eval_samples_per_second": 56.577, | |
| "eval_steps_per_second": 3.57, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 4.097684210526316, | |
| "grad_norm": 0.8192425966262817, | |
| "learning_rate": 2.0026202507954332e-05, | |
| "loss": 2.9449, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 4.125754385964912, | |
| "grad_norm": 0.8179590702056885, | |
| "learning_rate": 1.940233327094641e-05, | |
| "loss": 2.9534, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 4.153824561403509, | |
| "grad_norm": 0.8264966607093811, | |
| "learning_rate": 1.8778464033938487e-05, | |
| "loss": 2.957, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 4.181894736842105, | |
| "grad_norm": 0.8411971926689148, | |
| "learning_rate": 1.8154594796930563e-05, | |
| "loss": 2.9503, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 4.209964912280702, | |
| "grad_norm": 0.818305253982544, | |
| "learning_rate": 1.7530725559922642e-05, | |
| "loss": 2.9621, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 4.209964912280702, | |
| "eval_accuracy": 0.4239245812568591, | |
| "eval_bleu": 0.14536464175084168, | |
| "eval_loss": 3.1270124912261963, | |
| "eval_perplexity": 22.805745303809648, | |
| "eval_runtime": 12.3884, | |
| "eval_samples_per_second": 92.102, | |
| "eval_steps_per_second": 5.812, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 4.2380350877192985, | |
| "grad_norm": 0.8167079091072083, | |
| "learning_rate": 1.6906856322914718e-05, | |
| "loss": 2.944, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 4.266105263157895, | |
| "grad_norm": 0.837181031703949, | |
| "learning_rate": 1.6282987085906794e-05, | |
| "loss": 2.9407, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 4.294175438596492, | |
| "grad_norm": 0.8356810808181763, | |
| "learning_rate": 1.5659117848898873e-05, | |
| "loss": 2.9364, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 4.322245614035087, | |
| "grad_norm": 0.8358649015426636, | |
| "learning_rate": 1.5035248611890947e-05, | |
| "loss": 2.9307, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 4.350315789473684, | |
| "grad_norm": 0.842452347278595, | |
| "learning_rate": 1.4411379374883025e-05, | |
| "loss": 2.9422, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 4.350315789473684, | |
| "eval_accuracy": 0.42498777032717266, | |
| "eval_bleu": 0.1467649753048137, | |
| "eval_loss": 3.1211211681365967, | |
| "eval_perplexity": 22.671784281214766, | |
| "eval_runtime": 20.2024, | |
| "eval_samples_per_second": 56.479, | |
| "eval_steps_per_second": 3.564, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 4.3783859649122805, | |
| "grad_norm": 0.8464500904083252, | |
| "learning_rate": 1.37875101378751e-05, | |
| "loss": 2.9435, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 4.406456140350877, | |
| "grad_norm": 0.8314975500106812, | |
| "learning_rate": 1.316364090086718e-05, | |
| "loss": 2.9405, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 4.434526315789474, | |
| "grad_norm": 0.8358827829360962, | |
| "learning_rate": 1.2539771663859257e-05, | |
| "loss": 2.9438, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 4.46259649122807, | |
| "grad_norm": 0.8394728899002075, | |
| "learning_rate": 1.1915902426851333e-05, | |
| "loss": 2.9233, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 4.490666666666667, | |
| "grad_norm": 0.839545488357544, | |
| "learning_rate": 1.1292033189843409e-05, | |
| "loss": 2.9224, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 4.490666666666667, | |
| "eval_accuracy": 0.4256748594765614, | |
| "eval_bleu": 0.1454115201862865, | |
| "eval_loss": 3.1149473190307617, | |
| "eval_perplexity": 22.53224330181115, | |
| "eval_runtime": 20.2018, | |
| "eval_samples_per_second": 56.48, | |
| "eval_steps_per_second": 3.564, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 4.518736842105263, | |
| "grad_norm": 0.847416341304779, | |
| "learning_rate": 1.0668163952835486e-05, | |
| "loss": 2.9372, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 4.54680701754386, | |
| "grad_norm": 0.8180661201477051, | |
| "learning_rate": 1.0044294715827562e-05, | |
| "loss": 2.9422, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 4.574877192982456, | |
| "grad_norm": 0.8318551182746887, | |
| "learning_rate": 9.42042547881964e-06, | |
| "loss": 2.9373, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 4.602947368421052, | |
| "grad_norm": 0.8445524573326111, | |
| "learning_rate": 8.796556241811717e-06, | |
| "loss": 2.9212, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 4.631017543859649, | |
| "grad_norm": 0.8220964670181274, | |
| "learning_rate": 8.172687004803793e-06, | |
| "loss": 2.9475, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 4.631017543859649, | |
| "eval_accuracy": 0.4264193488416722, | |
| "eval_bleu": 0.149677834207122, | |
| "eval_loss": 3.1084439754486084, | |
| "eval_perplexity": 22.38618383417341, | |
| "eval_runtime": 20.2407, | |
| "eval_samples_per_second": 56.372, | |
| "eval_steps_per_second": 3.557, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 4.659087719298245, | |
| "grad_norm": 0.8326128721237183, | |
| "learning_rate": 7.548817767795871e-06, | |
| "loss": 2.9335, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 4.687157894736842, | |
| "grad_norm": 0.822635293006897, | |
| "learning_rate": 6.924948530787947e-06, | |
| "loss": 2.9389, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 4.7152280701754385, | |
| "grad_norm": 0.8398991227149963, | |
| "learning_rate": 6.301079293780024e-06, | |
| "loss": 2.9219, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 4.743298245614035, | |
| "grad_norm": 0.8308337330818176, | |
| "learning_rate": 5.677210056772101e-06, | |
| "loss": 2.9374, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 4.771368421052632, | |
| "grad_norm": 0.8378835320472717, | |
| "learning_rate": 5.053340819764177e-06, | |
| "loss": 2.9318, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 4.771368421052632, | |
| "eval_accuracy": 0.4270327601022238, | |
| "eval_bleu": 0.14678838838741357, | |
| "eval_loss": 3.104135036468506, | |
| "eval_perplexity": 22.289930657494395, | |
| "eval_runtime": 10.9973, | |
| "eval_samples_per_second": 103.753, | |
| "eval_steps_per_second": 6.547, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 4.799438596491228, | |
| "grad_norm": 0.8561663031578064, | |
| "learning_rate": 4.429471582756255e-06, | |
| "loss": 2.9288, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 4.827508771929825, | |
| "grad_norm": 0.8296898007392883, | |
| "learning_rate": 3.805602345748331e-06, | |
| "loss": 2.9205, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 4.855578947368421, | |
| "grad_norm": 0.8513033390045166, | |
| "learning_rate": 3.1817331087404082e-06, | |
| "loss": 2.9328, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 4.883649122807018, | |
| "grad_norm": 0.8511661887168884, | |
| "learning_rate": 2.557863871732485e-06, | |
| "loss": 2.9392, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 4.911719298245614, | |
| "grad_norm": 0.8221142292022705, | |
| "learning_rate": 1.933994634724562e-06, | |
| "loss": 2.9268, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 4.911719298245614, | |
| "eval_accuracy": 0.4273857285929322, | |
| "eval_bleu": 0.1460515137764717, | |
| "eval_loss": 3.101016044616699, | |
| "eval_perplexity": 22.220516852214956, | |
| "eval_runtime": 10.9797, | |
| "eval_samples_per_second": 103.919, | |
| "eval_steps_per_second": 6.558, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 4.93978947368421, | |
| "grad_norm": 0.8456584811210632, | |
| "learning_rate": 1.3101253977166387e-06, | |
| "loss": 2.9326, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 4.967859649122807, | |
| "grad_norm": 0.8384252786636353, | |
| "learning_rate": 6.862561607087155e-07, | |
| "loss": 2.9198, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 4.995929824561403, | |
| "grad_norm": 0.8392898440361023, | |
| "learning_rate": 6.238692370079231e-08, | |
| "loss": 2.9187, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 4.998736842105263, | |
| "step": 17810, | |
| "total_flos": 1.0584067483285586e+18, | |
| "train_loss": 3.584141966119902, | |
| "train_runtime": 28526.973, | |
| "train_samples_per_second": 19.98, | |
| "train_steps_per_second": 0.624 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 17810, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "EarlyStoppingCallback": { | |
| "args": { | |
| "early_stopping_patience": 2, | |
| "early_stopping_threshold": 0.0 | |
| }, | |
| "attributes": { | |
| "early_stopping_patience_counter": 0 | |
| } | |
| }, | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0584067483285586e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |