{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 410, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012195121951219513, "grad_norm": 20.605318069458008, "learning_rate": 0.0, "loss": 1.7928, "step": 1 }, { "epoch": 0.024390243902439025, "grad_norm": 20.76787567138672, "learning_rate": 2.439024390243903e-07, "loss": 1.7386, "step": 2 }, { "epoch": 0.036585365853658534, "grad_norm": 21.81036949157715, "learning_rate": 4.878048780487805e-07, "loss": 1.8763, "step": 3 }, { "epoch": 0.04878048780487805, "grad_norm": 20.621498107910156, "learning_rate": 7.317073170731707e-07, "loss": 1.8537, "step": 4 }, { "epoch": 0.06097560975609756, "grad_norm": 22.98723793029785, "learning_rate": 9.75609756097561e-07, "loss": 1.8113, "step": 5 }, { "epoch": 0.07317073170731707, "grad_norm": 19.314804077148438, "learning_rate": 1.2195121951219514e-06, "loss": 1.7677, "step": 6 }, { "epoch": 0.08536585365853659, "grad_norm": 21.158281326293945, "learning_rate": 1.4634146341463414e-06, "loss": 1.7847, "step": 7 }, { "epoch": 0.0975609756097561, "grad_norm": 16.294034957885742, "learning_rate": 1.707317073170732e-06, "loss": 1.6678, "step": 8 }, { "epoch": 0.10975609756097561, "grad_norm": 16.788780212402344, "learning_rate": 1.951219512195122e-06, "loss": 1.6558, "step": 9 }, { "epoch": 0.12195121951219512, "grad_norm": 12.225774765014648, "learning_rate": 2.1951219512195125e-06, "loss": 1.2956, "step": 10 }, { "epoch": 0.13414634146341464, "grad_norm": 14.551143646240234, "learning_rate": 2.4390243902439027e-06, "loss": 1.5254, "step": 11 }, { "epoch": 0.14634146341463414, "grad_norm": 11.28449535369873, "learning_rate": 2.682926829268293e-06, "loss": 1.3579, "step": 12 }, { "epoch": 0.15853658536585366, "grad_norm": 7.676495552062988, "learning_rate": 2.926829268292683e-06, "loss": 1.2552, "step": 13 }, { "epoch": 0.17073170731707318, "grad_norm": 6.054831027984619, "learning_rate": 3.1707317073170736e-06, "loss": 1.0942, "step": 14 }, { "epoch": 0.18292682926829268, "grad_norm": 6.24427604675293, "learning_rate": 3.414634146341464e-06, "loss": 1.1486, "step": 15 }, { "epoch": 0.1951219512195122, "grad_norm": 5.555965900421143, "learning_rate": 3.6585365853658537e-06, "loss": 1.0225, "step": 16 }, { "epoch": 0.2073170731707317, "grad_norm": 4.953287124633789, "learning_rate": 3.902439024390244e-06, "loss": 1.0188, "step": 17 }, { "epoch": 0.21951219512195122, "grad_norm": 4.212824821472168, "learning_rate": 4.146341463414634e-06, "loss": 0.9555, "step": 18 }, { "epoch": 0.23170731707317074, "grad_norm": 4.176329135894775, "learning_rate": 4.390243902439025e-06, "loss": 0.9, "step": 19 }, { "epoch": 0.24390243902439024, "grad_norm": 4.0246734619140625, "learning_rate": 4.634146341463416e-06, "loss": 0.9001, "step": 20 }, { "epoch": 0.25609756097560976, "grad_norm": 4.022885322570801, "learning_rate": 4.8780487804878055e-06, "loss": 0.8557, "step": 21 }, { "epoch": 0.2682926829268293, "grad_norm": 3.9502739906311035, "learning_rate": 5.121951219512195e-06, "loss": 0.9313, "step": 22 }, { "epoch": 0.2804878048780488, "grad_norm": 3.4761359691619873, "learning_rate": 5.365853658536586e-06, "loss": 0.8559, "step": 23 }, { "epoch": 0.2926829268292683, "grad_norm": 3.896311044692993, "learning_rate": 5.609756097560977e-06, "loss": 0.9048, "step": 24 }, { "epoch": 0.3048780487804878, "grad_norm": 3.714123010635376, "learning_rate": 5.853658536585366e-06, "loss": 0.7699, "step": 25 }, { "epoch": 0.3170731707317073, "grad_norm": 4.503406524658203, "learning_rate": 6.0975609756097564e-06, "loss": 0.8699, "step": 26 }, { "epoch": 0.32926829268292684, "grad_norm": 3.643167734146118, "learning_rate": 6.341463414634147e-06, "loss": 0.8047, "step": 27 }, { "epoch": 0.34146341463414637, "grad_norm": 3.93937087059021, "learning_rate": 6.585365853658538e-06, "loss": 0.8064, "step": 28 }, { "epoch": 0.35365853658536583, "grad_norm": 3.669752836227417, "learning_rate": 6.829268292682928e-06, "loss": 0.7593, "step": 29 }, { "epoch": 0.36585365853658536, "grad_norm": 3.5783209800720215, "learning_rate": 7.0731707317073175e-06, "loss": 0.7464, "step": 30 }, { "epoch": 0.3780487804878049, "grad_norm": 3.4129626750946045, "learning_rate": 7.317073170731707e-06, "loss": 0.8218, "step": 31 }, { "epoch": 0.3902439024390244, "grad_norm": 3.250596761703491, "learning_rate": 7.560975609756098e-06, "loss": 0.8161, "step": 32 }, { "epoch": 0.4024390243902439, "grad_norm": 3.030006170272827, "learning_rate": 7.804878048780489e-06, "loss": 0.6851, "step": 33 }, { "epoch": 0.4146341463414634, "grad_norm": 3.556096076965332, "learning_rate": 8.048780487804879e-06, "loss": 0.8649, "step": 34 }, { "epoch": 0.4268292682926829, "grad_norm": 3.155592203140259, "learning_rate": 8.292682926829268e-06, "loss": 0.7146, "step": 35 }, { "epoch": 0.43902439024390244, "grad_norm": 2.923524856567383, "learning_rate": 8.536585365853658e-06, "loss": 0.7535, "step": 36 }, { "epoch": 0.45121951219512196, "grad_norm": 3.1197190284729004, "learning_rate": 8.78048780487805e-06, "loss": 0.7267, "step": 37 }, { "epoch": 0.4634146341463415, "grad_norm": 2.902597188949585, "learning_rate": 9.02439024390244e-06, "loss": 0.7113, "step": 38 }, { "epoch": 0.47560975609756095, "grad_norm": 3.2583975791931152, "learning_rate": 9.268292682926831e-06, "loss": 0.8452, "step": 39 }, { "epoch": 0.4878048780487805, "grad_norm": 3.5036613941192627, "learning_rate": 9.51219512195122e-06, "loss": 0.7932, "step": 40 }, { "epoch": 0.5, "grad_norm": 2.883305788040161, "learning_rate": 9.756097560975611e-06, "loss": 0.7578, "step": 41 }, { "epoch": 0.5121951219512195, "grad_norm": 2.8983325958251953, "learning_rate": 1e-05, "loss": 0.6646, "step": 42 }, { "epoch": 0.524390243902439, "grad_norm": 3.0411853790283203, "learning_rate": 9.999959340292497e-06, "loss": 0.743, "step": 43 }, { "epoch": 0.5365853658536586, "grad_norm": 3.015455484390259, "learning_rate": 9.999837361831269e-06, "loss": 0.6727, "step": 44 }, { "epoch": 0.5487804878048781, "grad_norm": 3.099972724914551, "learning_rate": 9.999634066600162e-06, "loss": 0.7748, "step": 45 }, { "epoch": 0.5609756097560976, "grad_norm": 2.834282875061035, "learning_rate": 9.999349457905545e-06, "loss": 0.6954, "step": 46 }, { "epoch": 0.573170731707317, "grad_norm": 3.012594223022461, "learning_rate": 9.998983540376262e-06, "loss": 0.8249, "step": 47 }, { "epoch": 0.5853658536585366, "grad_norm": 3.121540069580078, "learning_rate": 9.99853631996355e-06, "loss": 0.7512, "step": 48 }, { "epoch": 0.5975609756097561, "grad_norm": 2.814594030380249, "learning_rate": 9.99800780394095e-06, "loss": 0.749, "step": 49 }, { "epoch": 0.6097560975609756, "grad_norm": 2.8075897693634033, "learning_rate": 9.997398000904185e-06, "loss": 0.7249, "step": 50 }, { "epoch": 0.6219512195121951, "grad_norm": 3.2552330493927, "learning_rate": 9.996706920771024e-06, "loss": 0.7802, "step": 51 }, { "epoch": 0.6341463414634146, "grad_norm": 3.095428705215454, "learning_rate": 9.995934574781108e-06, "loss": 0.753, "step": 52 }, { "epoch": 0.6463414634146342, "grad_norm": 2.9792091846466064, "learning_rate": 9.995080975495786e-06, "loss": 0.7911, "step": 53 }, { "epoch": 0.6585365853658537, "grad_norm": 3.0372695922851562, "learning_rate": 9.994146136797893e-06, "loss": 0.7471, "step": 54 }, { "epoch": 0.6707317073170732, "grad_norm": 3.14581036567688, "learning_rate": 9.993130073891539e-06, "loss": 0.7912, "step": 55 }, { "epoch": 0.6829268292682927, "grad_norm": 2.859478235244751, "learning_rate": 9.992032803301852e-06, "loss": 0.6547, "step": 56 }, { "epoch": 0.6951219512195121, "grad_norm": 2.866575002670288, "learning_rate": 9.990854342874712e-06, "loss": 0.7098, "step": 57 }, { "epoch": 0.7073170731707317, "grad_norm": 3.036907434463501, "learning_rate": 9.98959471177646e-06, "loss": 0.8274, "step": 58 }, { "epoch": 0.7195121951219512, "grad_norm": 2.837873935699463, "learning_rate": 9.988253930493592e-06, "loss": 0.7151, "step": 59 }, { "epoch": 0.7317073170731707, "grad_norm": 2.6678829193115234, "learning_rate": 9.986832020832416e-06, "loss": 0.6541, "step": 60 }, { "epoch": 0.7439024390243902, "grad_norm": 2.9930105209350586, "learning_rate": 9.985329005918702e-06, "loss": 0.6892, "step": 61 }, { "epoch": 0.7560975609756098, "grad_norm": 2.858548164367676, "learning_rate": 9.983744910197315e-06, "loss": 0.6988, "step": 62 }, { "epoch": 0.7682926829268293, "grad_norm": 3.0590319633483887, "learning_rate": 9.982079759431797e-06, "loss": 0.6853, "step": 63 }, { "epoch": 0.7804878048780488, "grad_norm": 2.8750498294830322, "learning_rate": 9.980333580703968e-06, "loss": 0.7181, "step": 64 }, { "epoch": 0.7926829268292683, "grad_norm": 2.720283031463623, "learning_rate": 9.978506402413472e-06, "loss": 0.6339, "step": 65 }, { "epoch": 0.8048780487804879, "grad_norm": 2.936540126800537, "learning_rate": 9.976598254277324e-06, "loss": 0.7085, "step": 66 }, { "epoch": 0.8170731707317073, "grad_norm": 2.7820205688476562, "learning_rate": 9.974609167329425e-06, "loss": 0.6682, "step": 67 }, { "epoch": 0.8292682926829268, "grad_norm": 2.852302312850952, "learning_rate": 9.972539173920048e-06, "loss": 0.7067, "step": 68 }, { "epoch": 0.8414634146341463, "grad_norm": 2.763120651245117, "learning_rate": 9.970388307715326e-06, "loss": 0.6512, "step": 69 }, { "epoch": 0.8536585365853658, "grad_norm": 2.834955930709839, "learning_rate": 9.968156603696696e-06, "loss": 0.692, "step": 70 }, { "epoch": 0.8658536585365854, "grad_norm": 2.5952882766723633, "learning_rate": 9.965844098160326e-06, "loss": 0.6458, "step": 71 }, { "epoch": 0.8780487804878049, "grad_norm": 2.793827533721924, "learning_rate": 9.963450828716543e-06, "loss": 0.7312, "step": 72 }, { "epoch": 0.8902439024390244, "grad_norm": 2.7760300636291504, "learning_rate": 9.960976834289197e-06, "loss": 0.6733, "step": 73 }, { "epoch": 0.9024390243902439, "grad_norm": 3.0652453899383545, "learning_rate": 9.958422155115044e-06, "loss": 0.7255, "step": 74 }, { "epoch": 0.9146341463414634, "grad_norm": 2.7409512996673584, "learning_rate": 9.955786832743089e-06, "loss": 0.7146, "step": 75 }, { "epoch": 0.926829268292683, "grad_norm": 2.671405553817749, "learning_rate": 9.953070910033904e-06, "loss": 0.7051, "step": 76 }, { "epoch": 0.9390243902439024, "grad_norm": 3.065516233444214, "learning_rate": 9.95027443115894e-06, "loss": 0.7027, "step": 77 }, { "epoch": 0.9512195121951219, "grad_norm": 2.724518060684204, "learning_rate": 9.947397441599801e-06, "loss": 0.7046, "step": 78 }, { "epoch": 0.9634146341463414, "grad_norm": 2.762394428253174, "learning_rate": 9.944439988147509e-06, "loss": 0.6638, "step": 79 }, { "epoch": 0.975609756097561, "grad_norm": 2.7874350547790527, "learning_rate": 9.941402118901743e-06, "loss": 0.6985, "step": 80 }, { "epoch": 0.9878048780487805, "grad_norm": 2.785700798034668, "learning_rate": 9.938283883270051e-06, "loss": 0.6443, "step": 81 }, { "epoch": 1.0, "grad_norm": 2.859963893890381, "learning_rate": 9.935085331967054e-06, "loss": 0.6987, "step": 82 }, { "epoch": 1.0121951219512195, "grad_norm": 2.341641902923584, "learning_rate": 9.931806517013612e-06, "loss": 0.4309, "step": 83 }, { "epoch": 1.024390243902439, "grad_norm": 2.2350566387176514, "learning_rate": 9.928447491735994e-06, "loss": 0.3769, "step": 84 }, { "epoch": 1.0365853658536586, "grad_norm": 2.750514030456543, "learning_rate": 9.925008310764988e-06, "loss": 0.5076, "step": 85 }, { "epoch": 1.048780487804878, "grad_norm": 2.627335548400879, "learning_rate": 9.921489030035036e-06, "loss": 0.359, "step": 86 }, { "epoch": 1.0609756097560976, "grad_norm": 2.739978075027466, "learning_rate": 9.917889706783304e-06, "loss": 0.4735, "step": 87 }, { "epoch": 1.0731707317073171, "grad_norm": 3.0831549167633057, "learning_rate": 9.914210399548768e-06, "loss": 0.5604, "step": 88 }, { "epoch": 1.0853658536585367, "grad_norm": 3.0366146564483643, "learning_rate": 9.910451168171248e-06, "loss": 0.3986, "step": 89 }, { "epoch": 1.0975609756097562, "grad_norm": 2.8682730197906494, "learning_rate": 9.906612073790443e-06, "loss": 0.4118, "step": 90 }, { "epoch": 1.1097560975609757, "grad_norm": 2.9994473457336426, "learning_rate": 9.902693178844937e-06, "loss": 0.4581, "step": 91 }, { "epoch": 1.1219512195121952, "grad_norm": 3.4703030586242676, "learning_rate": 9.898694547071177e-06, "loss": 0.5222, "step": 92 }, { "epoch": 1.1341463414634148, "grad_norm": 2.6934309005737305, "learning_rate": 9.894616243502442e-06, "loss": 0.3656, "step": 93 }, { "epoch": 1.146341463414634, "grad_norm": 2.379758834838867, "learning_rate": 9.890458334467784e-06, "loss": 0.3277, "step": 94 }, { "epoch": 1.1585365853658536, "grad_norm": 2.7950727939605713, "learning_rate": 9.886220887590953e-06, "loss": 0.4012, "step": 95 }, { "epoch": 1.170731707317073, "grad_norm": 2.668951988220215, "learning_rate": 9.881903971789285e-06, "loss": 0.4384, "step": 96 }, { "epoch": 1.1829268292682926, "grad_norm": 2.785778522491455, "learning_rate": 9.877507657272596e-06, "loss": 0.4727, "step": 97 }, { "epoch": 1.1951219512195121, "grad_norm": 2.7798571586608887, "learning_rate": 9.873032015542027e-06, "loss": 0.4594, "step": 98 }, { "epoch": 1.2073170731707317, "grad_norm": 2.9862515926361084, "learning_rate": 9.868477119388897e-06, "loss": 0.4715, "step": 99 }, { "epoch": 1.2195121951219512, "grad_norm": 2.749171495437622, "learning_rate": 9.863843042893499e-06, "loss": 0.4125, "step": 100 }, { "epoch": 1.2317073170731707, "grad_norm": 2.4786319732666016, "learning_rate": 9.859129861423915e-06, "loss": 0.4036, "step": 101 }, { "epoch": 1.2439024390243902, "grad_norm": 2.724829912185669, "learning_rate": 9.854337651634773e-06, "loss": 0.4688, "step": 102 }, { "epoch": 1.2560975609756098, "grad_norm": 2.5419397354125977, "learning_rate": 9.849466491466017e-06, "loss": 0.4276, "step": 103 }, { "epoch": 1.2682926829268293, "grad_norm": 2.508129596710205, "learning_rate": 9.844516460141622e-06, "loss": 0.401, "step": 104 }, { "epoch": 1.2804878048780488, "grad_norm": 2.677839756011963, "learning_rate": 9.839487638168321e-06, "loss": 0.3839, "step": 105 }, { "epoch": 1.2926829268292683, "grad_norm": 2.811065912246704, "learning_rate": 9.834380107334284e-06, "loss": 0.3876, "step": 106 }, { "epoch": 1.3048780487804879, "grad_norm": 2.7741312980651855, "learning_rate": 9.829193950707798e-06, "loss": 0.4019, "step": 107 }, { "epoch": 1.3170731707317074, "grad_norm": 2.604609727859497, "learning_rate": 9.823929252635905e-06, "loss": 0.3753, "step": 108 }, { "epoch": 1.329268292682927, "grad_norm": 3.5267436504364014, "learning_rate": 9.818586098743038e-06, "loss": 0.5063, "step": 109 }, { "epoch": 1.3414634146341464, "grad_norm": 2.785386085510254, "learning_rate": 9.813164575929628e-06, "loss": 0.4035, "step": 110 }, { "epoch": 1.3536585365853657, "grad_norm": 2.7874209880828857, "learning_rate": 9.807664772370689e-06, "loss": 0.4387, "step": 111 }, { "epoch": 1.3658536585365852, "grad_norm": 2.616459369659424, "learning_rate": 9.80208677751438e-06, "loss": 0.4403, "step": 112 }, { "epoch": 1.3780487804878048, "grad_norm": 2.593151092529297, "learning_rate": 9.79643068208056e-06, "loss": 0.418, "step": 113 }, { "epoch": 1.3902439024390243, "grad_norm": 2.3522331714630127, "learning_rate": 9.7906965780593e-06, "loss": 0.3226, "step": 114 }, { "epoch": 1.4024390243902438, "grad_norm": 2.945878028869629, "learning_rate": 9.784884558709398e-06, "loss": 0.4744, "step": 115 }, { "epoch": 1.4146341463414633, "grad_norm": 2.6254990100860596, "learning_rate": 9.778994718556856e-06, "loss": 0.3656, "step": 116 }, { "epoch": 1.4268292682926829, "grad_norm": 2.6019349098205566, "learning_rate": 9.773027153393349e-06, "loss": 0.3957, "step": 117 }, { "epoch": 1.4390243902439024, "grad_norm": 2.8025217056274414, "learning_rate": 9.766981960274653e-06, "loss": 0.4242, "step": 118 }, { "epoch": 1.451219512195122, "grad_norm": 2.747736930847168, "learning_rate": 9.760859237519087e-06, "loss": 0.4247, "step": 119 }, { "epoch": 1.4634146341463414, "grad_norm": 2.8022918701171875, "learning_rate": 9.754659084705893e-06, "loss": 0.3984, "step": 120 }, { "epoch": 1.475609756097561, "grad_norm": 2.5835225582122803, "learning_rate": 9.748381602673633e-06, "loss": 0.4205, "step": 121 }, { "epoch": 1.4878048780487805, "grad_norm": 2.7356934547424316, "learning_rate": 9.742026893518541e-06, "loss": 0.4098, "step": 122 }, { "epoch": 1.5, "grad_norm": 2.6171412467956543, "learning_rate": 9.735595060592861e-06, "loss": 0.4281, "step": 123 }, { "epoch": 1.5121951219512195, "grad_norm": 2.646216630935669, "learning_rate": 9.729086208503174e-06, "loss": 0.4301, "step": 124 }, { "epoch": 1.524390243902439, "grad_norm": 3.031221866607666, "learning_rate": 9.722500443108687e-06, "loss": 0.5132, "step": 125 }, { "epoch": 1.5365853658536586, "grad_norm": 2.813753843307495, "learning_rate": 9.715837871519518e-06, "loss": 0.464, "step": 126 }, { "epoch": 1.548780487804878, "grad_norm": 2.7644271850585938, "learning_rate": 9.709098602094952e-06, "loss": 0.4589, "step": 127 }, { "epoch": 1.5609756097560976, "grad_norm": 2.8581771850585938, "learning_rate": 9.70228274444168e-06, "loss": 0.4659, "step": 128 }, { "epoch": 1.5731707317073171, "grad_norm": 2.6003692150115967, "learning_rate": 9.695390409412011e-06, "loss": 0.3784, "step": 129 }, { "epoch": 1.5853658536585367, "grad_norm": 2.455249547958374, "learning_rate": 9.688421709102076e-06, "loss": 0.4141, "step": 130 }, { "epoch": 1.5975609756097562, "grad_norm": 2.439664363861084, "learning_rate": 9.681376756850003e-06, "loss": 0.3995, "step": 131 }, { "epoch": 1.6097560975609757, "grad_norm": 2.6555984020233154, "learning_rate": 9.67425566723407e-06, "loss": 0.4611, "step": 132 }, { "epoch": 1.6219512195121952, "grad_norm": 2.4294567108154297, "learning_rate": 9.667058556070846e-06, "loss": 0.4316, "step": 133 }, { "epoch": 1.6341463414634148, "grad_norm": 2.5822300910949707, "learning_rate": 9.659785540413303e-06, "loss": 0.4274, "step": 134 }, { "epoch": 1.6463414634146343, "grad_norm": 2.7250919342041016, "learning_rate": 9.652436738548917e-06, "loss": 0.4271, "step": 135 }, { "epoch": 1.6585365853658538, "grad_norm": 2.6819536685943604, "learning_rate": 9.645012269997747e-06, "loss": 0.4141, "step": 136 }, { "epoch": 1.6707317073170733, "grad_norm": 2.830106496810913, "learning_rate": 9.637512255510475e-06, "loss": 0.466, "step": 137 }, { "epoch": 1.6829268292682928, "grad_norm": 2.6315557956695557, "learning_rate": 9.629936817066459e-06, "loss": 0.4085, "step": 138 }, { "epoch": 1.6951219512195121, "grad_norm": 2.916368246078491, "learning_rate": 9.622286077871748e-06, "loss": 0.4728, "step": 139 }, { "epoch": 1.7073170731707317, "grad_norm": 3.0268235206604004, "learning_rate": 9.614560162357065e-06, "loss": 0.4548, "step": 140 }, { "epoch": 1.7195121951219512, "grad_norm": 2.8294835090637207, "learning_rate": 9.606759196175799e-06, "loss": 0.4145, "step": 141 }, { "epoch": 1.7317073170731707, "grad_norm": 2.861173391342163, "learning_rate": 9.598883306201949e-06, "loss": 0.4283, "step": 142 }, { "epoch": 1.7439024390243902, "grad_norm": 2.8794517517089844, "learning_rate": 9.590932620528068e-06, "loss": 0.5036, "step": 143 }, { "epoch": 1.7560975609756098, "grad_norm": 2.633896589279175, "learning_rate": 9.58290726846318e-06, "loss": 0.4355, "step": 144 }, { "epoch": 1.7682926829268293, "grad_norm": 2.5964772701263428, "learning_rate": 9.57480738053067e-06, "loss": 0.443, "step": 145 }, { "epoch": 1.7804878048780488, "grad_norm": 2.5255353450775146, "learning_rate": 9.566633088466169e-06, "loss": 0.4135, "step": 146 }, { "epoch": 1.7926829268292683, "grad_norm": 2.3389077186584473, "learning_rate": 9.558384525215406e-06, "loss": 0.4233, "step": 147 }, { "epoch": 1.8048780487804879, "grad_norm": 2.570801019668579, "learning_rate": 9.550061824932047e-06, "loss": 0.4227, "step": 148 }, { "epoch": 1.8170731707317072, "grad_norm": 2.7482798099517822, "learning_rate": 9.54166512297552e-06, "loss": 0.4779, "step": 149 }, { "epoch": 1.8292682926829267, "grad_norm": 3.0880026817321777, "learning_rate": 9.533194555908796e-06, "loss": 0.5231, "step": 150 }, { "epoch": 1.8414634146341462, "grad_norm": 2.6744909286499023, "learning_rate": 9.524650261496195e-06, "loss": 0.4608, "step": 151 }, { "epoch": 1.8536585365853657, "grad_norm": 2.891713857650757, "learning_rate": 9.516032378701117e-06, "loss": 0.473, "step": 152 }, { "epoch": 1.8658536585365852, "grad_norm": 2.547239303588867, "learning_rate": 9.5073410476838e-06, "loss": 0.4051, "step": 153 }, { "epoch": 1.8780487804878048, "grad_norm": 2.723076581954956, "learning_rate": 9.498576409799034e-06, "loss": 0.4558, "step": 154 }, { "epoch": 1.8902439024390243, "grad_norm": 3.1596052646636963, "learning_rate": 9.489738607593867e-06, "loss": 0.4865, "step": 155 }, { "epoch": 1.9024390243902438, "grad_norm": 2.7183949947357178, "learning_rate": 9.480827784805278e-06, "loss": 0.5014, "step": 156 }, { "epoch": 1.9146341463414633, "grad_norm": 2.5864574909210205, "learning_rate": 9.471844086357848e-06, "loss": 0.4149, "step": 157 }, { "epoch": 1.9268292682926829, "grad_norm": 2.5046157836914062, "learning_rate": 9.462787658361394e-06, "loss": 0.3962, "step": 158 }, { "epoch": 1.9390243902439024, "grad_norm": 2.8331422805786133, "learning_rate": 9.453658648108604e-06, "loss": 0.3853, "step": 159 }, { "epoch": 1.951219512195122, "grad_norm": 2.512298822402954, "learning_rate": 9.444457204072632e-06, "loss": 0.4437, "step": 160 }, { "epoch": 1.9634146341463414, "grad_norm": 2.444852828979492, "learning_rate": 9.435183475904688e-06, "loss": 0.3504, "step": 161 }, { "epoch": 1.975609756097561, "grad_norm": 2.8331000804901123, "learning_rate": 9.425837614431601e-06, "loss": 0.4716, "step": 162 }, { "epoch": 1.9878048780487805, "grad_norm": 2.661059856414795, "learning_rate": 9.416419771653368e-06, "loss": 0.4385, "step": 163 }, { "epoch": 2.0, "grad_norm": 2.646305799484253, "learning_rate": 9.406930100740686e-06, "loss": 0.4184, "step": 164 }, { "epoch": 2.0121951219512195, "grad_norm": 2.712597608566284, "learning_rate": 9.397368756032445e-06, "loss": 0.2314, "step": 165 }, { "epoch": 2.024390243902439, "grad_norm": 2.586576461791992, "learning_rate": 9.387735893033244e-06, "loss": 0.1831, "step": 166 }, { "epoch": 2.0365853658536586, "grad_norm": 2.5278258323669434, "learning_rate": 9.378031668410836e-06, "loss": 0.2375, "step": 167 }, { "epoch": 2.048780487804878, "grad_norm": 2.541187047958374, "learning_rate": 9.368256239993597e-06, "loss": 0.1981, "step": 168 }, { "epoch": 2.0609756097560976, "grad_norm": 2.764477252960205, "learning_rate": 9.358409766767946e-06, "loss": 0.2029, "step": 169 }, { "epoch": 2.073170731707317, "grad_norm": 2.4784131050109863, "learning_rate": 9.348492408875779e-06, "loss": 0.1535, "step": 170 }, { "epoch": 2.0853658536585367, "grad_norm": 2.915125846862793, "learning_rate": 9.338504327611839e-06, "loss": 0.1598, "step": 171 }, { "epoch": 2.097560975609756, "grad_norm": 2.7254488468170166, "learning_rate": 9.328445685421113e-06, "loss": 0.1462, "step": 172 }, { "epoch": 2.1097560975609757, "grad_norm": 2.9409985542297363, "learning_rate": 9.318316645896182e-06, "loss": 0.203, "step": 173 }, { "epoch": 2.1219512195121952, "grad_norm": 2.588385820388794, "learning_rate": 9.308117373774555e-06, "loss": 0.1795, "step": 174 }, { "epoch": 2.1341463414634148, "grad_norm": 2.7931816577911377, "learning_rate": 9.297848034936007e-06, "loss": 0.1993, "step": 175 }, { "epoch": 2.1463414634146343, "grad_norm": 2.3102173805236816, "learning_rate": 9.287508796399858e-06, "loss": 0.1839, "step": 176 }, { "epoch": 2.158536585365854, "grad_norm": 2.3756439685821533, "learning_rate": 9.277099826322277e-06, "loss": 0.2063, "step": 177 }, { "epoch": 2.1707317073170733, "grad_norm": 2.2752017974853516, "learning_rate": 9.266621293993534e-06, "loss": 0.1699, "step": 178 }, { "epoch": 2.182926829268293, "grad_norm": 2.484127998352051, "learning_rate": 9.256073369835255e-06, "loss": 0.1763, "step": 179 }, { "epoch": 2.1951219512195124, "grad_norm": 2.3598098754882812, "learning_rate": 9.245456225397642e-06, "loss": 0.1677, "step": 180 }, { "epoch": 2.207317073170732, "grad_norm": 2.2330524921417236, "learning_rate": 9.23477003335669e-06, "loss": 0.185, "step": 181 }, { "epoch": 2.2195121951219514, "grad_norm": 2.439162492752075, "learning_rate": 9.224014967511378e-06, "loss": 0.1582, "step": 182 }, { "epoch": 2.231707317073171, "grad_norm": 2.601541042327881, "learning_rate": 9.213191202780835e-06, "loss": 0.1737, "step": 183 }, { "epoch": 2.2439024390243905, "grad_norm": 2.3318488597869873, "learning_rate": 9.20229891520151e-06, "loss": 0.1688, "step": 184 }, { "epoch": 2.2560975609756095, "grad_norm": 2.883798122406006, "learning_rate": 9.191338281924288e-06, "loss": 0.2094, "step": 185 }, { "epoch": 2.2682926829268295, "grad_norm": 2.4024503231048584, "learning_rate": 9.180309481211629e-06, "loss": 0.183, "step": 186 }, { "epoch": 2.2804878048780486, "grad_norm": 2.7932958602905273, "learning_rate": 9.169212692434658e-06, "loss": 0.2388, "step": 187 }, { "epoch": 2.292682926829268, "grad_norm": 2.345780372619629, "learning_rate": 9.158048096070249e-06, "loss": 0.1698, "step": 188 }, { "epoch": 2.3048780487804876, "grad_norm": 2.3633759021759033, "learning_rate": 9.14681587369809e-06, "loss": 0.1797, "step": 189 }, { "epoch": 2.317073170731707, "grad_norm": 2.4073266983032227, "learning_rate": 9.13551620799773e-06, "loss": 0.1744, "step": 190 }, { "epoch": 2.3292682926829267, "grad_norm": 2.4266092777252197, "learning_rate": 9.124149282745614e-06, "loss": 0.1874, "step": 191 }, { "epoch": 2.341463414634146, "grad_norm": 2.277799129486084, "learning_rate": 9.112715282812081e-06, "loss": 0.2014, "step": 192 }, { "epoch": 2.3536585365853657, "grad_norm": 2.5907177925109863, "learning_rate": 9.101214394158371e-06, "loss": 0.1911, "step": 193 }, { "epoch": 2.3658536585365852, "grad_norm": 2.6057519912719727, "learning_rate": 9.089646803833589e-06, "loss": 0.2172, "step": 194 }, { "epoch": 2.3780487804878048, "grad_norm": 2.3195533752441406, "learning_rate": 9.078012699971673e-06, "loss": 0.184, "step": 195 }, { "epoch": 2.3902439024390243, "grad_norm": 2.62652850151062, "learning_rate": 9.066312271788323e-06, "loss": 0.2185, "step": 196 }, { "epoch": 2.402439024390244, "grad_norm": 2.2538259029388428, "learning_rate": 9.054545709577939e-06, "loss": 0.1797, "step": 197 }, { "epoch": 2.4146341463414633, "grad_norm": 2.573920965194702, "learning_rate": 9.042713204710509e-06, "loss": 0.1791, "step": 198 }, { "epoch": 2.426829268292683, "grad_norm": 2.010896921157837, "learning_rate": 9.030814949628509e-06, "loss": 0.1471, "step": 199 }, { "epoch": 2.4390243902439024, "grad_norm": 2.5009727478027344, "learning_rate": 9.018851137843765e-06, "loss": 0.1805, "step": 200 }, { "epoch": 2.451219512195122, "grad_norm": 2.673194169998169, "learning_rate": 9.006821963934316e-06, "loss": 0.2134, "step": 201 }, { "epoch": 2.4634146341463414, "grad_norm": 2.851163387298584, "learning_rate": 8.994727623541237e-06, "loss": 0.1902, "step": 202 }, { "epoch": 2.475609756097561, "grad_norm": 3.064375877380371, "learning_rate": 8.982568313365467e-06, "loss": 0.2247, "step": 203 }, { "epoch": 2.4878048780487805, "grad_norm": 2.5090184211730957, "learning_rate": 8.970344231164602e-06, "loss": 0.2022, "step": 204 }, { "epoch": 2.5, "grad_norm": 2.2384963035583496, "learning_rate": 8.958055575749685e-06, "loss": 0.1954, "step": 205 }, { "epoch": 2.5121951219512195, "grad_norm": 2.3855085372924805, "learning_rate": 8.94570254698197e-06, "loss": 0.2088, "step": 206 }, { "epoch": 2.524390243902439, "grad_norm": 2.38485050201416, "learning_rate": 8.933285345769671e-06, "loss": 0.1926, "step": 207 }, { "epoch": 2.5365853658536586, "grad_norm": 2.5828115940093994, "learning_rate": 8.920804174064697e-06, "loss": 0.2452, "step": 208 }, { "epoch": 2.548780487804878, "grad_norm": 2.271554470062256, "learning_rate": 8.908259234859365e-06, "loss": 0.1858, "step": 209 }, { "epoch": 2.5609756097560976, "grad_norm": 2.114044189453125, "learning_rate": 8.895650732183094e-06, "loss": 0.1766, "step": 210 }, { "epoch": 2.573170731707317, "grad_norm": 2.3854148387908936, "learning_rate": 8.882978871099104e-06, "loss": 0.2026, "step": 211 }, { "epoch": 2.5853658536585367, "grad_norm": 2.409749746322632, "learning_rate": 8.870243857701054e-06, "loss": 0.2135, "step": 212 }, { "epoch": 2.597560975609756, "grad_norm": 2.269014596939087, "learning_rate": 8.857445899109716e-06, "loss": 0.173, "step": 213 }, { "epoch": 2.6097560975609757, "grad_norm": 2.1958768367767334, "learning_rate": 8.84458520346959e-06, "loss": 0.1803, "step": 214 }, { "epoch": 2.6219512195121952, "grad_norm": 2.2031567096710205, "learning_rate": 8.831661979945522e-06, "loss": 0.1701, "step": 215 }, { "epoch": 2.6341463414634148, "grad_norm": 2.523292303085327, "learning_rate": 8.818676438719314e-06, "loss": 0.1988, "step": 216 }, { "epoch": 2.6463414634146343, "grad_norm": 2.597362995147705, "learning_rate": 8.805628790986284e-06, "loss": 0.2264, "step": 217 }, { "epoch": 2.658536585365854, "grad_norm": 2.802621603012085, "learning_rate": 8.792519248951851e-06, "loss": 0.2293, "step": 218 }, { "epoch": 2.6707317073170733, "grad_norm": 2.707906484603882, "learning_rate": 8.779348025828071e-06, "loss": 0.2012, "step": 219 }, { "epoch": 2.682926829268293, "grad_norm": 2.630911350250244, "learning_rate": 8.766115335830178e-06, "loss": 0.1975, "step": 220 }, { "epoch": 2.6951219512195124, "grad_norm": 2.492384195327759, "learning_rate": 8.752821394173092e-06, "loss": 0.1893, "step": 221 }, { "epoch": 2.7073170731707314, "grad_norm": 2.3401095867156982, "learning_rate": 8.739466417067926e-06, "loss": 0.1769, "step": 222 }, { "epoch": 2.7195121951219514, "grad_norm": 2.6099853515625, "learning_rate": 8.726050621718462e-06, "loss": 0.1746, "step": 223 }, { "epoch": 2.7317073170731705, "grad_norm": 2.4008710384368896, "learning_rate": 8.71257422631763e-06, "loss": 0.222, "step": 224 }, { "epoch": 2.7439024390243905, "grad_norm": 2.5295095443725586, "learning_rate": 8.699037450043945e-06, "loss": 0.2196, "step": 225 }, { "epoch": 2.7560975609756095, "grad_norm": 2.4341542720794678, "learning_rate": 8.685440513057955e-06, "loss": 0.2019, "step": 226 }, { "epoch": 2.7682926829268295, "grad_norm": 2.379326343536377, "learning_rate": 8.671783636498652e-06, "loss": 0.2263, "step": 227 }, { "epoch": 2.7804878048780486, "grad_norm": 2.4653515815734863, "learning_rate": 8.658067042479877e-06, "loss": 0.197, "step": 228 }, { "epoch": 2.7926829268292686, "grad_norm": 2.4599173069000244, "learning_rate": 8.644290954086711e-06, "loss": 0.1995, "step": 229 }, { "epoch": 2.8048780487804876, "grad_norm": 2.559979200363159, "learning_rate": 8.630455595371846e-06, "loss": 0.2138, "step": 230 }, { "epoch": 2.817073170731707, "grad_norm": 2.173933267593384, "learning_rate": 8.616561191351934e-06, "loss": 0.1822, "step": 231 }, { "epoch": 2.8292682926829267, "grad_norm": 2.4872312545776367, "learning_rate": 8.602607968003935e-06, "loss": 0.1805, "step": 232 }, { "epoch": 2.841463414634146, "grad_norm": 2.255208730697632, "learning_rate": 8.588596152261447e-06, "loss": 0.1825, "step": 233 }, { "epoch": 2.8536585365853657, "grad_norm": 2.602861166000366, "learning_rate": 8.574525972010997e-06, "loss": 0.2079, "step": 234 }, { "epoch": 2.8658536585365852, "grad_norm": 2.6664066314697266, "learning_rate": 8.560397656088353e-06, "loss": 0.1909, "step": 235 }, { "epoch": 2.8780487804878048, "grad_norm": 3.140064001083374, "learning_rate": 8.546211434274791e-06, "loss": 0.1985, "step": 236 }, { "epoch": 2.8902439024390243, "grad_norm": 2.759251832962036, "learning_rate": 8.531967537293365e-06, "loss": 0.1862, "step": 237 }, { "epoch": 2.902439024390244, "grad_norm": 2.6289727687835693, "learning_rate": 8.517666196805142e-06, "loss": 0.207, "step": 238 }, { "epoch": 2.9146341463414633, "grad_norm": 2.671435594558716, "learning_rate": 8.503307645405461e-06, "loss": 0.2103, "step": 239 }, { "epoch": 2.926829268292683, "grad_norm": 2.4491989612579346, "learning_rate": 8.488892116620114e-06, "loss": 0.2086, "step": 240 }, { "epoch": 2.9390243902439024, "grad_norm": 2.1627562046051025, "learning_rate": 8.474419844901575e-06, "loss": 0.1785, "step": 241 }, { "epoch": 2.951219512195122, "grad_norm": 2.683394432067871, "learning_rate": 8.459891065625184e-06, "loss": 0.2746, "step": 242 }, { "epoch": 2.9634146341463414, "grad_norm": 2.3977043628692627, "learning_rate": 8.445306015085301e-06, "loss": 0.2042, "step": 243 }, { "epoch": 2.975609756097561, "grad_norm": 2.0520613193511963, "learning_rate": 8.430664930491485e-06, "loss": 0.1897, "step": 244 }, { "epoch": 2.9878048780487805, "grad_norm": 2.36509370803833, "learning_rate": 8.415968049964623e-06, "loss": 0.1859, "step": 245 }, { "epoch": 3.0, "grad_norm": 2.167886257171631, "learning_rate": 8.401215612533056e-06, "loss": 0.1665, "step": 246 }, { "epoch": 3.0121951219512195, "grad_norm": 1.8608198165893555, "learning_rate": 8.386407858128707e-06, "loss": 0.1037, "step": 247 }, { "epoch": 3.024390243902439, "grad_norm": 1.8207582235336304, "learning_rate": 8.371545027583154e-06, "loss": 0.0807, "step": 248 }, { "epoch": 3.0365853658536586, "grad_norm": 1.7909525632858276, "learning_rate": 8.356627362623742e-06, "loss": 0.0819, "step": 249 }, { "epoch": 3.048780487804878, "grad_norm": 2.130682945251465, "learning_rate": 8.341655105869622e-06, "loss": 0.1154, "step": 250 }, { "epoch": 3.0609756097560976, "grad_norm": 1.9704978466033936, "learning_rate": 8.326628500827826e-06, "loss": 0.0959, "step": 251 }, { "epoch": 3.073170731707317, "grad_norm": 2.402252197265625, "learning_rate": 8.311547791889307e-06, "loss": 0.1006, "step": 252 }, { "epoch": 3.0853658536585367, "grad_norm": 2.2904582023620605, "learning_rate": 8.296413224324944e-06, "loss": 0.0985, "step": 253 }, { "epoch": 3.097560975609756, "grad_norm": 2.511240005493164, "learning_rate": 8.281225044281578e-06, "loss": 0.0695, "step": 254 }, { "epoch": 3.1097560975609757, "grad_norm": 2.37315034866333, "learning_rate": 8.265983498777987e-06, "loss": 0.086, "step": 255 }, { "epoch": 3.1219512195121952, "grad_norm": 2.4025444984436035, "learning_rate": 8.25068883570089e-06, "loss": 0.0877, "step": 256 }, { "epoch": 3.1341463414634148, "grad_norm": 2.855544328689575, "learning_rate": 8.235341303800892e-06, "loss": 0.1104, "step": 257 }, { "epoch": 3.1463414634146343, "grad_norm": 2.7334654331207275, "learning_rate": 8.219941152688459e-06, "loss": 0.0996, "step": 258 }, { "epoch": 3.158536585365854, "grad_norm": 1.7848544120788574, "learning_rate": 8.204488632829848e-06, "loss": 0.0779, "step": 259 }, { "epoch": 3.1707317073170733, "grad_norm": 2.5994298458099365, "learning_rate": 8.188983995543031e-06, "loss": 0.1027, "step": 260 }, { "epoch": 3.182926829268293, "grad_norm": 2.1597657203674316, "learning_rate": 8.173427492993617e-06, "loss": 0.0974, "step": 261 }, { "epoch": 3.1951219512195124, "grad_norm": 2.6595215797424316, "learning_rate": 8.157819378190743e-06, "loss": 0.1053, "step": 262 }, { "epoch": 3.207317073170732, "grad_norm": 1.92975652217865, "learning_rate": 8.142159904982963e-06, "loss": 0.1003, "step": 263 }, { "epoch": 3.2195121951219514, "grad_norm": 1.939504861831665, "learning_rate": 8.126449328054115e-06, "loss": 0.0948, "step": 264 }, { "epoch": 3.231707317073171, "grad_norm": 2.238565444946289, "learning_rate": 8.110687902919185e-06, "loss": 0.1021, "step": 265 }, { "epoch": 3.2439024390243905, "grad_norm": 2.1030704975128174, "learning_rate": 8.094875885920148e-06, "loss": 0.0961, "step": 266 }, { "epoch": 3.2560975609756095, "grad_norm": 2.0035948753356934, "learning_rate": 8.079013534221798e-06, "loss": 0.0985, "step": 267 }, { "epoch": 3.2682926829268295, "grad_norm": 2.1001100540161133, "learning_rate": 8.063101105807566e-06, "loss": 0.1089, "step": 268 }, { "epoch": 3.2804878048780486, "grad_norm": 1.935497760772705, "learning_rate": 8.047138859475328e-06, "loss": 0.0882, "step": 269 }, { "epoch": 3.292682926829268, "grad_norm": 2.4864578247070312, "learning_rate": 8.031127054833192e-06, "loss": 0.1085, "step": 270 }, { "epoch": 3.3048780487804876, "grad_norm": 1.89180326461792, "learning_rate": 8.01506595229527e-06, "loss": 0.1096, "step": 271 }, { "epoch": 3.317073170731707, "grad_norm": 2.166079521179199, "learning_rate": 7.998955813077457e-06, "loss": 0.0717, "step": 272 }, { "epoch": 3.3292682926829267, "grad_norm": 2.1305079460144043, "learning_rate": 7.982796899193177e-06, "loss": 0.1042, "step": 273 }, { "epoch": 3.341463414634146, "grad_norm": 2.0318334102630615, "learning_rate": 7.966589473449109e-06, "loss": 0.0943, "step": 274 }, { "epoch": 3.3536585365853657, "grad_norm": 2.6421074867248535, "learning_rate": 7.95033379944093e-06, "loss": 0.1161, "step": 275 }, { "epoch": 3.3658536585365852, "grad_norm": 2.3139538764953613, "learning_rate": 7.934030141549024e-06, "loss": 0.1219, "step": 276 }, { "epoch": 3.3780487804878048, "grad_norm": 2.0743587017059326, "learning_rate": 7.917678764934169e-06, "loss": 0.1024, "step": 277 }, { "epoch": 3.3902439024390243, "grad_norm": 2.187187671661377, "learning_rate": 7.901279935533248e-06, "loss": 0.0864, "step": 278 }, { "epoch": 3.402439024390244, "grad_norm": 1.9640257358551025, "learning_rate": 7.8848339200549e-06, "loss": 0.0954, "step": 279 }, { "epoch": 3.4146341463414633, "grad_norm": 2.0996806621551514, "learning_rate": 7.868340985975195e-06, "loss": 0.0941, "step": 280 }, { "epoch": 3.426829268292683, "grad_norm": 2.0792341232299805, "learning_rate": 7.851801401533288e-06, "loss": 0.0908, "step": 281 }, { "epoch": 3.4390243902439024, "grad_norm": 2.0881197452545166, "learning_rate": 7.835215435727042e-06, "loss": 0.1059, "step": 282 }, { "epoch": 3.451219512195122, "grad_norm": 2.6827352046966553, "learning_rate": 7.818583358308664e-06, "loss": 0.1316, "step": 283 }, { "epoch": 3.4634146341463414, "grad_norm": 2.0524280071258545, "learning_rate": 7.801905439780317e-06, "loss": 0.0957, "step": 284 }, { "epoch": 3.475609756097561, "grad_norm": 2.184852361679077, "learning_rate": 7.785181951389718e-06, "loss": 0.1123, "step": 285 }, { "epoch": 3.4878048780487805, "grad_norm": 2.2295339107513428, "learning_rate": 7.76841316512572e-06, "loss": 0.1198, "step": 286 }, { "epoch": 3.5, "grad_norm": 2.101522922515869, "learning_rate": 7.751599353713906e-06, "loss": 0.0991, "step": 287 }, { "epoch": 3.5121951219512195, "grad_norm": 1.8743051290512085, "learning_rate": 7.734740790612137e-06, "loss": 0.0869, "step": 288 }, { "epoch": 3.524390243902439, "grad_norm": 1.9927822351455688, "learning_rate": 7.717837750006106e-06, "loss": 0.1094, "step": 289 }, { "epoch": 3.5365853658536586, "grad_norm": 2.079759359359741, "learning_rate": 7.700890506804895e-06, "loss": 0.1011, "step": 290 }, { "epoch": 3.548780487804878, "grad_norm": 2.3300259113311768, "learning_rate": 7.68389933663648e-06, "loss": 0.1374, "step": 291 }, { "epoch": 3.5609756097560976, "grad_norm": 2.1061301231384277, "learning_rate": 7.666864515843266e-06, "loss": 0.1123, "step": 292 }, { "epoch": 3.573170731707317, "grad_norm": 1.9325755834579468, "learning_rate": 7.649786321477585e-06, "loss": 0.1052, "step": 293 }, { "epoch": 3.5853658536585367, "grad_norm": 2.3022353649139404, "learning_rate": 7.632665031297193e-06, "loss": 0.102, "step": 294 }, { "epoch": 3.597560975609756, "grad_norm": 1.8938615322113037, "learning_rate": 7.615500923760748e-06, "loss": 0.1065, "step": 295 }, { "epoch": 3.6097560975609757, "grad_norm": 1.8526796102523804, "learning_rate": 7.59829427802329e-06, "loss": 0.0971, "step": 296 }, { "epoch": 3.6219512195121952, "grad_norm": 2.010892391204834, "learning_rate": 7.581045373931691e-06, "loss": 0.0932, "step": 297 }, { "epoch": 3.6341463414634148, "grad_norm": 2.140416383743286, "learning_rate": 7.563754492020108e-06, "loss": 0.0934, "step": 298 }, { "epoch": 3.6463414634146343, "grad_norm": 1.9991627931594849, "learning_rate": 7.54642191350542e-06, "loss": 0.1137, "step": 299 }, { "epoch": 3.658536585365854, "grad_norm": 1.98257577419281, "learning_rate": 7.5290479202826596e-06, "loss": 0.1058, "step": 300 }, { "epoch": 3.6707317073170733, "grad_norm": 1.9862565994262695, "learning_rate": 7.511632794920419e-06, "loss": 0.0977, "step": 301 }, { "epoch": 3.682926829268293, "grad_norm": 2.034688711166382, "learning_rate": 7.494176820656258e-06, "loss": 0.1248, "step": 302 }, { "epoch": 3.6951219512195124, "grad_norm": 1.8107631206512451, "learning_rate": 7.4766802813921016e-06, "loss": 0.0888, "step": 303 }, { "epoch": 3.7073170731707314, "grad_norm": 1.7797682285308838, "learning_rate": 7.4591434616896156e-06, "loss": 0.0971, "step": 304 }, { "epoch": 3.7195121951219514, "grad_norm": 1.8483872413635254, "learning_rate": 7.4415666467655835e-06, "loss": 0.1033, "step": 305 }, { "epoch": 3.7317073170731705, "grad_norm": 1.8434807062149048, "learning_rate": 7.423950122487269e-06, "loss": 0.0929, "step": 306 }, { "epoch": 3.7439024390243905, "grad_norm": 2.006572961807251, "learning_rate": 7.406294175367758e-06, "loss": 0.1034, "step": 307 }, { "epoch": 3.7560975609756095, "grad_norm": 2.015620708465576, "learning_rate": 7.388599092561315e-06, "loss": 0.1091, "step": 308 }, { "epoch": 3.7682926829268295, "grad_norm": 2.08795428276062, "learning_rate": 7.3708651618586925e-06, "loss": 0.0908, "step": 309 }, { "epoch": 3.7804878048780486, "grad_norm": 2.066549777984619, "learning_rate": 7.353092671682464e-06, "loss": 0.093, "step": 310 }, { "epoch": 3.7926829268292686, "grad_norm": 2.227687120437622, "learning_rate": 7.335281911082332e-06, "loss": 0.1042, "step": 311 }, { "epoch": 3.8048780487804876, "grad_norm": 2.5046164989471436, "learning_rate": 7.317433169730421e-06, "loss": 0.136, "step": 312 }, { "epoch": 3.817073170731707, "grad_norm": 2.0135955810546875, "learning_rate": 7.299546737916574e-06, "loss": 0.0942, "step": 313 }, { "epoch": 3.8292682926829267, "grad_norm": 2.3147573471069336, "learning_rate": 7.281622906543625e-06, "loss": 0.11, "step": 314 }, { "epoch": 3.841463414634146, "grad_norm": 2.515584707260132, "learning_rate": 7.26366196712267e-06, "loss": 0.1248, "step": 315 }, { "epoch": 3.8536585365853657, "grad_norm": 1.988805890083313, "learning_rate": 7.245664211768327e-06, "loss": 0.089, "step": 316 }, { "epoch": 3.8658536585365852, "grad_norm": 2.0414860248565674, "learning_rate": 7.227629933193983e-06, "loss": 0.0991, "step": 317 }, { "epoch": 3.8780487804878048, "grad_norm": 1.9820183515548706, "learning_rate": 7.209559424707034e-06, "loss": 0.1163, "step": 318 }, { "epoch": 3.8902439024390243, "grad_norm": 1.9290958642959595, "learning_rate": 7.191452980204119e-06, "loss": 0.1201, "step": 319 }, { "epoch": 3.902439024390244, "grad_norm": 1.9230592250823975, "learning_rate": 7.173310894166328e-06, "loss": 0.1138, "step": 320 }, { "epoch": 3.9146341463414633, "grad_norm": 1.6345875263214111, "learning_rate": 7.155133461654429e-06, "loss": 0.0935, "step": 321 }, { "epoch": 3.926829268292683, "grad_norm": 1.9335048198699951, "learning_rate": 7.136920978304056e-06, "loss": 0.1031, "step": 322 }, { "epoch": 3.9390243902439024, "grad_norm": 1.7330572605133057, "learning_rate": 7.118673740320907e-06, "loss": 0.0945, "step": 323 }, { "epoch": 3.951219512195122, "grad_norm": 1.8825818300247192, "learning_rate": 7.10039204447593e-06, "loss": 0.0966, "step": 324 }, { "epoch": 3.9634146341463414, "grad_norm": 2.1690921783447266, "learning_rate": 7.082076188100483e-06, "loss": 0.1348, "step": 325 }, { "epoch": 3.975609756097561, "grad_norm": 2.1976025104522705, "learning_rate": 7.063726469081511e-06, "loss": 0.1046, "step": 326 }, { "epoch": 3.9878048780487805, "grad_norm": 2.0651566982269287, "learning_rate": 7.045343185856701e-06, "loss": 0.0848, "step": 327 }, { "epoch": 4.0, "grad_norm": 2.3218235969543457, "learning_rate": 7.026926637409615e-06, "loss": 0.1261, "step": 328 }, { "epoch": 4.012195121951219, "grad_norm": 1.517854928970337, "learning_rate": 7.008477123264849e-06, "loss": 0.0424, "step": 329 }, { "epoch": 4.024390243902439, "grad_norm": 1.6785979270935059, "learning_rate": 6.989994943483136e-06, "loss": 0.053, "step": 330 }, { "epoch": 4.036585365853658, "grad_norm": 1.0940113067626953, "learning_rate": 6.971480398656488e-06, "loss": 0.0347, "step": 331 }, { "epoch": 4.048780487804878, "grad_norm": 1.434532880783081, "learning_rate": 6.952933789903299e-06, "loss": 0.0468, "step": 332 }, { "epoch": 4.060975609756097, "grad_norm": 1.7367973327636719, "learning_rate": 6.93435541886344e-06, "loss": 0.0439, "step": 333 }, { "epoch": 4.073170731707317, "grad_norm": 1.4013808965682983, "learning_rate": 6.915745587693365e-06, "loss": 0.0341, "step": 334 }, { "epoch": 4.085365853658536, "grad_norm": 1.7729628086090088, "learning_rate": 6.89710459906119e-06, "loss": 0.0524, "step": 335 }, { "epoch": 4.097560975609756, "grad_norm": 1.8550630807876587, "learning_rate": 6.878432756141775e-06, "loss": 0.0559, "step": 336 }, { "epoch": 4.109756097560975, "grad_norm": 1.9048420190811157, "learning_rate": 6.8597303626117886e-06, "loss": 0.0567, "step": 337 }, { "epoch": 4.121951219512195, "grad_norm": 2.3313469886779785, "learning_rate": 6.8409977226447685e-06, "loss": 0.0589, "step": 338 }, { "epoch": 4.134146341463414, "grad_norm": 1.5067005157470703, "learning_rate": 6.822235140906183e-06, "loss": 0.0415, "step": 339 }, { "epoch": 4.146341463414634, "grad_norm": 1.7281876802444458, "learning_rate": 6.803442922548462e-06, "loss": 0.0491, "step": 340 }, { "epoch": 4.158536585365853, "grad_norm": 1.7764736413955688, "learning_rate": 6.784621373206051e-06, "loss": 0.049, "step": 341 }, { "epoch": 4.170731707317073, "grad_norm": 2.0232222080230713, "learning_rate": 6.765770798990423e-06, "loss": 0.0524, "step": 342 }, { "epoch": 4.182926829268292, "grad_norm": 1.9550089836120605, "learning_rate": 6.746891506485112e-06, "loss": 0.0526, "step": 343 }, { "epoch": 4.195121951219512, "grad_norm": 2.0394773483276367, "learning_rate": 6.727983802740723e-06, "loss": 0.0546, "step": 344 }, { "epoch": 4.2073170731707314, "grad_norm": 1.6590560674667358, "learning_rate": 6.709047995269939e-06, "loss": 0.0422, "step": 345 }, { "epoch": 4.219512195121951, "grad_norm": 1.8558006286621094, "learning_rate": 6.690084392042514e-06, "loss": 0.0518, "step": 346 }, { "epoch": 4.2317073170731705, "grad_norm": 1.2415188550949097, "learning_rate": 6.671093301480276e-06, "loss": 0.0333, "step": 347 }, { "epoch": 4.2439024390243905, "grad_norm": 1.7380534410476685, "learning_rate": 6.6520750324520965e-06, "loss": 0.0556, "step": 348 }, { "epoch": 4.2560975609756095, "grad_norm": 1.4161667823791504, "learning_rate": 6.63302989426888e-06, "loss": 0.0414, "step": 349 }, { "epoch": 4.2682926829268295, "grad_norm": 1.6313724517822266, "learning_rate": 6.613958196678525e-06, "loss": 0.0757, "step": 350 }, { "epoch": 4.280487804878049, "grad_norm": 1.9501330852508545, "learning_rate": 6.594860249860888e-06, "loss": 0.0675, "step": 351 }, { "epoch": 4.2926829268292686, "grad_norm": 1.5222731828689575, "learning_rate": 6.575736364422747e-06, "loss": 0.0537, "step": 352 }, { "epoch": 4.304878048780488, "grad_norm": 1.367255687713623, "learning_rate": 6.55658685139273e-06, "loss": 0.0459, "step": 353 }, { "epoch": 4.317073170731708, "grad_norm": 1.4813297986984253, "learning_rate": 6.5374120222162815e-06, "loss": 0.06, "step": 354 }, { "epoch": 4.329268292682927, "grad_norm": 1.5068612098693848, "learning_rate": 6.518212188750579e-06, "loss": 0.0514, "step": 355 }, { "epoch": 4.341463414634147, "grad_norm": 1.66206955909729, "learning_rate": 6.498987663259467e-06, "loss": 0.0675, "step": 356 }, { "epoch": 4.353658536585366, "grad_norm": 1.4990217685699463, "learning_rate": 6.479738758408379e-06, "loss": 0.0695, "step": 357 }, { "epoch": 4.365853658536586, "grad_norm": 1.5749341249465942, "learning_rate": 6.460465787259251e-06, "loss": 0.0508, "step": 358 }, { "epoch": 4.378048780487805, "grad_norm": 1.499898076057434, "learning_rate": 6.44116906326543e-06, "loss": 0.0591, "step": 359 }, { "epoch": 4.390243902439025, "grad_norm": 1.46736478805542, "learning_rate": 6.421848900266581e-06, "loss": 0.05, "step": 360 }, { "epoch": 4.402439024390244, "grad_norm": 1.4807460308074951, "learning_rate": 6.402505612483569e-06, "loss": 0.0523, "step": 361 }, { "epoch": 4.414634146341464, "grad_norm": 1.4587833881378174, "learning_rate": 6.383139514513368e-06, "loss": 0.0576, "step": 362 }, { "epoch": 4.426829268292683, "grad_norm": 1.4291479587554932, "learning_rate": 6.363750921323929e-06, "loss": 0.0479, "step": 363 }, { "epoch": 4.439024390243903, "grad_norm": 1.364157795906067, "learning_rate": 6.3443401482490615e-06, "loss": 0.0528, "step": 364 }, { "epoch": 4.451219512195122, "grad_norm": 2.088580369949341, "learning_rate": 6.32490751098331e-06, "loss": 0.0605, "step": 365 }, { "epoch": 4.463414634146342, "grad_norm": 1.5994398593902588, "learning_rate": 6.30545332557681e-06, "loss": 0.0525, "step": 366 }, { "epoch": 4.475609756097561, "grad_norm": 1.7937228679656982, "learning_rate": 6.2859779084301584e-06, "loss": 0.0517, "step": 367 }, { "epoch": 4.487804878048781, "grad_norm": 1.3765718936920166, "learning_rate": 6.266481576289263e-06, "loss": 0.041, "step": 368 }, { "epoch": 4.5, "grad_norm": 1.7616742849349976, "learning_rate": 6.246964646240186e-06, "loss": 0.0715, "step": 369 }, { "epoch": 4.512195121951219, "grad_norm": 1.496747374534607, "learning_rate": 6.227427435703997e-06, "loss": 0.0633, "step": 370 }, { "epoch": 4.524390243902439, "grad_norm": 1.53587007522583, "learning_rate": 6.207870262431599e-06, "loss": 0.0557, "step": 371 }, { "epoch": 4.536585365853659, "grad_norm": 1.664995789527893, "learning_rate": 6.188293444498573e-06, "loss": 0.0599, "step": 372 }, { "epoch": 4.548780487804878, "grad_norm": 1.8567813634872437, "learning_rate": 6.1686973002999935e-06, "loss": 0.0643, "step": 373 }, { "epoch": 4.560975609756097, "grad_norm": 2.01507568359375, "learning_rate": 6.149082148545258e-06, "loss": 0.0637, "step": 374 }, { "epoch": 4.573170731707317, "grad_norm": 1.800641417503357, "learning_rate": 6.129448308252899e-06, "loss": 0.0587, "step": 375 }, { "epoch": 4.585365853658536, "grad_norm": 2.0126662254333496, "learning_rate": 6.109796098745398e-06, "loss": 0.0669, "step": 376 }, { "epoch": 4.597560975609756, "grad_norm": 1.8245577812194824, "learning_rate": 6.090125839643991e-06, "loss": 0.0541, "step": 377 }, { "epoch": 4.609756097560975, "grad_norm": 1.3531700372695923, "learning_rate": 6.070437850863472e-06, "loss": 0.0445, "step": 378 }, { "epoch": 4.621951219512195, "grad_norm": 1.9308772087097168, "learning_rate": 6.0507324526069854e-06, "loss": 0.0608, "step": 379 }, { "epoch": 4.634146341463414, "grad_norm": 1.5027072429656982, "learning_rate": 6.031009965360824e-06, "loss": 0.0634, "step": 380 }, { "epoch": 4.646341463414634, "grad_norm": 1.3451308012008667, "learning_rate": 6.011270709889213e-06, "loss": 0.0411, "step": 381 }, { "epoch": 4.658536585365853, "grad_norm": 1.618082046508789, "learning_rate": 5.991515007229093e-06, "loss": 0.0575, "step": 382 }, { "epoch": 4.670731707317073, "grad_norm": 1.6030172109603882, "learning_rate": 5.971743178684901e-06, "loss": 0.0575, "step": 383 }, { "epoch": 4.682926829268292, "grad_norm": 1.582740306854248, "learning_rate": 5.951955545823342e-06, "loss": 0.0613, "step": 384 }, { "epoch": 4.695121951219512, "grad_norm": 1.7536263465881348, "learning_rate": 5.932152430468165e-06, "loss": 0.052, "step": 385 }, { "epoch": 4.7073170731707314, "grad_norm": 2.1995296478271484, "learning_rate": 5.912334154694919e-06, "loss": 0.0629, "step": 386 }, { "epoch": 4.719512195121951, "grad_norm": 1.8581688404083252, "learning_rate": 5.892501040825721e-06, "loss": 0.041, "step": 387 }, { "epoch": 4.7317073170731705, "grad_norm": 1.8024824857711792, "learning_rate": 5.872653411424017e-06, "loss": 0.0708, "step": 388 }, { "epoch": 4.7439024390243905, "grad_norm": 1.7822990417480469, "learning_rate": 5.85279158928933e-06, "loss": 0.0528, "step": 389 }, { "epoch": 4.7560975609756095, "grad_norm": 1.9106731414794922, "learning_rate": 5.832915897452008e-06, "loss": 0.0643, "step": 390 }, { "epoch": 4.7682926829268295, "grad_norm": 1.593004584312439, "learning_rate": 5.813026659167982e-06, "loss": 0.054, "step": 391 }, { "epoch": 4.780487804878049, "grad_norm": 1.8973208665847778, "learning_rate": 5.793124197913492e-06, "loss": 0.0737, "step": 392 }, { "epoch": 4.7926829268292686, "grad_norm": 1.9966886043548584, "learning_rate": 5.773208837379843e-06, "loss": 0.0634, "step": 393 }, { "epoch": 4.804878048780488, "grad_norm": 1.5227646827697754, "learning_rate": 5.753280901468126e-06, "loss": 0.0496, "step": 394 }, { "epoch": 4.817073170731708, "grad_norm": 1.6435083150863647, "learning_rate": 5.733340714283959e-06, "loss": 0.0664, "step": 395 }, { "epoch": 4.829268292682927, "grad_norm": 1.3312773704528809, "learning_rate": 5.713388600132217e-06, "loss": 0.0534, "step": 396 }, { "epoch": 4.841463414634147, "grad_norm": 1.868194580078125, "learning_rate": 5.693424883511748e-06, "loss": 0.0565, "step": 397 }, { "epoch": 4.853658536585366, "grad_norm": 1.5551823377609253, "learning_rate": 5.6734498891101005e-06, "loss": 0.0604, "step": 398 }, { "epoch": 4.865853658536586, "grad_norm": 1.8578870296478271, "learning_rate": 5.653463941798252e-06, "loss": 0.0728, "step": 399 }, { "epoch": 4.878048780487805, "grad_norm": 1.5294170379638672, "learning_rate": 5.633467366625306e-06, "loss": 0.0637, "step": 400 }, { "epoch": 4.890243902439025, "grad_norm": 1.2593622207641602, "learning_rate": 5.613460488813225e-06, "loss": 0.0512, "step": 401 }, { "epoch": 4.902439024390244, "grad_norm": 1.7771371603012085, "learning_rate": 5.593443633751527e-06, "loss": 0.0658, "step": 402 }, { "epoch": 4.914634146341464, "grad_norm": 1.5825587511062622, "learning_rate": 5.573417126992004e-06, "loss": 0.0671, "step": 403 }, { "epoch": 4.926829268292683, "grad_norm": 1.6244094371795654, "learning_rate": 5.553381294243413e-06, "loss": 0.0585, "step": 404 }, { "epoch": 4.939024390243903, "grad_norm": 1.501323938369751, "learning_rate": 5.5333364613662e-06, "loss": 0.0578, "step": 405 }, { "epoch": 4.951219512195122, "grad_norm": 1.5930196046829224, "learning_rate": 5.513282954367179e-06, "loss": 0.064, "step": 406 }, { "epoch": 4.963414634146341, "grad_norm": 1.4195719957351685, "learning_rate": 5.493221099394239e-06, "loss": 0.0443, "step": 407 }, { "epoch": 4.975609756097561, "grad_norm": 1.3484866619110107, "learning_rate": 5.473151222731044e-06, "loss": 0.0577, "step": 408 }, { "epoch": 4.987804878048781, "grad_norm": 1.677027940750122, "learning_rate": 5.453073650791724e-06, "loss": 0.0604, "step": 409 }, { "epoch": 5.0, "grad_norm": 1.7022733688354492, "learning_rate": 5.432988710115553e-06, "loss": 0.0674, "step": 410 } ], "logging_steps": 1, "max_steps": 820, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 1, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7614453848064.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }